mirror of
https://github.com/prometheus/prometheus.git
synced 2024-11-12 16:44:05 -08:00
golangci-lint: enable whitespace linter (#13905)
Signed-off-by: Matthieu MOREL <matthieu.morel35@gmail.com>
This commit is contained in:
parent
594b317ecc
commit
6f595c6762
|
@ -27,6 +27,7 @@ linters:
|
||||||
- unconvert
|
- unconvert
|
||||||
- unused
|
- unused
|
||||||
- usestdlibvars
|
- usestdlibvars
|
||||||
|
- whitespace
|
||||||
|
|
||||||
issues:
|
issues:
|
||||||
max-same-issues: 0
|
max-same-issues: 0
|
||||||
|
|
|
@ -78,7 +78,6 @@ func TestBackfillRuleIntegration(t *testing.T) {
|
||||||
// Execute the test more than once to simulate running the rule importer twice with the same data.
|
// Execute the test more than once to simulate running the rule importer twice with the same data.
|
||||||
// We expect duplicate blocks with the same series are created when run more than once.
|
// We expect duplicate blocks with the same series are created when run more than once.
|
||||||
for i := 0; i < tt.runcount; i++ {
|
for i := 0; i < tt.runcount; i++ {
|
||||||
|
|
||||||
ruleImporter, err := newTestRuleImporter(ctx, start, tmpDir, tt.samples, tt.maxBlockDuration)
|
ruleImporter, err := newTestRuleImporter(ctx, start, tmpDir, tt.samples, tt.maxBlockDuration)
|
||||||
require.NoError(t, err)
|
require.NoError(t, err)
|
||||||
path1 := filepath.Join(tmpDir, "test.file")
|
path1 := filepath.Join(tmpDir, "test.file")
|
||||||
|
|
|
@ -228,7 +228,6 @@ func targetsForApp(app *Application) []model.LabelSet {
|
||||||
}
|
}
|
||||||
|
|
||||||
targets = append(targets, target)
|
targets = append(targets, target)
|
||||||
|
|
||||||
}
|
}
|
||||||
return targets
|
return targets
|
||||||
}
|
}
|
||||||
|
|
|
@ -122,7 +122,6 @@ func (d *robotDiscovery) refresh(context.Context) ([]*targetgroup.Group, error)
|
||||||
labels[hetznerLabelPublicIPv6Network] = model.LabelValue(fmt.Sprintf("%s/%s", subnet.IP, subnet.Mask))
|
labels[hetznerLabelPublicIPv6Network] = model.LabelValue(fmt.Sprintf("%s/%s", subnet.IP, subnet.Mask))
|
||||||
break
|
break
|
||||||
}
|
}
|
||||||
|
|
||||||
}
|
}
|
||||||
targets[i] = labels
|
targets[i] = labels
|
||||||
}
|
}
|
||||||
|
|
|
@ -453,7 +453,6 @@ func targetsForApp(app *app) []model.LabelSet {
|
||||||
// Gather info about the app's 'tasks'. Each instance (container) is considered a task
|
// Gather info about the app's 'tasks'. Each instance (container) is considered a task
|
||||||
// and can be reachable at one or more host:port endpoints.
|
// and can be reachable at one or more host:port endpoints.
|
||||||
for _, t := range app.Tasks {
|
for _, t := range app.Tasks {
|
||||||
|
|
||||||
// There are no labels to gather if only Ports is defined. (eg. with host networking)
|
// There are no labels to gather if only Ports is defined. (eg. with host networking)
|
||||||
// Ports can only be gathered from the Task (not from the app) and are guaranteed
|
// Ports can only be gathered from the Task (not from the app) and are guaranteed
|
||||||
// to be the same across all tasks. If we haven't gathered any ports by now,
|
// to be the same across all tasks. If we haven't gathered any ports by now,
|
||||||
|
@ -464,7 +463,6 @@ func targetsForApp(app *app) []model.LabelSet {
|
||||||
|
|
||||||
// Iterate over the ports we gathered using one of the methods above.
|
// Iterate over the ports we gathered using one of the methods above.
|
||||||
for i, port := range ports {
|
for i, port := range ports {
|
||||||
|
|
||||||
// A zero port here means that either the portMapping has a zero port defined,
|
// A zero port here means that either the portMapping has a zero port defined,
|
||||||
// or there is a portDefinition with requirePorts set to false. This means the port
|
// or there is a portDefinition with requirePorts set to false. This means the port
|
||||||
// is auto-generated by Mesos and needs to be looked up in the task.
|
// is auto-generated by Mesos and needs to be looked up in the task.
|
||||||
|
@ -516,7 +514,6 @@ func extractPortMapping(portMappings []portMapping, containerNet bool) ([]uint32
|
||||||
labels := make([]map[string]string, len(portMappings))
|
labels := make([]map[string]string, len(portMappings))
|
||||||
|
|
||||||
for i := 0; i < len(portMappings); i++ {
|
for i := 0; i < len(portMappings); i++ {
|
||||||
|
|
||||||
labels[i] = portMappings[i].Labels
|
labels[i] = portMappings[i].Labels
|
||||||
|
|
||||||
if containerNet {
|
if containerNet {
|
||||||
|
|
|
@ -80,7 +80,6 @@ func (d *Discovery) refreshNodes(ctx context.Context) ([]*targetgroup.Group, err
|
||||||
labels[model.AddressLabel] = model.LabelValue(addr)
|
labels[model.AddressLabel] = model.LabelValue(addr)
|
||||||
|
|
||||||
tg.Targets = append(tg.Targets, labels)
|
tg.Targets = append(tg.Targets, labels)
|
||||||
|
|
||||||
}
|
}
|
||||||
return []*targetgroup.Group{tg}, nil
|
return []*targetgroup.Group{tg}, nil
|
||||||
}
|
}
|
||||||
|
|
|
@ -190,7 +190,6 @@ func (d *instanceDiscovery) refresh(ctx context.Context) ([]*targetgroup.Group,
|
||||||
|
|
||||||
targets = append(targets, labels)
|
targets = append(targets, labels)
|
||||||
}
|
}
|
||||||
|
|
||||||
}
|
}
|
||||||
|
|
||||||
return []*targetgroup.Group{{Source: "scaleway", Targets: targets}}, nil
|
return []*targetgroup.Group{{Source: "scaleway", Targets: targets}}, nil
|
||||||
|
|
|
@ -291,7 +291,6 @@ func parseServersetMember(data []byte, path string) (model.LabelSet, error) {
|
||||||
endpoint.Host)
|
endpoint.Host)
|
||||||
labels[serversetEndpointLabelPrefix+"_port_"+cleanName] = model.LabelValue(
|
labels[serversetEndpointLabelPrefix+"_port_"+cleanName] = model.LabelValue(
|
||||||
fmt.Sprintf("%d", endpoint.Port))
|
fmt.Sprintf("%d", endpoint.Port))
|
||||||
|
|
||||||
}
|
}
|
||||||
|
|
||||||
labels[serversetStatusLabel] = model.LabelValue(member.Status)
|
labels[serversetStatusLabel] = model.LabelValue(member.Status)
|
||||||
|
|
|
@ -342,7 +342,6 @@ func funcHoltWinters(vals []parser.Value, args parser.Expressions, enh *EvalNode
|
||||||
// Run the smoothing operation.
|
// Run the smoothing operation.
|
||||||
var x, y float64
|
var x, y float64
|
||||||
for i := 1; i < l; i++ {
|
for i := 1; i < l; i++ {
|
||||||
|
|
||||||
// Scale the raw value against the smoothing factor.
|
// Scale the raw value against the smoothing factor.
|
||||||
x = sf * samples.Floats[i].F
|
x = sf * samples.Floats[i].F
|
||||||
|
|
||||||
|
@ -1240,7 +1239,6 @@ func funcHistogramQuantile(vals []parser.Value, args parser.Expressions, enh *Ev
|
||||||
enh.signatureToMetricWithBuckets[string(enh.lblBuf)] = mb
|
enh.signatureToMetricWithBuckets[string(enh.lblBuf)] = mb
|
||||||
}
|
}
|
||||||
mb.buckets = append(mb.buckets, bucket{upperBound, sample.F})
|
mb.buckets = append(mb.buckets, bucket{upperBound, sample.F})
|
||||||
|
|
||||||
}
|
}
|
||||||
|
|
||||||
// Now deal with the histograms.
|
// Now deal with the histograms.
|
||||||
|
|
|
@ -258,7 +258,6 @@ func ParseSeriesDesc(input string) (labels labels.Labels, values []SequenceValue
|
||||||
|
|
||||||
labels = result.labels
|
labels = result.labels
|
||||||
values = result.values
|
values = result.values
|
||||||
|
|
||||||
}
|
}
|
||||||
|
|
||||||
if len(p.parseErrors) != 0 {
|
if len(p.parseErrors) != 0 {
|
||||||
|
|
|
@ -567,7 +567,6 @@ func (ev *evalCmd) compareResult(result parser.Value) error {
|
||||||
return fmt.Errorf("expected histogram value at index %v (t=%v) for %s to be %v, but got %v (result has %s)", i, actual.T, ev.metrics[hash], expected.H, actual.H, formatSeriesResult(s))
|
return fmt.Errorf("expected histogram value at index %v (t=%v) for %s to be %v, but got %v (result has %s)", i, actual.T, ev.metrics[hash], expected.H, actual.H, formatSeriesResult(s))
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
}
|
}
|
||||||
|
|
||||||
for hash := range ev.expected {
|
for hash := range ev.expected {
|
||||||
|
|
|
@ -336,7 +336,6 @@ func TestForStateAddSamples(t *testing.T) {
|
||||||
for _, aa := range rule.ActiveAlerts() {
|
for _, aa := range rule.ActiveAlerts() {
|
||||||
require.Zero(t, aa.Labels.Get(model.MetricNameLabel), "%s label set on active alert: %s", model.MetricNameLabel, aa.Labels)
|
require.Zero(t, aa.Labels.Get(model.MetricNameLabel), "%s label set on active alert: %s", model.MetricNameLabel, aa.Labels)
|
||||||
}
|
}
|
||||||
|
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
|
@ -1824,7 +1823,6 @@ func TestDependencyMapUpdatesOnGroupUpdate(t *testing.T) {
|
||||||
} else {
|
} else {
|
||||||
require.Equal(t, orig[h], depMap)
|
require.Equal(t, orig[h], depMap)
|
||||||
}
|
}
|
||||||
|
|
||||||
}
|
}
|
||||||
|
|
||||||
// Groups will be recreated when updated.
|
// Groups will be recreated when updated.
|
||||||
|
@ -1962,7 +1960,6 @@ func TestAsyncRuleEvaluation(t *testing.T) {
|
||||||
require.Less(t, time.Since(start).Seconds(), (time.Duration(ruleCount) * artificialDelay).Seconds())
|
require.Less(t, time.Since(start).Seconds(), (time.Duration(ruleCount) * artificialDelay).Seconds())
|
||||||
// Each rule produces one vector.
|
// Each rule produces one vector.
|
||||||
require.EqualValues(t, ruleCount, testutil.ToFloat64(group.metrics.GroupSamples))
|
require.EqualValues(t, ruleCount, testutil.ToFloat64(group.metrics.GroupSamples))
|
||||||
|
|
||||||
}
|
}
|
||||||
})
|
})
|
||||||
|
|
||||||
|
|
|
@ -185,7 +185,6 @@ func (m *Manager) reload() {
|
||||||
sp.Sync(groups)
|
sp.Sync(groups)
|
||||||
wg.Done()
|
wg.Done()
|
||||||
}(m.scrapePools[setName], groups)
|
}(m.scrapePools[setName], groups)
|
||||||
|
|
||||||
}
|
}
|
||||||
m.mtxScrape.Unlock()
|
m.mtxScrape.Unlock()
|
||||||
wg.Wait()
|
wg.Wait()
|
||||||
|
|
|
@ -583,7 +583,6 @@ func TestManagerTargetsUpdates(t *testing.T) {
|
||||||
|
|
||||||
tgSent := make(map[string][]*targetgroup.Group)
|
tgSent := make(map[string][]*targetgroup.Group)
|
||||||
for x := 0; x < 10; x++ {
|
for x := 0; x < 10; x++ {
|
||||||
|
|
||||||
tgSent[strconv.Itoa(x)] = []*targetgroup.Group{
|
tgSent[strconv.Itoa(x)] = []*targetgroup.Group{
|
||||||
{
|
{
|
||||||
Source: strconv.Itoa(x),
|
Source: strconv.Itoa(x),
|
||||||
|
|
|
@ -592,7 +592,6 @@ func TestMaxSchemaAppender(t *testing.T) {
|
||||||
_, err = app.AppendHistogram(0, lbls, ts, nil, fh)
|
_, err = app.AppendHistogram(0, lbls, ts, nil, fh)
|
||||||
require.Equal(t, c.expectSchema, fh.Schema)
|
require.Equal(t, c.expectSchema, fh.Schema)
|
||||||
require.NoError(t, err)
|
require.NoError(t, err)
|
||||||
|
|
||||||
} else {
|
} else {
|
||||||
h := c.h.Copy()
|
h := c.h.Copy()
|
||||||
_, err = app.AppendHistogram(0, lbls, ts, h, nil)
|
_, err = app.AppendHistogram(0, lbls, ts, h, nil)
|
||||||
|
|
|
@ -377,7 +377,6 @@ func TestMergeChunkQuerierWithNoVerticalChunkSeriesMerger(t *testing.T) {
|
||||||
actChks, actErr := ExpandChunks(actualSeries.Iterator(nil))
|
actChks, actErr := ExpandChunks(actualSeries.Iterator(nil))
|
||||||
require.Equal(t, expErr, actErr)
|
require.Equal(t, expErr, actErr)
|
||||||
require.Equal(t, expChks, actChks)
|
require.Equal(t, expChks, actChks)
|
||||||
|
|
||||||
}
|
}
|
||||||
require.NoError(t, merged.Err())
|
require.NoError(t, merged.Err())
|
||||||
require.False(t, tc.expected.Next(), "Expected Next() to be false")
|
require.False(t, tc.expected.Next(), "Expected Next() to be false")
|
||||||
|
|
|
@ -845,7 +845,6 @@ func (c *TestWriteClient) Store(_ context.Context, req []byte, _ int) error {
|
||||||
} else {
|
} else {
|
||||||
c.receivedHistograms[seriesName] = append(c.receivedHistograms[seriesName], histogram)
|
c.receivedHistograms[seriesName] = append(c.receivedHistograms[seriesName], histogram)
|
||||||
}
|
}
|
||||||
|
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
if c.withWaitGroup {
|
if c.withWaitGroup {
|
||||||
|
|
|
@ -135,7 +135,6 @@ func (h *writeHandler) write(ctx context.Context, req *prompb.WriteRequest) (err
|
||||||
}
|
}
|
||||||
return err
|
return err
|
||||||
}
|
}
|
||||||
|
|
||||||
}
|
}
|
||||||
|
|
||||||
for _, ep := range ts.Exemplars {
|
for _, ep := range ts.Exemplars {
|
||||||
|
|
|
@ -459,7 +459,6 @@ func TestLabelNamesWithMatchers(t *testing.T) {
|
||||||
"unique", fmt.Sprintf("value%d", i),
|
"unique", fmt.Sprintf("value%d", i),
|
||||||
), []chunks.Sample{sample{100, 0, nil, nil}}))
|
), []chunks.Sample{sample{100, 0, nil, nil}}))
|
||||||
}
|
}
|
||||||
|
|
||||||
}
|
}
|
||||||
|
|
||||||
blockDir := createBlock(t, tmpdir, seriesEntries)
|
blockDir := createBlock(t, tmpdir, seriesEntries)
|
||||||
|
|
|
@ -73,7 +73,6 @@ func readHistogramChunkLayoutSpans(b *bstreamReader) ([]histogram.Span, error) {
|
||||||
return nil, err
|
return nil, err
|
||||||
}
|
}
|
||||||
for i := 0; i < int(num); i++ {
|
for i := 0; i < int(num); i++ {
|
||||||
|
|
||||||
length, err := readVarbitUint(b)
|
length, err := readVarbitUint(b)
|
||||||
if err != nil {
|
if err != nil {
|
||||||
return nil, err
|
return nil, err
|
||||||
|
|
|
@ -1766,7 +1766,6 @@ func OverlappingBlocks(bm []BlockMeta) Overlaps {
|
||||||
// Fetch the critical overlapped time range foreach overlap groups.
|
// Fetch the critical overlapped time range foreach overlap groups.
|
||||||
overlapGroups := Overlaps{}
|
overlapGroups := Overlaps{}
|
||||||
for _, overlap := range overlaps {
|
for _, overlap := range overlaps {
|
||||||
|
|
||||||
minRange := TimeRange{Min: 0, Max: math.MaxInt64}
|
minRange := TimeRange{Min: 0, Max: math.MaxInt64}
|
||||||
for _, b := range overlap {
|
for _, b := range overlap {
|
||||||
if minRange.Max > b.MaxTime {
|
if minRange.Max > b.MaxTime {
|
||||||
|
|
|
@ -1275,7 +1275,6 @@ func (s *memSeries) appendPreprocessor(t int64, e chunkenc.Encoding, o chunkOpts
|
||||||
// encoding. So we cut a new chunk with the expected encoding.
|
// encoding. So we cut a new chunk with the expected encoding.
|
||||||
c = s.cutNewHeadChunk(t, e, o.chunkRange)
|
c = s.cutNewHeadChunk(t, e, o.chunkRange)
|
||||||
chunkCreated = true
|
chunkCreated = true
|
||||||
|
|
||||||
}
|
}
|
||||||
|
|
||||||
numSamples := c.chunk.NumSamples()
|
numSamples := c.chunk.NumSamples()
|
||||||
|
|
|
@ -1215,7 +1215,6 @@ func TestHeadDeleteSimple(t *testing.T) {
|
||||||
for _, smpl := range smplsAll {
|
for _, smpl := range smplsAll {
|
||||||
_, err := app.Append(0, lblsDefault, smpl.t, smpl.f)
|
_, err := app.Append(0, lblsDefault, smpl.t, smpl.f)
|
||||||
require.NoError(t, err)
|
require.NoError(t, err)
|
||||||
|
|
||||||
}
|
}
|
||||||
require.NoError(t, app.Commit())
|
require.NoError(t, app.Commit())
|
||||||
|
|
||||||
|
@ -1229,7 +1228,6 @@ func TestHeadDeleteSimple(t *testing.T) {
|
||||||
for _, smpl := range c.addSamples {
|
for _, smpl := range c.addSamples {
|
||||||
_, err := app.Append(0, lblsDefault, smpl.t, smpl.f)
|
_, err := app.Append(0, lblsDefault, smpl.t, smpl.f)
|
||||||
require.NoError(t, err)
|
require.NoError(t, err)
|
||||||
|
|
||||||
}
|
}
|
||||||
require.NoError(t, app.Commit())
|
require.NoError(t, app.Commit())
|
||||||
|
|
||||||
|
@ -3851,7 +3849,6 @@ func TestChunkSnapshot(t *testing.T) {
|
||||||
}
|
}
|
||||||
|
|
||||||
{ // Additional data to only include in WAL and m-mapped chunks and not snapshot. This mimics having an old snapshot on disk.
|
{ // Additional data to only include in WAL and m-mapped chunks and not snapshot. This mimics having an old snapshot on disk.
|
||||||
|
|
||||||
// Add more samples.
|
// Add more samples.
|
||||||
app := head.Appender(context.Background())
|
app := head.Appender(context.Background())
|
||||||
for i := 1; i <= numSeries; i++ {
|
for i := 1; i <= numSeries; i++ {
|
||||||
|
|
|
@ -1323,7 +1323,6 @@ func DeleteChunkSnapshots(dir string, maxIndex, maxOffset int) error {
|
||||||
errs.Add(err)
|
errs.Add(err)
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
}
|
}
|
||||||
return errs.Err()
|
return errs.Err()
|
||||||
}
|
}
|
||||||
|
|
|
@ -1528,7 +1528,6 @@ func (r *Reader) LabelValues(ctx context.Context, name string, matchers ...*labe
|
||||||
values = append(values, k)
|
values = append(values, k)
|
||||||
}
|
}
|
||||||
return values, nil
|
return values, nil
|
||||||
|
|
||||||
}
|
}
|
||||||
e, ok := r.postings[name]
|
e, ok := r.postings[name]
|
||||||
if !ok {
|
if !ok {
|
||||||
|
|
|
@ -844,7 +844,6 @@ func (p *populateWithDelChunkSeriesIterator) Next() bool {
|
||||||
return true
|
return true
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
}
|
}
|
||||||
return false
|
return false
|
||||||
}
|
}
|
||||||
|
|
|
@ -213,7 +213,6 @@ func (w *Watcher) setMetrics() {
|
||||||
w.samplesSentPreTailing = w.metrics.samplesSentPreTailing.WithLabelValues(w.name)
|
w.samplesSentPreTailing = w.metrics.samplesSentPreTailing.WithLabelValues(w.name)
|
||||||
w.currentSegmentMetric = w.metrics.currentSegment.WithLabelValues(w.name)
|
w.currentSegmentMetric = w.metrics.currentSegment.WithLabelValues(w.name)
|
||||||
w.notificationsSkipped = w.metrics.notificationsSkipped.WithLabelValues(w.name)
|
w.notificationsSkipped = w.metrics.notificationsSkipped.WithLabelValues(w.name)
|
||||||
|
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
|
|
Loading…
Reference in a new issue