// Copyright 2017 The Prometheus Authors // Licensed under the Apache License, Version 2.0 (the "License"); // you may not use this file except in compliance with the License. // You may obtain a copy of the License at // // http://www.apache.org/licenses/LICENSE-2.0 // // Unless required by applicable law or agreed to in writing, software // distributed under the License is distributed on an "AS IS" BASIS, // WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. // See the License for the specific language governing permissions and // limitations under the License. package tsdb import ( "context" "encoding/binary" "errors" "fmt" "hash/crc32" "math/rand" "os" "path/filepath" "sort" "strconv" "testing" "github.com/go-kit/log" prom_testutil "github.com/prometheus/client_golang/prometheus/testutil" "github.com/stretchr/testify/require" "github.com/prometheus/prometheus/model/histogram" "github.com/prometheus/prometheus/model/labels" "github.com/prometheus/prometheus/storage" "github.com/prometheus/prometheus/tsdb/chunkenc" "github.com/prometheus/prometheus/tsdb/chunks" "github.com/prometheus/prometheus/tsdb/fileutil" "github.com/prometheus/prometheus/tsdb/tsdbutil" "github.com/prometheus/prometheus/tsdb/wlog" ) // In Prometheus 2.1.0 we had a bug where the meta.json version was falsely bumped // to 2. We had a migration in place resetting it to 1 but we should move immediately to // version 3 next time to avoid confusion and issues. func TestBlockMetaMustNeverBeVersion2(t *testing.T) { dir := t.TempDir() _, err := writeMetaFile(log.NewNopLogger(), dir, &BlockMeta{}) require.NoError(t, err) meta, _, err := readMetaFile(dir) require.NoError(t, err) require.NotEqual(t, 2, meta.Version, "meta.json version must never be 2") } func TestSetCompactionFailed(t *testing.T) { tmpdir := t.TempDir() blockDir := createBlock(t, tmpdir, genSeries(1, 1, 0, 1)) b, err := OpenBlock(nil, blockDir, nil) require.NoError(t, err) require.Equal(t, false, b.meta.Compaction.Failed) require.NoError(t, b.setCompactionFailed()) require.Equal(t, true, b.meta.Compaction.Failed) require.NoError(t, b.Close()) b, err = OpenBlock(nil, blockDir, nil) require.NoError(t, err) require.Equal(t, true, b.meta.Compaction.Failed) require.NoError(t, b.Close()) } func TestCreateBlock(t *testing.T) { tmpdir := t.TempDir() b, err := OpenBlock(nil, createBlock(t, tmpdir, genSeries(1, 1, 0, 10)), nil) require.NoError(t, err) require.NoError(t, b.Close()) } func BenchmarkOpenBlock(b *testing.B) { tmpdir := b.TempDir() blockDir := createBlock(b, tmpdir, genSeries(1e6, 20, 0, 10)) b.Run("benchmark", func(b *testing.B) { for i := 0; i < b.N; i++ { block, err := OpenBlock(nil, blockDir, nil) require.NoError(b, err) require.NoError(b, block.Close()) } }) } func TestCorruptedChunk(t *testing.T) { for _, tc := range []struct { name string corrFunc func(f *os.File) // Func that applies the corruption. openErr error iterErr error }{ { name: "invalid header size", corrFunc: func(f *os.File) { require.NoError(t, f.Truncate(1)) }, openErr: errors.New("invalid segment header in segment 0: invalid size"), }, { name: "invalid magic number", corrFunc: func(f *os.File) { magicChunksOffset := int64(0) _, err := f.Seek(magicChunksOffset, 0) require.NoError(t, err) // Set invalid magic number. b := make([]byte, chunks.MagicChunksSize) binary.BigEndian.PutUint32(b[:chunks.MagicChunksSize], 0x00000000) n, err := f.Write(b) require.NoError(t, err) require.Equal(t, chunks.MagicChunksSize, n) }, openErr: errors.New("invalid magic number 0"), }, { name: "invalid chunk format version", corrFunc: func(f *os.File) { chunksFormatVersionOffset := int64(4) _, err := f.Seek(chunksFormatVersionOffset, 0) require.NoError(t, err) // Set invalid chunk format version. b := make([]byte, chunks.ChunksFormatVersionSize) b[0] = 0 n, err := f.Write(b) require.NoError(t, err) require.Equal(t, chunks.ChunksFormatVersionSize, n) }, openErr: errors.New("invalid chunk format version 0"), }, { name: "chunk not enough bytes to read the chunk length", corrFunc: func(f *os.File) { // Truncate one byte after the segment header. require.NoError(t, f.Truncate(chunks.SegmentHeaderSize+1)) }, iterErr: errors.New("cannot populate chunk 8 from block 00000000000000000000000000: segment doesn't include enough bytes to read the chunk size data field - required:13, available:9"), }, { name: "chunk not enough bytes to read the data", corrFunc: func(f *os.File) { fi, err := f.Stat() require.NoError(t, err) require.NoError(t, f.Truncate(fi.Size()-1)) }, iterErr: errors.New("cannot populate chunk 8 from block 00000000000000000000000000: segment doesn't include enough bytes to read the chunk - required:26, available:25"), }, { name: "checksum mismatch", corrFunc: func(f *os.File) { fi, err := f.Stat() require.NoError(t, err) // Get the chunk data end offset. chkEndOffset := int(fi.Size()) - crc32.Size // Seek to the last byte of chunk data and modify it. _, err = f.Seek(int64(chkEndOffset-1), 0) require.NoError(t, err) n, err := f.Write([]byte("x")) require.NoError(t, err) require.Equal(t, n, 1) }, iterErr: errors.New("cannot populate chunk 8 from block 00000000000000000000000000: checksum mismatch expected:cfc0526c, actual:34815eae"), }, } { t.Run(tc.name, func(t *testing.T) { tmpdir := t.TempDir() series := storage.NewListSeries(labels.FromStrings("a", "b"), []tsdbutil.Sample{sample{1, 1, nil, nil}}) blockDir := createBlock(t, tmpdir, []storage.Series{series}) files, err := sequenceFiles(chunkDir(blockDir)) require.NoError(t, err) require.Greater(t, len(files), 0, "No chunk created.") f, err := os.OpenFile(files[0], os.O_RDWR, 0o666) require.NoError(t, err) // Apply corruption function. tc.corrFunc(f) require.NoError(t, f.Close()) // Check open err. b, err := OpenBlock(nil, blockDir, nil) if tc.openErr != nil { require.Equal(t, tc.openErr.Error(), err.Error()) return } defer func() { require.NoError(t, b.Close()) }() querier, err := NewBlockQuerier(b, 0, 1) require.NoError(t, err) defer func() { require.NoError(t, querier.Close()) }() set := querier.Select(false, nil, labels.MustNewMatcher(labels.MatchEqual, "a", "b")) // Check chunk errors during iter time. require.True(t, set.Next()) it := set.At().Iterator(nil) require.Equal(t, chunkenc.ValNone, it.Next()) require.Equal(t, tc.iterErr.Error(), it.Err().Error()) }) } } func TestLabelValuesWithMatchers(t *testing.T) { tmpdir := t.TempDir() var seriesEntries []storage.Series for i := 0; i < 100; i++ { seriesEntries = append(seriesEntries, storage.NewListSeries(labels.FromStrings( "tens", fmt.Sprintf("value%d", i/10), "unique", fmt.Sprintf("value%d", i), ), []tsdbutil.Sample{sample{100, 0, nil, nil}})) } blockDir := createBlock(t, tmpdir, seriesEntries) files, err := sequenceFiles(chunkDir(blockDir)) require.NoError(t, err) require.Greater(t, len(files), 0, "No chunk created.") // Check open err. block, err := OpenBlock(nil, blockDir, nil) require.NoError(t, err) defer func() { require.NoError(t, block.Close()) }() indexReader, err := block.Index() require.NoError(t, err) defer func() { require.NoError(t, indexReader.Close()) }() testCases := []struct { name string labelName string matchers []*labels.Matcher expectedValues []string }{ { name: "get tens based on unique id", labelName: "tens", matchers: []*labels.Matcher{labels.MustNewMatcher(labels.MatchEqual, "unique", "value35")}, expectedValues: []string{"value3"}, }, { name: "get unique ids based on a ten", labelName: "unique", matchers: []*labels.Matcher{labels.MustNewMatcher(labels.MatchEqual, "tens", "value1")}, expectedValues: []string{"value10", "value11", "value12", "value13", "value14", "value15", "value16", "value17", "value18", "value19"}, }, { name: "get tens by pattern matching on unique id", labelName: "tens", matchers: []*labels.Matcher{labels.MustNewMatcher(labels.MatchRegexp, "unique", "value[5-7]5")}, expectedValues: []string{"value5", "value6", "value7"}, }, { name: "get tens by matching for absence of unique label", labelName: "tens", matchers: []*labels.Matcher{labels.MustNewMatcher(labels.MatchNotEqual, "unique", "")}, expectedValues: []string{"value0", "value1", "value2", "value3", "value4", "value5", "value6", "value7", "value8", "value9"}, }, } for _, tt := range testCases { t.Run(tt.name, func(t *testing.T) { actualValues, err := indexReader.SortedLabelValues(tt.labelName, tt.matchers...) require.NoError(t, err) require.Equal(t, tt.expectedValues, actualValues) actualValues, err = indexReader.LabelValues(tt.labelName, tt.matchers...) sort.Strings(actualValues) require.NoError(t, err) require.Equal(t, tt.expectedValues, actualValues) }) } } // TestBlockSize ensures that the block size is calculated correctly. func TestBlockSize(t *testing.T) { tmpdir := t.TempDir() var ( blockInit *Block expSizeInit int64 blockDirInit string err error ) // Create a block and compare the reported size vs actual disk size. { blockDirInit = createBlock(t, tmpdir, genSeries(10, 1, 1, 100)) blockInit, err = OpenBlock(nil, blockDirInit, nil) require.NoError(t, err) defer func() { require.NoError(t, blockInit.Close()) }() expSizeInit = blockInit.Size() actSizeInit, err := fileutil.DirSize(blockInit.Dir()) require.NoError(t, err) require.Equal(t, expSizeInit, actSizeInit) } // Delete some series and check the sizes again. { require.NoError(t, blockInit.Delete(1, 10, labels.MustNewMatcher(labels.MatchRegexp, "", ".*"))) expAfterDelete := blockInit.Size() require.Greater(t, expAfterDelete, expSizeInit, "after a delete the block size should be bigger as the tombstone file should grow %v > %v", expAfterDelete, expSizeInit) actAfterDelete, err := fileutil.DirSize(blockDirInit) require.NoError(t, err) require.Equal(t, expAfterDelete, actAfterDelete, "after a delete reported block size doesn't match actual disk size") c, err := NewLeveledCompactor(context.Background(), nil, log.NewNopLogger(), []int64{0}, nil, nil, true) require.NoError(t, err) blockDirAfterCompact, err := c.Compact(tmpdir, []string{blockInit.Dir()}, nil) require.NoError(t, err) blockAfterCompact, err := OpenBlock(nil, filepath.Join(tmpdir, blockDirAfterCompact.String()), nil) require.NoError(t, err) defer func() { require.NoError(t, blockAfterCompact.Close()) }() expAfterCompact := blockAfterCompact.Size() actAfterCompact, err := fileutil.DirSize(blockAfterCompact.Dir()) require.NoError(t, err) require.Greater(t, actAfterDelete, actAfterCompact, "after a delete and compaction the block size should be smaller %v,%v", actAfterDelete, actAfterCompact) require.Equal(t, expAfterCompact, actAfterCompact, "after a delete and compaction reported block size doesn't match actual disk size") } } func TestReadIndexFormatV1(t *testing.T) { /* The block here was produced at the commit 706602daed1487f7849990678b4ece4599745905 used in 2.0.0 with: db, _ := Open("v1db", nil, nil, nil) app := db.Appender() app.Add(labels.FromStrings("foo", "bar"), 1, 2) app.Add(labels.FromStrings("foo", "baz"), 3, 4) app.Add(labels.FromStrings("foo", "meh"), 1000*3600*4, 4) // Not in the block. // Make sure we've enough values for the lack of sorting of postings offsets to show up. for i := 0; i < 100; i++ { app.Add(labels.FromStrings("bar", strconv.FormatInt(int64(i), 10)), 0, 0) } app.Commit() db.compact() db.Close() */ blockDir := filepath.Join("testdata", "index_format_v1") block, err := OpenBlock(nil, blockDir, nil) require.NoError(t, err) t.Cleanup(func() { require.NoError(t, block.Close()) }) q, err := NewBlockQuerier(block, 0, 1000) require.NoError(t, err) require.Equal(t, query(t, q, labels.MustNewMatcher(labels.MatchEqual, "foo", "bar")), map[string][]tsdbutil.Sample{`{foo="bar"}`: {sample{t: 1, f: 2}}}) q, err = NewBlockQuerier(block, 0, 1000) require.NoError(t, err) require.Equal(t, query(t, q, labels.MustNewMatcher(labels.MatchNotRegexp, "foo", "^.?$")), map[string][]tsdbutil.Sample{ `{foo="bar"}`: {sample{t: 1, f: 2}}, `{foo="baz"}`: {sample{t: 3, f: 4}}, }) } func BenchmarkLabelValuesWithMatchers(b *testing.B) { tmpdir := b.TempDir() var seriesEntries []storage.Series metricCount := 1000000 for i := 0; i < metricCount; i++ { // Note these series are not created in sort order: 'value2' sorts after 'value10'. // This makes a big difference to the benchmark timing. seriesEntries = append(seriesEntries, storage.NewListSeries(labels.FromStrings( "a_unique", fmt.Sprintf("value%d", i), "b_tens", fmt.Sprintf("value%d", i/(metricCount/10)), "c_ninety", fmt.Sprintf("value%d", i/(metricCount/10)/9), // "0" for the first 90%, then "1" ), []tsdbutil.Sample{sample{100, 0, nil, nil}})) } blockDir := createBlock(b, tmpdir, seriesEntries) files, err := sequenceFiles(chunkDir(blockDir)) require.NoError(b, err) require.Greater(b, len(files), 0, "No chunk created.") // Check open err. block, err := OpenBlock(nil, blockDir, nil) require.NoError(b, err) defer func() { require.NoError(b, block.Close()) }() indexReader, err := block.Index() require.NoError(b, err) defer func() { require.NoError(b, indexReader.Close()) }() matchers := []*labels.Matcher{labels.MustNewMatcher(labels.MatchEqual, "c_ninety", "value0")} b.ResetTimer() b.ReportAllocs() for benchIdx := 0; benchIdx < b.N; benchIdx++ { actualValues, err := indexReader.LabelValues("b_tens", matchers...) require.NoError(b, err) require.Equal(b, 9, len(actualValues)) } } func TestLabelNamesWithMatchers(t *testing.T) { tmpdir := t.TempDir() var seriesEntries []storage.Series for i := 0; i < 100; i++ { seriesEntries = append(seriesEntries, storage.NewListSeries(labels.FromStrings( "unique", fmt.Sprintf("value%d", i), ), []tsdbutil.Sample{sample{100, 0, nil, nil}})) if i%10 == 0 { seriesEntries = append(seriesEntries, storage.NewListSeries(labels.FromStrings( "tens", fmt.Sprintf("value%d", i/10), "unique", fmt.Sprintf("value%d", i), ), []tsdbutil.Sample{sample{100, 0, nil, nil}})) } if i%20 == 0 { seriesEntries = append(seriesEntries, storage.NewListSeries(labels.FromStrings( "tens", fmt.Sprintf("value%d", i/10), "twenties", fmt.Sprintf("value%d", i/20), "unique", fmt.Sprintf("value%d", i), ), []tsdbutil.Sample{sample{100, 0, nil, nil}})) } } blockDir := createBlock(t, tmpdir, seriesEntries) files, err := sequenceFiles(chunkDir(blockDir)) require.NoError(t, err) require.Greater(t, len(files), 0, "No chunk created.") // Check open err. block, err := OpenBlock(nil, blockDir, nil) require.NoError(t, err) t.Cleanup(func() { require.NoError(t, block.Close()) }) indexReader, err := block.Index() require.NoError(t, err) t.Cleanup(func() { require.NoError(t, indexReader.Close()) }) testCases := []struct { name string labelName string matchers []*labels.Matcher expectedNames []string }{ { name: "get with non-empty unique: all", matchers: []*labels.Matcher{labels.MustNewMatcher(labels.MatchNotEqual, "unique", "")}, expectedNames: []string{"tens", "twenties", "unique"}, }, { name: "get with unique ending in 1: only unique", matchers: []*labels.Matcher{labels.MustNewMatcher(labels.MatchRegexp, "unique", "value.*1")}, expectedNames: []string{"unique"}, }, { name: "get with unique = value20: all", matchers: []*labels.Matcher{labels.MustNewMatcher(labels.MatchEqual, "unique", "value20")}, expectedNames: []string{"tens", "twenties", "unique"}, }, { name: "get tens = 1: unique & tens", matchers: []*labels.Matcher{labels.MustNewMatcher(labels.MatchEqual, "tens", "value1")}, expectedNames: []string{"tens", "unique"}, }, } for _, tt := range testCases { t.Run(tt.name, func(t *testing.T) { actualNames, err := indexReader.LabelNames(tt.matchers...) require.NoError(t, err) require.Equal(t, tt.expectedNames, actualNames) }) } } // createBlock creates a block with given set of series and returns its dir. func createBlock(tb testing.TB, dir string, series []storage.Series) string { blockDir, err := CreateBlock(series, dir, 0, log.NewNopLogger()) require.NoError(tb, err) return blockDir } func createBlockFromHead(tb testing.TB, dir string, head *Head) string { compactor, err := NewLeveledCompactor(context.Background(), nil, log.NewNopLogger(), []int64{1000000}, nil, nil, true) require.NoError(tb, err) require.NoError(tb, os.MkdirAll(dir, 0o777)) // Add +1 millisecond to block maxt because block intervals are half-open: [b.MinTime, b.MaxTime). // Because of this block intervals are always +1 than the total samples it includes. ulid, err := compactor.Write(dir, head, head.MinTime(), head.MaxTime()+1, nil) require.NoError(tb, err) return filepath.Join(dir, ulid.String()) } func createHead(tb testing.TB, w *wlog.WL, series []storage.Series, chunkDir string) *Head { opts := DefaultHeadOptions() opts.ChunkDirRoot = chunkDir head, err := NewHead(nil, nil, w, nil, opts, nil) require.NoError(tb, err) var it chunkenc.Iterator ctx := context.Background() app := head.Appender(ctx) for _, s := range series { ref := storage.SeriesRef(0) it = s.Iterator(it) lset := s.Labels() typ := it.Next() lastTyp := typ for ; typ != chunkenc.ValNone; typ = it.Next() { if lastTyp != typ { // The behaviour of appender is undefined if samples of different types // are appended to the same series in a single Commit(). require.NoError(tb, app.Commit()) app = head.Appender(ctx) } switch typ { case chunkenc.ValFloat: t, v := it.At() ref, err = app.Append(ref, lset, t, v) case chunkenc.ValHistogram: t, h := it.AtHistogram() ref, err = app.AppendHistogram(ref, lset, t, h, nil) case chunkenc.ValFloatHistogram: t, fh := it.AtFloatHistogram() ref, err = app.AppendHistogram(ref, lset, t, nil, fh) default: err = fmt.Errorf("unknown sample type %s", typ.String()) } require.NoError(tb, err) lastTyp = typ } require.NoError(tb, it.Err()) } require.NoError(tb, app.Commit()) return head } func createHeadWithOOOSamples(tb testing.TB, w *wlog.WL, series []storage.Series, chunkDir string, oooSampleFrequency int) *Head { opts := DefaultHeadOptions() opts.ChunkDirRoot = chunkDir opts.OutOfOrderTimeWindow.Store(10000000000) head, err := NewHead(nil, nil, w, nil, opts, nil) require.NoError(tb, err) oooSampleLabels := make([]labels.Labels, 0, len(series)) oooSamples := make([]tsdbutil.SampleSlice, 0, len(series)) var it chunkenc.Iterator totalSamples := 0 app := head.Appender(context.Background()) for _, s := range series { ref := storage.SeriesRef(0) it = s.Iterator(it) lset := s.Labels() os := tsdbutil.SampleSlice{} count := 0 for it.Next() == chunkenc.ValFloat { totalSamples++ count++ t, v := it.At() if count%oooSampleFrequency == 0 { os = append(os, sample{t: t, f: v}) continue } ref, err = app.Append(ref, lset, t, v) require.NoError(tb, err) } require.NoError(tb, it.Err()) if len(os) > 0 { oooSampleLabels = append(oooSampleLabels, lset) oooSamples = append(oooSamples, os) } } require.NoError(tb, app.Commit()) oooSamplesAppended := 0 require.Equal(tb, float64(0), prom_testutil.ToFloat64(head.metrics.outOfOrderSamplesAppended)) app = head.Appender(context.Background()) for i, lset := range oooSampleLabels { ref := storage.SeriesRef(0) for _, sample := range oooSamples[i] { ref, err = app.Append(ref, lset, sample.T(), sample.F()) require.NoError(tb, err) oooSamplesAppended++ } } require.NoError(tb, app.Commit()) actOOOAppended := prom_testutil.ToFloat64(head.metrics.outOfOrderSamplesAppended) require.GreaterOrEqual(tb, actOOOAppended, float64(oooSamplesAppended-len(series))) require.LessOrEqual(tb, actOOOAppended, float64(oooSamplesAppended)) require.Equal(tb, float64(totalSamples), prom_testutil.ToFloat64(head.metrics.samplesAppended)) return head } const ( defaultLabelName = "labelName" defaultLabelValue = "labelValue" ) // genSeries generates series of float64 samples with a given number of labels and values. func genSeries(totalSeries, labelCount int, mint, maxt int64) []storage.Series { return genSeriesFromSampleGenerator(totalSeries, labelCount, mint, maxt, 1, func(ts int64) tsdbutil.Sample { return sample{t: ts, f: rand.Float64()} }) } // genHistogramSeries generates series of histogram samples with a given number of labels and values. func genHistogramSeries(totalSeries, labelCount int, mint, maxt, step int64, floatHistogram bool) []storage.Series { return genSeriesFromSampleGenerator(totalSeries, labelCount, mint, maxt, step, func(ts int64) tsdbutil.Sample { h := &histogram.Histogram{ Count: 5 + uint64(ts*4), ZeroCount: 2 + uint64(ts), ZeroThreshold: 0.001, Sum: 18.4 * rand.Float64(), Schema: 1, PositiveSpans: []histogram.Span{ {Offset: 0, Length: 2}, {Offset: 1, Length: 2}, }, PositiveBuckets: []int64{ts + 1, 1, -1, 0}, } if ts != mint { // By setting the counter reset hint to "no counter // reset" for all histograms but the first, we cover the // most common cases. If the series is manipulated later // or spans more than one block when ingested into the // storage, the hint has to be adjusted. Note that the // storage itself treats this particular hint the same // as "unknown". h.CounterResetHint = histogram.NotCounterReset } if floatHistogram { return sample{t: ts, fh: h.ToFloat()} } return sample{t: ts, h: h} }) } // genHistogramAndFloatSeries generates series of mixed histogram and float64 samples with a given number of labels and values. func genHistogramAndFloatSeries(totalSeries, labelCount int, mint, maxt, step int64, floatHistogram bool) []storage.Series { floatSample := false count := 0 return genSeriesFromSampleGenerator(totalSeries, labelCount, mint, maxt, step, func(ts int64) tsdbutil.Sample { count++ var s sample if floatSample { s = sample{t: ts, f: rand.Float64()} } else { h := &histogram.Histogram{ Count: 5 + uint64(ts*4), ZeroCount: 2 + uint64(ts), ZeroThreshold: 0.001, Sum: 18.4 * rand.Float64(), Schema: 1, PositiveSpans: []histogram.Span{ {Offset: 0, Length: 2}, {Offset: 1, Length: 2}, }, PositiveBuckets: []int64{ts + 1, 1, -1, 0}, } if count > 1 && count%5 != 1 { // Same rationale for this as above in // genHistogramSeries, just that we have to be // smarter to find out if the previous sample // was a histogram, too. h.CounterResetHint = histogram.NotCounterReset } if floatHistogram { s = sample{t: ts, fh: h.ToFloat()} } else { s = sample{t: ts, h: h} } } if count%5 == 0 { // Flip the sample type for every 5 samples. floatSample = !floatSample } return s }) } func genSeriesFromSampleGenerator(totalSeries, labelCount int, mint, maxt, step int64, generator func(ts int64) tsdbutil.Sample) []storage.Series { if totalSeries == 0 || labelCount == 0 { return nil } series := make([]storage.Series, totalSeries) for i := 0; i < totalSeries; i++ { lbls := make(map[string]string, labelCount) lbls[defaultLabelName] = strconv.Itoa(i) for j := 1; len(lbls) < labelCount; j++ { lbls[defaultLabelName+strconv.Itoa(j)] = defaultLabelValue + strconv.Itoa(j) } samples := make([]tsdbutil.Sample, 0, (maxt-mint)/step+1) for t := mint; t < maxt; t += step { samples = append(samples, generator(t)) } series[i] = storage.NewListSeries(labels.FromMap(lbls), samples) } return series } // populateSeries generates series from given labels, mint and maxt. func populateSeries(lbls []map[string]string, mint, maxt int64) []storage.Series { if len(lbls) == 0 { return nil } series := make([]storage.Series, 0, len(lbls)) for _, lbl := range lbls { if len(lbl) == 0 { continue } samples := make([]tsdbutil.Sample, 0, maxt-mint+1) for t := mint; t <= maxt; t++ { samples = append(samples, sample{t: t, f: rand.Float64()}) } series = append(series, storage.NewListSeries(labels.FromMap(lbl), samples)) } return series }