mirror of
https://github.com/prometheus/prometheus.git
synced 2025-02-02 08:31:11 -08:00
Fix tests
This commit is contained in:
parent
51cf003517
commit
0cd4063a02
|
@ -15,9 +15,10 @@ package tsdb
|
||||||
|
|
||||||
import (
|
import (
|
||||||
"context"
|
"context"
|
||||||
"crypto/rand"
|
crand "crypto/rand"
|
||||||
"fmt"
|
"fmt"
|
||||||
"math"
|
"math"
|
||||||
|
"math/rand"
|
||||||
"os"
|
"os"
|
||||||
"path"
|
"path"
|
||||||
"path/filepath"
|
"path/filepath"
|
||||||
|
@ -443,9 +444,9 @@ func TestCompactionFailWillCleanUpTempDir(t *testing.T) {
|
||||||
tmpdir := t.TempDir()
|
tmpdir := t.TempDir()
|
||||||
|
|
||||||
shardedBlocks := []shardedBlock{
|
shardedBlocks := []shardedBlock{
|
||||||
{meta: &BlockMeta{ULID: ulid.MustNew(ulid.Now(), rand.Reader)}},
|
{meta: &BlockMeta{ULID: ulid.MustNew(ulid.Now(), crand.Reader)}},
|
||||||
{meta: &BlockMeta{ULID: ulid.MustNew(ulid.Now(), rand.Reader)}},
|
{meta: &BlockMeta{ULID: ulid.MustNew(ulid.Now(), crand.Reader)}},
|
||||||
{meta: &BlockMeta{ULID: ulid.MustNew(ulid.Now(), rand.Reader)}},
|
{meta: &BlockMeta{ULID: ulid.MustNew(ulid.Now(), crand.Reader)}},
|
||||||
}
|
}
|
||||||
|
|
||||||
require.Error(t, compactor.write(tmpdir, shardedBlocks, erringBReader{}))
|
require.Error(t, compactor.write(tmpdir, shardedBlocks, erringBReader{}))
|
||||||
|
@ -639,7 +640,7 @@ func TestCompaction_CompactEmptyBlocks(t *testing.T) {
|
||||||
|
|
||||||
for _, r := range ranges {
|
for _, r := range ranges {
|
||||||
// Generate blocks using index and chunk writer. CreateBlock would not return valid block for 0 series.
|
// Generate blocks using index and chunk writer. CreateBlock would not return valid block for 0 series.
|
||||||
id := ulid.MustNew(ulid.Now(), rand.Reader)
|
id := ulid.MustNew(ulid.Now(), crand.Reader)
|
||||||
m := &BlockMeta{
|
m := &BlockMeta{
|
||||||
ULID: id,
|
ULID: id,
|
||||||
MinTime: r[0],
|
MinTime: r[0],
|
||||||
|
@ -1689,7 +1690,7 @@ func TestHeadCompactionWithHistograms(t *testing.T) {
|
||||||
// Compaction.
|
// Compaction.
|
||||||
mint := head.MinTime()
|
mint := head.MinTime()
|
||||||
maxt := head.MaxTime() + 1 // Block intervals are half-open: [b.MinTime, b.MaxTime).
|
maxt := head.MaxTime() + 1 // Block intervals are half-open: [b.MinTime, b.MaxTime).
|
||||||
compactor, err := NewLeveledCompactor(context.Background(), nil, nil, []int64{DefaultBlockDuration}, chunkenc.NewPool(), nil)
|
compactor, err := NewLeveledCompactor(context.Background(), nil, nil, []int64{DefaultBlockDuration}, chunkenc.NewPool(), nil, true)
|
||||||
require.NoError(t, err)
|
require.NoError(t, err)
|
||||||
id, err := compactor.Write(head.opts.ChunkDirRoot, head, mint, maxt, nil)
|
id, err := compactor.Write(head.opts.ChunkDirRoot, head, mint, maxt, nil)
|
||||||
require.NoError(t, err)
|
require.NoError(t, err)
|
||||||
|
@ -1835,7 +1836,7 @@ func TestSparseHistogramSpaceSavings(t *testing.T) {
|
||||||
// Sparse head compaction.
|
// Sparse head compaction.
|
||||||
mint := sparseHead.MinTime()
|
mint := sparseHead.MinTime()
|
||||||
maxt := sparseHead.MaxTime() + 1 // Block intervals are half-open: [b.MinTime, b.MaxTime).
|
maxt := sparseHead.MaxTime() + 1 // Block intervals are half-open: [b.MinTime, b.MaxTime).
|
||||||
compactor, err := NewLeveledCompactor(context.Background(), nil, nil, []int64{DefaultBlockDuration}, chunkenc.NewPool(), nil)
|
compactor, err := NewLeveledCompactor(context.Background(), nil, nil, []int64{DefaultBlockDuration}, chunkenc.NewPool(), nil, true)
|
||||||
require.NoError(t, err)
|
require.NoError(t, err)
|
||||||
sparseULID, err = compactor.Write(sparseHead.opts.ChunkDirRoot, sparseHead, mint, maxt, nil)
|
sparseULID, err = compactor.Write(sparseHead.opts.ChunkDirRoot, sparseHead, mint, maxt, nil)
|
||||||
require.NoError(t, err)
|
require.NoError(t, err)
|
||||||
|
@ -1887,7 +1888,7 @@ func TestSparseHistogramSpaceSavings(t *testing.T) {
|
||||||
// Old head compaction.
|
// Old head compaction.
|
||||||
mint := oldHead.MinTime()
|
mint := oldHead.MinTime()
|
||||||
maxt := oldHead.MaxTime() + 1 // Block intervals are half-open: [b.MinTime, b.MaxTime).
|
maxt := oldHead.MaxTime() + 1 // Block intervals are half-open: [b.MinTime, b.MaxTime).
|
||||||
compactor, err := NewLeveledCompactor(context.Background(), nil, nil, []int64{DefaultBlockDuration}, chunkenc.NewPool(), nil)
|
compactor, err := NewLeveledCompactor(context.Background(), nil, nil, []int64{DefaultBlockDuration}, chunkenc.NewPool(), nil, true)
|
||||||
require.NoError(t, err)
|
require.NoError(t, err)
|
||||||
oldULID, err = compactor.Write(oldHead.opts.ChunkDirRoot, oldHead, mint, maxt, nil)
|
oldULID, err = compactor.Write(oldHead.opts.ChunkDirRoot, oldHead, mint, maxt, nil)
|
||||||
require.NoError(t, err)
|
require.NoError(t, err)
|
||||||
|
|
|
@ -1329,7 +1329,8 @@ func TestMemSeries_appendHistogram(t *testing.T) {
|
||||||
}()
|
}()
|
||||||
chunkRange := int64(1000)
|
chunkRange := int64(1000)
|
||||||
|
|
||||||
s := newMemSeries(labels.Labels{}, 1, defaultIsolationDisabled)
|
lbls := labels.Labels{}
|
||||||
|
s := newMemSeries(lbls, 1, lbls.Hash(), 0, defaultIsolationDisabled)
|
||||||
|
|
||||||
histograms := GenerateTestHistograms(4)
|
histograms := GenerateTestHistograms(4)
|
||||||
histogramWithOneMoreBucket := histograms[3].Copy()
|
histogramWithOneMoreBucket := histograms[3].Copy()
|
||||||
|
|
Loading…
Reference in a new issue