mirror of
https://github.com/prometheus/prometheus.git
synced 2025-03-05 20:59:13 -08:00
Merge pull request #13089 from prometheus/beorn7/release
Merge release-2.48 back into main
This commit is contained in:
commit
79f4e45d64
|
@ -1,5 +1,11 @@
|
||||||
# Changelog
|
# Changelog
|
||||||
|
|
||||||
|
## 2.48.0-rc.2 / 2023-11-02
|
||||||
|
|
||||||
|
* [ENHANCEMENT] Scraping: Add configuration option for tracking staleness of scraped timestamps. #13060
|
||||||
|
* [BUGFIX] Storage: Fix crash caused by incorrect mixed samples handling. #13055
|
||||||
|
* [BUGFIX] TSDB: Fix compactor failures by adding min time to histogram chunks. #13062
|
||||||
|
|
||||||
## 2.48.0-rc.1 / 2023-10-24
|
## 2.48.0-rc.1 / 2023-10-24
|
||||||
|
|
||||||
* [BUGFIX] PromQL: Reduce inefficiency introduced by warnings/annotations and temporarily remove possible non-counter warnings. #13012
|
* [BUGFIX] PromQL: Reduce inefficiency introduced by warnings/annotations and temporarily remove possible non-counter warnings. #13012
|
||||||
|
|
|
@ -38,6 +38,7 @@ import (
|
||||||
"github.com/prometheus/prometheus/tsdb/chunks"
|
"github.com/prometheus/prometheus/tsdb/chunks"
|
||||||
"github.com/prometheus/prometheus/tsdb/fileutil"
|
"github.com/prometheus/prometheus/tsdb/fileutil"
|
||||||
"github.com/prometheus/prometheus/tsdb/tombstones"
|
"github.com/prometheus/prometheus/tsdb/tombstones"
|
||||||
|
"github.com/prometheus/prometheus/tsdb/tsdbutil"
|
||||||
"github.com/prometheus/prometheus/tsdb/wlog"
|
"github.com/prometheus/prometheus/tsdb/wlog"
|
||||||
)
|
)
|
||||||
|
|
||||||
|
@ -933,6 +934,32 @@ func TestCompaction_populateBlock(t *testing.T) {
|
||||||
},
|
},
|
||||||
},
|
},
|
||||||
},
|
},
|
||||||
|
{
|
||||||
|
// Regression test for populateWithDelChunkSeriesIterator failing to set minTime on chunks.
|
||||||
|
title: "Populate from mixed type series and expect sample inside the interval only.",
|
||||||
|
compactMinTime: 1,
|
||||||
|
compactMaxTime: 11,
|
||||||
|
inputSeriesSamples: [][]seriesSamples{
|
||||||
|
{
|
||||||
|
{
|
||||||
|
lset: map[string]string{"a": "1"},
|
||||||
|
chunks: [][]sample{
|
||||||
|
{{t: 0, h: tsdbutil.GenerateTestHistogram(0)}, {t: 1, h: tsdbutil.GenerateTestHistogram(1)}},
|
||||||
|
{{t: 10, f: 1}, {t: 11, f: 2}},
|
||||||
|
},
|
||||||
|
},
|
||||||
|
},
|
||||||
|
},
|
||||||
|
expSeriesSamples: []seriesSamples{
|
||||||
|
{
|
||||||
|
lset: map[string]string{"a": "1"},
|
||||||
|
chunks: [][]sample{
|
||||||
|
{{t: 1, h: tsdbutil.GenerateTestHistogram(1)}},
|
||||||
|
{{t: 10, f: 1}},
|
||||||
|
},
|
||||||
|
},
|
||||||
|
},
|
||||||
|
},
|
||||||
} {
|
} {
|
||||||
t.Run(tc.title, func(t *testing.T) {
|
t.Run(tc.title, func(t *testing.T) {
|
||||||
blocks := make([]BlockReader, 0, len(tc.inputSeriesSamples))
|
blocks := make([]BlockReader, 0, len(tc.inputSeriesSamples))
|
||||||
|
@ -974,12 +1001,23 @@ func TestCompaction_populateBlock(t *testing.T) {
|
||||||
firstTs int64 = math.MaxInt64
|
firstTs int64 = math.MaxInt64
|
||||||
s sample
|
s sample
|
||||||
)
|
)
|
||||||
for iter.Next() == chunkenc.ValFloat {
|
for vt := iter.Next(); vt != chunkenc.ValNone; vt = iter.Next() {
|
||||||
|
switch vt {
|
||||||
|
case chunkenc.ValFloat:
|
||||||
s.t, s.f = iter.At()
|
s.t, s.f = iter.At()
|
||||||
|
samples = append(samples, s)
|
||||||
|
case chunkenc.ValHistogram:
|
||||||
|
s.t, s.h = iter.AtHistogram()
|
||||||
|
samples = append(samples, s)
|
||||||
|
case chunkenc.ValFloatHistogram:
|
||||||
|
s.t, s.fh = iter.AtFloatHistogram()
|
||||||
|
samples = append(samples, s)
|
||||||
|
default:
|
||||||
|
require.Fail(t, "unexpected value type")
|
||||||
|
}
|
||||||
if firstTs == math.MaxInt64 {
|
if firstTs == math.MaxInt64 {
|
||||||
firstTs = s.t
|
firstTs = s.t
|
||||||
}
|
}
|
||||||
samples = append(samples, s)
|
|
||||||
}
|
}
|
||||||
|
|
||||||
// Check if chunk has correct min, max times.
|
// Check if chunk has correct min, max times.
|
||||||
|
|
|
@ -840,6 +840,7 @@ func (p *populateWithDelChunkSeriesIterator) Next() bool {
|
||||||
}
|
}
|
||||||
return false
|
return false
|
||||||
}
|
}
|
||||||
|
p.curr.MinTime = p.currDelIter.AtT()
|
||||||
|
|
||||||
// Re-encode the chunk if iterator is provider. This means that it has
|
// Re-encode the chunk if iterator is provider. This means that it has
|
||||||
// some samples to be deleted or chunk is opened.
|
// some samples to be deleted or chunk is opened.
|
||||||
|
@ -855,7 +856,6 @@ func (p *populateWithDelChunkSeriesIterator) Next() bool {
|
||||||
if app, err = newChunk.Appender(); err != nil {
|
if app, err = newChunk.Appender(); err != nil {
|
||||||
break
|
break
|
||||||
}
|
}
|
||||||
|
|
||||||
for vt := valueType; vt != chunkenc.ValNone; vt = p.currDelIter.Next() {
|
for vt := valueType; vt != chunkenc.ValNone; vt = p.currDelIter.Next() {
|
||||||
if vt != chunkenc.ValHistogram {
|
if vt != chunkenc.ValHistogram {
|
||||||
err = fmt.Errorf("found value type %v in histogram chunk", vt)
|
err = fmt.Errorf("found value type %v in histogram chunk", vt)
|
||||||
|
@ -873,15 +873,12 @@ func (p *populateWithDelChunkSeriesIterator) Next() bool {
|
||||||
if app, err = newChunk.Appender(); err != nil {
|
if app, err = newChunk.Appender(); err != nil {
|
||||||
break
|
break
|
||||||
}
|
}
|
||||||
var v float64
|
for vt := valueType; vt != chunkenc.ValNone; vt = p.currDelIter.Next() {
|
||||||
t, v = p.currDelIter.At()
|
|
||||||
p.curr.MinTime = t
|
|
||||||
app.Append(t, v)
|
|
||||||
for vt := p.currDelIter.Next(); vt != chunkenc.ValNone; vt = p.currDelIter.Next() {
|
|
||||||
if vt != chunkenc.ValFloat {
|
if vt != chunkenc.ValFloat {
|
||||||
err = fmt.Errorf("found value type %v in float chunk", vt)
|
err = fmt.Errorf("found value type %v in float chunk", vt)
|
||||||
break
|
break
|
||||||
}
|
}
|
||||||
|
var v float64
|
||||||
t, v = p.currDelIter.At()
|
t, v = p.currDelIter.At()
|
||||||
app.Append(t, v)
|
app.Append(t, v)
|
||||||
}
|
}
|
||||||
|
@ -890,7 +887,6 @@ func (p *populateWithDelChunkSeriesIterator) Next() bool {
|
||||||
if app, err = newChunk.Appender(); err != nil {
|
if app, err = newChunk.Appender(); err != nil {
|
||||||
break
|
break
|
||||||
}
|
}
|
||||||
|
|
||||||
for vt := valueType; vt != chunkenc.ValNone; vt = p.currDelIter.Next() {
|
for vt := valueType; vt != chunkenc.ValNone; vt = p.currDelIter.Next() {
|
||||||
if vt != chunkenc.ValFloatHistogram {
|
if vt != chunkenc.ValFloatHistogram {
|
||||||
err = fmt.Errorf("found value type %v in histogram chunk", vt)
|
err = fmt.Errorf("found value type %v in histogram chunk", vt)
|
||||||
|
|
|
@ -132,12 +132,35 @@ func createIdxChkReaders(t *testing.T, tc []seriesSamples) (IndexReader, ChunkRe
|
||||||
Ref: chunkRef,
|
Ref: chunkRef,
|
||||||
})
|
})
|
||||||
|
|
||||||
|
switch {
|
||||||
|
case chk[0].fh != nil:
|
||||||
|
chunk := chunkenc.NewFloatHistogramChunk()
|
||||||
|
app, _ := chunk.Appender()
|
||||||
|
for _, smpl := range chk {
|
||||||
|
require.NotNil(t, smpl.fh, "chunk can only contain one type of sample")
|
||||||
|
_, _, _, err := app.AppendFloatHistogram(nil, smpl.t, smpl.fh, true)
|
||||||
|
require.NoError(t, err, "chunk should be appendable")
|
||||||
|
}
|
||||||
|
chkReader[chunkRef] = chunk
|
||||||
|
case chk[0].h != nil:
|
||||||
|
chunk := chunkenc.NewHistogramChunk()
|
||||||
|
app, _ := chunk.Appender()
|
||||||
|
for _, smpl := range chk {
|
||||||
|
require.NotNil(t, smpl.h, "chunk can only contain one type of sample")
|
||||||
|
_, _, _, err := app.AppendHistogram(nil, smpl.t, smpl.h, true)
|
||||||
|
require.NoError(t, err, "chunk should be appendable")
|
||||||
|
}
|
||||||
|
chkReader[chunkRef] = chunk
|
||||||
|
default:
|
||||||
chunk := chunkenc.NewXORChunk()
|
chunk := chunkenc.NewXORChunk()
|
||||||
app, _ := chunk.Appender()
|
app, _ := chunk.Appender()
|
||||||
for _, smpl := range chk {
|
for _, smpl := range chk {
|
||||||
|
require.Nil(t, smpl.h, "chunk can only contain one type of sample")
|
||||||
|
require.Nil(t, smpl.fh, "chunk can only contain one type of sample")
|
||||||
app.Append(smpl.t, smpl.f)
|
app.Append(smpl.t, smpl.f)
|
||||||
}
|
}
|
||||||
chkReader[chunkRef] = chunk
|
chkReader[chunkRef] = chunk
|
||||||
|
}
|
||||||
chunkRef++
|
chunkRef++
|
||||||
}
|
}
|
||||||
ls := labels.FromMap(s.lset)
|
ls := labels.FromMap(s.lset)
|
||||||
|
@ -693,12 +716,16 @@ func (r *fakeChunksReader) Chunk(meta chunks.Meta) (chunkenc.Chunk, error) {
|
||||||
}
|
}
|
||||||
|
|
||||||
func TestPopulateWithTombSeriesIterators(t *testing.T) {
|
func TestPopulateWithTombSeriesIterators(t *testing.T) {
|
||||||
|
type minMaxTimes struct {
|
||||||
|
minTime, maxTime int64
|
||||||
|
}
|
||||||
cases := []struct {
|
cases := []struct {
|
||||||
name string
|
name string
|
||||||
chks [][]chunks.Sample
|
chks [][]chunks.Sample
|
||||||
|
|
||||||
expected []chunks.Sample
|
expected []chunks.Sample
|
||||||
expectedChks []chunks.Meta
|
expectedChks []chunks.Meta
|
||||||
|
expectedMinMaxTimes []minMaxTimes
|
||||||
|
|
||||||
intervals tombstones.Intervals
|
intervals tombstones.Intervals
|
||||||
|
|
||||||
|
@ -717,6 +744,7 @@ func TestPopulateWithTombSeriesIterators(t *testing.T) {
|
||||||
expectedChks: []chunks.Meta{
|
expectedChks: []chunks.Meta{
|
||||||
assureChunkFromSamples(t, []chunks.Sample{}),
|
assureChunkFromSamples(t, []chunks.Sample{}),
|
||||||
},
|
},
|
||||||
|
expectedMinMaxTimes: []minMaxTimes{{0, 0}},
|
||||||
},
|
},
|
||||||
{
|
{
|
||||||
name: "three empty chunks", // This should never happen.
|
name: "three empty chunks", // This should never happen.
|
||||||
|
@ -727,6 +755,7 @@ func TestPopulateWithTombSeriesIterators(t *testing.T) {
|
||||||
assureChunkFromSamples(t, []chunks.Sample{}),
|
assureChunkFromSamples(t, []chunks.Sample{}),
|
||||||
assureChunkFromSamples(t, []chunks.Sample{}),
|
assureChunkFromSamples(t, []chunks.Sample{}),
|
||||||
},
|
},
|
||||||
|
expectedMinMaxTimes: []minMaxTimes{{0, 0}, {0, 0}, {0, 0}},
|
||||||
},
|
},
|
||||||
{
|
{
|
||||||
name: "one chunk",
|
name: "one chunk",
|
||||||
|
@ -742,6 +771,7 @@ func TestPopulateWithTombSeriesIterators(t *testing.T) {
|
||||||
sample{1, 2, nil, nil}, sample{2, 3, nil, nil}, sample{3, 5, nil, nil}, sample{6, 1, nil, nil},
|
sample{1, 2, nil, nil}, sample{2, 3, nil, nil}, sample{3, 5, nil, nil}, sample{6, 1, nil, nil},
|
||||||
}),
|
}),
|
||||||
},
|
},
|
||||||
|
expectedMinMaxTimes: []minMaxTimes{{1, 6}},
|
||||||
},
|
},
|
||||||
{
|
{
|
||||||
name: "two full chunks",
|
name: "two full chunks",
|
||||||
|
@ -761,6 +791,7 @@ func TestPopulateWithTombSeriesIterators(t *testing.T) {
|
||||||
sample{7, 89, nil, nil}, sample{9, 8, nil, nil},
|
sample{7, 89, nil, nil}, sample{9, 8, nil, nil},
|
||||||
}),
|
}),
|
||||||
},
|
},
|
||||||
|
expectedMinMaxTimes: []minMaxTimes{{1, 6}, {7, 9}},
|
||||||
},
|
},
|
||||||
{
|
{
|
||||||
name: "three full chunks",
|
name: "three full chunks",
|
||||||
|
@ -784,6 +815,7 @@ func TestPopulateWithTombSeriesIterators(t *testing.T) {
|
||||||
sample{10, 22, nil, nil}, sample{203, 3493, nil, nil},
|
sample{10, 22, nil, nil}, sample{203, 3493, nil, nil},
|
||||||
}),
|
}),
|
||||||
},
|
},
|
||||||
|
expectedMinMaxTimes: []minMaxTimes{{1, 6}, {7, 9}, {10, 203}},
|
||||||
},
|
},
|
||||||
// Seek cases.
|
// Seek cases.
|
||||||
{
|
{
|
||||||
|
@ -854,6 +886,7 @@ func TestPopulateWithTombSeriesIterators(t *testing.T) {
|
||||||
sample{7, 89, nil, nil},
|
sample{7, 89, nil, nil},
|
||||||
}),
|
}),
|
||||||
},
|
},
|
||||||
|
expectedMinMaxTimes: []minMaxTimes{{3, 6}, {7, 7}},
|
||||||
},
|
},
|
||||||
{
|
{
|
||||||
name: "two chunks with trimmed middle sample of first chunk",
|
name: "two chunks with trimmed middle sample of first chunk",
|
||||||
|
@ -874,6 +907,7 @@ func TestPopulateWithTombSeriesIterators(t *testing.T) {
|
||||||
sample{7, 89, nil, nil}, sample{9, 8, nil, nil},
|
sample{7, 89, nil, nil}, sample{9, 8, nil, nil},
|
||||||
}),
|
}),
|
||||||
},
|
},
|
||||||
|
expectedMinMaxTimes: []minMaxTimes{{1, 6}, {7, 9}},
|
||||||
},
|
},
|
||||||
{
|
{
|
||||||
name: "two chunks with deletion across two chunks",
|
name: "two chunks with deletion across two chunks",
|
||||||
|
@ -894,6 +928,7 @@ func TestPopulateWithTombSeriesIterators(t *testing.T) {
|
||||||
sample{9, 8, nil, nil},
|
sample{9, 8, nil, nil},
|
||||||
}),
|
}),
|
||||||
},
|
},
|
||||||
|
expectedMinMaxTimes: []minMaxTimes{{1, 3}, {9, 9}},
|
||||||
},
|
},
|
||||||
// Deletion with seek.
|
// Deletion with seek.
|
||||||
{
|
{
|
||||||
|
@ -934,9 +969,33 @@ func TestPopulateWithTombSeriesIterators(t *testing.T) {
|
||||||
sample{6, 0, tsdbutil.SetHistogramNotCounterReset(tsdbutil.GenerateTestHistogram(6)), nil},
|
sample{6, 0, tsdbutil.SetHistogramNotCounterReset(tsdbutil.GenerateTestHistogram(6)), nil},
|
||||||
}),
|
}),
|
||||||
},
|
},
|
||||||
|
expectedMinMaxTimes: []minMaxTimes{{1, 6}},
|
||||||
},
|
},
|
||||||
{
|
{
|
||||||
name: "one histogram chunk intersect with deletion interval",
|
name: "one histogram chunk intersect with earlier deletion interval",
|
||||||
|
chks: [][]chunks.Sample{
|
||||||
|
{
|
||||||
|
sample{1, 0, tsdbutil.GenerateTestHistogram(1), nil},
|
||||||
|
sample{2, 0, tsdbutil.GenerateTestHistogram(2), nil},
|
||||||
|
sample{3, 0, tsdbutil.GenerateTestHistogram(3), nil},
|
||||||
|
sample{6, 0, tsdbutil.GenerateTestHistogram(6), nil},
|
||||||
|
},
|
||||||
|
},
|
||||||
|
intervals: tombstones.Intervals{{Mint: 1, Maxt: 2}},
|
||||||
|
expected: []chunks.Sample{
|
||||||
|
sample{3, 0, tsdbutil.SetHistogramNotCounterReset(tsdbutil.GenerateTestHistogram(3)), nil},
|
||||||
|
sample{6, 0, tsdbutil.SetHistogramNotCounterReset(tsdbutil.GenerateTestHistogram(6)), nil},
|
||||||
|
},
|
||||||
|
expectedChks: []chunks.Meta{
|
||||||
|
assureChunkFromSamples(t, []chunks.Sample{
|
||||||
|
sample{3, 0, tsdbutil.SetHistogramNotCounterReset(tsdbutil.GenerateTestHistogram(3)), nil},
|
||||||
|
sample{6, 0, tsdbutil.SetHistogramNotCounterReset(tsdbutil.GenerateTestHistogram(6)), nil},
|
||||||
|
}),
|
||||||
|
},
|
||||||
|
expectedMinMaxTimes: []minMaxTimes{{3, 6}},
|
||||||
|
},
|
||||||
|
{
|
||||||
|
name: "one histogram chunk intersect with later deletion interval",
|
||||||
chks: [][]chunks.Sample{
|
chks: [][]chunks.Sample{
|
||||||
{
|
{
|
||||||
sample{1, 0, tsdbutil.GenerateTestHistogram(1), nil},
|
sample{1, 0, tsdbutil.GenerateTestHistogram(1), nil},
|
||||||
|
@ -958,6 +1017,7 @@ func TestPopulateWithTombSeriesIterators(t *testing.T) {
|
||||||
sample{3, 0, tsdbutil.SetHistogramNotCounterReset(tsdbutil.GenerateTestHistogram(3)), nil},
|
sample{3, 0, tsdbutil.SetHistogramNotCounterReset(tsdbutil.GenerateTestHistogram(3)), nil},
|
||||||
}),
|
}),
|
||||||
},
|
},
|
||||||
|
expectedMinMaxTimes: []minMaxTimes{{1, 3}},
|
||||||
},
|
},
|
||||||
{
|
{
|
||||||
name: "one float histogram chunk",
|
name: "one float histogram chunk",
|
||||||
|
@ -983,9 +1043,33 @@ func TestPopulateWithTombSeriesIterators(t *testing.T) {
|
||||||
sample{6, 0, nil, tsdbutil.SetFloatHistogramNotCounterReset(tsdbutil.GenerateTestFloatHistogram(6))},
|
sample{6, 0, nil, tsdbutil.SetFloatHistogramNotCounterReset(tsdbutil.GenerateTestFloatHistogram(6))},
|
||||||
}),
|
}),
|
||||||
},
|
},
|
||||||
|
expectedMinMaxTimes: []minMaxTimes{{1, 6}},
|
||||||
},
|
},
|
||||||
{
|
{
|
||||||
name: "one float histogram chunk intersect with deletion interval",
|
name: "one float histogram chunk intersect with earlier deletion interval",
|
||||||
|
chks: [][]chunks.Sample{
|
||||||
|
{
|
||||||
|
sample{1, 0, nil, tsdbutil.GenerateTestFloatHistogram(1)},
|
||||||
|
sample{2, 0, nil, tsdbutil.GenerateTestFloatHistogram(2)},
|
||||||
|
sample{3, 0, nil, tsdbutil.GenerateTestFloatHistogram(3)},
|
||||||
|
sample{6, 0, nil, tsdbutil.GenerateTestFloatHistogram(6)},
|
||||||
|
},
|
||||||
|
},
|
||||||
|
intervals: tombstones.Intervals{{Mint: 1, Maxt: 2}},
|
||||||
|
expected: []chunks.Sample{
|
||||||
|
sample{3, 0, nil, tsdbutil.SetFloatHistogramNotCounterReset(tsdbutil.GenerateTestFloatHistogram(3))},
|
||||||
|
sample{6, 0, nil, tsdbutil.SetFloatHistogramNotCounterReset(tsdbutil.GenerateTestFloatHistogram(6))},
|
||||||
|
},
|
||||||
|
expectedChks: []chunks.Meta{
|
||||||
|
assureChunkFromSamples(t, []chunks.Sample{
|
||||||
|
sample{3, 0, nil, tsdbutil.SetFloatHistogramNotCounterReset(tsdbutil.GenerateTestFloatHistogram(3))},
|
||||||
|
sample{6, 0, nil, tsdbutil.SetFloatHistogramNotCounterReset(tsdbutil.GenerateTestFloatHistogram(6))},
|
||||||
|
}),
|
||||||
|
},
|
||||||
|
expectedMinMaxTimes: []minMaxTimes{{3, 6}},
|
||||||
|
},
|
||||||
|
{
|
||||||
|
name: "one float histogram chunk intersect with later deletion interval",
|
||||||
chks: [][]chunks.Sample{
|
chks: [][]chunks.Sample{
|
||||||
{
|
{
|
||||||
sample{1, 0, nil, tsdbutil.GenerateTestFloatHistogram(1)},
|
sample{1, 0, nil, tsdbutil.GenerateTestFloatHistogram(1)},
|
||||||
|
@ -1007,6 +1091,7 @@ func TestPopulateWithTombSeriesIterators(t *testing.T) {
|
||||||
sample{3, 0, nil, tsdbutil.SetFloatHistogramNotCounterReset(tsdbutil.GenerateTestFloatHistogram(3))},
|
sample{3, 0, nil, tsdbutil.SetFloatHistogramNotCounterReset(tsdbutil.GenerateTestFloatHistogram(3))},
|
||||||
}),
|
}),
|
||||||
},
|
},
|
||||||
|
expectedMinMaxTimes: []minMaxTimes{{1, 3}},
|
||||||
},
|
},
|
||||||
{
|
{
|
||||||
name: "one gauge histogram chunk",
|
name: "one gauge histogram chunk",
|
||||||
|
@ -1032,9 +1117,33 @@ func TestPopulateWithTombSeriesIterators(t *testing.T) {
|
||||||
sample{6, 0, tsdbutil.GenerateTestGaugeHistogram(6), nil},
|
sample{6, 0, tsdbutil.GenerateTestGaugeHistogram(6), nil},
|
||||||
}),
|
}),
|
||||||
},
|
},
|
||||||
|
expectedMinMaxTimes: []minMaxTimes{{1, 6}},
|
||||||
},
|
},
|
||||||
{
|
{
|
||||||
name: "one gauge histogram chunk intersect with deletion interval",
|
name: "one gauge histogram chunk intersect with earlier deletion interval",
|
||||||
|
chks: [][]chunks.Sample{
|
||||||
|
{
|
||||||
|
sample{1, 0, tsdbutil.GenerateTestGaugeHistogram(1), nil},
|
||||||
|
sample{2, 0, tsdbutil.GenerateTestGaugeHistogram(2), nil},
|
||||||
|
sample{3, 0, tsdbutil.GenerateTestGaugeHistogram(3), nil},
|
||||||
|
sample{6, 0, tsdbutil.GenerateTestGaugeHistogram(6), nil},
|
||||||
|
},
|
||||||
|
},
|
||||||
|
intervals: tombstones.Intervals{{Mint: 1, Maxt: 2}},
|
||||||
|
expected: []chunks.Sample{
|
||||||
|
sample{3, 0, tsdbutil.GenerateTestGaugeHistogram(3), nil},
|
||||||
|
sample{6, 0, tsdbutil.GenerateTestGaugeHistogram(6), nil},
|
||||||
|
},
|
||||||
|
expectedChks: []chunks.Meta{
|
||||||
|
assureChunkFromSamples(t, []chunks.Sample{
|
||||||
|
sample{3, 0, tsdbutil.GenerateTestGaugeHistogram(3), nil},
|
||||||
|
sample{6, 0, tsdbutil.GenerateTestGaugeHistogram(6), nil},
|
||||||
|
}),
|
||||||
|
},
|
||||||
|
expectedMinMaxTimes: []minMaxTimes{{3, 6}},
|
||||||
|
},
|
||||||
|
{
|
||||||
|
name: "one gauge histogram chunk intersect with later deletion interval",
|
||||||
chks: [][]chunks.Sample{
|
chks: [][]chunks.Sample{
|
||||||
{
|
{
|
||||||
sample{1, 0, tsdbutil.GenerateTestGaugeHistogram(1), nil},
|
sample{1, 0, tsdbutil.GenerateTestGaugeHistogram(1), nil},
|
||||||
|
@ -1056,6 +1165,7 @@ func TestPopulateWithTombSeriesIterators(t *testing.T) {
|
||||||
sample{3, 0, tsdbutil.GenerateTestGaugeHistogram(3), nil},
|
sample{3, 0, tsdbutil.GenerateTestGaugeHistogram(3), nil},
|
||||||
}),
|
}),
|
||||||
},
|
},
|
||||||
|
expectedMinMaxTimes: []minMaxTimes{{1, 3}},
|
||||||
},
|
},
|
||||||
{
|
{
|
||||||
name: "one gauge float histogram",
|
name: "one gauge float histogram",
|
||||||
|
@ -1081,9 +1191,33 @@ func TestPopulateWithTombSeriesIterators(t *testing.T) {
|
||||||
sample{6, 0, nil, tsdbutil.GenerateTestGaugeFloatHistogram(6)},
|
sample{6, 0, nil, tsdbutil.GenerateTestGaugeFloatHistogram(6)},
|
||||||
}),
|
}),
|
||||||
},
|
},
|
||||||
|
expectedMinMaxTimes: []minMaxTimes{{1, 6}},
|
||||||
},
|
},
|
||||||
{
|
{
|
||||||
name: "one gauge float histogram chunk intersect with deletion interval",
|
name: "one gauge float histogram chunk intersect with earlier deletion interval",
|
||||||
|
chks: [][]chunks.Sample{
|
||||||
|
{
|
||||||
|
sample{1, 0, nil, tsdbutil.GenerateTestGaugeFloatHistogram(1)},
|
||||||
|
sample{2, 0, nil, tsdbutil.GenerateTestGaugeFloatHistogram(2)},
|
||||||
|
sample{3, 0, nil, tsdbutil.GenerateTestGaugeFloatHistogram(3)},
|
||||||
|
sample{6, 0, nil, tsdbutil.GenerateTestGaugeFloatHistogram(6)},
|
||||||
|
},
|
||||||
|
},
|
||||||
|
intervals: tombstones.Intervals{{Mint: 1, Maxt: 2}},
|
||||||
|
expected: []chunks.Sample{
|
||||||
|
sample{3, 0, nil, tsdbutil.GenerateTestGaugeFloatHistogram(3)},
|
||||||
|
sample{6, 0, nil, tsdbutil.GenerateTestGaugeFloatHistogram(6)},
|
||||||
|
},
|
||||||
|
expectedChks: []chunks.Meta{
|
||||||
|
assureChunkFromSamples(t, []chunks.Sample{
|
||||||
|
sample{3, 0, nil, tsdbutil.GenerateTestGaugeFloatHistogram(3)},
|
||||||
|
sample{6, 0, nil, tsdbutil.GenerateTestGaugeFloatHistogram(6)},
|
||||||
|
}),
|
||||||
|
},
|
||||||
|
expectedMinMaxTimes: []minMaxTimes{{3, 6}},
|
||||||
|
},
|
||||||
|
{
|
||||||
|
name: "one gauge float histogram chunk intersect with later deletion interval",
|
||||||
chks: [][]chunks.Sample{
|
chks: [][]chunks.Sample{
|
||||||
{
|
{
|
||||||
sample{1, 0, nil, tsdbutil.GenerateTestGaugeFloatHistogram(1)},
|
sample{1, 0, nil, tsdbutil.GenerateTestGaugeFloatHistogram(1)},
|
||||||
|
@ -1105,6 +1239,134 @@ func TestPopulateWithTombSeriesIterators(t *testing.T) {
|
||||||
sample{3, 0, nil, tsdbutil.GenerateTestGaugeFloatHistogram(3)},
|
sample{3, 0, nil, tsdbutil.GenerateTestGaugeFloatHistogram(3)},
|
||||||
}),
|
}),
|
||||||
},
|
},
|
||||||
|
expectedMinMaxTimes: []minMaxTimes{{1, 3}},
|
||||||
|
},
|
||||||
|
{
|
||||||
|
name: "three full mixed chunks",
|
||||||
|
chks: [][]chunks.Sample{
|
||||||
|
{sample{1, 2, nil, nil}, sample{2, 3, nil, nil}, sample{3, 5, nil, nil}, sample{6, 1, nil, nil}},
|
||||||
|
{
|
||||||
|
sample{7, 0, tsdbutil.GenerateTestGaugeHistogram(89), nil},
|
||||||
|
sample{9, 0, tsdbutil.GenerateTestGaugeHistogram(8), nil},
|
||||||
|
},
|
||||||
|
{
|
||||||
|
sample{10, 0, nil, tsdbutil.GenerateTestGaugeFloatHistogram(22)},
|
||||||
|
sample{203, 0, nil, tsdbutil.GenerateTestGaugeFloatHistogram(3493)},
|
||||||
|
},
|
||||||
|
},
|
||||||
|
|
||||||
|
expected: []chunks.Sample{
|
||||||
|
sample{1, 2, nil, nil}, sample{2, 3, nil, nil}, sample{3, 5, nil, nil}, sample{6, 1, nil, nil}, sample{7, 0, tsdbutil.GenerateTestGaugeHistogram(89), nil}, sample{9, 0, tsdbutil.GenerateTestGaugeHistogram(8), nil}, sample{10, 0, nil, tsdbutil.GenerateTestGaugeFloatHistogram(22)}, sample{203, 0, nil, tsdbutil.GenerateTestGaugeFloatHistogram(3493)},
|
||||||
|
},
|
||||||
|
expectedChks: []chunks.Meta{
|
||||||
|
assureChunkFromSamples(t, []chunks.Sample{
|
||||||
|
sample{1, 2, nil, nil}, sample{2, 3, nil, nil}, sample{3, 5, nil, nil}, sample{6, 1, nil, nil},
|
||||||
|
}),
|
||||||
|
assureChunkFromSamples(t, []chunks.Sample{
|
||||||
|
sample{7, 0, tsdbutil.GenerateTestGaugeHistogram(89), nil},
|
||||||
|
sample{9, 0, tsdbutil.GenerateTestGaugeHistogram(8), nil},
|
||||||
|
}),
|
||||||
|
assureChunkFromSamples(t, []chunks.Sample{
|
||||||
|
sample{10, 0, nil, tsdbutil.GenerateTestGaugeFloatHistogram(22)},
|
||||||
|
sample{203, 0, nil, tsdbutil.GenerateTestGaugeFloatHistogram(3493)},
|
||||||
|
}),
|
||||||
|
},
|
||||||
|
expectedMinMaxTimes: []minMaxTimes{{1, 6}, {7, 9}, {10, 203}},
|
||||||
|
},
|
||||||
|
{
|
||||||
|
name: "three full mixed chunks in different order",
|
||||||
|
chks: [][]chunks.Sample{
|
||||||
|
{
|
||||||
|
sample{7, 0, tsdbutil.GenerateTestGaugeHistogram(89), nil},
|
||||||
|
sample{9, 0, tsdbutil.GenerateTestGaugeHistogram(8), nil},
|
||||||
|
},
|
||||||
|
{sample{11, 2, nil, nil}, sample{12, 3, nil, nil}, sample{13, 5, nil, nil}, sample{16, 1, nil, nil}},
|
||||||
|
{
|
||||||
|
sample{100, 0, nil, tsdbutil.GenerateTestGaugeFloatHistogram(22)},
|
||||||
|
sample{203, 0, nil, tsdbutil.GenerateTestGaugeFloatHistogram(3493)},
|
||||||
|
},
|
||||||
|
},
|
||||||
|
|
||||||
|
expected: []chunks.Sample{
|
||||||
|
sample{7, 0, tsdbutil.GenerateTestGaugeHistogram(89), nil}, sample{9, 0, tsdbutil.GenerateTestGaugeHistogram(8), nil}, sample{11, 2, nil, nil}, sample{12, 3, nil, nil}, sample{13, 5, nil, nil}, sample{16, 1, nil, nil}, sample{100, 0, nil, tsdbutil.GenerateTestGaugeFloatHistogram(22)}, sample{203, 0, nil, tsdbutil.GenerateTestGaugeFloatHistogram(3493)},
|
||||||
|
},
|
||||||
|
expectedChks: []chunks.Meta{
|
||||||
|
assureChunkFromSamples(t, []chunks.Sample{
|
||||||
|
sample{7, 0, tsdbutil.GenerateTestGaugeHistogram(89), nil},
|
||||||
|
sample{9, 0, tsdbutil.GenerateTestGaugeHistogram(8), nil},
|
||||||
|
}),
|
||||||
|
assureChunkFromSamples(t, []chunks.Sample{
|
||||||
|
sample{11, 2, nil, nil}, sample{12, 3, nil, nil}, sample{13, 5, nil, nil}, sample{16, 1, nil, nil},
|
||||||
|
}),
|
||||||
|
assureChunkFromSamples(t, []chunks.Sample{
|
||||||
|
sample{100, 0, nil, tsdbutil.GenerateTestGaugeFloatHistogram(22)},
|
||||||
|
sample{203, 0, nil, tsdbutil.GenerateTestGaugeFloatHistogram(3493)},
|
||||||
|
}),
|
||||||
|
},
|
||||||
|
expectedMinMaxTimes: []minMaxTimes{{7, 9}, {11, 16}, {100, 203}},
|
||||||
|
},
|
||||||
|
{
|
||||||
|
name: "three full mixed chunks in different order intersect with deletion interval",
|
||||||
|
chks: [][]chunks.Sample{
|
||||||
|
{
|
||||||
|
sample{7, 0, tsdbutil.GenerateTestGaugeHistogram(89), nil},
|
||||||
|
sample{9, 0, tsdbutil.GenerateTestGaugeHistogram(8), nil},
|
||||||
|
},
|
||||||
|
{sample{11, 2, nil, nil}, sample{12, 3, nil, nil}, sample{13, 5, nil, nil}, sample{16, 1, nil, nil}},
|
||||||
|
{
|
||||||
|
sample{100, 0, nil, tsdbutil.GenerateTestGaugeFloatHistogram(22)},
|
||||||
|
sample{203, 0, nil, tsdbutil.GenerateTestGaugeFloatHistogram(3493)},
|
||||||
|
},
|
||||||
|
},
|
||||||
|
intervals: tombstones.Intervals{{Mint: 8, Maxt: 11}, {Mint: 15, Maxt: 150}},
|
||||||
|
|
||||||
|
expected: []chunks.Sample{
|
||||||
|
sample{7, 0, tsdbutil.GenerateTestGaugeHistogram(89), nil}, sample{12, 3, nil, nil}, sample{13, 5, nil, nil}, sample{203, 0, nil, tsdbutil.GenerateTestGaugeFloatHistogram(3493)},
|
||||||
|
},
|
||||||
|
expectedChks: []chunks.Meta{
|
||||||
|
assureChunkFromSamples(t, []chunks.Sample{
|
||||||
|
sample{7, 0, tsdbutil.GenerateTestGaugeHistogram(89), nil},
|
||||||
|
}),
|
||||||
|
assureChunkFromSamples(t, []chunks.Sample{
|
||||||
|
sample{12, 3, nil, nil}, sample{13, 5, nil, nil},
|
||||||
|
}),
|
||||||
|
assureChunkFromSamples(t, []chunks.Sample{
|
||||||
|
sample{203, 0, nil, tsdbutil.GenerateTestGaugeFloatHistogram(3493)},
|
||||||
|
}),
|
||||||
|
},
|
||||||
|
expectedMinMaxTimes: []minMaxTimes{{7, 7}, {12, 13}, {203, 203}},
|
||||||
|
},
|
||||||
|
{
|
||||||
|
name: "three full mixed chunks overlapping",
|
||||||
|
chks: [][]chunks.Sample{
|
||||||
|
{
|
||||||
|
sample{7, 0, tsdbutil.GenerateTestGaugeHistogram(89), nil},
|
||||||
|
sample{12, 0, tsdbutil.GenerateTestGaugeHistogram(8), nil},
|
||||||
|
},
|
||||||
|
{sample{11, 2, nil, nil}, sample{12, 3, nil, nil}, sample{13, 5, nil, nil}, sample{16, 1, nil, nil}},
|
||||||
|
{
|
||||||
|
sample{10, 0, nil, tsdbutil.GenerateTestGaugeFloatHistogram(22)},
|
||||||
|
sample{203, 0, nil, tsdbutil.GenerateTestGaugeFloatHistogram(3493)},
|
||||||
|
},
|
||||||
|
},
|
||||||
|
|
||||||
|
expected: []chunks.Sample{
|
||||||
|
sample{7, 0, tsdbutil.GenerateTestGaugeHistogram(89), nil}, sample{12, 0, tsdbutil.GenerateTestGaugeHistogram(8), nil}, sample{11, 2, nil, nil}, sample{12, 3, nil, nil}, sample{13, 5, nil, nil}, sample{16, 1, nil, nil}, sample{10, 0, nil, tsdbutil.GenerateTestGaugeFloatHistogram(22)}, sample{203, 0, nil, tsdbutil.GenerateTestGaugeFloatHistogram(3493)},
|
||||||
|
},
|
||||||
|
expectedChks: []chunks.Meta{
|
||||||
|
assureChunkFromSamples(t, []chunks.Sample{
|
||||||
|
sample{7, 0, tsdbutil.GenerateTestGaugeHistogram(89), nil},
|
||||||
|
sample{12, 0, tsdbutil.GenerateTestGaugeHistogram(8), nil},
|
||||||
|
}),
|
||||||
|
assureChunkFromSamples(t, []chunks.Sample{
|
||||||
|
sample{11, 2, nil, nil}, sample{12, 3, nil, nil}, sample{13, 5, nil, nil}, sample{16, 1, nil, nil},
|
||||||
|
}),
|
||||||
|
assureChunkFromSamples(t, []chunks.Sample{
|
||||||
|
sample{10, 0, nil, tsdbutil.GenerateTestGaugeFloatHistogram(22)},
|
||||||
|
sample{203, 0, nil, tsdbutil.GenerateTestGaugeFloatHistogram(3493)},
|
||||||
|
}),
|
||||||
|
},
|
||||||
|
expectedMinMaxTimes: []minMaxTimes{{7, 12}, {11, 16}, {10, 203}},
|
||||||
},
|
},
|
||||||
}
|
}
|
||||||
for _, tc := range cases {
|
for _, tc := range cases {
|
||||||
|
@ -1146,6 +1408,11 @@ func TestPopulateWithTombSeriesIterators(t *testing.T) {
|
||||||
rmChunkRefs(expandedResult)
|
rmChunkRefs(expandedResult)
|
||||||
rmChunkRefs(tc.expectedChks)
|
rmChunkRefs(tc.expectedChks)
|
||||||
require.Equal(t, tc.expectedChks, expandedResult)
|
require.Equal(t, tc.expectedChks, expandedResult)
|
||||||
|
|
||||||
|
for i, meta := range expandedResult {
|
||||||
|
require.Equal(t, tc.expectedMinMaxTimes[i].minTime, meta.MinTime)
|
||||||
|
require.Equal(t, tc.expectedMinMaxTimes[i].maxTime, meta.MaxTime)
|
||||||
|
}
|
||||||
})
|
})
|
||||||
})
|
})
|
||||||
}
|
}
|
||||||
|
@ -1157,67 +1424,213 @@ func rmChunkRefs(chks []chunks.Meta) {
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
|
func checkCurrVal(t *testing.T, valType chunkenc.ValueType, it *populateWithDelSeriesIterator, expectedTs, expectedValue int) {
|
||||||
|
switch valType {
|
||||||
|
case chunkenc.ValFloat:
|
||||||
|
ts, v := it.At()
|
||||||
|
require.Equal(t, int64(expectedTs), ts)
|
||||||
|
require.Equal(t, float64(expectedValue), v)
|
||||||
|
case chunkenc.ValHistogram:
|
||||||
|
ts, h := it.AtHistogram()
|
||||||
|
require.Equal(t, int64(expectedTs), ts)
|
||||||
|
h.CounterResetHint = histogram.UnknownCounterReset
|
||||||
|
require.Equal(t, tsdbutil.GenerateTestHistogram(expectedValue), h)
|
||||||
|
case chunkenc.ValFloatHistogram:
|
||||||
|
ts, h := it.AtFloatHistogram()
|
||||||
|
require.Equal(t, int64(expectedTs), ts)
|
||||||
|
h.CounterResetHint = histogram.UnknownCounterReset
|
||||||
|
require.Equal(t, tsdbutil.GenerateTestFloatHistogram(expectedValue), h)
|
||||||
|
default:
|
||||||
|
panic("unexpected value type")
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
// Regression for: https://github.com/prometheus/tsdb/pull/97
|
// Regression for: https://github.com/prometheus/tsdb/pull/97
|
||||||
func TestPopulateWithDelSeriesIterator_DoubleSeek(t *testing.T) {
|
func TestPopulateWithDelSeriesIterator_DoubleSeek(t *testing.T) {
|
||||||
f, chkMetas := createFakeReaderAndNotPopulatedChunks(
|
cases := []struct {
|
||||||
[]chunks.Sample{},
|
name string
|
||||||
[]chunks.Sample{sample{1, 1, nil, nil}, sample{2, 2, nil, nil}, sample{3, 3, nil, nil}},
|
valType chunkenc.ValueType
|
||||||
[]chunks.Sample{sample{4, 4, nil, nil}, sample{5, 5, nil, nil}},
|
chks [][]chunks.Sample
|
||||||
)
|
}{
|
||||||
|
{
|
||||||
|
name: "float",
|
||||||
|
valType: chunkenc.ValFloat,
|
||||||
|
chks: [][]chunks.Sample{
|
||||||
|
{},
|
||||||
|
{sample{1, 1, nil, nil}, sample{2, 2, nil, nil}, sample{3, 3, nil, nil}},
|
||||||
|
{sample{4, 4, nil, nil}, sample{5, 5, nil, nil}},
|
||||||
|
},
|
||||||
|
},
|
||||||
|
{
|
||||||
|
name: "histogram",
|
||||||
|
valType: chunkenc.ValHistogram,
|
||||||
|
chks: [][]chunks.Sample{
|
||||||
|
{},
|
||||||
|
{sample{1, 0, tsdbutil.GenerateTestHistogram(1), nil}, sample{2, 0, tsdbutil.GenerateTestHistogram(2), nil}, sample{3, 0, tsdbutil.GenerateTestHistogram(3), nil}},
|
||||||
|
{sample{4, 0, tsdbutil.GenerateTestHistogram(4), nil}, sample{5, 0, tsdbutil.GenerateTestHistogram(5), nil}},
|
||||||
|
},
|
||||||
|
},
|
||||||
|
{
|
||||||
|
name: "float histogram",
|
||||||
|
valType: chunkenc.ValFloatHistogram,
|
||||||
|
chks: [][]chunks.Sample{
|
||||||
|
{},
|
||||||
|
{sample{1, 0, nil, tsdbutil.GenerateTestFloatHistogram(1)}, sample{2, 0, nil, tsdbutil.GenerateTestFloatHistogram(2)}, sample{3, 0, nil, tsdbutil.GenerateTestFloatHistogram(3)}},
|
||||||
|
{sample{4, 0, nil, tsdbutil.GenerateTestFloatHistogram(4)}, sample{5, 0, nil, tsdbutil.GenerateTestFloatHistogram(5)}},
|
||||||
|
},
|
||||||
|
},
|
||||||
|
}
|
||||||
|
|
||||||
|
for _, tc := range cases {
|
||||||
|
t.Run(tc.name, func(t *testing.T) {
|
||||||
|
f, chkMetas := createFakeReaderAndNotPopulatedChunks(tc.chks...)
|
||||||
it := &populateWithDelSeriesIterator{}
|
it := &populateWithDelSeriesIterator{}
|
||||||
it.reset(ulid.ULID{}, f, chkMetas, nil)
|
it.reset(ulid.ULID{}, f, chkMetas, nil)
|
||||||
require.Equal(t, chunkenc.ValFloat, it.Seek(1))
|
require.Equal(t, tc.valType, it.Seek(1))
|
||||||
require.Equal(t, chunkenc.ValFloat, it.Seek(2))
|
require.Equal(t, tc.valType, it.Seek(2))
|
||||||
require.Equal(t, chunkenc.ValFloat, it.Seek(2))
|
require.Equal(t, tc.valType, it.Seek(2))
|
||||||
ts, v := it.At()
|
checkCurrVal(t, tc.valType, it, 2, 2)
|
||||||
require.Equal(t, int64(2), ts)
|
require.Equal(t, int64(0), chkMetas[0].MinTime)
|
||||||
require.Equal(t, float64(2), v)
|
require.Equal(t, int64(1), chkMetas[1].MinTime)
|
||||||
|
require.Equal(t, int64(4), chkMetas[2].MinTime)
|
||||||
|
})
|
||||||
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
// Regression when seeked chunks were still found via binary search and we always
|
// Regression when seeked chunks were still found via binary search and we always
|
||||||
// skipped to the end when seeking a value in the current chunk.
|
// skipped to the end when seeking a value in the current chunk.
|
||||||
func TestPopulateWithDelSeriesIterator_SeekInCurrentChunk(t *testing.T) {
|
func TestPopulateWithDelSeriesIterator_SeekInCurrentChunk(t *testing.T) {
|
||||||
f, chkMetas := createFakeReaderAndNotPopulatedChunks(
|
cases := []struct {
|
||||||
[]chunks.Sample{},
|
name string
|
||||||
[]chunks.Sample{sample{1, 2, nil, nil}, sample{3, 4, nil, nil}, sample{5, 6, nil, nil}, sample{7, 8, nil, nil}},
|
valType chunkenc.ValueType
|
||||||
[]chunks.Sample{},
|
chks [][]chunks.Sample
|
||||||
)
|
}{
|
||||||
|
{
|
||||||
|
name: "float",
|
||||||
|
valType: chunkenc.ValFloat,
|
||||||
|
chks: [][]chunks.Sample{
|
||||||
|
{},
|
||||||
|
{sample{1, 2, nil, nil}, sample{3, 4, nil, nil}, sample{5, 6, nil, nil}, sample{7, 8, nil, nil}},
|
||||||
|
{},
|
||||||
|
},
|
||||||
|
},
|
||||||
|
{
|
||||||
|
name: "histogram",
|
||||||
|
valType: chunkenc.ValHistogram,
|
||||||
|
chks: [][]chunks.Sample{
|
||||||
|
{},
|
||||||
|
{sample{1, 0, tsdbutil.GenerateTestHistogram(2), nil}, sample{3, 0, tsdbutil.GenerateTestHistogram(4), nil}, sample{5, 0, tsdbutil.GenerateTestHistogram(6), nil}, sample{7, 0, tsdbutil.GenerateTestHistogram(8), nil}},
|
||||||
|
{},
|
||||||
|
},
|
||||||
|
},
|
||||||
|
{
|
||||||
|
name: "float histogram",
|
||||||
|
valType: chunkenc.ValFloatHistogram,
|
||||||
|
chks: [][]chunks.Sample{
|
||||||
|
{},
|
||||||
|
{sample{1, 0, nil, tsdbutil.GenerateTestFloatHistogram(2)}, sample{3, 0, nil, tsdbutil.GenerateTestFloatHistogram(4)}, sample{5, 0, nil, tsdbutil.GenerateTestFloatHistogram(6)}, sample{7, 0, nil, tsdbutil.GenerateTestFloatHistogram(8)}},
|
||||||
|
{},
|
||||||
|
},
|
||||||
|
},
|
||||||
|
}
|
||||||
|
|
||||||
|
for _, tc := range cases {
|
||||||
|
t.Run(tc.name, func(t *testing.T) {
|
||||||
|
f, chkMetas := createFakeReaderAndNotPopulatedChunks(tc.chks...)
|
||||||
it := &populateWithDelSeriesIterator{}
|
it := &populateWithDelSeriesIterator{}
|
||||||
it.reset(ulid.ULID{}, f, chkMetas, nil)
|
it.reset(ulid.ULID{}, f, chkMetas, nil)
|
||||||
require.Equal(t, chunkenc.ValFloat, it.Next())
|
require.Equal(t, tc.valType, it.Next())
|
||||||
ts, v := it.At()
|
checkCurrVal(t, tc.valType, it, 1, 2)
|
||||||
require.Equal(t, int64(1), ts)
|
require.Equal(t, tc.valType, it.Seek(4))
|
||||||
require.Equal(t, float64(2), v)
|
checkCurrVal(t, tc.valType, it, 5, 6)
|
||||||
|
require.Equal(t, int64(0), chkMetas[0].MinTime)
|
||||||
require.Equal(t, chunkenc.ValFloat, it.Seek(4))
|
require.Equal(t, int64(1), chkMetas[1].MinTime)
|
||||||
ts, v = it.At()
|
require.Equal(t, int64(0), chkMetas[2].MinTime)
|
||||||
require.Equal(t, int64(5), ts)
|
})
|
||||||
require.Equal(t, float64(6), v)
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
func TestPopulateWithDelSeriesIterator_SeekWithMinTime(t *testing.T) {
|
func TestPopulateWithDelSeriesIterator_SeekWithMinTime(t *testing.T) {
|
||||||
f, chkMetas := createFakeReaderAndNotPopulatedChunks(
|
cases := []struct {
|
||||||
[]chunks.Sample{sample{1, 6, nil, nil}, sample{5, 6, nil, nil}, sample{6, 8, nil, nil}},
|
name string
|
||||||
)
|
valType chunkenc.ValueType
|
||||||
|
chks [][]chunks.Sample
|
||||||
|
}{
|
||||||
|
{
|
||||||
|
name: "float",
|
||||||
|
valType: chunkenc.ValFloat,
|
||||||
|
chks: [][]chunks.Sample{
|
||||||
|
{sample{1, 6, nil, nil}, sample{5, 6, nil, nil}, sample{6, 8, nil, nil}},
|
||||||
|
},
|
||||||
|
},
|
||||||
|
{
|
||||||
|
name: "histogram",
|
||||||
|
valType: chunkenc.ValHistogram,
|
||||||
|
chks: [][]chunks.Sample{
|
||||||
|
{sample{1, 0, tsdbutil.GenerateTestHistogram(6), nil}, sample{5, 0, tsdbutil.GenerateTestHistogram(6), nil}, sample{6, 0, tsdbutil.GenerateTestHistogram(8), nil}},
|
||||||
|
},
|
||||||
|
},
|
||||||
|
{
|
||||||
|
name: "float histogram",
|
||||||
|
valType: chunkenc.ValFloatHistogram,
|
||||||
|
chks: [][]chunks.Sample{
|
||||||
|
{sample{1, 0, nil, tsdbutil.GenerateTestFloatHistogram(6)}, sample{5, 0, nil, tsdbutil.GenerateTestFloatHistogram(6)}, sample{6, 0, nil, tsdbutil.GenerateTestFloatHistogram(8)}},
|
||||||
|
},
|
||||||
|
},
|
||||||
|
}
|
||||||
|
|
||||||
|
for _, tc := range cases {
|
||||||
|
t.Run(tc.name, func(t *testing.T) {
|
||||||
|
f, chkMetas := createFakeReaderAndNotPopulatedChunks(tc.chks...)
|
||||||
it := &populateWithDelSeriesIterator{}
|
it := &populateWithDelSeriesIterator{}
|
||||||
it.reset(ulid.ULID{}, f, chkMetas, nil)
|
it.reset(ulid.ULID{}, f, chkMetas, nil)
|
||||||
require.Equal(t, chunkenc.ValNone, it.Seek(7))
|
require.Equal(t, chunkenc.ValNone, it.Seek(7))
|
||||||
require.Equal(t, chunkenc.ValFloat, it.Seek(3))
|
require.Equal(t, tc.valType, it.Seek(3))
|
||||||
|
require.Equal(t, int64(1), chkMetas[0].MinTime)
|
||||||
|
})
|
||||||
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
// Regression when calling Next() with a time bounded to fit within two samples.
|
// Regression when calling Next() with a time bounded to fit within two samples.
|
||||||
// Seek gets called and advances beyond the max time, which was just accepted as a valid sample.
|
// Seek gets called and advances beyond the max time, which was just accepted as a valid sample.
|
||||||
func TestPopulateWithDelSeriesIterator_NextWithMinTime(t *testing.T) {
|
func TestPopulateWithDelSeriesIterator_NextWithMinTime(t *testing.T) {
|
||||||
f, chkMetas := createFakeReaderAndNotPopulatedChunks(
|
cases := []struct {
|
||||||
[]chunks.Sample{sample{1, 6, nil, nil}, sample{5, 6, nil, nil}, sample{7, 8, nil, nil}},
|
name string
|
||||||
)
|
valType chunkenc.ValueType
|
||||||
|
chks [][]chunks.Sample
|
||||||
|
}{
|
||||||
|
{
|
||||||
|
name: "float",
|
||||||
|
valType: chunkenc.ValFloat,
|
||||||
|
chks: [][]chunks.Sample{
|
||||||
|
{sample{1, 6, nil, nil}, sample{5, 6, nil, nil}, sample{7, 8, nil, nil}},
|
||||||
|
},
|
||||||
|
},
|
||||||
|
{
|
||||||
|
name: "histogram",
|
||||||
|
valType: chunkenc.ValHistogram,
|
||||||
|
chks: [][]chunks.Sample{
|
||||||
|
{sample{1, 0, tsdbutil.GenerateTestHistogram(6), nil}, sample{5, 0, tsdbutil.GenerateTestHistogram(6), nil}, sample{7, 0, tsdbutil.GenerateTestHistogram(8), nil}},
|
||||||
|
},
|
||||||
|
},
|
||||||
|
{
|
||||||
|
name: "float histogram",
|
||||||
|
valType: chunkenc.ValFloatHistogram,
|
||||||
|
chks: [][]chunks.Sample{
|
||||||
|
{sample{1, 0, nil, tsdbutil.GenerateTestFloatHistogram(6)}, sample{5, 0, nil, tsdbutil.GenerateTestFloatHistogram(6)}, sample{7, 0, nil, tsdbutil.GenerateTestFloatHistogram(8)}},
|
||||||
|
},
|
||||||
|
},
|
||||||
|
}
|
||||||
|
|
||||||
|
for _, tc := range cases {
|
||||||
|
t.Run(tc.name, func(t *testing.T) {
|
||||||
|
f, chkMetas := createFakeReaderAndNotPopulatedChunks(tc.chks...)
|
||||||
it := &populateWithDelSeriesIterator{}
|
it := &populateWithDelSeriesIterator{}
|
||||||
it.reset(ulid.ULID{}, f, chkMetas, tombstones.Intervals{{Mint: math.MinInt64, Maxt: 2}}.Add(tombstones.Interval{Mint: 4, Maxt: math.MaxInt64}))
|
it.reset(ulid.ULID{}, f, chkMetas, tombstones.Intervals{{Mint: math.MinInt64, Maxt: 2}}.Add(tombstones.Interval{Mint: 4, Maxt: math.MaxInt64}))
|
||||||
require.Equal(t, chunkenc.ValNone, it.Next())
|
require.Equal(t, chunkenc.ValNone, it.Next())
|
||||||
|
require.Equal(t, int64(1), chkMetas[0].MinTime)
|
||||||
|
})
|
||||||
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
// Test the cost of merging series sets for different number of merged sets and their size.
|
// Test the cost of merging series sets for different number of merged sets and their size.
|
||||||
|
|
|
@ -1,6 +1,6 @@
|
||||||
{
|
{
|
||||||
"name": "@prometheus-io/codemirror-promql",
|
"name": "@prometheus-io/codemirror-promql",
|
||||||
"version": "0.48.0-rc.1",
|
"version": "0.48.0-rc.2",
|
||||||
"description": "a CodeMirror mode for the PromQL language",
|
"description": "a CodeMirror mode for the PromQL language",
|
||||||
"types": "dist/esm/index.d.ts",
|
"types": "dist/esm/index.d.ts",
|
||||||
"module": "dist/esm/index.js",
|
"module": "dist/esm/index.js",
|
||||||
|
@ -29,7 +29,7 @@
|
||||||
},
|
},
|
||||||
"homepage": "https://github.com/prometheus/prometheus/blob/main/web/ui/module/codemirror-promql/README.md",
|
"homepage": "https://github.com/prometheus/prometheus/blob/main/web/ui/module/codemirror-promql/README.md",
|
||||||
"dependencies": {
|
"dependencies": {
|
||||||
"@prometheus-io/lezer-promql": "0.48.0-rc.1",
|
"@prometheus-io/lezer-promql": "0.48.0-rc.2",
|
||||||
"lru-cache": "^7.18.3"
|
"lru-cache": "^7.18.3"
|
||||||
},
|
},
|
||||||
"devDependencies": {
|
"devDependencies": {
|
||||||
|
|
|
@ -1,6 +1,6 @@
|
||||||
{
|
{
|
||||||
"name": "@prometheus-io/lezer-promql",
|
"name": "@prometheus-io/lezer-promql",
|
||||||
"version": "0.48.0-rc.1",
|
"version": "0.48.0-rc.2",
|
||||||
"description": "lezer-based PromQL grammar",
|
"description": "lezer-based PromQL grammar",
|
||||||
"main": "dist/index.cjs",
|
"main": "dist/index.cjs",
|
||||||
"type": "module",
|
"type": "module",
|
||||||
|
|
18
web/ui/package-lock.json
generated
18
web/ui/package-lock.json
generated
|
@ -1,12 +1,12 @@
|
||||||
{
|
{
|
||||||
"name": "prometheus-io",
|
"name": "prometheus-io",
|
||||||
"version": "0.48.0-rc.1",
|
"version": "0.48.0-rc.2",
|
||||||
"lockfileVersion": 2,
|
"lockfileVersion": 2,
|
||||||
"requires": true,
|
"requires": true,
|
||||||
"packages": {
|
"packages": {
|
||||||
"": {
|
"": {
|
||||||
"name": "prometheus-io",
|
"name": "prometheus-io",
|
||||||
"version": "0.48.0-rc.1",
|
"version": "0.48.0-rc.2",
|
||||||
"workspaces": [
|
"workspaces": [
|
||||||
"react-app",
|
"react-app",
|
||||||
"module/*"
|
"module/*"
|
||||||
|
@ -30,10 +30,10 @@
|
||||||
},
|
},
|
||||||
"module/codemirror-promql": {
|
"module/codemirror-promql": {
|
||||||
"name": "@prometheus-io/codemirror-promql",
|
"name": "@prometheus-io/codemirror-promql",
|
||||||
"version": "0.48.0-rc.1",
|
"version": "0.48.0-rc.2",
|
||||||
"license": "Apache-2.0",
|
"license": "Apache-2.0",
|
||||||
"dependencies": {
|
"dependencies": {
|
||||||
"@prometheus-io/lezer-promql": "0.48.0-rc.1",
|
"@prometheus-io/lezer-promql": "0.48.0-rc.2",
|
||||||
"lru-cache": "^7.18.3"
|
"lru-cache": "^7.18.3"
|
||||||
},
|
},
|
||||||
"devDependencies": {
|
"devDependencies": {
|
||||||
|
@ -70,7 +70,7 @@
|
||||||
},
|
},
|
||||||
"module/lezer-promql": {
|
"module/lezer-promql": {
|
||||||
"name": "@prometheus-io/lezer-promql",
|
"name": "@prometheus-io/lezer-promql",
|
||||||
"version": "0.48.0-rc.1",
|
"version": "0.48.0-rc.2",
|
||||||
"license": "Apache-2.0",
|
"license": "Apache-2.0",
|
||||||
"devDependencies": {
|
"devDependencies": {
|
||||||
"@lezer/generator": "^1.2.3",
|
"@lezer/generator": "^1.2.3",
|
||||||
|
@ -20764,7 +20764,7 @@
|
||||||
},
|
},
|
||||||
"react-app": {
|
"react-app": {
|
||||||
"name": "@prometheus-io/app",
|
"name": "@prometheus-io/app",
|
||||||
"version": "0.48.0-rc.1",
|
"version": "0.48.0-rc.2",
|
||||||
"dependencies": {
|
"dependencies": {
|
||||||
"@codemirror/autocomplete": "^6.7.1",
|
"@codemirror/autocomplete": "^6.7.1",
|
||||||
"@codemirror/commands": "^6.2.4",
|
"@codemirror/commands": "^6.2.4",
|
||||||
|
@ -20782,7 +20782,7 @@
|
||||||
"@lezer/lr": "^1.3.6",
|
"@lezer/lr": "^1.3.6",
|
||||||
"@nexucis/fuzzy": "^0.4.1",
|
"@nexucis/fuzzy": "^0.4.1",
|
||||||
"@nexucis/kvsearch": "^0.8.1",
|
"@nexucis/kvsearch": "^0.8.1",
|
||||||
"@prometheus-io/codemirror-promql": "0.48.0-rc.1",
|
"@prometheus-io/codemirror-promql": "0.48.0-rc.2",
|
||||||
"bootstrap": "^4.6.2",
|
"bootstrap": "^4.6.2",
|
||||||
"css.escape": "^1.5.1",
|
"css.escape": "^1.5.1",
|
||||||
"downshift": "^7.6.0",
|
"downshift": "^7.6.0",
|
||||||
|
@ -23422,7 +23422,7 @@
|
||||||
"@lezer/lr": "^1.3.6",
|
"@lezer/lr": "^1.3.6",
|
||||||
"@nexucis/fuzzy": "^0.4.1",
|
"@nexucis/fuzzy": "^0.4.1",
|
||||||
"@nexucis/kvsearch": "^0.8.1",
|
"@nexucis/kvsearch": "^0.8.1",
|
||||||
"@prometheus-io/codemirror-promql": "0.48.0-rc.1",
|
"@prometheus-io/codemirror-promql": "0.48.0-rc.2",
|
||||||
"@testing-library/react-hooks": "^7.0.2",
|
"@testing-library/react-hooks": "^7.0.2",
|
||||||
"@types/enzyme": "^3.10.13",
|
"@types/enzyme": "^3.10.13",
|
||||||
"@types/flot": "0.0.32",
|
"@types/flot": "0.0.32",
|
||||||
|
@ -23486,7 +23486,7 @@
|
||||||
"@lezer/common": "^1.0.3",
|
"@lezer/common": "^1.0.3",
|
||||||
"@lezer/highlight": "^1.1.6",
|
"@lezer/highlight": "^1.1.6",
|
||||||
"@lezer/lr": "^1.3.6",
|
"@lezer/lr": "^1.3.6",
|
||||||
"@prometheus-io/lezer-promql": "0.48.0-rc.1",
|
"@prometheus-io/lezer-promql": "0.48.0-rc.2",
|
||||||
"isomorphic-fetch": "^3.0.0",
|
"isomorphic-fetch": "^3.0.0",
|
||||||
"lru-cache": "^7.18.3",
|
"lru-cache": "^7.18.3",
|
||||||
"nock": "^13.3.1"
|
"nock": "^13.3.1"
|
||||||
|
|
|
@ -28,5 +28,5 @@
|
||||||
"ts-jest": "^29.1.0",
|
"ts-jest": "^29.1.0",
|
||||||
"typescript": "^4.9.5"
|
"typescript": "^4.9.5"
|
||||||
},
|
},
|
||||||
"version": "0.48.0-rc.1"
|
"version": "0.48.0-rc.2"
|
||||||
}
|
}
|
||||||
|
|
|
@ -1,6 +1,6 @@
|
||||||
{
|
{
|
||||||
"name": "@prometheus-io/app",
|
"name": "@prometheus-io/app",
|
||||||
"version": "0.48.0-rc.1",
|
"version": "0.48.0-rc.2",
|
||||||
"private": true,
|
"private": true,
|
||||||
"dependencies": {
|
"dependencies": {
|
||||||
"@codemirror/autocomplete": "^6.7.1",
|
"@codemirror/autocomplete": "^6.7.1",
|
||||||
|
@ -19,7 +19,7 @@
|
||||||
"@lezer/lr": "^1.3.6",
|
"@lezer/lr": "^1.3.6",
|
||||||
"@nexucis/fuzzy": "^0.4.1",
|
"@nexucis/fuzzy": "^0.4.1",
|
||||||
"@nexucis/kvsearch": "^0.8.1",
|
"@nexucis/kvsearch": "^0.8.1",
|
||||||
"@prometheus-io/codemirror-promql": "0.48.0-rc.1",
|
"@prometheus-io/codemirror-promql": "0.48.0-rc.2",
|
||||||
"bootstrap": "^4.6.2",
|
"bootstrap": "^4.6.2",
|
||||||
"css.escape": "^1.5.1",
|
"css.escape": "^1.5.1",
|
||||||
"downshift": "^7.6.0",
|
"downshift": "^7.6.0",
|
||||||
|
|
Loading…
Reference in a new issue