2017-07-07 04:46:41 -07:00
|
|
|
|
// Copyright 2017 The Prometheus Authors
|
|
|
|
|
// Licensed under the Apache License, Version 2.0 (the "License");
|
|
|
|
|
// you may not use this file except in compliance with the License.
|
|
|
|
|
// You may obtain a copy of the License at
|
|
|
|
|
//
|
|
|
|
|
// http://www.apache.org/licenses/LICENSE-2.0
|
|
|
|
|
//
|
|
|
|
|
// Unless required by applicable law or agreed to in writing, software
|
|
|
|
|
// distributed under the License is distributed on an "AS IS" BASIS,
|
|
|
|
|
// WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
|
|
|
|
|
// See the License for the specific language governing permissions and
|
|
|
|
|
// limitations under the License.
|
|
|
|
|
|
2017-07-06 12:29:26 -07:00
|
|
|
|
package tsdb
|
|
|
|
|
|
|
|
|
|
import (
|
2017-11-21 03:15:02 -08:00
|
|
|
|
"io/ioutil"
|
2018-10-12 02:45:19 -07:00
|
|
|
|
"math"
|
2017-11-21 03:15:02 -08:00
|
|
|
|
"os"
|
2019-01-29 00:26:01 -08:00
|
|
|
|
"path"
|
2017-11-21 03:15:02 -08:00
|
|
|
|
"path/filepath"
|
2017-07-06 12:29:26 -07:00
|
|
|
|
"testing"
|
2018-11-20 02:34:26 -08:00
|
|
|
|
"time"
|
2017-07-06 12:29:26 -07:00
|
|
|
|
|
2017-11-21 03:15:02 -08:00
|
|
|
|
"github.com/go-kit/kit/log"
|
|
|
|
|
"github.com/pkg/errors"
|
2019-01-29 00:26:01 -08:00
|
|
|
|
prom_testutil "github.com/prometheus/client_golang/prometheus/testutil"
|
2018-11-20 02:34:26 -08:00
|
|
|
|
dto "github.com/prometheus/client_model/go"
|
2018-10-12 02:45:19 -07:00
|
|
|
|
"github.com/prometheus/tsdb/chunks"
|
2018-11-20 02:34:26 -08:00
|
|
|
|
"github.com/prometheus/tsdb/labels"
|
2017-12-06 17:06:14 -08:00
|
|
|
|
"github.com/prometheus/tsdb/testutil"
|
2017-07-06 12:29:26 -07:00
|
|
|
|
)
|
|
|
|
|
|
|
|
|
|
func TestSplitByRange(t *testing.T) {
|
|
|
|
|
cases := []struct {
|
|
|
|
|
trange int64
|
2017-07-13 07:13:59 -07:00
|
|
|
|
ranges [][2]int64
|
|
|
|
|
output [][][2]int64
|
2017-07-06 12:29:26 -07:00
|
|
|
|
}{
|
|
|
|
|
{
|
|
|
|
|
trange: 60,
|
2017-07-13 07:13:59 -07:00
|
|
|
|
ranges: [][2]int64{{0, 10}},
|
|
|
|
|
output: [][][2]int64{
|
|
|
|
|
{{0, 10}},
|
|
|
|
|
},
|
2017-07-06 12:29:26 -07:00
|
|
|
|
},
|
|
|
|
|
{
|
|
|
|
|
trange: 60,
|
2017-07-13 07:13:59 -07:00
|
|
|
|
ranges: [][2]int64{{0, 60}},
|
|
|
|
|
output: [][][2]int64{
|
|
|
|
|
{{0, 60}},
|
|
|
|
|
},
|
2017-07-06 12:29:26 -07:00
|
|
|
|
},
|
|
|
|
|
{
|
|
|
|
|
trange: 60,
|
2017-07-13 07:13:59 -07:00
|
|
|
|
ranges: [][2]int64{{0, 10}, {9, 15}, {30, 60}},
|
|
|
|
|
output: [][][2]int64{
|
|
|
|
|
{{0, 10}, {9, 15}, {30, 60}},
|
|
|
|
|
},
|
2017-07-06 12:29:26 -07:00
|
|
|
|
},
|
|
|
|
|
{
|
|
|
|
|
trange: 60,
|
2017-07-13 07:13:59 -07:00
|
|
|
|
ranges: [][2]int64{{70, 90}, {125, 130}, {130, 180}, {1000, 1001}},
|
|
|
|
|
output: [][][2]int64{
|
|
|
|
|
{{70, 90}},
|
|
|
|
|
{{125, 130}, {130, 180}},
|
|
|
|
|
{{1000, 1001}},
|
|
|
|
|
},
|
2017-07-06 12:29:26 -07:00
|
|
|
|
},
|
2017-07-13 07:13:59 -07:00
|
|
|
|
// Mis-aligned or too-large blocks are ignored.
|
2017-07-06 12:29:26 -07:00
|
|
|
|
{
|
|
|
|
|
trange: 60,
|
2017-07-13 07:13:59 -07:00
|
|
|
|
ranges: [][2]int64{{50, 70}, {70, 80}},
|
|
|
|
|
output: [][][2]int64{
|
|
|
|
|
{{70, 80}},
|
|
|
|
|
},
|
2017-07-06 12:29:26 -07:00
|
|
|
|
},
|
2017-07-13 07:13:59 -07:00
|
|
|
|
{
|
|
|
|
|
trange: 72,
|
|
|
|
|
ranges: [][2]int64{{0, 144}, {144, 216}, {216, 288}},
|
|
|
|
|
output: [][][2]int64{
|
|
|
|
|
{{144, 216}},
|
|
|
|
|
{{216, 288}},
|
|
|
|
|
},
|
|
|
|
|
},
|
|
|
|
|
// Various awkward edge cases easy to hit with negative numbers.
|
2017-07-06 12:29:26 -07:00
|
|
|
|
{
|
|
|
|
|
trange: 60,
|
2017-07-13 07:13:59 -07:00
|
|
|
|
ranges: [][2]int64{{-10, -5}},
|
|
|
|
|
output: [][][2]int64{
|
|
|
|
|
{{-10, -5}},
|
|
|
|
|
},
|
|
|
|
|
},
|
|
|
|
|
{
|
|
|
|
|
trange: 60,
|
|
|
|
|
ranges: [][2]int64{{-60, -50}, {-10, -5}},
|
|
|
|
|
output: [][][2]int64{
|
|
|
|
|
{{-60, -50}, {-10, -5}},
|
|
|
|
|
},
|
|
|
|
|
},
|
|
|
|
|
{
|
|
|
|
|
trange: 60,
|
|
|
|
|
ranges: [][2]int64{{-60, -50}, {-10, -5}, {0, 15}},
|
|
|
|
|
output: [][][2]int64{
|
|
|
|
|
{{-60, -50}, {-10, -5}},
|
|
|
|
|
{{0, 15}},
|
|
|
|
|
},
|
2017-07-06 12:29:26 -07:00
|
|
|
|
},
|
|
|
|
|
}
|
|
|
|
|
|
|
|
|
|
for _, c := range cases {
|
2017-07-13 07:13:59 -07:00
|
|
|
|
// Transform input range tuples into dirMetas.
|
2017-07-06 12:29:26 -07:00
|
|
|
|
blocks := make([]dirMeta, 0, len(c.ranges))
|
|
|
|
|
for _, r := range c.ranges {
|
|
|
|
|
blocks = append(blocks, dirMeta{
|
|
|
|
|
meta: &BlockMeta{
|
|
|
|
|
MinTime: r[0],
|
|
|
|
|
MaxTime: r[1],
|
|
|
|
|
},
|
|
|
|
|
})
|
|
|
|
|
}
|
|
|
|
|
|
2017-07-13 07:13:59 -07:00
|
|
|
|
// Transform output range tuples into dirMetas.
|
|
|
|
|
exp := make([][]dirMeta, len(c.output))
|
|
|
|
|
for i, group := range c.output {
|
|
|
|
|
for _, r := range group {
|
|
|
|
|
exp[i] = append(exp[i], dirMeta{
|
|
|
|
|
meta: &BlockMeta{MinTime: r[0], MaxTime: r[1]},
|
|
|
|
|
})
|
|
|
|
|
}
|
|
|
|
|
}
|
|
|
|
|
|
2017-12-06 17:06:14 -08:00
|
|
|
|
testutil.Equals(t, exp, splitByRange(blocks, c.trange))
|
2017-07-06 12:29:26 -07:00
|
|
|
|
}
|
|
|
|
|
}
|
2017-08-13 01:41:08 -07:00
|
|
|
|
|
|
|
|
|
// See https://github.com/prometheus/prometheus/issues/3064
|
|
|
|
|
func TestNoPanicFor0Tombstones(t *testing.T) {
|
|
|
|
|
metas := []dirMeta{
|
|
|
|
|
{
|
|
|
|
|
dir: "1",
|
|
|
|
|
meta: &BlockMeta{
|
|
|
|
|
MinTime: 0,
|
|
|
|
|
MaxTime: 100,
|
|
|
|
|
},
|
|
|
|
|
},
|
|
|
|
|
{
|
|
|
|
|
dir: "2",
|
|
|
|
|
meta: &BlockMeta{
|
|
|
|
|
MinTime: 101,
|
|
|
|
|
MaxTime: 200,
|
|
|
|
|
},
|
|
|
|
|
},
|
|
|
|
|
}
|
|
|
|
|
|
2017-09-01 02:46:46 -07:00
|
|
|
|
c, err := NewLeveledCompactor(nil, nil, []int64{50}, nil)
|
2017-12-06 17:06:14 -08:00
|
|
|
|
testutil.Ok(t, err)
|
2017-09-01 02:46:46 -07:00
|
|
|
|
|
2017-08-13 01:41:08 -07:00
|
|
|
|
c.plan(metas)
|
|
|
|
|
}
|
2017-09-01 02:46:46 -07:00
|
|
|
|
|
|
|
|
|
func TestLeveledCompactor_plan(t *testing.T) {
|
2018-03-13 08:30:00 -07:00
|
|
|
|
// This mimicks our default ExponentialBlockRanges with min block size equals to 20.
|
2017-09-01 02:46:46 -07:00
|
|
|
|
compactor, err := NewLeveledCompactor(nil, nil, []int64{
|
|
|
|
|
20,
|
|
|
|
|
60,
|
2018-03-13 07:11:02 -07:00
|
|
|
|
180,
|
|
|
|
|
540,
|
|
|
|
|
1620,
|
2017-09-01 02:46:46 -07:00
|
|
|
|
}, nil)
|
2017-12-06 17:06:14 -08:00
|
|
|
|
testutil.Ok(t, err)
|
2017-09-01 02:46:46 -07:00
|
|
|
|
|
|
|
|
|
cases := []struct {
|
|
|
|
|
metas []dirMeta
|
|
|
|
|
expected []string
|
|
|
|
|
}{
|
|
|
|
|
{
|
|
|
|
|
metas: []dirMeta{
|
|
|
|
|
metaRange("1", 0, 20, nil),
|
|
|
|
|
},
|
|
|
|
|
expected: nil,
|
|
|
|
|
},
|
2018-03-13 05:30:27 -07:00
|
|
|
|
// We should wait for four blocks of size 20 to appear before compacting.
|
2017-09-01 02:46:46 -07:00
|
|
|
|
{
|
|
|
|
|
metas: []dirMeta{
|
|
|
|
|
metaRange("1", 0, 20, nil),
|
|
|
|
|
metaRange("2", 20, 40, nil),
|
|
|
|
|
},
|
|
|
|
|
expected: nil,
|
|
|
|
|
},
|
2018-03-13 05:30:27 -07:00
|
|
|
|
// We should wait for a next block of size 20 to appear before compacting
|
2018-03-13 07:11:02 -07:00
|
|
|
|
// the existing ones. We have three, but we ignore the fresh one from WAl.
|
2018-03-13 05:30:27 -07:00
|
|
|
|
{
|
|
|
|
|
metas: []dirMeta{
|
|
|
|
|
metaRange("1", 0, 20, nil),
|
|
|
|
|
metaRange("2", 20, 40, nil),
|
|
|
|
|
metaRange("3", 40, 60, nil),
|
|
|
|
|
},
|
|
|
|
|
expected: nil,
|
|
|
|
|
},
|
2017-09-01 02:46:46 -07:00
|
|
|
|
// Block to fill the entire parent range appeared – should be compacted.
|
|
|
|
|
{
|
|
|
|
|
metas: []dirMeta{
|
|
|
|
|
metaRange("1", 0, 20, nil),
|
|
|
|
|
metaRange("2", 20, 40, nil),
|
|
|
|
|
metaRange("3", 40, 60, nil),
|
2018-03-13 05:30:27 -07:00
|
|
|
|
metaRange("4", 60, 80, nil),
|
2017-09-01 02:46:46 -07:00
|
|
|
|
},
|
|
|
|
|
expected: []string{"1", "2", "3"},
|
|
|
|
|
},
|
2018-03-13 05:30:27 -07:00
|
|
|
|
// Block for the next parent range appeared with gap with size 20. Nothing will happen in the first one
|
|
|
|
|
// anymore but we ignore fresh one still, so no compaction.
|
|
|
|
|
{
|
|
|
|
|
metas: []dirMeta{
|
|
|
|
|
metaRange("1", 0, 20, nil),
|
|
|
|
|
metaRange("2", 20, 40, nil),
|
|
|
|
|
metaRange("3", 60, 80, nil),
|
|
|
|
|
},
|
|
|
|
|
expected: nil,
|
|
|
|
|
},
|
|
|
|
|
// Block for the next parent range appeared, and we have a gap with size 20 between second and third block.
|
|
|
|
|
// We will not get this missed gap anymore and we should compact just these two.
|
2017-09-01 02:46:46 -07:00
|
|
|
|
{
|
|
|
|
|
metas: []dirMeta{
|
|
|
|
|
metaRange("1", 0, 20, nil),
|
|
|
|
|
metaRange("2", 20, 40, nil),
|
|
|
|
|
metaRange("3", 60, 80, nil),
|
2018-03-13 05:30:27 -07:00
|
|
|
|
metaRange("4", 80, 100, nil),
|
2017-09-01 02:46:46 -07:00
|
|
|
|
},
|
|
|
|
|
expected: []string{"1", "2"},
|
|
|
|
|
},
|
|
|
|
|
{
|
2018-03-13 07:11:02 -07:00
|
|
|
|
// We have 20, 20, 20, 60, 60 range blocks. "5" is marked as fresh one.
|
2017-09-01 02:46:46 -07:00
|
|
|
|
metas: []dirMeta{
|
|
|
|
|
metaRange("1", 0, 20, nil),
|
|
|
|
|
metaRange("2", 20, 40, nil),
|
|
|
|
|
metaRange("3", 40, 60, nil),
|
|
|
|
|
metaRange("4", 60, 120, nil),
|
|
|
|
|
metaRange("5", 120, 180, nil),
|
|
|
|
|
},
|
2018-03-13 07:11:02 -07:00
|
|
|
|
expected: []string{"1", "2", "3"},
|
2017-09-01 02:46:46 -07:00
|
|
|
|
},
|
|
|
|
|
{
|
2018-03-13 07:11:02 -07:00
|
|
|
|
// We have 20, 60, 20, 60, 240 range blocks. We can compact 20 + 60 + 60.
|
2017-09-01 02:46:46 -07:00
|
|
|
|
metas: []dirMeta{
|
|
|
|
|
metaRange("2", 20, 40, nil),
|
|
|
|
|
metaRange("4", 60, 120, nil),
|
2018-03-13 07:11:02 -07:00
|
|
|
|
metaRange("5", 960, 980, nil), // Fresh one.
|
|
|
|
|
metaRange("6", 120, 180, nil),
|
|
|
|
|
metaRange("7", 720, 960, nil),
|
2018-03-13 05:30:27 -07:00
|
|
|
|
},
|
2018-03-13 07:11:02 -07:00
|
|
|
|
expected: []string{"2", "4", "6"},
|
2017-09-01 02:46:46 -07:00
|
|
|
|
},
|
2018-03-13 05:30:27 -07:00
|
|
|
|
// Do not select large blocks that have many tombstones when there is no fresh block.
|
2017-09-01 02:46:46 -07:00
|
|
|
|
{
|
|
|
|
|
metas: []dirMeta{
|
2018-03-13 07:11:02 -07:00
|
|
|
|
metaRange("1", 0, 540, &BlockStats{
|
2017-09-01 02:46:46 -07:00
|
|
|
|
NumSeries: 10,
|
|
|
|
|
NumTombstones: 3,
|
|
|
|
|
}),
|
|
|
|
|
},
|
2018-03-13 05:30:27 -07:00
|
|
|
|
expected: nil,
|
|
|
|
|
},
|
|
|
|
|
// Select large blocks that have many tombstones when fresh appears.
|
|
|
|
|
{
|
|
|
|
|
metas: []dirMeta{
|
2018-03-13 07:11:02 -07:00
|
|
|
|
metaRange("1", 0, 540, &BlockStats{
|
2018-03-13 05:30:27 -07:00
|
|
|
|
NumSeries: 10,
|
|
|
|
|
NumTombstones: 3,
|
|
|
|
|
}),
|
2018-03-13 07:11:02 -07:00
|
|
|
|
metaRange("2", 540, 560, nil),
|
2018-03-13 05:30:27 -07:00
|
|
|
|
},
|
2017-09-01 02:46:46 -07:00
|
|
|
|
expected: []string{"1"},
|
|
|
|
|
},
|
2018-03-13 05:30:27 -07:00
|
|
|
|
// For small blocks, do not compact tombstones, even when fresh appears.
|
2017-09-01 02:46:46 -07:00
|
|
|
|
{
|
|
|
|
|
metas: []dirMeta{
|
2018-03-13 07:11:02 -07:00
|
|
|
|
metaRange("1", 0, 60, &BlockStats{
|
2017-09-01 02:46:46 -07:00
|
|
|
|
NumSeries: 10,
|
|
|
|
|
NumTombstones: 3,
|
|
|
|
|
}),
|
2018-03-13 07:11:02 -07:00
|
|
|
|
metaRange("2", 60, 80, nil),
|
2017-09-01 02:46:46 -07:00
|
|
|
|
},
|
|
|
|
|
expected: nil,
|
|
|
|
|
},
|
|
|
|
|
// Regression test: we were stuck in a compact loop where we always recompacted
|
|
|
|
|
// the same block when tombstones and series counts were zero.
|
|
|
|
|
{
|
|
|
|
|
metas: []dirMeta{
|
2018-03-13 07:11:02 -07:00
|
|
|
|
metaRange("1", 0, 540, &BlockStats{
|
2017-09-01 02:46:46 -07:00
|
|
|
|
NumSeries: 0,
|
|
|
|
|
NumTombstones: 0,
|
|
|
|
|
}),
|
2018-03-13 07:11:02 -07:00
|
|
|
|
metaRange("2", 540, 560, nil),
|
2017-09-01 02:46:46 -07:00
|
|
|
|
},
|
|
|
|
|
expected: nil,
|
|
|
|
|
},
|
2018-03-13 08:30:00 -07:00
|
|
|
|
// Regression test: we were wrongly assuming that new block is fresh from WAL when its ULID is newest.
|
2018-03-13 07:11:02 -07:00
|
|
|
|
// We need to actually look on max time instead.
|
|
|
|
|
//
|
|
|
|
|
// With previous, wrong approach "8" block was ignored, so we were wrongly compacting 5 and 7 and introducing
|
|
|
|
|
// block overlaps.
|
|
|
|
|
{
|
|
|
|
|
metas: []dirMeta{
|
|
|
|
|
metaRange("5", 0, 360, nil),
|
|
|
|
|
metaRange("6", 540, 560, nil), // Fresh one.
|
|
|
|
|
metaRange("7", 360, 420, nil),
|
|
|
|
|
metaRange("8", 420, 540, nil),
|
|
|
|
|
},
|
|
|
|
|
expected: []string{"7", "8"},
|
|
|
|
|
},
|
2017-09-01 02:46:46 -07:00
|
|
|
|
}
|
|
|
|
|
|
2017-12-06 16:30:24 -08:00
|
|
|
|
for _, c := range cases {
|
2018-03-13 05:30:27 -07:00
|
|
|
|
if !t.Run("", func(t *testing.T) {
|
|
|
|
|
res, err := compactor.plan(c.metas)
|
|
|
|
|
testutil.Ok(t, err)
|
2017-09-01 02:46:46 -07:00
|
|
|
|
|
2018-03-13 05:30:27 -07:00
|
|
|
|
testutil.Equals(t, c.expected, res)
|
|
|
|
|
}) {
|
|
|
|
|
return
|
|
|
|
|
}
|
2017-09-01 02:46:46 -07:00
|
|
|
|
}
|
|
|
|
|
}
|
2017-11-21 03:15:02 -08:00
|
|
|
|
|
|
|
|
|
func TestRangeWithFailedCompactionWontGetSelected(t *testing.T) {
|
|
|
|
|
compactor, err := NewLeveledCompactor(nil, nil, []int64{
|
|
|
|
|
20,
|
|
|
|
|
60,
|
|
|
|
|
240,
|
|
|
|
|
720,
|
|
|
|
|
2160,
|
|
|
|
|
}, nil)
|
2017-12-06 17:06:14 -08:00
|
|
|
|
testutil.Ok(t, err)
|
2017-11-21 03:15:02 -08:00
|
|
|
|
|
|
|
|
|
cases := []struct {
|
|
|
|
|
metas []dirMeta
|
|
|
|
|
}{
|
|
|
|
|
{
|
|
|
|
|
metas: []dirMeta{
|
|
|
|
|
metaRange("1", 0, 20, nil),
|
|
|
|
|
metaRange("2", 20, 40, nil),
|
|
|
|
|
metaRange("3", 40, 60, nil),
|
2018-03-13 07:11:02 -07:00
|
|
|
|
metaRange("4", 60, 80, nil),
|
2017-11-21 03:15:02 -08:00
|
|
|
|
},
|
|
|
|
|
},
|
|
|
|
|
{
|
|
|
|
|
metas: []dirMeta{
|
|
|
|
|
metaRange("1", 0, 20, nil),
|
|
|
|
|
metaRange("2", 20, 40, nil),
|
|
|
|
|
metaRange("3", 60, 80, nil),
|
2018-03-13 07:11:02 -07:00
|
|
|
|
metaRange("4", 80, 100, nil),
|
2017-11-21 03:15:02 -08:00
|
|
|
|
},
|
|
|
|
|
},
|
|
|
|
|
{
|
|
|
|
|
metas: []dirMeta{
|
|
|
|
|
metaRange("1", 0, 20, nil),
|
|
|
|
|
metaRange("2", 20, 40, nil),
|
|
|
|
|
metaRange("3", 40, 60, nil),
|
|
|
|
|
metaRange("4", 60, 120, nil),
|
|
|
|
|
metaRange("5", 120, 180, nil),
|
2018-03-13 07:11:02 -07:00
|
|
|
|
metaRange("6", 180, 200, nil),
|
2017-11-21 03:15:02 -08:00
|
|
|
|
},
|
|
|
|
|
},
|
|
|
|
|
}
|
|
|
|
|
|
|
|
|
|
for _, c := range cases {
|
|
|
|
|
c.metas[1].meta.Compaction.Failed = true
|
|
|
|
|
res, err := compactor.plan(c.metas)
|
2017-12-06 17:06:14 -08:00
|
|
|
|
testutil.Ok(t, err)
|
2017-11-21 03:15:02 -08:00
|
|
|
|
|
2017-12-06 17:06:14 -08:00
|
|
|
|
testutil.Equals(t, []string(nil), res)
|
2017-11-21 03:15:02 -08:00
|
|
|
|
}
|
|
|
|
|
}
|
|
|
|
|
|
|
|
|
|
func TestCompactionFailWillCleanUpTempDir(t *testing.T) {
|
|
|
|
|
compactor, err := NewLeveledCompactor(nil, log.NewNopLogger(), []int64{
|
|
|
|
|
20,
|
|
|
|
|
60,
|
|
|
|
|
240,
|
|
|
|
|
720,
|
|
|
|
|
2160,
|
|
|
|
|
}, nil)
|
2017-12-06 17:06:14 -08:00
|
|
|
|
testutil.Ok(t, err)
|
2017-11-21 03:15:02 -08:00
|
|
|
|
|
|
|
|
|
tmpdir, err := ioutil.TempDir("", "test")
|
2017-12-06 17:06:14 -08:00
|
|
|
|
testutil.Ok(t, err)
|
2018-02-23 07:04:50 -08:00
|
|
|
|
defer os.RemoveAll(tmpdir)
|
2017-11-21 03:15:02 -08:00
|
|
|
|
|
2017-12-06 17:06:14 -08:00
|
|
|
|
testutil.NotOk(t, compactor.write(tmpdir, &BlockMeta{}, erringBReader{}))
|
2017-11-21 03:15:02 -08:00
|
|
|
|
_, err = os.Stat(filepath.Join(tmpdir, BlockMeta{}.ULID.String()) + ".tmp")
|
2017-12-06 17:06:14 -08:00
|
|
|
|
testutil.Assert(t, os.IsNotExist(err), "directory is not cleaned up")
|
2017-11-21 03:15:02 -08:00
|
|
|
|
}
|
|
|
|
|
|
|
|
|
|
func metaRange(name string, mint, maxt int64, stats *BlockStats) dirMeta {
|
|
|
|
|
meta := &BlockMeta{MinTime: mint, MaxTime: maxt}
|
|
|
|
|
if stats != nil {
|
|
|
|
|
meta.Stats = *stats
|
|
|
|
|
}
|
|
|
|
|
return dirMeta{
|
|
|
|
|
dir: name,
|
|
|
|
|
meta: meta,
|
|
|
|
|
}
|
|
|
|
|
}
|
|
|
|
|
|
|
|
|
|
type erringBReader struct{}
|
|
|
|
|
|
|
|
|
|
func (erringBReader) Index() (IndexReader, error) { return nil, errors.New("index") }
|
|
|
|
|
func (erringBReader) Chunks() (ChunkReader, error) { return nil, errors.New("chunks") }
|
|
|
|
|
func (erringBReader) Tombstones() (TombstoneReader, error) { return nil, errors.New("tombstones") }
|
2018-10-12 02:45:19 -07:00
|
|
|
|
|
|
|
|
|
type nopChunkWriter struct{}
|
|
|
|
|
|
|
|
|
|
func (nopChunkWriter) WriteChunks(chunks ...chunks.Meta) error { return nil }
|
|
|
|
|
func (nopChunkWriter) Close() error { return nil }
|
|
|
|
|
|
|
|
|
|
func TestCompaction_populateBlock(t *testing.T) {
|
|
|
|
|
var populateBlocksCases = []struct {
|
|
|
|
|
title string
|
|
|
|
|
inputSeriesSamples [][]seriesSamples
|
|
|
|
|
compactMinTime int64
|
|
|
|
|
compactMaxTime int64 // When not defined the test runner sets a default of math.MaxInt64.
|
|
|
|
|
|
|
|
|
|
expSeriesSamples []seriesSamples
|
|
|
|
|
expErr error
|
|
|
|
|
}{
|
|
|
|
|
{
|
|
|
|
|
title: "Populate block from empty input should return error.",
|
|
|
|
|
inputSeriesSamples: [][]seriesSamples{},
|
|
|
|
|
expErr: errors.New("cannot populate block from no readers"),
|
|
|
|
|
},
|
|
|
|
|
{
|
|
|
|
|
// Populate from single block without chunks. We expect these kind of series being ignored.
|
|
|
|
|
inputSeriesSamples: [][]seriesSamples{
|
|
|
|
|
{
|
|
|
|
|
{
|
|
|
|
|
lset: map[string]string{"a": "b"},
|
|
|
|
|
},
|
|
|
|
|
},
|
|
|
|
|
},
|
|
|
|
|
},
|
|
|
|
|
{
|
|
|
|
|
title: "Populate from single block. We expect the same samples at the output.",
|
|
|
|
|
inputSeriesSamples: [][]seriesSamples{
|
|
|
|
|
{
|
|
|
|
|
{
|
|
|
|
|
lset: map[string]string{"a": "b"},
|
|
|
|
|
chunks: [][]sample{{{t: 0}, {t: 10}}, {{t: 11}, {t: 20}}},
|
|
|
|
|
},
|
|
|
|
|
},
|
|
|
|
|
},
|
|
|
|
|
expSeriesSamples: []seriesSamples{
|
|
|
|
|
{
|
|
|
|
|
lset: map[string]string{"a": "b"},
|
|
|
|
|
chunks: [][]sample{{{t: 0}, {t: 10}}, {{t: 11}, {t: 20}}},
|
|
|
|
|
},
|
|
|
|
|
},
|
|
|
|
|
},
|
|
|
|
|
{
|
|
|
|
|
title: "Populate from two blocks.",
|
|
|
|
|
inputSeriesSamples: [][]seriesSamples{
|
|
|
|
|
{
|
|
|
|
|
{
|
|
|
|
|
lset: map[string]string{"a": "b"},
|
|
|
|
|
chunks: [][]sample{{{t: 0}, {t: 10}}, {{t: 11}, {t: 20}}},
|
|
|
|
|
},
|
|
|
|
|
{
|
|
|
|
|
lset: map[string]string{"a": "c"},
|
|
|
|
|
chunks: [][]sample{{{t: 1}, {t: 9}}, {{t: 10}, {t: 19}}},
|
|
|
|
|
},
|
|
|
|
|
{
|
|
|
|
|
// no-chunk series should be dropped.
|
|
|
|
|
lset: map[string]string{"a": "empty"},
|
|
|
|
|
},
|
|
|
|
|
},
|
|
|
|
|
{
|
|
|
|
|
{
|
|
|
|
|
lset: map[string]string{"a": "b"},
|
|
|
|
|
chunks: [][]sample{{{t: 21}, {t: 30}}},
|
|
|
|
|
},
|
|
|
|
|
{
|
|
|
|
|
lset: map[string]string{"a": "c"},
|
|
|
|
|
chunks: [][]sample{{{t: 40}, {t: 45}}},
|
|
|
|
|
},
|
|
|
|
|
},
|
|
|
|
|
},
|
|
|
|
|
expSeriesSamples: []seriesSamples{
|
|
|
|
|
{
|
|
|
|
|
lset: map[string]string{"a": "b"},
|
|
|
|
|
chunks: [][]sample{{{t: 0}, {t: 10}}, {{t: 11}, {t: 20}}, {{t: 21}, {t: 30}}},
|
|
|
|
|
},
|
|
|
|
|
{
|
|
|
|
|
lset: map[string]string{"a": "c"},
|
|
|
|
|
chunks: [][]sample{{{t: 1}, {t: 9}}, {{t: 10}, {t: 19}}, {{t: 40}, {t: 45}}},
|
|
|
|
|
},
|
|
|
|
|
},
|
|
|
|
|
},
|
|
|
|
|
{
|
|
|
|
|
title: "Populate from two blocks showing that order is maintained.",
|
|
|
|
|
inputSeriesSamples: [][]seriesSamples{
|
|
|
|
|
{
|
|
|
|
|
{
|
|
|
|
|
lset: map[string]string{"a": "b"},
|
|
|
|
|
chunks: [][]sample{{{t: 21}, {t: 30}}},
|
|
|
|
|
},
|
|
|
|
|
{
|
|
|
|
|
lset: map[string]string{"a": "c"},
|
|
|
|
|
chunks: [][]sample{{{t: 40}, {t: 45}}},
|
|
|
|
|
},
|
|
|
|
|
},
|
|
|
|
|
{
|
|
|
|
|
{
|
|
|
|
|
lset: map[string]string{"a": "b"},
|
|
|
|
|
chunks: [][]sample{{{t: 0}, {t: 10}}, {{t: 11}, {t: 20}}},
|
|
|
|
|
},
|
|
|
|
|
{
|
|
|
|
|
lset: map[string]string{"a": "c"},
|
|
|
|
|
chunks: [][]sample{{{t: 1}, {t: 9}}, {{t: 10}, {t: 19}}},
|
|
|
|
|
},
|
|
|
|
|
},
|
|
|
|
|
},
|
|
|
|
|
expSeriesSamples: []seriesSamples{
|
|
|
|
|
{
|
|
|
|
|
lset: map[string]string{"a": "b"},
|
|
|
|
|
chunks: [][]sample{{{t: 21}, {t: 30}}, {{t: 0}, {t: 10}}, {{t: 11}, {t: 20}}},
|
|
|
|
|
},
|
|
|
|
|
{
|
|
|
|
|
lset: map[string]string{"a": "c"},
|
|
|
|
|
chunks: [][]sample{{{t: 40}, {t: 45}}, {{t: 1}, {t: 9}}, {{t: 10}, {t: 19}}},
|
|
|
|
|
},
|
|
|
|
|
},
|
|
|
|
|
},
|
|
|
|
|
{
|
|
|
|
|
title: "Populate from two blocks showing that order or series is sorted.",
|
|
|
|
|
inputSeriesSamples: [][]seriesSamples{
|
|
|
|
|
{
|
|
|
|
|
{
|
|
|
|
|
lset: map[string]string{"a": "4"},
|
|
|
|
|
chunks: [][]sample{{{t: 5}, {t: 7}}},
|
|
|
|
|
},
|
|
|
|
|
{
|
|
|
|
|
lset: map[string]string{"a": "3"},
|
|
|
|
|
chunks: [][]sample{{{t: 5}, {t: 6}}},
|
|
|
|
|
},
|
|
|
|
|
{
|
|
|
|
|
lset: map[string]string{"a": "same"},
|
|
|
|
|
chunks: [][]sample{{{t: 1}, {t: 4}}},
|
|
|
|
|
},
|
|
|
|
|
},
|
|
|
|
|
{
|
|
|
|
|
{
|
|
|
|
|
lset: map[string]string{"a": "2"},
|
|
|
|
|
chunks: [][]sample{{{t: 1}, {t: 3}}},
|
|
|
|
|
},
|
|
|
|
|
{
|
|
|
|
|
lset: map[string]string{"a": "1"},
|
|
|
|
|
chunks: [][]sample{{{t: 1}, {t: 2}}},
|
|
|
|
|
},
|
|
|
|
|
{
|
|
|
|
|
lset: map[string]string{"a": "same"},
|
|
|
|
|
chunks: [][]sample{{{t: 5}, {t: 8}}},
|
|
|
|
|
},
|
|
|
|
|
},
|
|
|
|
|
},
|
|
|
|
|
expSeriesSamples: []seriesSamples{
|
|
|
|
|
{
|
|
|
|
|
lset: map[string]string{"a": "1"},
|
|
|
|
|
chunks: [][]sample{{{t: 1}, {t: 2}}},
|
|
|
|
|
},
|
|
|
|
|
{
|
|
|
|
|
lset: map[string]string{"a": "2"},
|
|
|
|
|
chunks: [][]sample{{{t: 1}, {t: 3}}},
|
|
|
|
|
},
|
|
|
|
|
{
|
|
|
|
|
lset: map[string]string{"a": "3"},
|
|
|
|
|
chunks: [][]sample{{{t: 5}, {t: 6}}},
|
|
|
|
|
},
|
|
|
|
|
{
|
|
|
|
|
lset: map[string]string{"a": "4"},
|
|
|
|
|
chunks: [][]sample{{{t: 5}, {t: 7}}},
|
|
|
|
|
},
|
|
|
|
|
{
|
|
|
|
|
lset: map[string]string{"a": "same"},
|
|
|
|
|
chunks: [][]sample{{{t: 1}, {t: 4}}, {{t: 5}, {t: 8}}},
|
|
|
|
|
},
|
|
|
|
|
},
|
|
|
|
|
},
|
|
|
|
|
{
|
|
|
|
|
// This should not happened because head block is making sure the chunks are not crossing block boundaries.
|
|
|
|
|
title: "Populate from single block containing chunk outside of compact meta time range.",
|
|
|
|
|
inputSeriesSamples: [][]seriesSamples{
|
|
|
|
|
{
|
|
|
|
|
{
|
|
|
|
|
lset: map[string]string{"a": "b"},
|
|
|
|
|
chunks: [][]sample{{{t: 1}, {t: 2}}, {{t: 10}, {t: 30}}},
|
|
|
|
|
},
|
|
|
|
|
},
|
|
|
|
|
},
|
|
|
|
|
compactMinTime: 0,
|
|
|
|
|
compactMaxTime: 20,
|
|
|
|
|
expErr: errors.New("found chunk with minTime: 10 maxTime: 30 outside of compacted minTime: 0 maxTime: 20"),
|
|
|
|
|
},
|
|
|
|
|
{
|
|
|
|
|
// Introduced by https://github.com/prometheus/tsdb/issues/347.
|
|
|
|
|
title: "Populate from single block containing extra chunk",
|
|
|
|
|
inputSeriesSamples: [][]seriesSamples{
|
|
|
|
|
{
|
|
|
|
|
{
|
|
|
|
|
lset: map[string]string{"a": "issue347"},
|
|
|
|
|
chunks: [][]sample{{{t: 1}, {t: 2}}, {{t: 10}, {t: 20}}},
|
|
|
|
|
},
|
|
|
|
|
},
|
|
|
|
|
},
|
|
|
|
|
compactMinTime: 0,
|
|
|
|
|
compactMaxTime: 10,
|
|
|
|
|
expErr: errors.New("found chunk with minTime: 10 maxTime: 20 outside of compacted minTime: 0 maxTime: 10"),
|
|
|
|
|
},
|
|
|
|
|
{
|
|
|
|
|
// No special deduplication expected.
|
|
|
|
|
title: "Populate from two blocks containing duplicated chunk.",
|
|
|
|
|
inputSeriesSamples: [][]seriesSamples{
|
|
|
|
|
{
|
|
|
|
|
{
|
|
|
|
|
lset: map[string]string{"a": "b"},
|
|
|
|
|
chunks: [][]sample{{{t: 1}, {t: 2}}, {{t: 10}, {t: 20}}},
|
|
|
|
|
},
|
|
|
|
|
},
|
|
|
|
|
{
|
|
|
|
|
{
|
|
|
|
|
lset: map[string]string{"a": "b"},
|
|
|
|
|
chunks: [][]sample{{{t: 10}, {t: 20}}},
|
|
|
|
|
},
|
|
|
|
|
},
|
|
|
|
|
},
|
|
|
|
|
expSeriesSamples: []seriesSamples{
|
|
|
|
|
{
|
|
|
|
|
lset: map[string]string{"a": "b"},
|
|
|
|
|
chunks: [][]sample{{{t: 1}, {t: 2}}, {{t: 10}, {t: 20}}, {{t: 10}, {t: 20}}},
|
|
|
|
|
},
|
|
|
|
|
},
|
|
|
|
|
},
|
|
|
|
|
}
|
|
|
|
|
|
|
|
|
|
for _, tc := range populateBlocksCases {
|
|
|
|
|
if ok := t.Run(tc.title, func(t *testing.T) {
|
|
|
|
|
blocks := make([]BlockReader, 0, len(tc.inputSeriesSamples))
|
|
|
|
|
for _, b := range tc.inputSeriesSamples {
|
|
|
|
|
ir, cr := createIdxChkReaders(b)
|
|
|
|
|
blocks = append(blocks, &mockBReader{ir: ir, cr: cr})
|
|
|
|
|
}
|
|
|
|
|
|
|
|
|
|
c, err := NewLeveledCompactor(nil, nil, []int64{0}, nil)
|
|
|
|
|
testutil.Ok(t, err)
|
|
|
|
|
|
|
|
|
|
meta := &BlockMeta{
|
|
|
|
|
MinTime: tc.compactMinTime,
|
|
|
|
|
MaxTime: tc.compactMaxTime,
|
|
|
|
|
}
|
|
|
|
|
if meta.MaxTime == 0 {
|
|
|
|
|
meta.MaxTime = math.MaxInt64
|
|
|
|
|
}
|
|
|
|
|
|
|
|
|
|
iw := &mockIndexWriter{}
|
|
|
|
|
err = c.populateBlock(blocks, meta, iw, nopChunkWriter{})
|
|
|
|
|
if tc.expErr != nil {
|
|
|
|
|
testutil.NotOk(t, err)
|
|
|
|
|
testutil.Equals(t, tc.expErr.Error(), err.Error())
|
|
|
|
|
return
|
|
|
|
|
}
|
|
|
|
|
testutil.Ok(t, err)
|
|
|
|
|
|
|
|
|
|
testutil.Equals(t, tc.expSeriesSamples, iw.series)
|
|
|
|
|
|
|
|
|
|
// Check if stats are calculated properly.
|
|
|
|
|
s := BlockStats{
|
|
|
|
|
NumSeries: uint64(len(tc.expSeriesSamples)),
|
|
|
|
|
}
|
|
|
|
|
for _, series := range tc.expSeriesSamples {
|
|
|
|
|
s.NumChunks += uint64(len(series.chunks))
|
|
|
|
|
for _, chk := range series.chunks {
|
|
|
|
|
s.NumSamples += uint64(len(chk))
|
|
|
|
|
}
|
|
|
|
|
}
|
|
|
|
|
testutil.Equals(t, s, meta.Stats)
|
|
|
|
|
}); !ok {
|
|
|
|
|
return
|
|
|
|
|
}
|
|
|
|
|
}
|
|
|
|
|
}
|
2018-11-20 02:34:26 -08:00
|
|
|
|
|
|
|
|
|
// TestDisableAutoCompactions checks that we can
|
|
|
|
|
// disable and enable the auto compaction.
|
|
|
|
|
// This is needed for unit tests that rely on
|
|
|
|
|
// checking state before and after a compaction.
|
|
|
|
|
func TestDisableAutoCompactions(t *testing.T) {
|
|
|
|
|
db, close := openTestDB(t, nil)
|
|
|
|
|
defer close()
|
|
|
|
|
defer db.Close()
|
|
|
|
|
|
|
|
|
|
blockRange := DefaultOptions.BlockRanges[0]
|
|
|
|
|
label := labels.FromStrings("foo", "bar")
|
|
|
|
|
|
|
|
|
|
// Trigger a compaction to check that it was skipped and
|
|
|
|
|
// no new blocks were created when compaction is disabled.
|
|
|
|
|
db.DisableCompactions()
|
|
|
|
|
app := db.Appender()
|
|
|
|
|
for i := int64(0); i < 3; i++ {
|
|
|
|
|
_, err := app.Add(label, i*blockRange, 0)
|
|
|
|
|
testutil.Ok(t, err)
|
|
|
|
|
_, err = app.Add(label, i*blockRange+1000, 0)
|
|
|
|
|
testutil.Ok(t, err)
|
|
|
|
|
}
|
|
|
|
|
testutil.Ok(t, app.Commit())
|
|
|
|
|
|
|
|
|
|
select {
|
|
|
|
|
case db.compactc <- struct{}{}:
|
|
|
|
|
default:
|
|
|
|
|
}
|
|
|
|
|
|
|
|
|
|
m := &dto.Metric{}
|
|
|
|
|
for x := 0; x < 10; x++ {
|
|
|
|
|
db.metrics.compactionsSkipped.Write(m)
|
|
|
|
|
if *m.Counter.Value > float64(0) {
|
|
|
|
|
break
|
|
|
|
|
}
|
|
|
|
|
time.Sleep(10 * time.Millisecond)
|
|
|
|
|
}
|
|
|
|
|
|
|
|
|
|
testutil.Assert(t, *m.Counter.Value > float64(0), "No compaction was skipped after the set timeout.")
|
|
|
|
|
testutil.Equals(t, 0, len(db.blocks))
|
|
|
|
|
|
|
|
|
|
// Enable the compaction, trigger it and check that the block is persisted.
|
|
|
|
|
db.EnableCompactions()
|
|
|
|
|
select {
|
|
|
|
|
case db.compactc <- struct{}{}:
|
|
|
|
|
default:
|
|
|
|
|
}
|
2018-12-12 03:49:03 -08:00
|
|
|
|
for x := 0; x < 100; x++ {
|
2018-11-20 02:34:26 -08:00
|
|
|
|
if len(db.Blocks()) > 0 {
|
|
|
|
|
break
|
|
|
|
|
}
|
2018-11-20 14:33:14 -08:00
|
|
|
|
time.Sleep(100 * time.Millisecond)
|
2018-11-20 02:34:26 -08:00
|
|
|
|
}
|
|
|
|
|
testutil.Assert(t, len(db.Blocks()) > 0, "No block was persisted after the set timeout.")
|
|
|
|
|
}
|
2019-01-29 00:26:01 -08:00
|
|
|
|
|
|
|
|
|
// TestDeleteCompactionBlockAfterFailedReload ensures that a failed reload imidiately after a compaction
|
|
|
|
|
// deletes the resulting block to avoid creatings blocks with the same time range.
|
|
|
|
|
func TestDeleteCompactionBlockAfterFailedReload(t *testing.T) {
|
|
|
|
|
|
|
|
|
|
tests := map[string]func(*DB) int{
|
|
|
|
|
"Test Head Compaction": func(db *DB) int {
|
|
|
|
|
rangeToTriggerCompaction := db.opts.BlockRanges[0]/2*3 - 1
|
|
|
|
|
defaultLabel := labels.FromStrings("foo", "bar")
|
|
|
|
|
|
|
|
|
|
// Add some data to the head that is enough to trigger a compaction.
|
|
|
|
|
app := db.Appender()
|
|
|
|
|
_, err := app.Add(defaultLabel, 1, 0)
|
|
|
|
|
testutil.Ok(t, err)
|
|
|
|
|
_, err = app.Add(defaultLabel, 2, 0)
|
|
|
|
|
testutil.Ok(t, err)
|
|
|
|
|
_, err = app.Add(defaultLabel, 3+rangeToTriggerCompaction, 0)
|
|
|
|
|
testutil.Ok(t, err)
|
|
|
|
|
testutil.Ok(t, app.Commit())
|
|
|
|
|
|
|
|
|
|
return 1
|
|
|
|
|
},
|
|
|
|
|
"Test Block Compaction": func(db *DB) int {
|
|
|
|
|
expBlocks := []*BlockMeta{
|
|
|
|
|
{MinTime: 0, MaxTime: 100},
|
|
|
|
|
{MinTime: 100, MaxTime: 150},
|
|
|
|
|
{MinTime: 150, MaxTime: 200},
|
|
|
|
|
}
|
|
|
|
|
for _, m := range expBlocks {
|
|
|
|
|
createBlock(t, db.Dir(), genSeries(1, 1, m.MinTime, m.MaxTime))
|
|
|
|
|
}
|
|
|
|
|
testutil.Ok(t, db.reload())
|
|
|
|
|
testutil.Equals(t, len(expBlocks), len(db.Blocks()), "unexpected block count after a reload")
|
|
|
|
|
|
|
|
|
|
return len(expBlocks) + 1
|
|
|
|
|
},
|
|
|
|
|
}
|
|
|
|
|
|
|
|
|
|
for title, bootStrap := range tests {
|
|
|
|
|
t.Run(title, func(t *testing.T) {
|
|
|
|
|
db, close := openTestDB(t, &Options{
|
|
|
|
|
BlockRanges: []int64{1, 100},
|
|
|
|
|
})
|
|
|
|
|
defer close()
|
|
|
|
|
defer db.Close()
|
|
|
|
|
db.DisableCompactions()
|
|
|
|
|
|
|
|
|
|
expBlocks := bootStrap(db)
|
|
|
|
|
|
|
|
|
|
// Create a block that will trigger the reloard to fail.
|
|
|
|
|
blockPath := createBlock(t, db.Dir(), genSeries(1, 1, 200, 300))
|
|
|
|
|
lastBlockIndex := path.Join(blockPath, indexFilename)
|
|
|
|
|
actBlocks, err := blockDirs(db.Dir())
|
|
|
|
|
testutil.Ok(t, err)
|
|
|
|
|
testutil.Equals(t, expBlocks, len(actBlocks))
|
|
|
|
|
testutil.Ok(t, os.RemoveAll(lastBlockIndex)) // Corrupt the block by removing the index file.
|
|
|
|
|
|
|
|
|
|
testutil.Equals(t, 0.0, prom_testutil.ToFloat64(db.metrics.reloadsFailed), "initial 'failed db reload' count metrics mismatch")
|
|
|
|
|
testutil.Equals(t, 0.0, prom_testutil.ToFloat64(db.compactor.(*LeveledCompactor).metrics.ran), "initial `compactions` count metric mismatch")
|
|
|
|
|
|
|
|
|
|
// Do the compaction and check the metrics.
|
|
|
|
|
// Since the most recent block is not included in the compaction,
|
|
|
|
|
// the compaction should succeed, but the reload should fail and
|
|
|
|
|
// the new block created from the compaction should be deleted.
|
|
|
|
|
db.EnableCompactions()
|
|
|
|
|
testutil.NotOk(t, db.compact())
|
|
|
|
|
testutil.Equals(t, 1.0, prom_testutil.ToFloat64(db.metrics.reloadsFailed), "'failed db reload' count metrics mismatch")
|
|
|
|
|
testutil.Equals(t, 1.0, prom_testutil.ToFloat64(db.compactor.(*LeveledCompactor).metrics.ran), "`compaction` count metric mismatch")
|
|
|
|
|
actBlocks, err = blockDirs(db.Dir())
|
|
|
|
|
testutil.Ok(t, err)
|
|
|
|
|
testutil.Equals(t, expBlocks, len(actBlocks))
|
|
|
|
|
})
|
|
|
|
|
}
|
|
|
|
|
}
|