mirror of
https://github.com/prometheus/prometheus.git
synced 2024-12-26 22:19:40 -08:00
[Cleanup] TSDB: Remove old deprecated WAL implementation
Deprecated since 2018. Signed-off-by: Bryan Boreham <bjboreham@gmail.com>
This commit is contained in:
parent
312e3fd728
commit
87edf1f960
|
@ -779,10 +779,6 @@ func open(dir string, l log.Logger, r prometheus.Registerer, opts *Options, rngs
|
||||||
walDir := filepath.Join(dir, "wal")
|
walDir := filepath.Join(dir, "wal")
|
||||||
wblDir := filepath.Join(dir, wlog.WblDirName)
|
wblDir := filepath.Join(dir, wlog.WblDirName)
|
||||||
|
|
||||||
// Migrate old WAL if one exists.
|
|
||||||
if err := MigrateWAL(l, walDir); err != nil {
|
|
||||||
return nil, fmt.Errorf("migrate WAL: %w", err)
|
|
||||||
}
|
|
||||||
for _, tmpDir := range []string{walDir, dir} {
|
for _, tmpDir := range []string{walDir, dir} {
|
||||||
// Remove tmp dirs.
|
// Remove tmp dirs.
|
||||||
if err := removeBestEffortTmpDirs(l, tmpDir); err != nil {
|
if err := removeBestEffortTmpDirs(l, tmpDir); err != nil {
|
||||||
|
|
|
@ -3598,7 +3598,7 @@ func testChunkQuerierShouldNotPanicIfHeadChunkIsTruncatedWhileReadingQueriedChun
|
||||||
// just to iterate through the bytes slice. We don't really care the reason why
|
// just to iterate through the bytes slice. We don't really care the reason why
|
||||||
// we read this data, we just need to read it to make sure the memory address
|
// we read this data, we just need to read it to make sure the memory address
|
||||||
// of the []byte is still valid.
|
// of the []byte is still valid.
|
||||||
chkCRC32 := newCRC32()
|
chkCRC32 := crc32.New(crc32.MakeTable(crc32.Castagnoli))
|
||||||
for _, chunk := range chunks {
|
for _, chunk := range chunks {
|
||||||
chkCRC32.Reset()
|
chkCRC32.Reset()
|
||||||
_, err := chkCRC32.Write(chunk.Bytes())
|
_, err := chkCRC32.Write(chunk.Bytes())
|
||||||
|
|
1303
tsdb/wal.go
1303
tsdb/wal.go
File diff suppressed because it is too large
Load diff
553
tsdb/wal_test.go
553
tsdb/wal_test.go
|
@ -1,553 +0,0 @@
|
||||||
// Copyright 2017 The Prometheus Authors
|
|
||||||
// Licensed under the Apache License, Version 2.0 (the "License");
|
|
||||||
// you may not use this file except in compliance with the License.
|
|
||||||
// You may obtain a copy of the License at
|
|
||||||
//
|
|
||||||
// http://www.apache.org/licenses/LICENSE-2.0
|
|
||||||
//
|
|
||||||
// Unless required by applicable law or agreed to in writing, software
|
|
||||||
// distributed under the License is distributed on an "AS IS" BASIS,
|
|
||||||
// WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
|
|
||||||
// See the License for the specific language governing permissions and
|
|
||||||
// limitations under the License.
|
|
||||||
|
|
||||||
//go:build !windows
|
|
||||||
|
|
||||||
package tsdb
|
|
||||||
|
|
||||||
import (
|
|
||||||
"encoding/binary"
|
|
||||||
"io"
|
|
||||||
"math/rand"
|
|
||||||
"os"
|
|
||||||
"path"
|
|
||||||
"path/filepath"
|
|
||||||
"testing"
|
|
||||||
"time"
|
|
||||||
|
|
||||||
"github.com/go-kit/log"
|
|
||||||
"github.com/stretchr/testify/require"
|
|
||||||
|
|
||||||
"github.com/prometheus/prometheus/model/labels"
|
|
||||||
"github.com/prometheus/prometheus/storage"
|
|
||||||
"github.com/prometheus/prometheus/tsdb/chunks"
|
|
||||||
"github.com/prometheus/prometheus/tsdb/record"
|
|
||||||
"github.com/prometheus/prometheus/tsdb/tombstones"
|
|
||||||
"github.com/prometheus/prometheus/tsdb/wlog"
|
|
||||||
"github.com/prometheus/prometheus/util/testutil"
|
|
||||||
)
|
|
||||||
|
|
||||||
func TestSegmentWAL_cut(t *testing.T) {
|
|
||||||
tmpdir := t.TempDir()
|
|
||||||
|
|
||||||
// This calls cut() implicitly the first time without a previous tail.
|
|
||||||
w, err := OpenSegmentWAL(tmpdir, nil, 0, nil)
|
|
||||||
require.NoError(t, err)
|
|
||||||
|
|
||||||
require.NoError(t, w.write(WALEntrySeries, 1, []byte("Hello World!!")))
|
|
||||||
|
|
||||||
require.NoError(t, w.cut())
|
|
||||||
|
|
||||||
// Cutting creates a new file.
|
|
||||||
require.Len(t, w.files, 2)
|
|
||||||
|
|
||||||
require.NoError(t, w.write(WALEntrySeries, 1, []byte("Hello World!!")))
|
|
||||||
|
|
||||||
require.NoError(t, w.Close())
|
|
||||||
|
|
||||||
for _, of := range w.files {
|
|
||||||
f, err := os.Open(of.Name())
|
|
||||||
require.NoError(t, err)
|
|
||||||
|
|
||||||
// Verify header data.
|
|
||||||
metab := make([]byte, 8)
|
|
||||||
_, err = f.Read(metab)
|
|
||||||
require.NoError(t, err)
|
|
||||||
require.Equal(t, WALMagic, binary.BigEndian.Uint32(metab[:4]))
|
|
||||||
require.Equal(t, WALFormatDefault, metab[4])
|
|
||||||
|
|
||||||
// We cannot actually check for correct pre-allocation as it is
|
|
||||||
// optional per filesystem and handled transparently.
|
|
||||||
et, flag, b, err := newWALReader(nil, nil).entry(f)
|
|
||||||
require.NoError(t, err)
|
|
||||||
require.Equal(t, WALEntrySeries, et)
|
|
||||||
require.Equal(t, byte(walSeriesSimple), flag)
|
|
||||||
require.Equal(t, []byte("Hello World!!"), b)
|
|
||||||
}
|
|
||||||
}
|
|
||||||
|
|
||||||
func TestSegmentWAL_Truncate(t *testing.T) {
|
|
||||||
const (
|
|
||||||
numMetrics = 20000
|
|
||||||
batch = 100
|
|
||||||
)
|
|
||||||
series, err := labels.ReadLabels(filepath.Join("testdata", "20kseries.json"), numMetrics)
|
|
||||||
require.NoError(t, err)
|
|
||||||
|
|
||||||
dir := t.TempDir()
|
|
||||||
|
|
||||||
w, err := OpenSegmentWAL(dir, nil, 0, nil)
|
|
||||||
require.NoError(t, err)
|
|
||||||
defer func(wal *SegmentWAL) { require.NoError(t, wal.Close()) }(w)
|
|
||||||
w.segmentSize = 10000
|
|
||||||
|
|
||||||
for i := 0; i < numMetrics; i += batch {
|
|
||||||
var rs []record.RefSeries
|
|
||||||
|
|
||||||
for j, s := range series[i : i+batch] {
|
|
||||||
rs = append(rs, record.RefSeries{Labels: s, Ref: chunks.HeadSeriesRef(i+j) + 1})
|
|
||||||
}
|
|
||||||
err := w.LogSeries(rs)
|
|
||||||
require.NoError(t, err)
|
|
||||||
}
|
|
||||||
|
|
||||||
// We mark the 2nd half of the files with a min timestamp that should discard
|
|
||||||
// them from the selection of compactable files.
|
|
||||||
for i, f := range w.files[len(w.files)/2:] {
|
|
||||||
f.maxTime = int64(1000 + i)
|
|
||||||
}
|
|
||||||
// All series in those files must be preserved regarding of the provided postings list.
|
|
||||||
boundarySeries := w.files[len(w.files)/2].minSeries
|
|
||||||
|
|
||||||
// We truncate while keeping every 2nd series.
|
|
||||||
keep := map[chunks.HeadSeriesRef]struct{}{}
|
|
||||||
for i := 1; i <= numMetrics; i += 2 {
|
|
||||||
keep[chunks.HeadSeriesRef(i)] = struct{}{}
|
|
||||||
}
|
|
||||||
keepf := func(id chunks.HeadSeriesRef) bool {
|
|
||||||
_, ok := keep[id]
|
|
||||||
return ok
|
|
||||||
}
|
|
||||||
|
|
||||||
err = w.Truncate(1000, keepf)
|
|
||||||
require.NoError(t, err)
|
|
||||||
|
|
||||||
var expected []record.RefSeries
|
|
||||||
|
|
||||||
for i := 1; i <= numMetrics; i++ {
|
|
||||||
if i%2 == 1 || chunks.HeadSeriesRef(i) >= boundarySeries {
|
|
||||||
expected = append(expected, record.RefSeries{Ref: chunks.HeadSeriesRef(i), Labels: series[i-1]})
|
|
||||||
}
|
|
||||||
}
|
|
||||||
|
|
||||||
// Call Truncate once again to see whether we can read the written file without
|
|
||||||
// creating a new WAL.
|
|
||||||
err = w.Truncate(1000, keepf)
|
|
||||||
require.NoError(t, err)
|
|
||||||
require.NoError(t, w.Close())
|
|
||||||
|
|
||||||
// The same again with a new WAL.
|
|
||||||
w, err = OpenSegmentWAL(dir, nil, 0, nil)
|
|
||||||
require.NoError(t, err)
|
|
||||||
defer func(wal *SegmentWAL) { require.NoError(t, wal.Close()) }(w)
|
|
||||||
|
|
||||||
var readSeries []record.RefSeries
|
|
||||||
r := w.Reader()
|
|
||||||
|
|
||||||
require.NoError(t, r.Read(func(s []record.RefSeries) {
|
|
||||||
readSeries = append(readSeries, s...)
|
|
||||||
}, nil, nil))
|
|
||||||
|
|
||||||
testutil.RequireEqual(t, expected, readSeries)
|
|
||||||
}
|
|
||||||
|
|
||||||
// Symmetrical test of reading and writing to the WAL via its main interface.
|
|
||||||
func TestSegmentWAL_Log_Restore(t *testing.T) {
|
|
||||||
const (
|
|
||||||
numMetrics = 50
|
|
||||||
iterations = 5
|
|
||||||
stepSize = 5
|
|
||||||
)
|
|
||||||
// Generate testing data. It does not make semantic sense but
|
|
||||||
// for the purpose of this test.
|
|
||||||
series, err := labels.ReadLabels(filepath.Join("testdata", "20kseries.json"), numMetrics)
|
|
||||||
require.NoError(t, err)
|
|
||||||
|
|
||||||
dir := t.TempDir()
|
|
||||||
|
|
||||||
var (
|
|
||||||
recordedSeries [][]record.RefSeries
|
|
||||||
recordedSamples [][]record.RefSample
|
|
||||||
recordedDeletes [][]tombstones.Stone
|
|
||||||
)
|
|
||||||
var totalSamples int
|
|
||||||
|
|
||||||
// Open WAL a bunch of times, validate all previous data can be read,
|
|
||||||
// write more data to it, close it.
|
|
||||||
for k := 0; k < numMetrics; k += numMetrics / iterations {
|
|
||||||
w, err := OpenSegmentWAL(dir, nil, 0, nil)
|
|
||||||
require.NoError(t, err)
|
|
||||||
|
|
||||||
// Set smaller segment size so we can actually write several files.
|
|
||||||
w.segmentSize = 1000 * 1000
|
|
||||||
|
|
||||||
r := w.Reader()
|
|
||||||
|
|
||||||
var (
|
|
||||||
resultSeries [][]record.RefSeries
|
|
||||||
resultSamples [][]record.RefSample
|
|
||||||
resultDeletes [][]tombstones.Stone
|
|
||||||
)
|
|
||||||
|
|
||||||
serf := func(series []record.RefSeries) {
|
|
||||||
if len(series) > 0 {
|
|
||||||
clsets := make([]record.RefSeries, len(series))
|
|
||||||
copy(clsets, series)
|
|
||||||
resultSeries = append(resultSeries, clsets)
|
|
||||||
}
|
|
||||||
}
|
|
||||||
smplf := func(smpls []record.RefSample) {
|
|
||||||
if len(smpls) > 0 {
|
|
||||||
csmpls := make([]record.RefSample, len(smpls))
|
|
||||||
copy(csmpls, smpls)
|
|
||||||
resultSamples = append(resultSamples, csmpls)
|
|
||||||
}
|
|
||||||
}
|
|
||||||
|
|
||||||
delf := func(stones []tombstones.Stone) {
|
|
||||||
if len(stones) > 0 {
|
|
||||||
cst := make([]tombstones.Stone, len(stones))
|
|
||||||
copy(cst, stones)
|
|
||||||
resultDeletes = append(resultDeletes, cst)
|
|
||||||
}
|
|
||||||
}
|
|
||||||
|
|
||||||
require.NoError(t, r.Read(serf, smplf, delf))
|
|
||||||
|
|
||||||
testutil.RequireEqual(t, recordedSamples, resultSamples)
|
|
||||||
testutil.RequireEqual(t, recordedSeries, resultSeries)
|
|
||||||
testutil.RequireEqual(t, recordedDeletes, resultDeletes)
|
|
||||||
|
|
||||||
series := series[k : k+(numMetrics/iterations)]
|
|
||||||
|
|
||||||
// Insert in batches and generate different amounts of samples for each.
|
|
||||||
for i := 0; i < len(series); i += stepSize {
|
|
||||||
var samples []record.RefSample
|
|
||||||
var stones []tombstones.Stone
|
|
||||||
|
|
||||||
for j := 0; j < i*10; j++ {
|
|
||||||
samples = append(samples, record.RefSample{
|
|
||||||
Ref: chunks.HeadSeriesRef(j % 10000),
|
|
||||||
T: int64(j * 2),
|
|
||||||
V: rand.Float64(),
|
|
||||||
})
|
|
||||||
}
|
|
||||||
|
|
||||||
for j := 0; j < i*20; j++ {
|
|
||||||
ts := rand.Int63()
|
|
||||||
stones = append(stones, tombstones.Stone{Ref: storage.SeriesRef(rand.Uint64()), Intervals: tombstones.Intervals{{Mint: ts, Maxt: ts + rand.Int63n(10000)}}})
|
|
||||||
}
|
|
||||||
|
|
||||||
lbls := series[i : i+stepSize]
|
|
||||||
series := make([]record.RefSeries, 0, len(series))
|
|
||||||
for j, l := range lbls {
|
|
||||||
series = append(series, record.RefSeries{
|
|
||||||
Ref: chunks.HeadSeriesRef(i + j),
|
|
||||||
Labels: l,
|
|
||||||
})
|
|
||||||
}
|
|
||||||
|
|
||||||
require.NoError(t, w.LogSeries(series))
|
|
||||||
require.NoError(t, w.LogSamples(samples))
|
|
||||||
require.NoError(t, w.LogDeletes(stones))
|
|
||||||
|
|
||||||
if len(lbls) > 0 {
|
|
||||||
recordedSeries = append(recordedSeries, series)
|
|
||||||
}
|
|
||||||
if len(samples) > 0 {
|
|
||||||
recordedSamples = append(recordedSamples, samples)
|
|
||||||
totalSamples += len(samples)
|
|
||||||
}
|
|
||||||
if len(stones) > 0 {
|
|
||||||
recordedDeletes = append(recordedDeletes, stones)
|
|
||||||
}
|
|
||||||
}
|
|
||||||
|
|
||||||
require.NoError(t, w.Close())
|
|
||||||
}
|
|
||||||
}
|
|
||||||
|
|
||||||
func TestWALRestoreCorrupted_invalidSegment(t *testing.T) {
|
|
||||||
dir := t.TempDir()
|
|
||||||
|
|
||||||
wal, err := OpenSegmentWAL(dir, nil, 0, nil)
|
|
||||||
require.NoError(t, err)
|
|
||||||
defer func(wal *SegmentWAL) { require.NoError(t, wal.Close()) }(wal)
|
|
||||||
|
|
||||||
_, err = wal.createSegmentFile(filepath.Join(dir, "000000"))
|
|
||||||
require.NoError(t, err)
|
|
||||||
f, err := wal.createSegmentFile(filepath.Join(dir, "000001"))
|
|
||||||
require.NoError(t, err)
|
|
||||||
f2, err := wal.createSegmentFile(filepath.Join(dir, "000002"))
|
|
||||||
require.NoError(t, err)
|
|
||||||
require.NoError(t, f2.Close())
|
|
||||||
|
|
||||||
// Make header of second segment invalid.
|
|
||||||
_, err = f.WriteAt([]byte{1, 2, 3, 4}, 0)
|
|
||||||
require.NoError(t, err)
|
|
||||||
require.NoError(t, f.Close())
|
|
||||||
|
|
||||||
require.NoError(t, wal.Close())
|
|
||||||
|
|
||||||
wal, err = OpenSegmentWAL(dir, log.NewLogfmtLogger(os.Stderr), 0, nil)
|
|
||||||
require.NoError(t, err)
|
|
||||||
defer func(wal *SegmentWAL) { require.NoError(t, wal.Close()) }(wal)
|
|
||||||
|
|
||||||
files, err := os.ReadDir(dir)
|
|
||||||
require.NoError(t, err)
|
|
||||||
fns := []string{}
|
|
||||||
for _, f := range files {
|
|
||||||
fns = append(fns, f.Name())
|
|
||||||
}
|
|
||||||
require.Equal(t, []string{"000000"}, fns)
|
|
||||||
}
|
|
||||||
|
|
||||||
// Test reading from a WAL that has been corrupted through various means.
|
|
||||||
func TestWALRestoreCorrupted(t *testing.T) {
|
|
||||||
cases := []struct {
|
|
||||||
name string
|
|
||||||
f func(*testing.T, *SegmentWAL)
|
|
||||||
}{
|
|
||||||
{
|
|
||||||
name: "truncate_checksum",
|
|
||||||
f: func(t *testing.T, w *SegmentWAL) {
|
|
||||||
f, err := os.OpenFile(w.files[0].Name(), os.O_WRONLY, 0o666)
|
|
||||||
require.NoError(t, err)
|
|
||||||
defer f.Close()
|
|
||||||
|
|
||||||
off, err := f.Seek(0, io.SeekEnd)
|
|
||||||
require.NoError(t, err)
|
|
||||||
|
|
||||||
require.NoError(t, f.Truncate(off-1))
|
|
||||||
},
|
|
||||||
},
|
|
||||||
{
|
|
||||||
name: "truncate_body",
|
|
||||||
f: func(t *testing.T, w *SegmentWAL) {
|
|
||||||
f, err := os.OpenFile(w.files[0].Name(), os.O_WRONLY, 0o666)
|
|
||||||
require.NoError(t, err)
|
|
||||||
defer f.Close()
|
|
||||||
|
|
||||||
off, err := f.Seek(0, io.SeekEnd)
|
|
||||||
require.NoError(t, err)
|
|
||||||
|
|
||||||
require.NoError(t, f.Truncate(off-8))
|
|
||||||
},
|
|
||||||
},
|
|
||||||
{
|
|
||||||
name: "body_content",
|
|
||||||
f: func(t *testing.T, w *SegmentWAL) {
|
|
||||||
f, err := os.OpenFile(w.files[0].Name(), os.O_WRONLY, 0o666)
|
|
||||||
require.NoError(t, err)
|
|
||||||
defer f.Close()
|
|
||||||
|
|
||||||
off, err := f.Seek(0, io.SeekEnd)
|
|
||||||
require.NoError(t, err)
|
|
||||||
|
|
||||||
// Write junk before checksum starts.
|
|
||||||
_, err = f.WriteAt([]byte{1, 2, 3, 4}, off-8)
|
|
||||||
require.NoError(t, err)
|
|
||||||
},
|
|
||||||
},
|
|
||||||
{
|
|
||||||
name: "checksum",
|
|
||||||
f: func(t *testing.T, w *SegmentWAL) {
|
|
||||||
f, err := os.OpenFile(w.files[0].Name(), os.O_WRONLY, 0o666)
|
|
||||||
require.NoError(t, err)
|
|
||||||
defer f.Close()
|
|
||||||
|
|
||||||
off, err := f.Seek(0, io.SeekEnd)
|
|
||||||
require.NoError(t, err)
|
|
||||||
|
|
||||||
// Write junk into checksum
|
|
||||||
_, err = f.WriteAt([]byte{1, 2, 3, 4}, off-4)
|
|
||||||
require.NoError(t, err)
|
|
||||||
},
|
|
||||||
},
|
|
||||||
}
|
|
||||||
for _, c := range cases {
|
|
||||||
t.Run(c.name, func(t *testing.T) {
|
|
||||||
// Generate testing data. It does not make semantic sense but
|
|
||||||
// for the purpose of this test.
|
|
||||||
dir := t.TempDir()
|
|
||||||
|
|
||||||
w, err := OpenSegmentWAL(dir, nil, 0, nil)
|
|
||||||
require.NoError(t, err)
|
|
||||||
defer func(wal *SegmentWAL) { require.NoError(t, wal.Close()) }(w)
|
|
||||||
|
|
||||||
require.NoError(t, w.LogSamples([]record.RefSample{{T: 1, V: 2}}))
|
|
||||||
require.NoError(t, w.LogSamples([]record.RefSample{{T: 2, V: 3}}))
|
|
||||||
|
|
||||||
require.NoError(t, w.cut())
|
|
||||||
|
|
||||||
// Sleep 2 seconds to avoid error where cut and test "cases" function may write or
|
|
||||||
// truncate the file out of orders as "cases" are not synchronized with cut.
|
|
||||||
// Hopefully cut will complete by 2 seconds.
|
|
||||||
time.Sleep(2 * time.Second)
|
|
||||||
|
|
||||||
require.NoError(t, w.LogSamples([]record.RefSample{{T: 3, V: 4}}))
|
|
||||||
require.NoError(t, w.LogSamples([]record.RefSample{{T: 5, V: 6}}))
|
|
||||||
|
|
||||||
require.NoError(t, w.Close())
|
|
||||||
|
|
||||||
// cut() truncates and fsyncs the first segment async. If it happens after
|
|
||||||
// the corruption we apply below, the corruption will be overwritten again.
|
|
||||||
// Fire and forget a sync to avoid flakiness.
|
|
||||||
w.files[0].Sync()
|
|
||||||
// Corrupt the second entry in the first file.
|
|
||||||
// After re-opening we must be able to read the first entry
|
|
||||||
// and the rest, including the second file, must be truncated for clean further
|
|
||||||
// writes.
|
|
||||||
c.f(t, w)
|
|
||||||
|
|
||||||
logger := log.NewLogfmtLogger(os.Stderr)
|
|
||||||
|
|
||||||
w2, err := OpenSegmentWAL(dir, logger, 0, nil)
|
|
||||||
require.NoError(t, err)
|
|
||||||
defer func(wal *SegmentWAL) { require.NoError(t, wal.Close()) }(w2)
|
|
||||||
|
|
||||||
r := w2.Reader()
|
|
||||||
|
|
||||||
serf := func(l []record.RefSeries) {
|
|
||||||
require.Empty(t, l)
|
|
||||||
}
|
|
||||||
|
|
||||||
// Weird hack to check order of reads.
|
|
||||||
i := 0
|
|
||||||
samplef := func(s []record.RefSample) {
|
|
||||||
if i == 0 {
|
|
||||||
require.Equal(t, []record.RefSample{{T: 1, V: 2}}, s)
|
|
||||||
i++
|
|
||||||
} else {
|
|
||||||
require.Equal(t, []record.RefSample{{T: 99, V: 100}}, s)
|
|
||||||
}
|
|
||||||
}
|
|
||||||
|
|
||||||
require.NoError(t, r.Read(serf, samplef, nil))
|
|
||||||
|
|
||||||
require.NoError(t, w2.LogSamples([]record.RefSample{{T: 99, V: 100}}))
|
|
||||||
require.NoError(t, w2.Close())
|
|
||||||
|
|
||||||
// We should see the first valid entry and the new one, everything after
|
|
||||||
// is truncated.
|
|
||||||
w3, err := OpenSegmentWAL(dir, logger, 0, nil)
|
|
||||||
require.NoError(t, err)
|
|
||||||
defer func(wal *SegmentWAL) { require.NoError(t, wal.Close()) }(w3)
|
|
||||||
|
|
||||||
r = w3.Reader()
|
|
||||||
|
|
||||||
i = 0
|
|
||||||
require.NoError(t, r.Read(serf, samplef, nil))
|
|
||||||
})
|
|
||||||
}
|
|
||||||
}
|
|
||||||
|
|
||||||
func TestMigrateWAL_Empty(t *testing.T) {
|
|
||||||
// The migration procedure must properly deal with a zero-length segment,
|
|
||||||
// which is valid in the new format.
|
|
||||||
dir := t.TempDir()
|
|
||||||
|
|
||||||
wdir := path.Join(dir, "wal")
|
|
||||||
|
|
||||||
// Initialize empty WAL.
|
|
||||||
w, err := wlog.New(nil, nil, wdir, wlog.CompressionNone)
|
|
||||||
require.NoError(t, err)
|
|
||||||
require.NoError(t, w.Close())
|
|
||||||
|
|
||||||
require.NoError(t, MigrateWAL(nil, wdir))
|
|
||||||
}
|
|
||||||
|
|
||||||
func TestMigrateWAL_Fuzz(t *testing.T) {
|
|
||||||
dir := t.TempDir()
|
|
||||||
|
|
||||||
wdir := path.Join(dir, "wal")
|
|
||||||
|
|
||||||
// Should pass if no WAL exists yet.
|
|
||||||
require.NoError(t, MigrateWAL(nil, wdir))
|
|
||||||
|
|
||||||
oldWAL, err := OpenSegmentWAL(wdir, nil, time.Minute, nil)
|
|
||||||
require.NoError(t, err)
|
|
||||||
|
|
||||||
// Write some data.
|
|
||||||
require.NoError(t, oldWAL.LogSeries([]record.RefSeries{
|
|
||||||
{Ref: 100, Labels: labels.FromStrings("abc", "def", "123", "456")},
|
|
||||||
{Ref: 1, Labels: labels.FromStrings("abc", "def2", "1234", "4567")},
|
|
||||||
}))
|
|
||||||
require.NoError(t, oldWAL.LogSamples([]record.RefSample{
|
|
||||||
{Ref: 1, T: 100, V: 200},
|
|
||||||
{Ref: 2, T: 300, V: 400},
|
|
||||||
}))
|
|
||||||
require.NoError(t, oldWAL.LogSeries([]record.RefSeries{
|
|
||||||
{Ref: 200, Labels: labels.FromStrings("xyz", "def", "foo", "bar")},
|
|
||||||
}))
|
|
||||||
require.NoError(t, oldWAL.LogSamples([]record.RefSample{
|
|
||||||
{Ref: 3, T: 100, V: 200},
|
|
||||||
{Ref: 4, T: 300, V: 400},
|
|
||||||
}))
|
|
||||||
require.NoError(t, oldWAL.LogDeletes([]tombstones.Stone{
|
|
||||||
{Ref: 1, Intervals: []tombstones.Interval{{Mint: 100, Maxt: 200}}},
|
|
||||||
}))
|
|
||||||
|
|
||||||
require.NoError(t, oldWAL.Close())
|
|
||||||
|
|
||||||
// Perform migration.
|
|
||||||
require.NoError(t, MigrateWAL(nil, wdir))
|
|
||||||
|
|
||||||
w, err := wlog.New(nil, nil, wdir, wlog.CompressionNone)
|
|
||||||
require.NoError(t, err)
|
|
||||||
|
|
||||||
// We can properly write some new data after migration.
|
|
||||||
var enc record.Encoder
|
|
||||||
require.NoError(t, w.Log(enc.Samples([]record.RefSample{
|
|
||||||
{Ref: 500, T: 1, V: 1},
|
|
||||||
}, nil)))
|
|
||||||
|
|
||||||
require.NoError(t, w.Close())
|
|
||||||
|
|
||||||
// Read back all data.
|
|
||||||
sr, err := wlog.NewSegmentsReader(wdir)
|
|
||||||
require.NoError(t, err)
|
|
||||||
|
|
||||||
r := wlog.NewReader(sr)
|
|
||||||
var res []interface{}
|
|
||||||
dec := record.NewDecoder(labels.NewSymbolTable())
|
|
||||||
|
|
||||||
for r.Next() {
|
|
||||||
rec := r.Record()
|
|
||||||
|
|
||||||
switch dec.Type(rec) {
|
|
||||||
case record.Series:
|
|
||||||
s, err := dec.Series(rec, nil)
|
|
||||||
require.NoError(t, err)
|
|
||||||
res = append(res, s)
|
|
||||||
case record.Samples:
|
|
||||||
s, err := dec.Samples(rec, nil)
|
|
||||||
require.NoError(t, err)
|
|
||||||
res = append(res, s)
|
|
||||||
case record.Tombstones:
|
|
||||||
s, err := dec.Tombstones(rec, nil)
|
|
||||||
require.NoError(t, err)
|
|
||||||
res = append(res, s)
|
|
||||||
default:
|
|
||||||
require.Fail(t, "unknown record type %d", dec.Type(rec))
|
|
||||||
}
|
|
||||||
}
|
|
||||||
require.NoError(t, r.Err())
|
|
||||||
|
|
||||||
testutil.RequireEqual(t, []interface{}{
|
|
||||||
[]record.RefSeries{
|
|
||||||
{Ref: 100, Labels: labels.FromStrings("abc", "def", "123", "456")},
|
|
||||||
{Ref: 1, Labels: labels.FromStrings("abc", "def2", "1234", "4567")},
|
|
||||||
},
|
|
||||||
[]record.RefSample{{Ref: 1, T: 100, V: 200}, {Ref: 2, T: 300, V: 400}},
|
|
||||||
[]record.RefSeries{
|
|
||||||
{Ref: 200, Labels: labels.FromStrings("xyz", "def", "foo", "bar")},
|
|
||||||
},
|
|
||||||
[]record.RefSample{{Ref: 3, T: 100, V: 200}, {Ref: 4, T: 300, V: 400}},
|
|
||||||
[]tombstones.Stone{{Ref: 1, Intervals: []tombstones.Interval{{Mint: 100, Maxt: 200}}}},
|
|
||||||
[]record.RefSample{{Ref: 500, T: 1, V: 1}},
|
|
||||||
}, res)
|
|
||||||
|
|
||||||
// Migrating an already migrated WAL shouldn't do anything.
|
|
||||||
require.NoError(t, MigrateWAL(nil, wdir))
|
|
||||||
}
|
|
Loading…
Reference in a new issue