mirror of
https://github.com/prometheus/prometheus.git
synced 2024-12-27 14:39:40 -08:00
0f760f63dd
Several things done here: - Set `max-issues-per-linter` to 0 so that we actually see all linter warnings and not just 50 per linter. (As we also set `max-same-issues` to 0, I assume this was the intention from the beginning.) - Stop using the golangci-lint default excludes (by setting `exclude-use-default: false`. Those are too generous and don't match our style conventions. (I have re-added some of the excludes explicitly in this commit. See below.) - Re-add the `errcheck` exclusion we have used so far via the defaults. - Exclude the signature requirement `govet` has for `Seek` methods because we use non-standard `Seek` methods a lot. (But we keep other requirements, while the default excludes completely disabled the check for common method segnatures.) - Exclude warnings about missing doc comments on exported symbols. (We used to be pretty adamant about doc comments, but stopped that at some point in the past. By now, we have about 500 missing doc comments. We may consider reintroducing this check, but that's outside of the scope of this commit. The default excludes of golangci-lint essentially ignore doc comments completely.) - By stop using the default excludes, we now get warnings back on malformed doc comments. That's the most impactful change in this commit. It does not enforce doc comments (again), but _if_ there is a doc comment, it has to have the recommended form. (Most of the changes in this commit are fixing this form.) - Improve wording/spelling of some comments in .golangci.yml, and remove an outdated comment. - Leave `package-comments` inactive, but add a TODO asking if we should change that. - Add a new sub-linter `comment-spacings` (and fix corresponding comments), which avoids missing spaces after the leading `//`. Signed-off-by: beorn7 <beorn@grafana.com>
1142 lines
35 KiB
Go
1142 lines
35 KiB
Go
// Copyright 2020 The Prometheus Authors
|
|
// Licensed under the Apache License, Version 2.0 (the "License");
|
|
// you may not use this file except in compliance with the License.
|
|
// You may obtain a copy of the License at
|
|
//
|
|
// http://www.apache.org/licenses/LICENSE-2.0
|
|
//
|
|
// Unless required by applicable law or agreed to in writing, software
|
|
// distributed under the License is distributed on an "AS IS" BASIS,
|
|
// WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
|
|
// See the License for the specific language governing permissions and
|
|
// limitations under the License.
|
|
|
|
package chunks
|
|
|
|
import (
|
|
"bufio"
|
|
"encoding/binary"
|
|
"errors"
|
|
"fmt"
|
|
"hash"
|
|
"io"
|
|
"os"
|
|
"path/filepath"
|
|
"slices"
|
|
"strconv"
|
|
"sync"
|
|
|
|
"github.com/dennwc/varint"
|
|
"github.com/prometheus/client_golang/prometheus"
|
|
"go.uber.org/atomic"
|
|
|
|
"github.com/prometheus/prometheus/tsdb/chunkenc"
|
|
tsdb_errors "github.com/prometheus/prometheus/tsdb/errors"
|
|
"github.com/prometheus/prometheus/tsdb/fileutil"
|
|
)
|
|
|
|
// Head chunk file header fields constants.
|
|
const (
|
|
// MagicHeadChunks is 4 bytes at the beginning of a head chunk file.
|
|
MagicHeadChunks = 0x0130BC91
|
|
|
|
headChunksFormatV1 = 1
|
|
)
|
|
|
|
// ErrChunkDiskMapperClosed returned by any method indicates
|
|
// that the ChunkDiskMapper was closed.
|
|
var ErrChunkDiskMapperClosed = errors.New("ChunkDiskMapper closed")
|
|
|
|
const (
|
|
// MintMaxtSize is the size of the mint/maxt for head chunk file and chunks.
|
|
MintMaxtSize = 8
|
|
// SeriesRefSize is the size of series reference on disk.
|
|
SeriesRefSize = 8
|
|
// HeadChunkFileHeaderSize is the total size of the header for a head chunk file.
|
|
HeadChunkFileHeaderSize = SegmentHeaderSize
|
|
// MaxHeadChunkFileSize is the max size of a head chunk file.
|
|
MaxHeadChunkFileSize = 128 * 1024 * 1024 // 128 MiB.
|
|
// CRCSize is the size of crc32 sum on disk.
|
|
CRCSize = 4
|
|
// MaxHeadChunkMetaSize is the max size of an mmapped chunks minus the chunks data.
|
|
// Max because the uvarint size can be smaller.
|
|
MaxHeadChunkMetaSize = SeriesRefSize + 2*MintMaxtSize + ChunkEncodingSize + MaxChunkLengthFieldSize + CRCSize
|
|
// MinWriteBufferSize is the minimum write buffer size allowed.
|
|
MinWriteBufferSize = 64 * 1024 // 64KB.
|
|
// MaxWriteBufferSize is the maximum write buffer size allowed.
|
|
MaxWriteBufferSize = 8 * 1024 * 1024 // 8 MiB.
|
|
// DefaultWriteBufferSize is the default write buffer size.
|
|
DefaultWriteBufferSize = 4 * 1024 * 1024 // 4 MiB.
|
|
// DefaultWriteQueueSize is the default size of the in-memory queue used before flushing chunks to the disk.
|
|
// A value of 0 completely disables this feature.
|
|
DefaultWriteQueueSize = 0
|
|
)
|
|
|
|
// ChunkDiskMapperRef represents the location of a head chunk on disk.
|
|
// The upper 4 bytes hold the index of the head chunk file and
|
|
// the lower 4 bytes hold the byte offset in the head chunk file where the chunk starts.
|
|
type ChunkDiskMapperRef uint64
|
|
|
|
func newChunkDiskMapperRef(seq, offset uint64) ChunkDiskMapperRef {
|
|
return ChunkDiskMapperRef((seq << 32) | offset)
|
|
}
|
|
|
|
func (ref ChunkDiskMapperRef) Unpack() (seq, offset int) {
|
|
seq = int(ref >> 32)
|
|
offset = int((ref << 32) >> 32)
|
|
return seq, offset
|
|
}
|
|
|
|
func (ref ChunkDiskMapperRef) GreaterThanOrEqualTo(r ChunkDiskMapperRef) bool {
|
|
s1, o1 := ref.Unpack()
|
|
s2, o2 := r.Unpack()
|
|
return s1 > s2 || (s1 == s2 && o1 >= o2)
|
|
}
|
|
|
|
func (ref ChunkDiskMapperRef) GreaterThan(r ChunkDiskMapperRef) bool {
|
|
s1, o1 := ref.Unpack()
|
|
s2, o2 := r.Unpack()
|
|
return s1 > s2 || (s1 == s2 && o1 > o2)
|
|
}
|
|
|
|
// CorruptionErr is an error that's returned when corruption is encountered.
|
|
type CorruptionErr struct {
|
|
Dir string
|
|
FileIndex int
|
|
Err error
|
|
}
|
|
|
|
func (e *CorruptionErr) Error() string {
|
|
return fmt.Errorf("corruption in head chunk file %s: %w", segmentFile(e.Dir, e.FileIndex), e.Err).Error()
|
|
}
|
|
|
|
func (e *CorruptionErr) Unwrap() error {
|
|
return e.Err
|
|
}
|
|
|
|
// chunkPos keeps track of the position in the head chunk files.
|
|
// chunkPos is not thread-safe, a lock must be used to protect it.
|
|
type chunkPos struct {
|
|
seq uint64 // Index of chunk file.
|
|
offset uint64 // Offset within chunk file.
|
|
cutFile bool // When true then the next chunk will be written to a new file.
|
|
}
|
|
|
|
// getNextChunkRef takes a chunk and returns the chunk reference which will refer to it once it has been written.
|
|
// getNextChunkRef also decides whether a new file should be cut before writing this chunk, and it returns the decision via the second return value.
|
|
// The order of calling getNextChunkRef must be the order in which chunks are written to the disk.
|
|
func (f *chunkPos) getNextChunkRef(chk chunkenc.Chunk) (chkRef ChunkDiskMapperRef, cutFile bool) {
|
|
chkLen := uint64(len(chk.Bytes()))
|
|
bytesToWrite := f.bytesToWriteForChunk(chkLen)
|
|
|
|
if f.shouldCutNewFile(bytesToWrite) {
|
|
f.toNewFile()
|
|
f.cutFile = false
|
|
cutFile = true
|
|
}
|
|
|
|
chkOffset := f.offset
|
|
f.offset += bytesToWrite
|
|
|
|
return newChunkDiskMapperRef(f.seq, chkOffset), cutFile
|
|
}
|
|
|
|
// toNewFile updates the seq/offset position to point to the beginning of a new chunk file.
|
|
func (f *chunkPos) toNewFile() {
|
|
f.seq++
|
|
f.offset = SegmentHeaderSize
|
|
}
|
|
|
|
// cutFileOnNextChunk triggers that the next chunk will be written in to a new file.
|
|
// Not thread safe, a lock must be held when calling this.
|
|
func (f *chunkPos) cutFileOnNextChunk() {
|
|
f.cutFile = true
|
|
}
|
|
|
|
// setSeq sets the sequence number of the head chunk file.
|
|
func (f *chunkPos) setSeq(seq uint64) {
|
|
f.seq = seq
|
|
}
|
|
|
|
// shouldCutNewFile returns whether a new file should be cut based on the file size.
|
|
// Not thread safe, a lock must be held when calling this.
|
|
func (f *chunkPos) shouldCutNewFile(bytesToWrite uint64) bool {
|
|
if f.cutFile {
|
|
return true
|
|
}
|
|
|
|
return f.offset == 0 || // First head chunk file.
|
|
f.offset+bytesToWrite > MaxHeadChunkFileSize // Exceeds the max head chunk file size.
|
|
}
|
|
|
|
// bytesToWriteForChunk returns the number of bytes that will need to be written for the given chunk size,
|
|
// including all meta data before and after the chunk data.
|
|
// Head chunk format: https://github.com/prometheus/prometheus/blob/main/tsdb/docs/format/head_chunks.md#chunk
|
|
func (f *chunkPos) bytesToWriteForChunk(chkLen uint64) uint64 {
|
|
// Headers.
|
|
bytes := uint64(SeriesRefSize) + 2*MintMaxtSize + ChunkEncodingSize
|
|
|
|
// Size of chunk length encoded as uvarint.
|
|
bytes += uint64(varint.UvarintSize(chkLen))
|
|
|
|
// Chunk length.
|
|
bytes += chkLen
|
|
|
|
// crc32.
|
|
bytes += CRCSize
|
|
|
|
return bytes
|
|
}
|
|
|
|
// ChunkDiskMapper is for writing the Head block chunks to disk
|
|
// and access chunks via mmapped files.
|
|
type ChunkDiskMapper struct {
|
|
// Writer.
|
|
dir *os.File
|
|
writeBufferSize int
|
|
|
|
curFile *os.File // File being written to.
|
|
curFileSequence int // Index of current open file being appended to. 0 if no file is active.
|
|
curFileOffset atomic.Uint64 // Bytes written in current open file.
|
|
curFileMaxt int64 // Used for the size retention.
|
|
|
|
// The values in evtlPos represent the file position which will eventually be
|
|
// reached once the content of the write queue has been fully processed.
|
|
evtlPosMtx sync.Mutex
|
|
evtlPos chunkPos
|
|
|
|
byteBuf [MaxHeadChunkMetaSize]byte // Buffer used to write the header of the chunk.
|
|
chkWriter *bufio.Writer // Writer for the current open file.
|
|
crc32 hash.Hash
|
|
writePathMtx sync.Mutex
|
|
|
|
// Reader.
|
|
// The int key in the map is the file number on the disk.
|
|
mmappedChunkFiles map[int]*mmappedChunkFile // Contains the m-mapped files for each chunk file mapped with its index.
|
|
closers map[int]io.Closer // Closers for resources behind the byte slices.
|
|
readPathMtx sync.RWMutex // Mutex used to protect the above 2 maps.
|
|
pool chunkenc.Pool // This is used when fetching a chunk from the disk to allocate a chunk.
|
|
|
|
// Writer and Reader.
|
|
// We flush chunks to disk in batches. Hence, we store them in this buffer
|
|
// from which chunks are served till they are flushed and are ready for m-mapping.
|
|
chunkBuffer *chunkBuffer
|
|
|
|
// Whether the maxt field is set for all mmapped chunk files tracked within the mmappedChunkFiles map.
|
|
// This is done after iterating through all the chunks in those files using the IterateAllChunks method.
|
|
fileMaxtSet bool
|
|
|
|
writeQueue *chunkWriteQueue
|
|
|
|
closed bool
|
|
}
|
|
|
|
// mmappedChunkFile provides mmap access to an entire head chunks file that holds many chunks.
|
|
type mmappedChunkFile struct {
|
|
byteSlice ByteSlice
|
|
maxt int64 // Max timestamp among all of this file's chunks.
|
|
}
|
|
|
|
// NewChunkDiskMapper returns a new ChunkDiskMapper against the given directory
|
|
// using the default head chunk file duration.
|
|
// NOTE: 'IterateAllChunks' method needs to be called at least once after creating ChunkDiskMapper
|
|
// to set the maxt of all files.
|
|
func NewChunkDiskMapper(reg prometheus.Registerer, dir string, pool chunkenc.Pool, writeBufferSize, writeQueueSize int) (*ChunkDiskMapper, error) {
|
|
// Validate write buffer size.
|
|
if writeBufferSize < MinWriteBufferSize || writeBufferSize > MaxWriteBufferSize {
|
|
return nil, fmt.Errorf("ChunkDiskMapper write buffer size should be between %d and %d (actual: %d)", MinWriteBufferSize, MaxWriteBufferSize, writeBufferSize)
|
|
}
|
|
if writeBufferSize%1024 != 0 {
|
|
return nil, fmt.Errorf("ChunkDiskMapper write buffer size should be a multiple of 1024 (actual: %d)", writeBufferSize)
|
|
}
|
|
|
|
if err := os.MkdirAll(dir, 0o777); err != nil {
|
|
return nil, err
|
|
}
|
|
dirFile, err := fileutil.OpenDir(dir)
|
|
if err != nil {
|
|
return nil, err
|
|
}
|
|
|
|
m := &ChunkDiskMapper{
|
|
dir: dirFile,
|
|
pool: pool,
|
|
writeBufferSize: writeBufferSize,
|
|
crc32: newCRC32(),
|
|
chunkBuffer: newChunkBuffer(),
|
|
}
|
|
|
|
if writeQueueSize > 0 {
|
|
m.writeQueue = newChunkWriteQueue(reg, writeQueueSize, m.writeChunk)
|
|
}
|
|
|
|
if m.pool == nil {
|
|
m.pool = chunkenc.NewPool()
|
|
}
|
|
|
|
return m, m.openMMapFiles()
|
|
}
|
|
|
|
// Chunk encodings for out-of-order chunks.
|
|
// These encodings must be only used by the Head block for its internal bookkeeping.
|
|
const (
|
|
OutOfOrderMask = uint8(0b10000000)
|
|
)
|
|
|
|
func (cdm *ChunkDiskMapper) ApplyOutOfOrderMask(sourceEncoding chunkenc.Encoding) chunkenc.Encoding {
|
|
enc := uint8(sourceEncoding) | OutOfOrderMask
|
|
return chunkenc.Encoding(enc)
|
|
}
|
|
|
|
func (cdm *ChunkDiskMapper) IsOutOfOrderChunk(e chunkenc.Encoding) bool {
|
|
return (uint8(e) & OutOfOrderMask) != 0
|
|
}
|
|
|
|
func (cdm *ChunkDiskMapper) RemoveMasks(sourceEncoding chunkenc.Encoding) chunkenc.Encoding {
|
|
restored := uint8(sourceEncoding) & (^OutOfOrderMask)
|
|
return chunkenc.Encoding(restored)
|
|
}
|
|
|
|
// openMMapFiles opens all files within dir for mmapping.
|
|
func (cdm *ChunkDiskMapper) openMMapFiles() (returnErr error) {
|
|
cdm.mmappedChunkFiles = map[int]*mmappedChunkFile{}
|
|
cdm.closers = map[int]io.Closer{}
|
|
defer func() {
|
|
if returnErr != nil {
|
|
returnErr = tsdb_errors.NewMulti(returnErr, closeAllFromMap(cdm.closers)).Err()
|
|
|
|
cdm.mmappedChunkFiles = nil
|
|
cdm.closers = nil
|
|
}
|
|
}()
|
|
|
|
files, err := listChunkFiles(cdm.dir.Name())
|
|
if err != nil {
|
|
return err
|
|
}
|
|
|
|
files, err = repairLastChunkFile(files)
|
|
if err != nil {
|
|
return err
|
|
}
|
|
|
|
chkFileIndices := make([]int, 0, len(files))
|
|
for seq, fn := range files {
|
|
f, err := fileutil.OpenMmapFile(fn)
|
|
if err != nil {
|
|
return fmt.Errorf("mmap files, file: %s: %w", fn, err)
|
|
}
|
|
cdm.closers[seq] = f
|
|
cdm.mmappedChunkFiles[seq] = &mmappedChunkFile{byteSlice: realByteSlice(f.Bytes())}
|
|
chkFileIndices = append(chkFileIndices, seq)
|
|
}
|
|
|
|
// Check for gaps in the files.
|
|
slices.Sort(chkFileIndices)
|
|
if len(chkFileIndices) == 0 {
|
|
return nil
|
|
}
|
|
lastSeq := chkFileIndices[0]
|
|
for _, seq := range chkFileIndices[1:] {
|
|
if seq != lastSeq+1 {
|
|
return fmt.Errorf("found unsequential head chunk files %s (index: %d) and %s (index: %d)", files[lastSeq], lastSeq, files[seq], seq)
|
|
}
|
|
lastSeq = seq
|
|
}
|
|
|
|
for i, b := range cdm.mmappedChunkFiles {
|
|
if b.byteSlice.Len() < HeadChunkFileHeaderSize {
|
|
return fmt.Errorf("%s: invalid head chunk file header: %w", files[i], errInvalidSize)
|
|
}
|
|
// Verify magic number.
|
|
if m := binary.BigEndian.Uint32(b.byteSlice.Range(0, MagicChunksSize)); m != MagicHeadChunks {
|
|
return fmt.Errorf("%s: invalid magic number %x", files[i], m)
|
|
}
|
|
|
|
// Verify chunk format version.
|
|
if v := int(b.byteSlice.Range(MagicChunksSize, MagicChunksSize+ChunksFormatVersionSize)[0]); v != chunksFormatV1 {
|
|
return fmt.Errorf("%s: invalid chunk format version %d", files[i], v)
|
|
}
|
|
}
|
|
|
|
cdm.evtlPos.setSeq(uint64(lastSeq))
|
|
|
|
return nil
|
|
}
|
|
|
|
func listChunkFiles(dir string) (map[int]string, error) {
|
|
files, err := os.ReadDir(dir)
|
|
if err != nil {
|
|
return nil, err
|
|
}
|
|
res := map[int]string{}
|
|
for _, fi := range files {
|
|
seq, err := strconv.ParseUint(fi.Name(), 10, 64)
|
|
if err != nil {
|
|
continue
|
|
}
|
|
res[int(seq)] = filepath.Join(dir, fi.Name())
|
|
}
|
|
|
|
return res, nil
|
|
}
|
|
|
|
// HardLinkChunkFiles creates hardlinks for chunk files from src to dst.
|
|
// It does nothing if src doesn't exist and ensures dst is created if not.
|
|
func HardLinkChunkFiles(src, dst string) error {
|
|
_, err := os.Stat(src)
|
|
if os.IsNotExist(err) {
|
|
return nil
|
|
}
|
|
if err != nil {
|
|
return fmt.Errorf("check source chunks dir: %w", err)
|
|
}
|
|
if err := os.MkdirAll(dst, 0o777); err != nil {
|
|
return fmt.Errorf("set up destination chunks dir: %w", err)
|
|
}
|
|
files, err := listChunkFiles(src)
|
|
if err != nil {
|
|
return fmt.Errorf("list chunks: %w", err)
|
|
}
|
|
for _, filePath := range files {
|
|
_, fileName := filepath.Split(filePath)
|
|
err := os.Link(filepath.Join(src, fileName), filepath.Join(dst, fileName))
|
|
if err != nil {
|
|
return fmt.Errorf("hardlink a chunk: %w", err)
|
|
}
|
|
}
|
|
return nil
|
|
}
|
|
|
|
// repairLastChunkFile deletes the last file if it's empty.
|
|
// Because we don't fsync when creating these files, we could end
|
|
// up with an empty file at the end during an abrupt shutdown.
|
|
func repairLastChunkFile(files map[int]string) (_ map[int]string, returnErr error) {
|
|
lastFile := -1
|
|
for seq := range files {
|
|
if seq > lastFile {
|
|
lastFile = seq
|
|
}
|
|
}
|
|
|
|
if lastFile <= 0 {
|
|
return files, nil
|
|
}
|
|
|
|
f, err := os.Open(files[lastFile])
|
|
if err != nil {
|
|
return files, fmt.Errorf("open file during last head chunk file repair: %w", err)
|
|
}
|
|
|
|
buf := make([]byte, MagicChunksSize)
|
|
size, err := f.Read(buf)
|
|
if err != nil && !errors.Is(err, io.EOF) {
|
|
return files, fmt.Errorf("failed to read magic number during last head chunk file repair: %w", err)
|
|
}
|
|
if err := f.Close(); err != nil {
|
|
return files, fmt.Errorf("close file during last head chunk file repair: %w", err)
|
|
}
|
|
|
|
// We either don't have enough bytes for the magic number or the magic number is 0.
|
|
// NOTE: we should not check for wrong magic number here because that error
|
|
// needs to be sent up the function called (already done elsewhere)
|
|
// for proper repair mechanism to happen in the Head.
|
|
if size < MagicChunksSize || binary.BigEndian.Uint32(buf) == 0 {
|
|
// Corrupt file, hence remove it.
|
|
if err := os.RemoveAll(files[lastFile]); err != nil {
|
|
return files, fmt.Errorf("delete corrupted, empty head chunk file during last file repair: %w", err)
|
|
}
|
|
delete(files, lastFile)
|
|
}
|
|
|
|
return files, nil
|
|
}
|
|
|
|
// WriteChunk writes the chunk to disk.
|
|
// The returned chunk ref is the reference from where the chunk encoding starts for the chunk.
|
|
func (cdm *ChunkDiskMapper) WriteChunk(seriesRef HeadSeriesRef, mint, maxt int64, chk chunkenc.Chunk, isOOO bool, callback func(err error)) (chkRef ChunkDiskMapperRef) {
|
|
// cdm.evtlPosMtx must be held to serialize the calls to cdm.evtlPos.getNextChunkRef() and the writing of the chunk (either with or without queue).
|
|
cdm.evtlPosMtx.Lock()
|
|
defer cdm.evtlPosMtx.Unlock()
|
|
ref, cutFile := cdm.evtlPos.getNextChunkRef(chk)
|
|
|
|
if cdm.writeQueue != nil {
|
|
return cdm.writeChunkViaQueue(ref, isOOO, cutFile, seriesRef, mint, maxt, chk, callback)
|
|
}
|
|
|
|
err := cdm.writeChunk(seriesRef, mint, maxt, chk, ref, isOOO, cutFile)
|
|
if callback != nil {
|
|
callback(err)
|
|
}
|
|
|
|
return ref
|
|
}
|
|
|
|
func (cdm *ChunkDiskMapper) writeChunkViaQueue(ref ChunkDiskMapperRef, isOOO, cutFile bool, seriesRef HeadSeriesRef, mint, maxt int64, chk chunkenc.Chunk, callback func(err error)) (chkRef ChunkDiskMapperRef) {
|
|
var err error
|
|
if callback != nil {
|
|
defer func() {
|
|
if err != nil {
|
|
callback(err)
|
|
}
|
|
}()
|
|
}
|
|
|
|
err = cdm.writeQueue.addJob(chunkWriteJob{
|
|
cutFile: cutFile,
|
|
seriesRef: seriesRef,
|
|
mint: mint,
|
|
maxt: maxt,
|
|
chk: chk,
|
|
ref: ref,
|
|
isOOO: isOOO,
|
|
callback: callback,
|
|
})
|
|
|
|
return ref
|
|
}
|
|
|
|
func (cdm *ChunkDiskMapper) writeChunk(seriesRef HeadSeriesRef, mint, maxt int64, chk chunkenc.Chunk, ref ChunkDiskMapperRef, isOOO, cutFile bool) (err error) {
|
|
cdm.writePathMtx.Lock()
|
|
defer cdm.writePathMtx.Unlock()
|
|
|
|
if cdm.closed {
|
|
return ErrChunkDiskMapperClosed
|
|
}
|
|
|
|
if cutFile {
|
|
err := cdm.cutAndExpectRef(ref)
|
|
if err != nil {
|
|
return err
|
|
}
|
|
}
|
|
|
|
// if len(chk.Bytes())+MaxHeadChunkMetaSize >= writeBufferSize, it means that chunk >= the buffer size;
|
|
// so no need to flush here, as we have to flush at the end (to not keep partial chunks in buffer).
|
|
if len(chk.Bytes())+MaxHeadChunkMetaSize < cdm.writeBufferSize && cdm.chkWriter.Available() < MaxHeadChunkMetaSize+len(chk.Bytes()) {
|
|
if err := cdm.flushBuffer(); err != nil {
|
|
return err
|
|
}
|
|
}
|
|
|
|
cdm.crc32.Reset()
|
|
bytesWritten := 0
|
|
|
|
binary.BigEndian.PutUint64(cdm.byteBuf[bytesWritten:], uint64(seriesRef))
|
|
bytesWritten += SeriesRefSize
|
|
binary.BigEndian.PutUint64(cdm.byteBuf[bytesWritten:], uint64(mint))
|
|
bytesWritten += MintMaxtSize
|
|
binary.BigEndian.PutUint64(cdm.byteBuf[bytesWritten:], uint64(maxt))
|
|
bytesWritten += MintMaxtSize
|
|
enc := chk.Encoding()
|
|
if isOOO {
|
|
enc = cdm.ApplyOutOfOrderMask(enc)
|
|
}
|
|
cdm.byteBuf[bytesWritten] = byte(enc)
|
|
bytesWritten += ChunkEncodingSize
|
|
n := binary.PutUvarint(cdm.byteBuf[bytesWritten:], uint64(len(chk.Bytes())))
|
|
bytesWritten += n
|
|
|
|
if err := cdm.writeAndAppendToCRC32(cdm.byteBuf[:bytesWritten]); err != nil {
|
|
return err
|
|
}
|
|
if err := cdm.writeAndAppendToCRC32(chk.Bytes()); err != nil {
|
|
return err
|
|
}
|
|
if err := cdm.writeCRC32(); err != nil {
|
|
return err
|
|
}
|
|
|
|
if maxt > cdm.curFileMaxt {
|
|
cdm.curFileMaxt = maxt
|
|
}
|
|
|
|
cdm.chunkBuffer.put(ref, chk)
|
|
|
|
if len(chk.Bytes())+MaxHeadChunkMetaSize >= cdm.writeBufferSize {
|
|
// The chunk was bigger than the buffer itself.
|
|
// Flushing to not keep partial chunks in buffer.
|
|
if err := cdm.flushBuffer(); err != nil {
|
|
return err
|
|
}
|
|
}
|
|
|
|
return nil
|
|
}
|
|
|
|
// CutNewFile makes that a new file will be created the next time a chunk is written.
|
|
func (cdm *ChunkDiskMapper) CutNewFile() {
|
|
cdm.evtlPosMtx.Lock()
|
|
defer cdm.evtlPosMtx.Unlock()
|
|
|
|
cdm.evtlPos.cutFileOnNextChunk()
|
|
}
|
|
|
|
func (cdm *ChunkDiskMapper) IsQueueEmpty() bool {
|
|
if cdm.writeQueue == nil {
|
|
return true
|
|
}
|
|
|
|
return cdm.writeQueue.queueIsEmpty()
|
|
}
|
|
|
|
// cutAndExpectRef creates a new m-mapped file.
|
|
// The write lock should be held before calling this.
|
|
// It ensures that the position in the new file matches the given chunk reference, if not then it errors.
|
|
func (cdm *ChunkDiskMapper) cutAndExpectRef(chkRef ChunkDiskMapperRef) (err error) {
|
|
seq, offset, err := cdm.cut()
|
|
if err != nil {
|
|
return err
|
|
}
|
|
|
|
if expSeq, expOffset := chkRef.Unpack(); seq != expSeq || offset != expOffset {
|
|
return fmt.Errorf("expected newly cut file to have sequence:offset %d:%d, got %d:%d", expSeq, expOffset, seq, offset)
|
|
}
|
|
|
|
return nil
|
|
}
|
|
|
|
// cut creates a new m-mapped file. The write lock should be held before calling this.
|
|
// It returns the file sequence and the offset in that file to start writing chunks.
|
|
func (cdm *ChunkDiskMapper) cut() (seq, offset int, returnErr error) {
|
|
// Sync current tail to disk and close.
|
|
if err := cdm.finalizeCurFile(); err != nil {
|
|
return 0, 0, err
|
|
}
|
|
|
|
offset, newFile, seq, err := cutSegmentFile(cdm.dir, MagicHeadChunks, headChunksFormatV1, HeadChunkFilePreallocationSize)
|
|
if err != nil {
|
|
return 0, 0, err
|
|
}
|
|
|
|
defer func() {
|
|
// The file should not be closed if there is no error,
|
|
// its kept open in the ChunkDiskMapper.
|
|
if returnErr != nil {
|
|
returnErr = tsdb_errors.NewMulti(returnErr, newFile.Close()).Err()
|
|
}
|
|
}()
|
|
|
|
cdm.curFileOffset.Store(uint64(offset))
|
|
|
|
if cdm.curFile != nil {
|
|
cdm.readPathMtx.Lock()
|
|
cdm.mmappedChunkFiles[cdm.curFileSequence].maxt = cdm.curFileMaxt
|
|
cdm.readPathMtx.Unlock()
|
|
}
|
|
|
|
mmapFile, err := fileutil.OpenMmapFileWithSize(newFile.Name(), MaxHeadChunkFileSize)
|
|
if err != nil {
|
|
return 0, 0, err
|
|
}
|
|
|
|
cdm.readPathMtx.Lock()
|
|
cdm.curFileSequence = seq
|
|
cdm.curFile = newFile
|
|
if cdm.chkWriter != nil {
|
|
cdm.chkWriter.Reset(newFile)
|
|
} else {
|
|
cdm.chkWriter = bufio.NewWriterSize(newFile, cdm.writeBufferSize)
|
|
}
|
|
|
|
cdm.closers[cdm.curFileSequence] = mmapFile
|
|
cdm.mmappedChunkFiles[cdm.curFileSequence] = &mmappedChunkFile{byteSlice: realByteSlice(mmapFile.Bytes())}
|
|
cdm.readPathMtx.Unlock()
|
|
|
|
cdm.curFileMaxt = 0
|
|
|
|
return seq, offset, nil
|
|
}
|
|
|
|
// finalizeCurFile writes all pending data to the current tail file,
|
|
// truncates its size, and closes it.
|
|
func (cdm *ChunkDiskMapper) finalizeCurFile() error {
|
|
if cdm.curFile == nil {
|
|
return nil
|
|
}
|
|
|
|
if err := cdm.flushBuffer(); err != nil {
|
|
return err
|
|
}
|
|
|
|
if err := cdm.curFile.Sync(); err != nil {
|
|
return err
|
|
}
|
|
|
|
return cdm.curFile.Close()
|
|
}
|
|
|
|
func (cdm *ChunkDiskMapper) write(b []byte) error {
|
|
n, err := cdm.chkWriter.Write(b)
|
|
cdm.curFileOffset.Add(uint64(n))
|
|
return err
|
|
}
|
|
|
|
func (cdm *ChunkDiskMapper) writeAndAppendToCRC32(b []byte) error {
|
|
if err := cdm.write(b); err != nil {
|
|
return err
|
|
}
|
|
_, err := cdm.crc32.Write(b)
|
|
return err
|
|
}
|
|
|
|
func (cdm *ChunkDiskMapper) writeCRC32() error {
|
|
return cdm.write(cdm.crc32.Sum(cdm.byteBuf[:0]))
|
|
}
|
|
|
|
// flushBuffer flushes the current in-memory chunks.
|
|
// Assumes that writePathMtx is _write_ locked before calling this method.
|
|
func (cdm *ChunkDiskMapper) flushBuffer() error {
|
|
if err := cdm.chkWriter.Flush(); err != nil {
|
|
return err
|
|
}
|
|
cdm.chunkBuffer.clear()
|
|
return nil
|
|
}
|
|
|
|
// Chunk returns a chunk from a given reference.
|
|
func (cdm *ChunkDiskMapper) Chunk(ref ChunkDiskMapperRef) (chunkenc.Chunk, error) {
|
|
cdm.readPathMtx.RLock()
|
|
// We hold this read lock for the entire duration because if Close()
|
|
// is called, the data in the byte slice will get corrupted as the mmapped
|
|
// file will be closed.
|
|
defer cdm.readPathMtx.RUnlock()
|
|
|
|
if cdm.closed {
|
|
return nil, ErrChunkDiskMapperClosed
|
|
}
|
|
|
|
if cdm.writeQueue != nil {
|
|
chunk := cdm.writeQueue.get(ref)
|
|
if chunk != nil {
|
|
return chunk, nil
|
|
}
|
|
}
|
|
|
|
sgmIndex, chkStart := ref.Unpack()
|
|
// We skip the series ref and the mint/maxt beforehand.
|
|
chkStart += SeriesRefSize + (2 * MintMaxtSize)
|
|
|
|
// If it is the current open file, then the chunks can be in the buffer too.
|
|
if sgmIndex == cdm.curFileSequence {
|
|
chunk := cdm.chunkBuffer.get(ref)
|
|
if chunk != nil {
|
|
return chunk, nil
|
|
}
|
|
}
|
|
|
|
mmapFile, ok := cdm.mmappedChunkFiles[sgmIndex]
|
|
if !ok {
|
|
if sgmIndex > cdm.curFileSequence {
|
|
return nil, &CorruptionErr{
|
|
Dir: cdm.dir.Name(),
|
|
FileIndex: -1,
|
|
Err: fmt.Errorf("head chunk file index %d more than current open file", sgmIndex),
|
|
}
|
|
}
|
|
return nil, &CorruptionErr{
|
|
Dir: cdm.dir.Name(),
|
|
FileIndex: sgmIndex,
|
|
Err: fmt.Errorf("head chunk file index %d does not exist on disk", sgmIndex),
|
|
}
|
|
}
|
|
|
|
if chkStart+MaxChunkLengthFieldSize > mmapFile.byteSlice.Len() {
|
|
return nil, &CorruptionErr{
|
|
Dir: cdm.dir.Name(),
|
|
FileIndex: sgmIndex,
|
|
Err: fmt.Errorf("head chunk file doesn't include enough bytes to read the chunk size data field - required:%v, available:%v", chkStart+MaxChunkLengthFieldSize, mmapFile.byteSlice.Len()),
|
|
}
|
|
}
|
|
|
|
// Encoding.
|
|
chkEnc := mmapFile.byteSlice.Range(chkStart, chkStart+ChunkEncodingSize)[0]
|
|
sourceChkEnc := chunkenc.Encoding(chkEnc)
|
|
// Extract the encoding from the byte. ChunkDiskMapper uses only the last 7 bits for the encoding.
|
|
chkEnc = byte(cdm.RemoveMasks(sourceChkEnc))
|
|
// Data length.
|
|
// With the minimum chunk length this should never cause us reading
|
|
// over the end of the slice.
|
|
chkDataLenStart := chkStart + ChunkEncodingSize
|
|
c := mmapFile.byteSlice.Range(chkDataLenStart, chkDataLenStart+MaxChunkLengthFieldSize)
|
|
chkDataLen, n := binary.Uvarint(c)
|
|
if n <= 0 {
|
|
return nil, &CorruptionErr{
|
|
Dir: cdm.dir.Name(),
|
|
FileIndex: sgmIndex,
|
|
Err: fmt.Errorf("reading chunk length failed with %d", n),
|
|
}
|
|
}
|
|
|
|
// Verify the chunk data end.
|
|
chkDataEnd := chkDataLenStart + n + int(chkDataLen)
|
|
if chkDataEnd > mmapFile.byteSlice.Len() {
|
|
return nil, &CorruptionErr{
|
|
Dir: cdm.dir.Name(),
|
|
FileIndex: sgmIndex,
|
|
Err: fmt.Errorf("head chunk file doesn't include enough bytes to read the chunk - required:%v, available:%v", chkDataEnd, mmapFile.byteSlice.Len()),
|
|
}
|
|
}
|
|
|
|
// Check the CRC.
|
|
sum := mmapFile.byteSlice.Range(chkDataEnd, chkDataEnd+CRCSize)
|
|
if err := checkCRC32(mmapFile.byteSlice.Range(chkStart-(SeriesRefSize+2*MintMaxtSize), chkDataEnd), sum); err != nil {
|
|
return nil, &CorruptionErr{
|
|
Dir: cdm.dir.Name(),
|
|
FileIndex: sgmIndex,
|
|
Err: err,
|
|
}
|
|
}
|
|
|
|
// The chunk data itself.
|
|
chkData := mmapFile.byteSlice.Range(chkDataEnd-int(chkDataLen), chkDataEnd)
|
|
|
|
// Make a copy of the chunk data to prevent a panic occurring because the returned
|
|
// chunk data slice references an mmap-ed file which could be closed after the
|
|
// function returns but while the chunk is still in use.
|
|
chkDataCopy := make([]byte, len(chkData))
|
|
copy(chkDataCopy, chkData)
|
|
|
|
chk, err := cdm.pool.Get(chunkenc.Encoding(chkEnc), chkDataCopy)
|
|
if err != nil {
|
|
return nil, &CorruptionErr{
|
|
Dir: cdm.dir.Name(),
|
|
FileIndex: sgmIndex,
|
|
Err: err,
|
|
}
|
|
}
|
|
return chk, nil
|
|
}
|
|
|
|
// IterateAllChunks iterates all mmappedChunkFiles (in order of head chunk file name/number) and all the chunks within it
|
|
// and runs the provided function with information about each chunk. It returns on the first error encountered.
|
|
// NOTE: This method needs to be called at least once after creating ChunkDiskMapper
|
|
// to set the maxt of all files.
|
|
func (cdm *ChunkDiskMapper) IterateAllChunks(f func(seriesRef HeadSeriesRef, chunkRef ChunkDiskMapperRef, mint, maxt int64, numSamples uint16, encoding chunkenc.Encoding, isOOO bool) error) (err error) {
|
|
cdm.writePathMtx.Lock()
|
|
defer cdm.writePathMtx.Unlock()
|
|
|
|
defer func() {
|
|
cdm.fileMaxtSet = true
|
|
}()
|
|
|
|
// Iterate files in ascending order.
|
|
segIDs := make([]int, 0, len(cdm.mmappedChunkFiles))
|
|
for seg := range cdm.mmappedChunkFiles {
|
|
segIDs = append(segIDs, seg)
|
|
}
|
|
slices.Sort(segIDs)
|
|
for _, segID := range segIDs {
|
|
mmapFile := cdm.mmappedChunkFiles[segID]
|
|
fileEnd := mmapFile.byteSlice.Len()
|
|
if segID == cdm.curFileSequence {
|
|
fileEnd = int(cdm.curFileSize())
|
|
}
|
|
idx := HeadChunkFileHeaderSize
|
|
for idx < fileEnd {
|
|
if fileEnd-idx < MaxHeadChunkMetaSize {
|
|
// Check for all 0s which marks the end of the file.
|
|
allZeros := true
|
|
for _, b := range mmapFile.byteSlice.Range(idx, fileEnd) {
|
|
if b != byte(0) {
|
|
allZeros = false
|
|
break
|
|
}
|
|
}
|
|
if allZeros {
|
|
// End of segment chunk file content.
|
|
break
|
|
}
|
|
return &CorruptionErr{
|
|
Dir: cdm.dir.Name(),
|
|
FileIndex: segID,
|
|
Err: fmt.Errorf("head chunk file has some unread data, but doesn't include enough bytes to read the chunk header"+
|
|
" - required:%v, available:%v, file:%d", idx+MaxHeadChunkMetaSize, fileEnd, segID),
|
|
}
|
|
}
|
|
chunkRef := newChunkDiskMapperRef(uint64(segID), uint64(idx))
|
|
|
|
startIdx := idx
|
|
seriesRef := HeadSeriesRef(binary.BigEndian.Uint64(mmapFile.byteSlice.Range(idx, idx+SeriesRefSize)))
|
|
idx += SeriesRefSize
|
|
mint := int64(binary.BigEndian.Uint64(mmapFile.byteSlice.Range(idx, idx+MintMaxtSize)))
|
|
idx += MintMaxtSize
|
|
maxt := int64(binary.BigEndian.Uint64(mmapFile.byteSlice.Range(idx, idx+MintMaxtSize)))
|
|
idx += MintMaxtSize
|
|
|
|
// We preallocate file to help with m-mapping (especially windows systems).
|
|
// As series ref always starts from 1, we assume it being 0 to be the end of the actual file data.
|
|
// We are not considering possible file corruption that can cause it to be 0.
|
|
// Additionally we are checking mint and maxt just to be sure.
|
|
if seriesRef == 0 && mint == 0 && maxt == 0 {
|
|
break
|
|
}
|
|
|
|
chkEnc := chunkenc.Encoding(mmapFile.byteSlice.Range(idx, idx+ChunkEncodingSize)[0])
|
|
idx += ChunkEncodingSize
|
|
dataLen, n := binary.Uvarint(mmapFile.byteSlice.Range(idx, idx+MaxChunkLengthFieldSize))
|
|
idx += n
|
|
|
|
numSamples := binary.BigEndian.Uint16(mmapFile.byteSlice.Range(idx, idx+2))
|
|
idx += int(dataLen) // Skip the data.
|
|
|
|
// In the beginning we only checked for the chunk meta size.
|
|
// Now that we have added the chunk data length, we check for sufficient bytes again.
|
|
if idx+CRCSize > fileEnd {
|
|
return &CorruptionErr{
|
|
Dir: cdm.dir.Name(),
|
|
FileIndex: segID,
|
|
Err: fmt.Errorf("head chunk file doesn't include enough bytes to read the chunk header - required:%v, available:%v, file:%d", idx+CRCSize, fileEnd, segID),
|
|
}
|
|
}
|
|
|
|
// Check CRC.
|
|
sum := mmapFile.byteSlice.Range(idx, idx+CRCSize)
|
|
if err := checkCRC32(mmapFile.byteSlice.Range(startIdx, idx), sum); err != nil {
|
|
return &CorruptionErr{
|
|
Dir: cdm.dir.Name(),
|
|
FileIndex: segID,
|
|
Err: err,
|
|
}
|
|
}
|
|
idx += CRCSize
|
|
|
|
if maxt > mmapFile.maxt {
|
|
mmapFile.maxt = maxt
|
|
}
|
|
isOOO := cdm.IsOutOfOrderChunk(chkEnc)
|
|
// Extract the encoding from the byte. ChunkDiskMapper uses only the last 7 bits for the encoding.
|
|
chkEnc = cdm.RemoveMasks(chkEnc)
|
|
if err := f(seriesRef, chunkRef, mint, maxt, numSamples, chkEnc, isOOO); err != nil {
|
|
var cerr *CorruptionErr
|
|
if errors.As(err, &cerr) {
|
|
cerr.Dir = cdm.dir.Name()
|
|
cerr.FileIndex = segID
|
|
return cerr
|
|
}
|
|
return err
|
|
}
|
|
}
|
|
|
|
if idx > fileEnd {
|
|
// It should be equal to the slice length.
|
|
return &CorruptionErr{
|
|
Dir: cdm.dir.Name(),
|
|
FileIndex: segID,
|
|
Err: fmt.Errorf("head chunk file doesn't include enough bytes to read the last chunk data - required:%v, available:%v, file:%d", idx, fileEnd, segID),
|
|
}
|
|
}
|
|
}
|
|
|
|
return nil
|
|
}
|
|
|
|
// Truncate deletes the head chunk files with numbers less than the given fileNo.
|
|
func (cdm *ChunkDiskMapper) Truncate(fileNo uint32) error {
|
|
cdm.readPathMtx.RLock()
|
|
|
|
// Sort the file indices, else if files deletion fails in between,
|
|
// it can lead to unsequential files as the map is not sorted.
|
|
chkFileIndices := make([]int, 0, len(cdm.mmappedChunkFiles))
|
|
for seq := range cdm.mmappedChunkFiles {
|
|
chkFileIndices = append(chkFileIndices, seq)
|
|
}
|
|
slices.Sort(chkFileIndices)
|
|
|
|
var removedFiles []int
|
|
for _, seq := range chkFileIndices {
|
|
if seq == cdm.curFileSequence || uint32(seq) >= fileNo {
|
|
break
|
|
}
|
|
removedFiles = append(removedFiles, seq)
|
|
}
|
|
cdm.readPathMtx.RUnlock()
|
|
|
|
errs := tsdb_errors.NewMulti()
|
|
// Cut a new file only if the current file has some chunks.
|
|
if cdm.curFileSize() > HeadChunkFileHeaderSize {
|
|
// There is a known race condition here because between the check of curFileSize() and the call to CutNewFile()
|
|
// a new file could already be cut, this is acceptable because it will simply result in an empty file which
|
|
// won't do any harm.
|
|
cdm.CutNewFile()
|
|
}
|
|
pendingDeletes, err := cdm.deleteFiles(removedFiles)
|
|
errs.Add(err)
|
|
|
|
if len(chkFileIndices) == len(removedFiles) {
|
|
// All files were deleted. Reset the current sequence.
|
|
cdm.evtlPosMtx.Lock()
|
|
|
|
// We can safely reset the sequence only if the write queue is empty. If it's not empty,
|
|
// then there may be a job in the queue that will create a new segment file with an ID
|
|
// generated before the sequence reset.
|
|
//
|
|
// The queueIsEmpty() function must be called while holding the cdm.evtlPosMtx to avoid
|
|
// a race condition with WriteChunk().
|
|
if cdm.writeQueue == nil || cdm.writeQueue.queueIsEmpty() {
|
|
if err == nil {
|
|
cdm.evtlPos.setSeq(0)
|
|
} else {
|
|
// In case of error, set it to the last file number on the disk that was not deleted.
|
|
cdm.evtlPos.setSeq(uint64(pendingDeletes[len(pendingDeletes)-1]))
|
|
}
|
|
}
|
|
|
|
cdm.evtlPosMtx.Unlock()
|
|
}
|
|
|
|
return errs.Err()
|
|
}
|
|
|
|
// deleteFiles deletes the given file sequences in order of the sequence.
|
|
// In case of an error, it returns the sorted file sequences that were not deleted from the _disk_.
|
|
func (cdm *ChunkDiskMapper) deleteFiles(removedFiles []int) ([]int, error) {
|
|
slices.Sort(removedFiles) // To delete them in order.
|
|
cdm.readPathMtx.Lock()
|
|
for _, seq := range removedFiles {
|
|
if err := cdm.closers[seq].Close(); err != nil {
|
|
cdm.readPathMtx.Unlock()
|
|
return removedFiles, err
|
|
}
|
|
delete(cdm.mmappedChunkFiles, seq)
|
|
delete(cdm.closers, seq)
|
|
}
|
|
cdm.readPathMtx.Unlock()
|
|
|
|
// We actually delete the files separately to not block the readPathMtx for long.
|
|
for i, seq := range removedFiles {
|
|
if err := os.Remove(segmentFile(cdm.dir.Name(), seq)); err != nil {
|
|
return removedFiles[i:], err
|
|
}
|
|
}
|
|
|
|
return nil, nil
|
|
}
|
|
|
|
// DeleteCorrupted deletes all the head chunk files after the one which had the corruption
|
|
// (including the corrupt file).
|
|
func (cdm *ChunkDiskMapper) DeleteCorrupted(originalErr error) error {
|
|
var cerr *CorruptionErr
|
|
if !errors.As(originalErr, &cerr) {
|
|
return fmt.Errorf("cannot handle error: %w", originalErr)
|
|
}
|
|
|
|
// Delete all the head chunk files following the corrupt head chunk file.
|
|
segs := []int{}
|
|
cdm.readPathMtx.RLock()
|
|
lastSeq := 0
|
|
for seg := range cdm.mmappedChunkFiles {
|
|
switch {
|
|
case seg >= cerr.FileIndex:
|
|
segs = append(segs, seg)
|
|
case seg > lastSeq:
|
|
lastSeq = seg
|
|
}
|
|
}
|
|
cdm.readPathMtx.RUnlock()
|
|
|
|
pendingDeletes, err := cdm.deleteFiles(segs)
|
|
cdm.evtlPosMtx.Lock()
|
|
if err == nil {
|
|
cdm.evtlPos.setSeq(uint64(lastSeq))
|
|
} else {
|
|
// In case of error, set it to the last file number on the disk that was not deleted.
|
|
cdm.evtlPos.setSeq(uint64(pendingDeletes[len(pendingDeletes)-1]))
|
|
}
|
|
cdm.evtlPosMtx.Unlock()
|
|
|
|
return err
|
|
}
|
|
|
|
// Size returns the size of the chunk files.
|
|
func (cdm *ChunkDiskMapper) Size() (int64, error) {
|
|
return fileutil.DirSize(cdm.dir.Name())
|
|
}
|
|
|
|
func (cdm *ChunkDiskMapper) curFileSize() uint64 {
|
|
return cdm.curFileOffset.Load()
|
|
}
|
|
|
|
// Close closes all the open files in ChunkDiskMapper.
|
|
// It is not longer safe to access chunks from this struct after calling Close.
|
|
func (cdm *ChunkDiskMapper) Close() error {
|
|
// Locking the eventual position lock blocks WriteChunk()
|
|
cdm.evtlPosMtx.Lock()
|
|
defer cdm.evtlPosMtx.Unlock()
|
|
|
|
if cdm.writeQueue != nil {
|
|
cdm.writeQueue.stop()
|
|
}
|
|
|
|
// 'WriteChunk' locks writePathMtx first and then readPathMtx for cutting head chunk file.
|
|
// The lock order should not be reversed here else it can cause deadlocks.
|
|
cdm.writePathMtx.Lock()
|
|
defer cdm.writePathMtx.Unlock()
|
|
cdm.readPathMtx.Lock()
|
|
defer cdm.readPathMtx.Unlock()
|
|
|
|
if cdm.closed {
|
|
return nil
|
|
}
|
|
cdm.closed = true
|
|
|
|
errs := tsdb_errors.NewMulti(
|
|
closeAllFromMap(cdm.closers),
|
|
cdm.finalizeCurFile(),
|
|
cdm.dir.Close(),
|
|
)
|
|
cdm.mmappedChunkFiles = map[int]*mmappedChunkFile{}
|
|
cdm.closers = map[int]io.Closer{}
|
|
|
|
return errs.Err()
|
|
}
|
|
|
|
func closeAllFromMap(cs map[int]io.Closer) error {
|
|
errs := tsdb_errors.NewMulti()
|
|
for _, c := range cs {
|
|
errs.Add(c.Close())
|
|
}
|
|
return errs.Err()
|
|
}
|
|
|
|
const inBufferShards = 128 // 128 is a randomly chosen number.
|
|
|
|
// chunkBuffer is a thread safe lookup table for chunks by their ref.
|
|
type chunkBuffer struct {
|
|
inBufferChunks [inBufferShards]map[ChunkDiskMapperRef]chunkenc.Chunk
|
|
inBufferChunksMtxs [inBufferShards]sync.RWMutex
|
|
}
|
|
|
|
func newChunkBuffer() *chunkBuffer {
|
|
cb := &chunkBuffer{}
|
|
for i := 0; i < inBufferShards; i++ {
|
|
cb.inBufferChunks[i] = make(map[ChunkDiskMapperRef]chunkenc.Chunk)
|
|
}
|
|
return cb
|
|
}
|
|
|
|
func (cb *chunkBuffer) put(ref ChunkDiskMapperRef, chk chunkenc.Chunk) {
|
|
shardIdx := ref % inBufferShards
|
|
|
|
cb.inBufferChunksMtxs[shardIdx].Lock()
|
|
cb.inBufferChunks[shardIdx][ref] = chk
|
|
cb.inBufferChunksMtxs[shardIdx].Unlock()
|
|
}
|
|
|
|
func (cb *chunkBuffer) get(ref ChunkDiskMapperRef) chunkenc.Chunk {
|
|
shardIdx := ref % inBufferShards
|
|
|
|
cb.inBufferChunksMtxs[shardIdx].RLock()
|
|
defer cb.inBufferChunksMtxs[shardIdx].RUnlock()
|
|
|
|
return cb.inBufferChunks[shardIdx][ref]
|
|
}
|
|
|
|
func (cb *chunkBuffer) clear() {
|
|
for i := 0; i < inBufferShards; i++ {
|
|
cb.inBufferChunksMtxs[i].Lock()
|
|
cb.inBufferChunks[i] = make(map[ChunkDiskMapperRef]chunkenc.Chunk)
|
|
cb.inBufferChunksMtxs[i].Unlock()
|
|
}
|
|
}
|