From 8b84ee5ee6bbd326d54ed189801aa1f0415c0aed Mon Sep 17 00:00:00 2001 From: Fabian Reinartz Date: Thu, 22 Dec 2016 23:33:32 +0100 Subject: [PATCH] storage: remove old storage This removes all old storage files and only keeps interfaces to still allow the code to compile. --- cmd/prometheus/config.go | 75 +- cmd/prometheus/main.go | 10 - storage/local/chunk/chunk.go | 488 ----- storage/local/chunk/chunk_test.go | 49 - storage/local/chunk/delta.go | 379 ---- storage/local/chunk/delta_helpers.go | 84 - storage/local/chunk/delta_test.go | 116 -- storage/local/chunk/doubledelta.go | 525 ----- storage/local/chunk/instrumentation.go | 100 - storage/local/chunk/varbit.go | 1210 ----------- storage/local/chunk/varbit_helpers.go | 75 - storage/local/chunk/varbit_test.go | 52 - storage/local/codable/codable.go | 447 ----- storage/local/codable/codable_test.go | 165 -- storage/local/crashrecovery.go | 548 ----- storage/local/fixtures/b0/04b821ca50ba26.db | Bin 62460 -> 0 bytes storage/local/fixtures/b0/37c21e884e4fc5.db | Bin 48927 -> 0 bytes storage/local/fixtures/b0/37de1e884e5469.db | Bin 51009 -> 0 bytes storage/local/heads.go | 251 --- storage/local/index/index.go | 296 --- storage/local/index/interface.go | 61 - storage/local/index/leveldb.go | 210 -- storage/local/instrumentation.go | 46 - storage/local/interface.go | 11 + storage/local/locker.go | 79 - storage/local/locker_test.go | 58 - storage/local/mapper.go | 218 -- storage/local/mapper_test.go | 292 --- storage/local/noop_storage.go | 100 - storage/local/persistence.go | 1578 --------------- storage/local/persistence_test.go | 1256 ------------ storage/local/series.go | 737 ------- storage/local/series_test.go | 63 - storage/local/storage.go | 1774 ---------------- storage/local/storage_test.go | 2005 ------------------- storage/local/storagetool/main.go | 59 - storage/local/test_helpers.go | 72 - 37 files changed, 13 insertions(+), 13476 deletions(-) delete mode 100644 storage/local/chunk/chunk.go delete mode 100644 storage/local/chunk/chunk_test.go delete mode 100644 storage/local/chunk/delta.go delete mode 100644 storage/local/chunk/delta_helpers.go delete mode 100644 storage/local/chunk/delta_test.go delete mode 100644 storage/local/chunk/doubledelta.go delete mode 100644 storage/local/chunk/instrumentation.go delete mode 100644 storage/local/chunk/varbit.go delete mode 100644 storage/local/chunk/varbit_helpers.go delete mode 100644 storage/local/chunk/varbit_test.go delete mode 100644 storage/local/codable/codable.go delete mode 100644 storage/local/codable/codable_test.go delete mode 100644 storage/local/crashrecovery.go delete mode 100644 storage/local/fixtures/b0/04b821ca50ba26.db delete mode 100644 storage/local/fixtures/b0/37c21e884e4fc5.db delete mode 100644 storage/local/fixtures/b0/37de1e884e5469.db delete mode 100644 storage/local/heads.go delete mode 100644 storage/local/index/index.go delete mode 100644 storage/local/index/interface.go delete mode 100644 storage/local/index/leveldb.go delete mode 100644 storage/local/instrumentation.go delete mode 100644 storage/local/locker.go delete mode 100644 storage/local/locker_test.go delete mode 100644 storage/local/mapper.go delete mode 100644 storage/local/mapper_test.go delete mode 100644 storage/local/noop_storage.go delete mode 100644 storage/local/persistence.go delete mode 100644 storage/local/persistence_test.go delete mode 100644 storage/local/series.go delete mode 100644 storage/local/series_test.go delete mode 100644 storage/local/storage.go delete mode 100644 storage/local/storage_test.go delete mode 100644 storage/local/storagetool/main.go delete mode 100644 storage/local/test_helpers.go diff --git a/cmd/prometheus/config.go b/cmd/prometheus/config.go index d0662ba9c3..e7ff539b4b 100644 --- a/cmd/prometheus/config.go +++ b/cmd/prometheus/config.go @@ -29,9 +29,6 @@ import ( "github.com/prometheus/common/log" "github.com/prometheus/prometheus/notifier" "github.com/prometheus/prometheus/promql" - "github.com/prometheus/prometheus/storage/local" - "github.com/prometheus/prometheus/storage/local/chunk" - "github.com/prometheus/prometheus/storage/local/index" "github.com/prometheus/prometheus/web" ) @@ -43,7 +40,7 @@ var cfg = struct { printVersion bool configFile string - storage local.MemorySeriesStorageOptions + localStoragePath string localStorageEngine string notifier notifier.Options notifierTimeout time.Duration @@ -61,9 +58,6 @@ func init() { cfg.fs = flag.NewFlagSet(os.Args[0], flag.ContinueOnError) cfg.fs.Usage = usage - // Set additional defaults. - cfg.storage.SyncStrategy = local.Adaptive - cfg.fs.BoolVar( &cfg.printVersion, "version", false, "Print version information.", @@ -117,74 +111,9 @@ func init() { // Storage. cfg.fs.StringVar( - &cfg.storage.PersistenceStoragePath, "storage.local.path", "data", + &cfg.localStoragePath, "storage.local.path", "data", "Base path for metrics storage.", ) - cfg.fs.IntVar( - &cfg.storage.MemoryChunks, "storage.local.memory-chunks", 1024*1024, - "How many chunks to keep in memory. While the size of a chunk is 1kiB, the total memory usage will be significantly higher than this value * 1kiB. Furthermore, for various reasons, more chunks might have to be kept in memory temporarily. Sample ingestion will be throttled if the configured value is exceeded by more than 10%.", - ) - cfg.fs.DurationVar( - &cfg.storage.PersistenceRetentionPeriod, "storage.local.retention", 15*24*time.Hour, - "How long to retain samples in the local storage.", - ) - cfg.fs.IntVar( - &cfg.storage.MaxChunksToPersist, "storage.local.max-chunks-to-persist", 512*1024, - "How many chunks can be waiting for persistence before sample ingestion will be throttled. Many chunks waiting to be persisted will increase the checkpoint size.", - ) - cfg.fs.DurationVar( - &cfg.storage.CheckpointInterval, "storage.local.checkpoint-interval", 5*time.Minute, - "The period at which the in-memory metrics and the chunks not yet persisted to series files are checkpointed.", - ) - cfg.fs.IntVar( - &cfg.storage.CheckpointDirtySeriesLimit, "storage.local.checkpoint-dirty-series-limit", 5000, - "If approx. that many time series are in a state that would require a recovery operation after a crash, a checkpoint is triggered, even if the checkpoint interval hasn't passed yet. A recovery operation requires a disk seek. The default limit intends to keep the recovery time below 1min even on spinning disks. With SSD, recovery is much faster, so you might want to increase this value in that case to avoid overly frequent checkpoints.", - ) - cfg.fs.Var( - &cfg.storage.SyncStrategy, "storage.local.series-sync-strategy", - "When to sync series files after modification. Possible values: 'never', 'always', 'adaptive'. Sync'ing slows down storage performance but reduces the risk of data loss in case of an OS crash. With the 'adaptive' strategy, series files are sync'd for as long as the storage is not too much behind on chunk persistence.", - ) - cfg.fs.Float64Var( - &cfg.storage.MinShrinkRatio, "storage.local.series-file-shrink-ratio", 0.1, - "A series file is only truncated (to delete samples that have exceeded the retention period) if it shrinks by at least the provided ratio. This saves I/O operations while causing only a limited storage space overhead. If 0 or smaller, truncation will be performed even for a single dropped chunk, while 1 or larger will effectively prevent any truncation.", - ) - cfg.fs.BoolVar( - &cfg.storage.Dirty, "storage.local.dirty", false, - "If set, the local storage layer will perform crash recovery even if the last shutdown appears to be clean.", - ) - cfg.fs.BoolVar( - &cfg.storage.PedanticChecks, "storage.local.pedantic-checks", false, - "If set, a crash recovery will perform checks on each series file. This might take a very long time.", - ) - cfg.fs.Var( - &chunk.DefaultEncoding, "storage.local.chunk-encoding-version", - "Which chunk encoding version to use for newly created chunks. Currently supported is 0 (delta encoding), 1 (double-delta encoding), and 2 (double-delta encoding with variable bit-width).", - ) - // Index cache sizes. - cfg.fs.IntVar( - &index.FingerprintMetricCacheSize, "storage.local.index-cache-size.fingerprint-to-metric", index.FingerprintMetricCacheSize, - "The size in bytes for the fingerprint to metric index cache.", - ) - cfg.fs.IntVar( - &index.FingerprintTimeRangeCacheSize, "storage.local.index-cache-size.fingerprint-to-timerange", index.FingerprintTimeRangeCacheSize, - "The size in bytes for the metric time range index cache.", - ) - cfg.fs.IntVar( - &index.LabelNameLabelValuesCacheSize, "storage.local.index-cache-size.label-name-to-label-values", index.LabelNameLabelValuesCacheSize, - "The size in bytes for the label name to label values index cache.", - ) - cfg.fs.IntVar( - &index.LabelPairFingerprintsCacheSize, "storage.local.index-cache-size.label-pair-to-fingerprints", index.LabelPairFingerprintsCacheSize, - "The size in bytes for the label pair to fingerprints index cache.", - ) - cfg.fs.IntVar( - &cfg.storage.NumMutexes, "storage.local.num-fingerprint-mutexes", 4096, - "The number of mutexes used for fingerprint locking.", - ) - cfg.fs.StringVar( - &cfg.localStorageEngine, "storage.local.engine", "persisted", - "Local storage engine. Supported values are: 'persisted' (full local storage with on-disk persistence) and 'none' (no local storage).", - ) // Alertmanager. cfg.fs.Var( diff --git a/cmd/prometheus/main.go b/cmd/prometheus/main.go index e7680783ed..bacecba6aa 100644 --- a/cmd/prometheus/main.go +++ b/cmd/prometheus/main.go @@ -83,16 +83,6 @@ func Main() int { ) var localStorage local.Storage - switch cfg.localStorageEngine { - case "persisted": - localStorage = local.NewMemorySeriesStorage(&cfg.storage) - sampleAppender = storage.Fanout{localStorage} - case "none": - localStorage = &local.NoopStorage{} - default: - log.Errorf("Invalid local storage engine %q", cfg.localStorageEngine) - return 1 - } reloadableRemoteStorage := remote.New() sampleAppender = append(sampleAppender, reloadableRemoteStorage) diff --git a/storage/local/chunk/chunk.go b/storage/local/chunk/chunk.go deleted file mode 100644 index 733028d16b..0000000000 --- a/storage/local/chunk/chunk.go +++ /dev/null @@ -1,488 +0,0 @@ -// Copyright 2014 The Prometheus Authors -// Licensed under the Apache License, Version 2.0 (the "License"); -// you may not use this file except in compliance with the License. -// You may obtain a copy of the License at -// -// http://www.apache.org/licenses/LICENSE-2.0 -// -// Unless required by applicable law or agreed to in writing, software -// distributed under the License is distributed on an "AS IS" BASIS, -// WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. -// See the License for the specific language governing permissions and -// limitations under the License. - -package chunk - -import ( - "container/list" - "errors" - "fmt" - "io" - "sort" - "sync" - "sync/atomic" - - "github.com/prometheus/common/model" - - "github.com/prometheus/prometheus/storage/metric" -) - -// ChunkLen is the length of a chunk in bytes. -const ChunkLen = 1024 - -// DefaultEncoding can be changed via a flag. -var DefaultEncoding = DoubleDelta - -var errChunkBoundsExceeded = errors.New("attempted access outside of chunk boundaries") - -// EvictRequest is a request to evict a chunk from memory. -type EvictRequest struct { - Desc *Desc - Evict bool -} - -// Encoding defines which encoding we are using, delta, doubledelta, or varbit -type Encoding byte - -// String implements flag.Value. -func (e Encoding) String() string { - return fmt.Sprintf("%d", e) -} - -// Set implements flag.Value. -func (e *Encoding) Set(s string) error { - switch s { - case "0": - *e = Delta - case "1": - *e = DoubleDelta - case "2": - *e = Varbit - default: - return fmt.Errorf("invalid chunk encoding: %s", s) - } - return nil -} - -const ( - // Delta encoding - Delta Encoding = iota - // DoubleDelta encoding - DoubleDelta - // Varbit encoding - Varbit -) - -// Desc contains meta-data for a chunk. Pay special attention to the -// documented requirements for calling its methods concurrently (WRT pinning and -// locking). The doc comments spell out the requirements for each method, but -// here is an overview and general explanation: -// -// Everything that changes the pinning of the underlying chunk or deals with its -// eviction is protected by a mutex. This affects the following methods: Pin, -// Unpin, RefCount, IsEvicted, MaybeEvict. These methods can be called at any -// time without further prerequisites. -// -// Another group of methods acts on (or sets) the underlying chunk. These -// methods involve no locking. They may only be called if the caller has pinned -// the chunk (to guarantee the chunk is not evicted concurrently). Also, the -// caller must make sure nobody else will call these methods concurrently, -// either by holding the sole reference to the Desc (usually during loading -// or creation) or by locking the fingerprint of the series the Desc -// belongs to. The affected methods are: Add, MaybePopulateLastTime, SetChunk. -// -// Finally, there are the special cases FirstTime and LastTime. LastTime requires -// to have locked the fingerprint of the series but the chunk does not need to -// be pinned. That's because the ChunkLastTime field in Desc gets populated -// upon completion of the chunk (when it is still pinned, and which happens -// while the series's fingerprint is locked). Once that has happened, calling -// LastTime does not require the chunk to be loaded anymore. Before that has -// happened, the chunk is pinned anyway. The ChunkFirstTime field in Desc -// is populated upon creation of a Desc, so it is alway safe to call -// FirstTime. The FirstTime method is arguably not needed and only there for -// consistency with LastTime. -type Desc struct { - sync.Mutex // Protects pinning. - C Chunk // nil if chunk is evicted. - rCnt int - ChunkFirstTime model.Time // Populated at creation. Immutable. - ChunkLastTime model.Time // Populated on closing of the chunk, model.Earliest if unset. - - // EvictListElement is nil if the chunk is not in the evict list. - // EvictListElement is _not_ protected by the Desc mutex. - // It must only be touched by the evict list handler in MemorySeriesStorage. - EvictListElement *list.Element -} - -// NewDesc creates a new Desc pointing to the provided chunk. The provided chunk -// is assumed to be not persisted yet. Therefore, the refCount of the new -// Desc is 1 (preventing eviction prior to persisting). -func NewDesc(c Chunk, firstTime model.Time) *Desc { - Ops.WithLabelValues(CreateAndPin).Inc() - atomic.AddInt64(&NumMemChunks, 1) - NumMemDescs.Inc() - return &Desc{ - C: c, - rCnt: 1, - ChunkFirstTime: firstTime, - ChunkLastTime: model.Earliest, - } -} - -// Add adds a sample pair to the underlying chunk. For safe concurrent access, -// The chunk must be pinned, and the caller must have locked the fingerprint of -// the series. -func (d *Desc) Add(s model.SamplePair) ([]Chunk, error) { - return d.C.Add(s) -} - -// Pin increments the refCount by one. Upon increment from 0 to 1, this -// Desc is removed from the evict list. To enable the latter, the -// evictRequests channel has to be provided. This method can be called -// concurrently at any time. -func (d *Desc) Pin(evictRequests chan<- EvictRequest) { - d.Lock() - defer d.Unlock() - - if d.rCnt == 0 { - // Remove ourselves from the evict list. - evictRequests <- EvictRequest{d, false} - } - d.rCnt++ -} - -// Unpin decrements the refCount by one. Upon decrement from 1 to 0, this -// Desc is added to the evict list. To enable the latter, the evictRequests -// channel has to be provided. This method can be called concurrently at any -// time. -func (d *Desc) Unpin(evictRequests chan<- EvictRequest) { - d.Lock() - defer d.Unlock() - - if d.rCnt == 0 { - panic("cannot unpin already unpinned chunk") - } - d.rCnt-- - if d.rCnt == 0 { - // Add ourselves to the back of the evict list. - evictRequests <- EvictRequest{d, true} - } -} - -// RefCount returns the number of pins. This method can be called concurrently -// at any time. -func (d *Desc) RefCount() int { - d.Lock() - defer d.Unlock() - - return d.rCnt -} - -// FirstTime returns the timestamp of the first sample in the chunk. This method -// can be called concurrently at any time. It only returns the immutable -// d.ChunkFirstTime without any locking. Arguably, this method is -// useless. However, it provides consistency with the LastTime method. -func (d *Desc) FirstTime() model.Time { - return d.ChunkFirstTime -} - -// LastTime returns the timestamp of the last sample in the chunk. For safe -// concurrent access, this method requires the fingerprint of the time series to -// be locked. -func (d *Desc) LastTime() (model.Time, error) { - if d.ChunkLastTime != model.Earliest || d.C == nil { - return d.ChunkLastTime, nil - } - return d.C.NewIterator().LastTimestamp() -} - -// MaybePopulateLastTime populates the ChunkLastTime from the underlying chunk -// if it has not yet happened. Call this method directly after having added the -// last sample to a chunk or after closing a head chunk due to age. For safe -// concurrent access, the chunk must be pinned, and the caller must have locked -// the fingerprint of the series. -func (d *Desc) MaybePopulateLastTime() error { - if d.ChunkLastTime == model.Earliest && d.C != nil { - t, err := d.C.NewIterator().LastTimestamp() - if err != nil { - return err - } - d.ChunkLastTime = t - } - return nil -} - -// IsEvicted returns whether the chunk is evicted. For safe concurrent access, -// the caller must have locked the fingerprint of the series. -func (d *Desc) IsEvicted() bool { - // Locking required here because we do not want the caller to force - // pinning the chunk first, so it could be evicted while this method is - // called. - d.Lock() - defer d.Unlock() - - return d.C == nil -} - -// SetChunk sets the underlying chunk. The caller must have locked the -// fingerprint of the series and must have "pre-pinned" the chunk (i.e. first -// call Pin and then set the chunk). -func (d *Desc) SetChunk(c Chunk) { - if d.C != nil { - panic("chunk already set") - } - d.C = c -} - -// MaybeEvict evicts the chunk if the refCount is 0. It returns whether the chunk -// is now evicted, which includes the case that the chunk was evicted even -// before this method was called. It can be called concurrently at any time. -func (d *Desc) MaybeEvict() bool { - d.Lock() - defer d.Unlock() - - if d.C == nil { - return true - } - if d.rCnt != 0 { - return false - } - if d.ChunkLastTime == model.Earliest { - // This must never happen. - panic("ChunkLastTime not populated for evicted chunk") - } - d.C = nil - Ops.WithLabelValues(Evict).Inc() - atomic.AddInt64(&NumMemChunks, -1) - return true -} - -// Chunk is the interface for all chunks. Chunks are generally not -// goroutine-safe. -type Chunk interface { - // Add adds a SamplePair to the chunks, performs any necessary - // re-encoding, and adds any necessary overflow chunks. It returns the - // new version of the original chunk, followed by overflow chunks, if - // any. The first chunk returned might be the same as the original one - // or a newly allocated version. In any case, take the returned chunk as - // the relevant one and discard the original chunk. - Add(sample model.SamplePair) ([]Chunk, error) - Clone() Chunk - FirstTime() model.Time - NewIterator() Iterator - Marshal(io.Writer) error - MarshalToBuf([]byte) error - Unmarshal(io.Reader) error - UnmarshalFromBuf([]byte) error - Encoding() Encoding - Utilization() float64 - - // Len returns the number of samples in the chunk. Implementations may be - // expensive. - Len() int -} - -// Iterator enables efficient access to the content of a chunk. It is -// generally not safe to use an Iterator concurrently with or after chunk -// mutation. -type Iterator interface { - // Gets the last timestamp in the chunk. - LastTimestamp() (model.Time, error) - // Whether a given timestamp is contained between first and last value - // in the chunk. - Contains(model.Time) (bool, error) - // Scans the next value in the chunk. Directly after the iterator has - // been created, the next value is the first value in the - // chunk. Otherwise, it is the value following the last value scanned or - // found (by one of the Find... methods). Returns false if either the - // end of the chunk is reached or an error has occurred. - Scan() bool - // Finds the most recent value at or before the provided time. Returns - // false if either the chunk contains no value at or before the provided - // time, or an error has occurred. - FindAtOrBefore(model.Time) bool - // Finds the oldest value at or after the provided time. Returns false - // if either the chunk contains no value at or after the provided time, - // or an error has occurred. - FindAtOrAfter(model.Time) bool - // Returns the last value scanned (by the scan method) or found (by one - // of the find... methods). It returns model.ZeroSamplePair before any of - // those methods were called. - Value() model.SamplePair - // Returns the last error encountered. In general, an error signals data - // corruption in the chunk and requires quarantining. - Err() error -} - -// RangeValues is a utility function that retrieves all values within the given -// range from an Iterator. -func RangeValues(it Iterator, in metric.Interval) ([]model.SamplePair, error) { - result := []model.SamplePair{} - if !it.FindAtOrAfter(in.OldestInclusive) { - return result, it.Err() - } - for !it.Value().Timestamp.After(in.NewestInclusive) { - result = append(result, it.Value()) - if !it.Scan() { - break - } - } - return result, it.Err() -} - -// addToOverflowChunk is a utility function that creates a new chunk as overflow -// chunk, adds the provided sample to it, and returns a chunk slice containing -// the provided old chunk followed by the new overflow chunk. -func addToOverflowChunk(c Chunk, s model.SamplePair) ([]Chunk, error) { - overflowChunks, err := New().Add(s) - if err != nil { - return nil, err - } - return []Chunk{c, overflowChunks[0]}, nil -} - -// transcodeAndAdd is a utility function that transcodes the dst chunk into the -// provided src chunk (plus the necessary overflow chunks) and then adds the -// provided sample. It returns the new chunks (transcoded plus overflow) with -// the new sample at the end. -func transcodeAndAdd(dst Chunk, src Chunk, s model.SamplePair) ([]Chunk, error) { - Ops.WithLabelValues(Transcode).Inc() - - var ( - head = dst - body, NewChunks []Chunk - err error - ) - - it := src.NewIterator() - for it.Scan() { - if NewChunks, err = head.Add(it.Value()); err != nil { - return nil, err - } - body = append(body, NewChunks[:len(NewChunks)-1]...) - head = NewChunks[len(NewChunks)-1] - } - if it.Err() != nil { - return nil, it.Err() - } - - if NewChunks, err = head.Add(s); err != nil { - return nil, err - } - return append(body, NewChunks...), nil -} - -// New creates a new chunk according to the encoding set by the -// DefaultEncoding flag. -func New() Chunk { - chunk, err := NewForEncoding(DefaultEncoding) - if err != nil { - panic(err) - } - return chunk -} - -// NewForEncoding allows configuring what chunk type you want -func NewForEncoding(encoding Encoding) (Chunk, error) { - switch encoding { - case Delta: - return newDeltaEncodedChunk(d1, d0, true, ChunkLen), nil - case DoubleDelta: - return newDoubleDeltaEncodedChunk(d1, d0, true, ChunkLen), nil - case Varbit: - return newVarbitChunk(varbitZeroEncoding), nil - default: - return nil, fmt.Errorf("unknown chunk encoding: %v", encoding) - } -} - -// indexAccessor allows accesses to samples by index. -type indexAccessor interface { - timestampAtIndex(int) model.Time - sampleValueAtIndex(int) model.SampleValue - err() error -} - -// indexAccessingChunkIterator is a chunk iterator for chunks for which an -// indexAccessor implementation exists. -type indexAccessingChunkIterator struct { - len int - pos int - lastValue model.SamplePair - acc indexAccessor -} - -func newIndexAccessingChunkIterator(len int, acc indexAccessor) *indexAccessingChunkIterator { - return &indexAccessingChunkIterator{ - len: len, - pos: -1, - lastValue: model.ZeroSamplePair, - acc: acc, - } -} - -// lastTimestamp implements Iterator. -func (it *indexAccessingChunkIterator) LastTimestamp() (model.Time, error) { - return it.acc.timestampAtIndex(it.len - 1), it.acc.err() -} - -// contains implements Iterator. -func (it *indexAccessingChunkIterator) Contains(t model.Time) (bool, error) { - return !t.Before(it.acc.timestampAtIndex(0)) && - !t.After(it.acc.timestampAtIndex(it.len-1)), it.acc.err() -} - -// scan implements Iterator. -func (it *indexAccessingChunkIterator) Scan() bool { - it.pos++ - if it.pos >= it.len { - return false - } - it.lastValue = model.SamplePair{ - Timestamp: it.acc.timestampAtIndex(it.pos), - Value: it.acc.sampleValueAtIndex(it.pos), - } - return it.acc.err() == nil -} - -// findAtOrBefore implements Iterator. -func (it *indexAccessingChunkIterator) FindAtOrBefore(t model.Time) bool { - i := sort.Search(it.len, func(i int) bool { - return it.acc.timestampAtIndex(i).After(t) - }) - if i == 0 || it.acc.err() != nil { - return false - } - it.pos = i - 1 - it.lastValue = model.SamplePair{ - Timestamp: it.acc.timestampAtIndex(i - 1), - Value: it.acc.sampleValueAtIndex(i - 1), - } - return true -} - -// findAtOrAfter implements Iterator. -func (it *indexAccessingChunkIterator) FindAtOrAfter(t model.Time) bool { - i := sort.Search(it.len, func(i int) bool { - return !it.acc.timestampAtIndex(i).Before(t) - }) - if i == it.len || it.acc.err() != nil { - return false - } - it.pos = i - it.lastValue = model.SamplePair{ - Timestamp: it.acc.timestampAtIndex(i), - Value: it.acc.sampleValueAtIndex(i), - } - return true -} - -// value implements Iterator. -func (it *indexAccessingChunkIterator) Value() model.SamplePair { - return it.lastValue -} - -// err implements Iterator. -func (it *indexAccessingChunkIterator) Err() error { - return it.acc.err() -} diff --git a/storage/local/chunk/chunk_test.go b/storage/local/chunk/chunk_test.go deleted file mode 100644 index 20ac73c7e6..0000000000 --- a/storage/local/chunk/chunk_test.go +++ /dev/null @@ -1,49 +0,0 @@ -// Copyright 2016 The Prometheus Authors -// Licensed under the Apache License, Version 2.0 (the "License"); -// you may not use this file except in compliance with the License. -// You may obtain a copy of the License at -// -// http://www.apache.org/licenses/LICENSE-2.0 -// -// Unless required by applicable law or agreed to in writing, software -// distributed under the License is distributed on an "AS IS" BASIS, -// WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. -// See the License for the specific language governing permissions and -// limitations under the License. - -// Note: this file has tests for code in both delta.go and doubledelta.go -- -// it may make sense to split those out later, but given that the tests are -// near-identical and share a helper, this feels simpler for now. - -package chunk - -import ( - "testing" - - "github.com/prometheus/common/model" -) - -func TestLen(t *testing.T) { - chunks := []Chunk{} - for _, encoding := range []Encoding{Delta, DoubleDelta, Varbit} { - c, err := NewForEncoding(encoding) - if err != nil { - t.Fatal(err) - } - chunks = append(chunks, c) - } - - for _, c := range chunks { - for i := 0; i <= 10; i++ { - if c.Len() != i { - t.Errorf("chunk type %s should have %d samples, had %d", c.Encoding(), i, c.Len()) - } - - cs, _ := c.Add(model.SamplePair{ - Timestamp: model.Time(i), - Value: model.SampleValue(i), - }) - c = cs[0] - } - } -} diff --git a/storage/local/chunk/delta.go b/storage/local/chunk/delta.go deleted file mode 100644 index 4e3fd06455..0000000000 --- a/storage/local/chunk/delta.go +++ /dev/null @@ -1,379 +0,0 @@ -// Copyright 2014 The Prometheus Authors -// Licensed under the Apache License, Version 2.0 (the "License"); -// you may not use this file except in compliance with the License. -// You may obtain a copy of the License at -// -// http://www.apache.org/licenses/LICENSE-2.0 -// -// Unless required by applicable law or agreed to in writing, software -// distributed under the License is distributed on an "AS IS" BASIS, -// WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. -// See the License for the specific language governing permissions and -// limitations under the License. - -package chunk - -import ( - "encoding/binary" - "fmt" - "io" - "math" - - "github.com/prometheus/common/model" -) - -// The 21-byte header of a delta-encoded chunk looks like: -// -// - time delta bytes: 1 bytes -// - value delta bytes: 1 bytes -// - is integer: 1 byte -// - base time: 8 bytes -// - base value: 8 bytes -// - used buf bytes: 2 bytes -const ( - deltaHeaderBytes = 21 - - deltaHeaderTimeBytesOffset = 0 - deltaHeaderValueBytesOffset = 1 - deltaHeaderIsIntOffset = 2 - deltaHeaderBaseTimeOffset = 3 - deltaHeaderBaseValueOffset = 11 - deltaHeaderBufLenOffset = 19 -) - -// A deltaEncodedChunk adaptively stores sample timestamps and values with a -// delta encoding of various types (int, float) and bit widths. However, once 8 -// bytes would be needed to encode a delta value, a fall-back to the absolute -// numbers happens (so that timestamps are saved directly as int64 and values as -// float64). It implements the chunk interface. -type deltaEncodedChunk []byte - -// newDeltaEncodedChunk returns a newly allocated deltaEncodedChunk. -func newDeltaEncodedChunk(tb, vb deltaBytes, isInt bool, length int) *deltaEncodedChunk { - if tb < 1 { - panic("need at least 1 time delta byte") - } - if length < deltaHeaderBytes+16 { - panic(fmt.Errorf( - "chunk length %d bytes is insufficient, need at least %d", - length, deltaHeaderBytes+16, - )) - } - c := make(deltaEncodedChunk, deltaHeaderIsIntOffset+1, length) - - c[deltaHeaderTimeBytesOffset] = byte(tb) - c[deltaHeaderValueBytesOffset] = byte(vb) - if vb < d8 && isInt { // Only use int for fewer than 8 value delta bytes. - c[deltaHeaderIsIntOffset] = 1 - } else { - c[deltaHeaderIsIntOffset] = 0 - } - - return &c -} - -// Add implements chunk. -func (c deltaEncodedChunk) Add(s model.SamplePair) ([]Chunk, error) { - // TODO(beorn7): Since we return &c, this method might cause an unnecessary allocation. - if c.Len() == 0 { - c = c[:deltaHeaderBytes] - binary.LittleEndian.PutUint64(c[deltaHeaderBaseTimeOffset:], uint64(s.Timestamp)) - binary.LittleEndian.PutUint64(c[deltaHeaderBaseValueOffset:], math.Float64bits(float64(s.Value))) - } - - remainingBytes := cap(c) - len(c) - sampleSize := c.sampleSize() - - // Do we generally have space for another sample in this chunk? If not, - // overflow into a new one. - if remainingBytes < sampleSize { - return addToOverflowChunk(&c, s) - } - - baseValue := c.baseValue() - dt := s.Timestamp - c.baseTime() - if dt < 0 { - return nil, fmt.Errorf("time delta is less than zero: %v", dt) - } - - dv := s.Value - baseValue - tb := c.timeBytes() - vb := c.valueBytes() - isInt := c.isInt() - - // If the new sample is incompatible with the current encoding, reencode the - // existing chunk data into new chunk(s). - - ntb, nvb, nInt := tb, vb, isInt - if isInt && !isInt64(dv) { - // int->float. - nvb = d4 - nInt = false - } else if !isInt && vb == d4 && baseValue+model.SampleValue(float32(dv)) != s.Value { - // float32->float64. - nvb = d8 - } else { - if tb < d8 { - // Maybe more bytes for timestamp. - ntb = max(tb, bytesNeededForUnsignedTimestampDelta(dt)) - } - if c.isInt() && vb < d8 { - // Maybe more bytes for sample value. - nvb = max(vb, bytesNeededForIntegerSampleValueDelta(dv)) - } - } - if tb != ntb || vb != nvb || isInt != nInt { - if len(c)*2 < cap(c) { - return transcodeAndAdd(newDeltaEncodedChunk(ntb, nvb, nInt, cap(c)), &c, s) - } - // Chunk is already half full. Better create a new one and save the transcoding efforts. - return addToOverflowChunk(&c, s) - } - - offset := len(c) - c = c[:offset+sampleSize] - - switch tb { - case d1: - c[offset] = byte(dt) - case d2: - binary.LittleEndian.PutUint16(c[offset:], uint16(dt)) - case d4: - binary.LittleEndian.PutUint32(c[offset:], uint32(dt)) - case d8: - // Store the absolute value (no delta) in case of d8. - binary.LittleEndian.PutUint64(c[offset:], uint64(s.Timestamp)) - default: - return nil, fmt.Errorf("invalid number of bytes for time delta: %d", tb) - } - - offset += int(tb) - - if c.isInt() { - switch vb { - case d0: - // No-op. Constant value is stored as base value. - case d1: - c[offset] = byte(int8(dv)) - case d2: - binary.LittleEndian.PutUint16(c[offset:], uint16(int16(dv))) - case d4: - binary.LittleEndian.PutUint32(c[offset:], uint32(int32(dv))) - // d8 must not happen. Those samples are encoded as float64. - default: - return nil, fmt.Errorf("invalid number of bytes for integer delta: %d", vb) - } - } else { - switch vb { - case d4: - binary.LittleEndian.PutUint32(c[offset:], math.Float32bits(float32(dv))) - case d8: - // Store the absolute value (no delta) in case of d8. - binary.LittleEndian.PutUint64(c[offset:], math.Float64bits(float64(s.Value))) - default: - return nil, fmt.Errorf("invalid number of bytes for floating point delta: %d", vb) - } - } - return []Chunk{&c}, nil -} - -// Clone implements chunk. -func (c deltaEncodedChunk) Clone() Chunk { - clone := make(deltaEncodedChunk, len(c), cap(c)) - copy(clone, c) - return &clone -} - -// FirstTime implements chunk. -func (c deltaEncodedChunk) FirstTime() model.Time { - return c.baseTime() -} - -// NewIterator implements chunk. -func (c *deltaEncodedChunk) NewIterator() Iterator { - return newIndexAccessingChunkIterator(c.Len(), &deltaEncodedIndexAccessor{ - c: *c, - baseT: c.baseTime(), - baseV: c.baseValue(), - tBytes: c.timeBytes(), - vBytes: c.valueBytes(), - isInt: c.isInt(), - }) -} - -// Marshal implements chunk. -func (c deltaEncodedChunk) Marshal(w io.Writer) error { - if len(c) > math.MaxUint16 { - panic("chunk buffer length would overflow a 16 bit uint.") - } - binary.LittleEndian.PutUint16(c[deltaHeaderBufLenOffset:], uint16(len(c))) - - n, err := w.Write(c[:cap(c)]) - if err != nil { - return err - } - if n != cap(c) { - return fmt.Errorf("wanted to write %d bytes, wrote %d", cap(c), n) - } - return nil -} - -// MarshalToBuf implements chunk. -func (c deltaEncodedChunk) MarshalToBuf(buf []byte) error { - if len(c) > math.MaxUint16 { - panic("chunk buffer length would overflow a 16 bit uint") - } - binary.LittleEndian.PutUint16(c[deltaHeaderBufLenOffset:], uint16(len(c))) - - n := copy(buf, c) - if n != len(c) { - return fmt.Errorf("wanted to copy %d bytes to buffer, copied %d", len(c), n) - } - return nil -} - -// Unmarshal implements chunk. -func (c *deltaEncodedChunk) Unmarshal(r io.Reader) error { - *c = (*c)[:cap(*c)] - if _, err := io.ReadFull(r, *c); err != nil { - return err - } - return c.setLen() -} - -// UnmarshalFromBuf implements chunk. -func (c *deltaEncodedChunk) UnmarshalFromBuf(buf []byte) error { - *c = (*c)[:cap(*c)] - copy(*c, buf) - return c.setLen() -} - -// setLen sets the length of the underlying slice and performs some sanity checks. -func (c *deltaEncodedChunk) setLen() error { - l := binary.LittleEndian.Uint16((*c)[deltaHeaderBufLenOffset:]) - if int(l) > cap(*c) { - return fmt.Errorf("delta chunk length exceeded during unmarshaling: %d", l) - } - if int(l) < deltaHeaderBytes { - return fmt.Errorf("delta chunk length less than header size: %d < %d", l, deltaHeaderBytes) - } - switch c.timeBytes() { - case d1, d2, d4, d8: - // Pass. - default: - return fmt.Errorf("invalid number of time bytes in delta chunk: %d", c.timeBytes()) - } - switch c.valueBytes() { - case d0, d1, d2, d4, d8: - // Pass. - default: - return fmt.Errorf("invalid number of value bytes in delta chunk: %d", c.valueBytes()) - } - *c = (*c)[:l] - return nil -} - -// Encoding implements chunk. -func (c deltaEncodedChunk) Encoding() Encoding { return Delta } - -// Utilization implements chunk. -func (c deltaEncodedChunk) Utilization() float64 { - return float64(len(c)) / float64(cap(c)) -} - -func (c deltaEncodedChunk) timeBytes() deltaBytes { - return deltaBytes(c[deltaHeaderTimeBytesOffset]) -} - -func (c deltaEncodedChunk) valueBytes() deltaBytes { - return deltaBytes(c[deltaHeaderValueBytesOffset]) -} - -func (c deltaEncodedChunk) isInt() bool { - return c[deltaHeaderIsIntOffset] == 1 -} - -func (c deltaEncodedChunk) baseTime() model.Time { - return model.Time(binary.LittleEndian.Uint64(c[deltaHeaderBaseTimeOffset:])) -} - -func (c deltaEncodedChunk) baseValue() model.SampleValue { - return model.SampleValue(math.Float64frombits(binary.LittleEndian.Uint64(c[deltaHeaderBaseValueOffset:]))) -} - -func (c deltaEncodedChunk) sampleSize() int { - return int(c.timeBytes() + c.valueBytes()) -} - -// Len implements Chunk. Runs in constant time. -func (c deltaEncodedChunk) Len() int { - if len(c) < deltaHeaderBytes { - return 0 - } - return (len(c) - deltaHeaderBytes) / c.sampleSize() -} - -// deltaEncodedIndexAccessor implements indexAccessor. -type deltaEncodedIndexAccessor struct { - c deltaEncodedChunk - baseT model.Time - baseV model.SampleValue - tBytes, vBytes deltaBytes - isInt bool - lastErr error -} - -func (acc *deltaEncodedIndexAccessor) err() error { - return acc.lastErr -} - -func (acc *deltaEncodedIndexAccessor) timestampAtIndex(idx int) model.Time { - offset := deltaHeaderBytes + idx*int(acc.tBytes+acc.vBytes) - - switch acc.tBytes { - case d1: - return acc.baseT + model.Time(uint8(acc.c[offset])) - case d2: - return acc.baseT + model.Time(binary.LittleEndian.Uint16(acc.c[offset:])) - case d4: - return acc.baseT + model.Time(binary.LittleEndian.Uint32(acc.c[offset:])) - case d8: - // Take absolute value for d8. - return model.Time(binary.LittleEndian.Uint64(acc.c[offset:])) - default: - acc.lastErr = fmt.Errorf("invalid number of bytes for time delta: %d", acc.tBytes) - return model.Earliest - } -} - -func (acc *deltaEncodedIndexAccessor) sampleValueAtIndex(idx int) model.SampleValue { - offset := deltaHeaderBytes + idx*int(acc.tBytes+acc.vBytes) + int(acc.tBytes) - - if acc.isInt { - switch acc.vBytes { - case d0: - return acc.baseV - case d1: - return acc.baseV + model.SampleValue(int8(acc.c[offset])) - case d2: - return acc.baseV + model.SampleValue(int16(binary.LittleEndian.Uint16(acc.c[offset:]))) - case d4: - return acc.baseV + model.SampleValue(int32(binary.LittleEndian.Uint32(acc.c[offset:]))) - // No d8 for ints. - default: - acc.lastErr = fmt.Errorf("invalid number of bytes for integer delta: %d", acc.vBytes) - return 0 - } - } else { - switch acc.vBytes { - case d4: - return acc.baseV + model.SampleValue(math.Float32frombits(binary.LittleEndian.Uint32(acc.c[offset:]))) - case d8: - // Take absolute value for d8. - return model.SampleValue(math.Float64frombits(binary.LittleEndian.Uint64(acc.c[offset:]))) - default: - acc.lastErr = fmt.Errorf("invalid number of bytes for floating point delta: %d", acc.vBytes) - return 0 - } - } -} diff --git a/storage/local/chunk/delta_helpers.go b/storage/local/chunk/delta_helpers.go deleted file mode 100644 index 81e5d18cb9..0000000000 --- a/storage/local/chunk/delta_helpers.go +++ /dev/null @@ -1,84 +0,0 @@ -// Copyright 2015 The Prometheus Authors -// Licensed under the Apache License, Version 2.0 (the "License"); -// you may not use this file except in compliance with the License. -// You may obtain a copy of the License at -// -// http://www.apache.org/licenses/LICENSE-2.0 -// -// Unless required by applicable law or agreed to in writing, software -// distributed under the License is distributed on an "AS IS" BASIS, -// WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. -// See the License for the specific language governing permissions and -// limitations under the License. - -package chunk - -import ( - "math" - - "github.com/prometheus/common/model" -) - -type deltaBytes byte - -const ( - d0 deltaBytes = 0 - d1 deltaBytes = 1 - d2 deltaBytes = 2 - d4 deltaBytes = 4 - d8 deltaBytes = 8 -) - -func bytesNeededForUnsignedTimestampDelta(deltaT model.Time) deltaBytes { - switch { - case deltaT > math.MaxUint32: - return d8 - case deltaT > math.MaxUint16: - return d4 - case deltaT > math.MaxUint8: - return d2 - default: - return d1 - } -} - -func bytesNeededForSignedTimestampDelta(deltaT model.Time) deltaBytes { - switch { - case deltaT > math.MaxInt32 || deltaT < math.MinInt32: - return d8 - case deltaT > math.MaxInt16 || deltaT < math.MinInt16: - return d4 - case deltaT > math.MaxInt8 || deltaT < math.MinInt8: - return d2 - default: - return d1 - } -} - -func bytesNeededForIntegerSampleValueDelta(deltaV model.SampleValue) deltaBytes { - switch { - case deltaV < math.MinInt32 || deltaV > math.MaxInt32: - return d8 - case deltaV < math.MinInt16 || deltaV > math.MaxInt16: - return d4 - case deltaV < math.MinInt8 || deltaV > math.MaxInt8: - return d2 - case deltaV != 0: - return d1 - default: - return d0 - } -} - -func max(a, b deltaBytes) deltaBytes { - if a > b { - return a - } - return b -} - -// isInt64 returns true if v can be represented as an int64. -func isInt64(v model.SampleValue) bool { - // Note: Using math.Modf is slower than the conversion approach below. - return model.SampleValue(int64(v)) == v -} diff --git a/storage/local/chunk/delta_test.go b/storage/local/chunk/delta_test.go deleted file mode 100644 index 09fd35c217..0000000000 --- a/storage/local/chunk/delta_test.go +++ /dev/null @@ -1,116 +0,0 @@ -// Copyright 2016 The Prometheus Authors -// Licensed under the Apache License, Version 2.0 (the "License"); -// you may not use this file except in compliance with the License. -// You may obtain a copy of the License at -// -// http://www.apache.org/licenses/LICENSE-2.0 -// -// Unless required by applicable law or agreed to in writing, software -// distributed under the License is distributed on an "AS IS" BASIS, -// WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. -// See the License for the specific language governing permissions and -// limitations under the License. - -// Note: this file has tests for code in both delta.go and doubledelta.go -- -// it may make sense to split those out later, but given that the tests are -// near-identical and share a helper, this feels simpler for now. - -package chunk - -import ( - "bytes" - "encoding/binary" - "strings" - "testing" - - "github.com/prometheus/common/model" -) - -func TestUnmarshalingCorruptedDeltaReturnsAnError(t *testing.T) { - - var verifyUnmarshallingError = func( - err error, - chunkTypeName string, - unmarshalMethod string, - expectedStr string, - ) { - - if err == nil { - t.Errorf("Failed to obtain an error when unmarshalling corrupt %s (from %s)", chunkTypeName, unmarshalMethod) - return - } - - if !strings.Contains(err.Error(), expectedStr) { - t.Errorf( - "'%s' not present in error when unmarshalling corrupt %s (from %s): '%s'", - expectedStr, - chunkTypeName, - unmarshalMethod, - err.Error()) - } - } - - cases := []struct { - chunkTypeName string - chunkConstructor func(deltaBytes, deltaBytes, bool, int) Chunk - minHeaderLen int - chunkLenPos int - timeBytesPos int - }{ - { - chunkTypeName: "deltaEncodedChunk", - chunkConstructor: func(a, b deltaBytes, c bool, d int) Chunk { - return newDeltaEncodedChunk(a, b, c, d) - }, - minHeaderLen: deltaHeaderBytes, - chunkLenPos: deltaHeaderBufLenOffset, - timeBytesPos: deltaHeaderTimeBytesOffset, - }, - { - chunkTypeName: "doubleDeltaEncodedChunk", - chunkConstructor: func(a, b deltaBytes, c bool, d int) Chunk { - return newDoubleDeltaEncodedChunk(a, b, c, d) - }, - minHeaderLen: doubleDeltaHeaderMinBytes, - chunkLenPos: doubleDeltaHeaderBufLenOffset, - timeBytesPos: doubleDeltaHeaderTimeBytesOffset, - }, - } - for _, c := range cases { - chunk := c.chunkConstructor(d1, d4, false, ChunkLen) - - cs, err := chunk.Add(model.SamplePair{ - Timestamp: model.Now(), - Value: model.SampleValue(100), - }) - if err != nil { - t.Fatalf("Couldn't add sample to empty %s: %s", c.chunkTypeName, err) - } - - buf := make([]byte, ChunkLen) - - cs[0].MarshalToBuf(buf) - - // Corrupt time byte to 0, which is illegal. - buf[c.timeBytesPos] = 0 - err = cs[0].UnmarshalFromBuf(buf) - verifyUnmarshallingError(err, c.chunkTypeName, "buf", "invalid number of time bytes") - - err = cs[0].Unmarshal(bytes.NewBuffer(buf)) - verifyUnmarshallingError(err, c.chunkTypeName, "Reader", "invalid number of time bytes") - - // Fix the corruption to go on. - buf[c.timeBytesPos] = byte(d1) - - // Corrupt the length to be every possible too-small value - for i := 0; i < c.minHeaderLen; i++ { - binary.LittleEndian.PutUint16(buf[c.chunkLenPos:], uint16(i)) - - err = cs[0].UnmarshalFromBuf(buf) - verifyUnmarshallingError(err, c.chunkTypeName, "buf", "header size") - - err = cs[0].Unmarshal(bytes.NewBuffer(buf)) - verifyUnmarshallingError(err, c.chunkTypeName, "Reader", "header size") - } - } -} diff --git a/storage/local/chunk/doubledelta.go b/storage/local/chunk/doubledelta.go deleted file mode 100644 index 249c99d545..0000000000 --- a/storage/local/chunk/doubledelta.go +++ /dev/null @@ -1,525 +0,0 @@ -// Copyright 2014 The Prometheus Authors -// Licensed under the Apache License, Version 2.0 (the "License"); -// you may not use this file except in compliance with the License. -// You may obtain a copy of the License at -// -// http://www.apache.org/licenses/LICENSE-2.0 -// -// Unless required by applicable law or agreed to in writing, software -// distributed under the License is distributed on an "AS IS" BASIS, -// WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. -// See the License for the specific language governing permissions and -// limitations under the License. - -package chunk - -import ( - "encoding/binary" - "fmt" - "io" - "math" - - "github.com/prometheus/common/model" -) - -// The 37-byte header of a delta-encoded chunk looks like: -// -// - used buf bytes: 2 bytes -// - time double-delta bytes: 1 bytes -// - value double-delta bytes: 1 bytes -// - is integer: 1 byte -// - base time: 8 bytes -// - base value: 8 bytes -// - base time delta: 8 bytes -// - base value delta: 8 bytes -const ( - doubleDeltaHeaderBytes = 37 - doubleDeltaHeaderMinBytes = 21 // header isn't full for chunk w/ one sample - - doubleDeltaHeaderBufLenOffset = 0 - doubleDeltaHeaderTimeBytesOffset = 2 - doubleDeltaHeaderValueBytesOffset = 3 - doubleDeltaHeaderIsIntOffset = 4 - doubleDeltaHeaderBaseTimeOffset = 5 - doubleDeltaHeaderBaseValueOffset = 13 - doubleDeltaHeaderBaseTimeDeltaOffset = 21 - doubleDeltaHeaderBaseValueDeltaOffset = 29 -) - -// A doubleDeltaEncodedChunk adaptively stores sample timestamps and values with -// a double-delta encoding of various types (int, float) and bit widths. A base -// value and timestamp and a base delta for each is saved in the header. The -// payload consists of double-deltas, i.e. deviations from the values and -// timestamps calculated by applying the base value and time and the base deltas. -// However, once 8 bytes would be needed to encode a double-delta value, a -// fall-back to the absolute numbers happens (so that timestamps are saved -// directly as int64 and values as float64). -// doubleDeltaEncodedChunk implements the chunk interface. -type doubleDeltaEncodedChunk []byte - -// newDoubleDeltaEncodedChunk returns a newly allocated doubleDeltaEncodedChunk. -func newDoubleDeltaEncodedChunk(tb, vb deltaBytes, isInt bool, length int) *doubleDeltaEncodedChunk { - if tb < 1 { - panic("need at least 1 time delta byte") - } - if length < doubleDeltaHeaderBytes+16 { - panic(fmt.Errorf( - "chunk length %d bytes is insufficient, need at least %d", - length, doubleDeltaHeaderBytes+16, - )) - } - c := make(doubleDeltaEncodedChunk, doubleDeltaHeaderIsIntOffset+1, length) - - c[doubleDeltaHeaderTimeBytesOffset] = byte(tb) - c[doubleDeltaHeaderValueBytesOffset] = byte(vb) - if vb < d8 && isInt { // Only use int for fewer than 8 value double-delta bytes. - c[doubleDeltaHeaderIsIntOffset] = 1 - } else { - c[doubleDeltaHeaderIsIntOffset] = 0 - } - return &c -} - -// Add implements chunk. -func (c doubleDeltaEncodedChunk) Add(s model.SamplePair) ([]Chunk, error) { - // TODO(beorn7): Since we return &c, this method might cause an unnecessary allocation. - if c.Len() == 0 { - return c.addFirstSample(s), nil - } - - tb := c.timeBytes() - vb := c.valueBytes() - - if c.Len() == 1 { - return c.addSecondSample(s, tb, vb) - } - - remainingBytes := cap(c) - len(c) - sampleSize := c.sampleSize() - - // Do we generally have space for another sample in this chunk? If not, - // overflow into a new one. - if remainingBytes < sampleSize { - return addToOverflowChunk(&c, s) - } - - projectedTime := c.baseTime() + model.Time(c.Len())*c.baseTimeDelta() - ddt := s.Timestamp - projectedTime - - projectedValue := c.baseValue() + model.SampleValue(c.Len())*c.baseValueDelta() - ddv := s.Value - projectedValue - - ntb, nvb, nInt := tb, vb, c.isInt() - // If the new sample is incompatible with the current encoding, reencode the - // existing chunk data into new chunk(s). - if c.isInt() && !isInt64(ddv) { - // int->float. - nvb = d4 - nInt = false - } else if !c.isInt() && vb == d4 && projectedValue+model.SampleValue(float32(ddv)) != s.Value { - // float32->float64. - nvb = d8 - } else { - if tb < d8 { - // Maybe more bytes for timestamp. - ntb = max(tb, bytesNeededForSignedTimestampDelta(ddt)) - } - if c.isInt() && vb < d8 { - // Maybe more bytes for sample value. - nvb = max(vb, bytesNeededForIntegerSampleValueDelta(ddv)) - } - } - if tb != ntb || vb != nvb || c.isInt() != nInt { - if len(c)*2 < cap(c) { - return transcodeAndAdd(newDoubleDeltaEncodedChunk(ntb, nvb, nInt, cap(c)), &c, s) - } - // Chunk is already half full. Better create a new one and save the transcoding efforts. - return addToOverflowChunk(&c, s) - } - - offset := len(c) - c = c[:offset+sampleSize] - - switch tb { - case d1: - c[offset] = byte(ddt) - case d2: - binary.LittleEndian.PutUint16(c[offset:], uint16(ddt)) - case d4: - binary.LittleEndian.PutUint32(c[offset:], uint32(ddt)) - case d8: - // Store the absolute value (no delta) in case of d8. - binary.LittleEndian.PutUint64(c[offset:], uint64(s.Timestamp)) - default: - return nil, fmt.Errorf("invalid number of bytes for time delta: %d", tb) - } - - offset += int(tb) - - if c.isInt() { - switch vb { - case d0: - // No-op. Constant delta is stored as base value. - case d1: - c[offset] = byte(int8(ddv)) - case d2: - binary.LittleEndian.PutUint16(c[offset:], uint16(int16(ddv))) - case d4: - binary.LittleEndian.PutUint32(c[offset:], uint32(int32(ddv))) - // d8 must not happen. Those samples are encoded as float64. - default: - return nil, fmt.Errorf("invalid number of bytes for integer delta: %d", vb) - } - } else { - switch vb { - case d4: - binary.LittleEndian.PutUint32(c[offset:], math.Float32bits(float32(ddv))) - case d8: - // Store the absolute value (no delta) in case of d8. - binary.LittleEndian.PutUint64(c[offset:], math.Float64bits(float64(s.Value))) - default: - return nil, fmt.Errorf("invalid number of bytes for floating point delta: %d", vb) - } - } - return []Chunk{&c}, nil -} - -// Clone implements chunk. -func (c doubleDeltaEncodedChunk) Clone() Chunk { - clone := make(doubleDeltaEncodedChunk, len(c), cap(c)) - copy(clone, c) - return &clone -} - -// FirstTime implements chunk. -func (c doubleDeltaEncodedChunk) FirstTime() model.Time { - return c.baseTime() -} - -// NewIterator( implements chunk. -func (c *doubleDeltaEncodedChunk) NewIterator() Iterator { - return newIndexAccessingChunkIterator(c.Len(), &doubleDeltaEncodedIndexAccessor{ - c: *c, - baseT: c.baseTime(), - baseΔT: c.baseTimeDelta(), - baseV: c.baseValue(), - baseΔV: c.baseValueDelta(), - tBytes: c.timeBytes(), - vBytes: c.valueBytes(), - isInt: c.isInt(), - }) -} - -// Marshal implements chunk. -func (c doubleDeltaEncodedChunk) Marshal(w io.Writer) error { - if len(c) > math.MaxUint16 { - panic("chunk buffer length would overflow a 16 bit uint") - } - binary.LittleEndian.PutUint16(c[doubleDeltaHeaderBufLenOffset:], uint16(len(c))) - - n, err := w.Write(c[:cap(c)]) - if err != nil { - return err - } - if n != cap(c) { - return fmt.Errorf("wanted to write %d bytes, wrote %d", cap(c), n) - } - return nil -} - -// MarshalToBuf implements chunk. -func (c doubleDeltaEncodedChunk) MarshalToBuf(buf []byte) error { - if len(c) > math.MaxUint16 { - panic("chunk buffer length would overflow a 16 bit uint") - } - binary.LittleEndian.PutUint16(c[doubleDeltaHeaderBufLenOffset:], uint16(len(c))) - - n := copy(buf, c) - if n != len(c) { - return fmt.Errorf("wanted to copy %d bytes to buffer, copied %d", len(c), n) - } - return nil -} - -// Unmarshal implements chunk. -func (c *doubleDeltaEncodedChunk) Unmarshal(r io.Reader) error { - *c = (*c)[:cap(*c)] - if _, err := io.ReadFull(r, *c); err != nil { - return err - } - return c.setLen() -} - -// UnmarshalFromBuf implements chunk. -func (c *doubleDeltaEncodedChunk) UnmarshalFromBuf(buf []byte) error { - *c = (*c)[:cap(*c)] - copy(*c, buf) - return c.setLen() -} - -// setLen sets the length of the underlying slice and performs some sanity checks. -func (c *doubleDeltaEncodedChunk) setLen() error { - l := binary.LittleEndian.Uint16((*c)[doubleDeltaHeaderBufLenOffset:]) - if int(l) > cap(*c) { - return fmt.Errorf("doubledelta chunk length exceeded during unmarshaling: %d", l) - } - if int(l) < doubleDeltaHeaderMinBytes { - return fmt.Errorf("doubledelta chunk length less than header size: %d < %d", l, doubleDeltaHeaderMinBytes) - } - switch c.timeBytes() { - case d1, d2, d4, d8: - // Pass. - default: - return fmt.Errorf("invalid number of time bytes in doubledelta chunk: %d", c.timeBytes()) - } - switch c.valueBytes() { - case d0, d1, d2, d4, d8: - // Pass. - default: - return fmt.Errorf("invalid number of value bytes in doubledelta chunk: %d", c.valueBytes()) - } - *c = (*c)[:l] - return nil -} - -// Encoding implements chunk. -func (c doubleDeltaEncodedChunk) Encoding() Encoding { return DoubleDelta } - -// Utilization implements chunk. -func (c doubleDeltaEncodedChunk) Utilization() float64 { - return float64(len(c)-doubleDeltaHeaderIsIntOffset-1) / float64(cap(c)) -} - -func (c doubleDeltaEncodedChunk) baseTime() model.Time { - return model.Time( - binary.LittleEndian.Uint64( - c[doubleDeltaHeaderBaseTimeOffset:], - ), - ) -} - -func (c doubleDeltaEncodedChunk) baseValue() model.SampleValue { - return model.SampleValue( - math.Float64frombits( - binary.LittleEndian.Uint64( - c[doubleDeltaHeaderBaseValueOffset:], - ), - ), - ) -} - -func (c doubleDeltaEncodedChunk) baseTimeDelta() model.Time { - if len(c) < doubleDeltaHeaderBaseTimeDeltaOffset+8 { - return 0 - } - return model.Time( - binary.LittleEndian.Uint64( - c[doubleDeltaHeaderBaseTimeDeltaOffset:], - ), - ) -} - -func (c doubleDeltaEncodedChunk) baseValueDelta() model.SampleValue { - if len(c) < doubleDeltaHeaderBaseValueDeltaOffset+8 { - return 0 - } - return model.SampleValue( - math.Float64frombits( - binary.LittleEndian.Uint64( - c[doubleDeltaHeaderBaseValueDeltaOffset:], - ), - ), - ) -} - -func (c doubleDeltaEncodedChunk) timeBytes() deltaBytes { - return deltaBytes(c[doubleDeltaHeaderTimeBytesOffset]) -} - -func (c doubleDeltaEncodedChunk) valueBytes() deltaBytes { - return deltaBytes(c[doubleDeltaHeaderValueBytesOffset]) -} - -func (c doubleDeltaEncodedChunk) sampleSize() int { - return int(c.timeBytes() + c.valueBytes()) -} - -// Len implements Chunk. Runs in constant time. -func (c doubleDeltaEncodedChunk) Len() int { - if len(c) <= doubleDeltaHeaderIsIntOffset+1 { - return 0 - } - if len(c) <= doubleDeltaHeaderBaseValueOffset+8 { - return 1 - } - return (len(c)-doubleDeltaHeaderBytes)/c.sampleSize() + 2 -} - -func (c doubleDeltaEncodedChunk) isInt() bool { - return c[doubleDeltaHeaderIsIntOffset] == 1 -} - -// addFirstSample is a helper method only used by c.add(). It adds timestamp and -// value as base time and value. -func (c doubleDeltaEncodedChunk) addFirstSample(s model.SamplePair) []Chunk { - c = c[:doubleDeltaHeaderBaseValueOffset+8] - binary.LittleEndian.PutUint64( - c[doubleDeltaHeaderBaseTimeOffset:], - uint64(s.Timestamp), - ) - binary.LittleEndian.PutUint64( - c[doubleDeltaHeaderBaseValueOffset:], - math.Float64bits(float64(s.Value)), - ) - return []Chunk{&c} -} - -// addSecondSample is a helper method only used by c.add(). It calculates the -// base delta from the provided sample and adds it to the chunk. -func (c doubleDeltaEncodedChunk) addSecondSample(s model.SamplePair, tb, vb deltaBytes) ([]Chunk, error) { - baseTimeDelta := s.Timestamp - c.baseTime() - if baseTimeDelta < 0 { - return nil, fmt.Errorf("base time delta is less than zero: %v", baseTimeDelta) - } - c = c[:doubleDeltaHeaderBytes] - if tb >= d8 || bytesNeededForUnsignedTimestampDelta(baseTimeDelta) >= d8 { - // If already the base delta needs d8 (or we are at d8 - // already, anyway), we better encode this timestamp - // directly rather than as a delta and switch everything - // to d8. - c[doubleDeltaHeaderTimeBytesOffset] = byte(d8) - binary.LittleEndian.PutUint64( - c[doubleDeltaHeaderBaseTimeDeltaOffset:], - uint64(s.Timestamp), - ) - } else { - binary.LittleEndian.PutUint64( - c[doubleDeltaHeaderBaseTimeDeltaOffset:], - uint64(baseTimeDelta), - ) - } - baseValue := c.baseValue() - baseValueDelta := s.Value - baseValue - if vb >= d8 || baseValue+baseValueDelta != s.Value { - // If we can't reproduce the original sample value (or - // if we are at d8 already, anyway), we better encode - // this value directly rather than as a delta and switch - // everything to d8. - c[doubleDeltaHeaderValueBytesOffset] = byte(d8) - c[doubleDeltaHeaderIsIntOffset] = 0 - binary.LittleEndian.PutUint64( - c[doubleDeltaHeaderBaseValueDeltaOffset:], - math.Float64bits(float64(s.Value)), - ) - } else { - binary.LittleEndian.PutUint64( - c[doubleDeltaHeaderBaseValueDeltaOffset:], - math.Float64bits(float64(baseValueDelta)), - ) - } - return []Chunk{&c}, nil -} - -// doubleDeltaEncodedIndexAccessor implements indexAccessor. -type doubleDeltaEncodedIndexAccessor struct { - c doubleDeltaEncodedChunk - baseT, baseΔT model.Time - baseV, baseΔV model.SampleValue - tBytes, vBytes deltaBytes - isInt bool - lastErr error -} - -func (acc *doubleDeltaEncodedIndexAccessor) err() error { - return acc.lastErr -} - -func (acc *doubleDeltaEncodedIndexAccessor) timestampAtIndex(idx int) model.Time { - if idx == 0 { - return acc.baseT - } - if idx == 1 { - // If time bytes are at d8, the time is saved directly rather - // than as a difference. - if acc.tBytes == d8 { - return acc.baseΔT - } - return acc.baseT + acc.baseΔT - } - - offset := doubleDeltaHeaderBytes + (idx-2)*int(acc.tBytes+acc.vBytes) - - switch acc.tBytes { - case d1: - return acc.baseT + - model.Time(idx)*acc.baseΔT + - model.Time(int8(acc.c[offset])) - case d2: - return acc.baseT + - model.Time(idx)*acc.baseΔT + - model.Time(int16(binary.LittleEndian.Uint16(acc.c[offset:]))) - case d4: - return acc.baseT + - model.Time(idx)*acc.baseΔT + - model.Time(int32(binary.LittleEndian.Uint32(acc.c[offset:]))) - case d8: - // Take absolute value for d8. - return model.Time(binary.LittleEndian.Uint64(acc.c[offset:])) - default: - acc.lastErr = fmt.Errorf("invalid number of bytes for time delta: %d", acc.tBytes) - return model.Earliest - } -} - -func (acc *doubleDeltaEncodedIndexAccessor) sampleValueAtIndex(idx int) model.SampleValue { - if idx == 0 { - return acc.baseV - } - if idx == 1 { - // If value bytes are at d8, the value is saved directly rather - // than as a difference. - if acc.vBytes == d8 { - return acc.baseΔV - } - return acc.baseV + acc.baseΔV - } - - offset := doubleDeltaHeaderBytes + (idx-2)*int(acc.tBytes+acc.vBytes) + int(acc.tBytes) - - if acc.isInt { - switch acc.vBytes { - case d0: - return acc.baseV + - model.SampleValue(idx)*acc.baseΔV - case d1: - return acc.baseV + - model.SampleValue(idx)*acc.baseΔV + - model.SampleValue(int8(acc.c[offset])) - case d2: - return acc.baseV + - model.SampleValue(idx)*acc.baseΔV + - model.SampleValue(int16(binary.LittleEndian.Uint16(acc.c[offset:]))) - case d4: - return acc.baseV + - model.SampleValue(idx)*acc.baseΔV + - model.SampleValue(int32(binary.LittleEndian.Uint32(acc.c[offset:]))) - // No d8 for ints. - default: - acc.lastErr = fmt.Errorf("invalid number of bytes for integer delta: %d", acc.vBytes) - return 0 - } - } else { - switch acc.vBytes { - case d4: - return acc.baseV + - model.SampleValue(idx)*acc.baseΔV + - model.SampleValue(math.Float32frombits(binary.LittleEndian.Uint32(acc.c[offset:]))) - case d8: - // Take absolute value for d8. - return model.SampleValue(math.Float64frombits(binary.LittleEndian.Uint64(acc.c[offset:]))) - default: - acc.lastErr = fmt.Errorf("invalid number of bytes for floating point delta: %d", acc.vBytes) - return 0 - } - } -} diff --git a/storage/local/chunk/instrumentation.go b/storage/local/chunk/instrumentation.go deleted file mode 100644 index 2a67fb5bd5..0000000000 --- a/storage/local/chunk/instrumentation.go +++ /dev/null @@ -1,100 +0,0 @@ -// Copyright 2014 The Prometheus Authors -// Licensed under the Apache License, Version 2.0 (the "License"); -// you may not use this file except in compliance with the License. -// You may obtain a copy of the License at -// -// http://www.apache.org/licenses/LICENSE-2.0 -// -// Unless required by applicable law or agreed to in writing, software -// distributed under the License is distributed on an "AS IS" BASIS, -// WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. -// See the License for the specific language governing permissions and -// limitations under the License. - -package chunk - -import "github.com/prometheus/client_golang/prometheus" - -// Usually, a separate file for instrumentation is frowned upon. Metrics should -// be close to where they are used. However, the metrics below are set all over -// the place, so we go for a separate instrumentation file in this case. -var ( - Ops = prometheus.NewCounterVec( - prometheus.CounterOpts{ - Namespace: namespace, - Subsystem: subsystem, - Name: "chunk_ops_total", - Help: "The total number of chunk operations by their type.", - }, - []string{OpTypeLabel}, - ) - DescOps = prometheus.NewCounterVec( - prometheus.CounterOpts{ - Namespace: namespace, - Subsystem: subsystem, - Name: "chunkdesc_ops_total", - Help: "The total number of chunk descriptor operations by their type.", - }, - []string{OpTypeLabel}, - ) - NumMemDescs = prometheus.NewGauge(prometheus.GaugeOpts{ - Namespace: namespace, - Subsystem: subsystem, - Name: "memory_chunkdescs", - Help: "The current number of chunk descriptors in memory.", - }) -) - -const ( - namespace = "prometheus" - subsystem = "local_storage" - - // OpTypeLabel is the label name for chunk operation types. - OpTypeLabel = "type" - - // Op-types for ChunkOps. - - // CreateAndPin is the label value for create-and-pin chunk ops. - CreateAndPin = "create" // A Desc creation with refCount=1. - // PersistAndUnpin is the label value for persist chunk ops. - PersistAndUnpin = "persist" - // Pin is the label value for pin chunk ops (excludes pin on creation). - Pin = "pin" - // Unpin is the label value for unpin chunk ops (excludes the unpin on persisting). - Unpin = "unpin" - // Clone is the label value for clone chunk ops. - Clone = "clone" - // Transcode is the label value for transcode chunk ops. - Transcode = "transcode" - // Drop is the label value for drop chunk ops. - Drop = "drop" - - // Op-types for ChunkOps and ChunkDescOps. - - // Evict is the label value for evict chunk desc ops. - Evict = "evict" - // Load is the label value for load chunk and chunk desc ops. - Load = "load" -) - -func init() { - prometheus.MustRegister(Ops) - prometheus.MustRegister(DescOps) - prometheus.MustRegister(NumMemDescs) -} - -var ( - // NumMemChunks is the total number of chunks in memory. This is a - // global counter, also used internally, so not implemented as - // metrics. Collected in MemorySeriesStorage.Collect. - // TODO(beorn7): As it is used internally, it is actually very bad style - // to have it as a global variable. - NumMemChunks int64 - - // NumMemChunksDesc is the metric descriptor for the above. - NumMemChunksDesc = prometheus.NewDesc( - prometheus.BuildFQName(namespace, subsystem, "memory_chunks"), - "The current number of chunks in memory, excluding cloned chunks (i.e. chunks without a descriptor).", - nil, nil, - ) -) diff --git a/storage/local/chunk/varbit.go b/storage/local/chunk/varbit.go deleted file mode 100644 index 2ec59efd34..0000000000 --- a/storage/local/chunk/varbit.go +++ /dev/null @@ -1,1210 +0,0 @@ -// Copyright 2016 The Prometheus Authors -// Licensed under the Apache License, Version 2.0 (the "License"); -// you may not use this file except in compliance with the License. -// You may obtain a copy of the License at -// -// http://www.apache.org/licenses/LICENSE-2.0 -// -// Unless required by applicable law or agreed to in writing, software -// distributed under the License is distributed on an "AS IS" BASIS, -// WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. -// See the License for the specific language governing permissions and -// limitations under the License. - -package chunk - -import ( - "encoding/binary" - "fmt" - "io" - "math" - - "github.com/prometheus/common/model" -) - -// The varbit chunk encoding is broadly similar to the double-delta -// chunks. However, it uses a number of different bit-widths to save the -// double-deltas (rather than 1, 2, or 4 bytes). Also, it doesn't use the delta -// of the first two samples of a chunk as the base delta, but uses a "sliding" -// delta, i.e. the delta of the two previous samples. Both differences make -// random access more expensive. Sample values can be encoded with the same -// double-delta scheme as timestamps, but different value encodings can be -// chosen adaptively, among them XOR encoding and "zero" encoding for constant -// sample values. Overall, the varbit encoding results in a much better -// compression ratio (~1.3 bytes per sample compared to ~3.3 bytes per sample -// with double-delta encoding, for typical data sets). -// -// Major parts of the varbit encoding are inspired by the following paper: -// Gorilla: A Fast, Scalable, In-Memory Time Series Database -// T. Pelkonen et al., Facebook Inc. -// http://www.vldb.org/pvldb/vol8/p1816-teller.pdf -// Note that there are significant differences, some due to the way Prometheus -// chunks work, others to optimize for the Prometheus use-case. -// -// Layout of a 1024 byte varbit chunk (big endian, wherever it matters): -// - first time (int64): 8 bytes bit 0000-0063 -// - first value (float64): 8 bytes bit 0064-0127 -// - last time (int64): 8 bytes bit 0128-0191 -// - last value (float64): 8 bytes bit 0192-0255 -// - first Δt (t1-t0, unsigned): 3 bytes bit 0256-0279 -// - flags (byte) 1 byte bit 0280-0287 -// - bit offset for next sample 2 bytes bit 0288-0303 -// - first Δv for value encoding 1, otherwise payload -// 4 bytes bit 0304-0335 -// - payload 973 bytes bit 0336-8119 -// The following only exists if the chunk is still open. Otherwise, it might be -// used by payload. -// - bit offset for current ΔΔt=0 count 2 bytes bit 8120-8135 -// - last Δt 3 bytes bit 8136-8159 -// - special bytes for value encoding 4 bytes bit 8160-8191 -// - for encoding 1: last Δv 4 bytes bit 8160-8191 -// - for encoding 2: count of -// - last leading zeros (1 byte) 1 byte bit 8160-8167 -// - last significant bits (1 byte) 1 byte bit 8168-8175 -// -// FLAGS -// -// The two least significant bits of the flags byte define the value encoding -// for the whole chunk, see below. The most significant byte of the flags byte -// is set if the chunk is closed. No samples can be added anymore to a closed -// chunk. Furthermore, the last value of a closed chunk is only saved in the -// header (last time, last value), while in a chunk that is still open, the last -// sample in the payload is the same sample as saved in the header. -// -// The remaining bits in the flags byte are currently unused. -// -// TIMESTAMP ENCODING -// -// The 1st timestamp is saved directly. -// -// The difference to the 2nd timestamp is saved as first Δt. 3 bytes is enough -// for about 4.5h. Since we close a chunk after sitting idle for 1h, this -// limitation has no practical consequences. Should, for whatever reason, a -// larger delta be required, the chunk would be closed, i.e. the new sample is -// added as the last sample to the chunk, and the next sample will be added to a -// new chunk. -// -// From the 3rd timestamp on, a double-delta (ΔΔt) is saved: -// (t_{n} - t_{n-1}) - (t_{n-1} - t_{n-2}) -// To perform that operation, the last Δt is saved at the end of the chunk for -// as long the chunk is not closed yet (see above). -// -// Most of the times, ΔΔt is zero, even with the ms-precision of -// Prometheus. Therefore, we save a ΔΔt of zero as a leading '0' bit followed by -// 7 bits counting the number of consecutive ΔΔt==0 (the count is offset by -1, -// so the range of 0 to 127 represents 1 to 128 repetitions). -// -// If ΔΔt != 0, we essentially apply the Gorilla encoding scheme (cf. section -// 4.1.1 in the paper) but with different bit buckets as Prometheus uses ms -// rather than s, and the default scrape interval is 1m rather than 4m). In -// particular: -// -// - If ΔΔt is between [-32,31], store '10' followed by a 6 bit value. This is -// for minor irregularities in the scrape interval. -// -// - If ΔΔt is between [-65536,65535], store '110' followed by a 17 bit -// value. This will typically happen if a scrape is missed completely. -// -// - If ΔΔt is betwees [-4194304,4194303], store '111' followed by a 23 bit -// value. This spans more than 1h, which is usually enough as we close a -// chunk anyway if it doesn't receive any sample in 1h. -// -// - Should we nevertheless encounter a larger ΔΔt, we simply close the chunk, -// add the new sample as the last of the chunk, and add subsequent samples to -// a new chunk. -// -// VALUE ENCODING -// -// Value encoding can change and is determined by the two least significant bits -// of the 'flags' byte at bit position 280. The encoding can be changed without -// transcoding upon adding the 3rd sample. After that, an encoding change -// results either in transcoding or in closing the chunk. -// -// The 1st sample value is always saved directly. The 2nd sample value is saved -// in the header as the last value. Upon saving the 3rd value, an encoding is -// chosen, and the chunk is prepared accordingly. -// -// The following value encodings exist (with their value in the flags byte): -// -// 0: "Zero encoding". -// -// In many time series, the value simply stays constant over a long time -// (e.g. the "up" time series). In that case, all sample values are determined -// by the 1st value, and no further value encoding is happening at all. The -// payload consists entirely of timestamps. -// -// 1: Integer double-delta encoding. -// -// Many Prometheus metrics are integer counters and change in a quite regular -// fashion, similar to timestamps. Thus, the same double-delta encoding can be -// applied. This encoding works like the timestamp encoding described above, but -// with different bit buckets and without counting of repeated ΔΔv=0. The case -// of ΔΔv=0 is represented by a single '0' bit for each occurrence. The first Δv -// is saved as an int32 at bit position 288. The most recent Δv is saved as an -// int32 at the end of the chunk (see above). If Δv cannot be represented as a -// 32 bit signed integer, no integer double-delta encoding can be applied. -// -// Bit buckets (lead-in bytes followed by (signed) value bits): -// - '0': 0 bit -// - '10': 6 bit -// - '110': 13 bit -// - '1110': 20 bit -// - '1111': 33 bit -// Since Δv is restricted to 32 bit, 33 bit are always enough for ΔΔv. -// -// 2: XOR encoding. -// -// This follows almost precisely the Gorilla value encoding (cf. section 4.1.2 -// of the paper). The last count of leading zeros and the last count of -// meaningful bits in the XOR value is saved at the end of the chunk for as long -// as the chunk is not closed yet (see above). Note, though, that the number of -// significant bits is saved as (count-1), i.e. a saved value of 0 means 1 -// significant bit, a saved value of 1 means 2, and so on. Also, we save the -// numbers of leading zeros and significant bits anew if they drop a -// lot. Otherwise, you can easily be locked in with a high number of significant -// bits. -// -// 3: Direct encoding. -// -// If the sample values are just random, it is most efficient to save sample -// values directly as float64. -// -// ZIPPING TIMESTAMPS AND VALUES TOGETHER -// -// Usually, encoded timestamps and encoded values simply alternate. There are -// two exceptions: -// -// (1) With the "zero encoding" for values, the payload only contains -// timestamps. -// -// (2) In a consecutive row of up to 128 ΔΔt=0 repeats, the count of timestamps -// determines how many sample values will follow directly after another. - -const ( - varbitMinLength = 128 - varbitMaxLength = 8191 - - // Useful byte offsets. - varbitFirstTimeOffset = 0 - varbitFirstValueOffset = 8 - varbitLastTimeOffset = 16 - varbitLastValueOffset = 24 - varbitFirstTimeDeltaOffset = 32 - varbitFlagOffset = 35 - varbitNextSampleBitOffsetOffset = 36 - varbitFirstValueDeltaOffset = 38 - // The following are in the "footer" and only usable if the chunk is - // still open. - varbitCountOffsetBitOffset = ChunkLen - 9 - varbitLastTimeDeltaOffset = ChunkLen - 7 - varbitLastValueDeltaOffset = ChunkLen - 4 - varbitLastLeadingZerosCountOffset = ChunkLen - 4 - varbitLastSignificantBitsCountOffset = ChunkLen - 3 - - varbitFirstSampleBitOffset uint16 = 0 // Symbolic, don't really read or write here. - varbitSecondSampleBitOffset uint16 = 1 // Symbolic, don't really read or write here. - // varbitThirdSampleBitOffset is a bit special. Depending on the encoding, there can - // be various things at this offset. It's most of the time symbolic, but in the best - // case (zero encoding for values), it will be the real offset for the 3rd sample. - varbitThirdSampleBitOffset uint16 = varbitFirstValueDeltaOffset * 8 - - // If the bit offset for the next sample is above this threshold, no new - // samples can be added to the chunk's payload (because the payload has - // already reached the footer). However, one more sample can be saved in - // the header as the last sample. - varbitNextSampleBitOffsetThreshold = 8 * varbitCountOffsetBitOffset - - varbitMaxTimeDelta = 1 << 24 // What fits into a 3-byte timestamp. -) - -type varbitValueEncoding byte - -const ( - varbitZeroEncoding varbitValueEncoding = iota - varbitIntDoubleDeltaEncoding - varbitXOREncoding - varbitDirectEncoding -) - -// varbitWorstCaseBitsPerSample provides the worst-case number of bits needed -// per sample with the various value encodings. The counts already include the -// up to 27 bits taken by a timestamp. -var varbitWorstCaseBitsPerSample = map[varbitValueEncoding]int{ - varbitZeroEncoding: 27 + 0, - varbitIntDoubleDeltaEncoding: 27 + 38, - varbitXOREncoding: 27 + 13 + 64, - varbitDirectEncoding: 27 + 64, -} - -// varbitChunk implements the chunk interface. -type varbitChunk []byte - -// newVarbitChunk returns a newly allocated varbitChunk. For simplicity, all -// varbit chunks must have the length as determined by the ChunkLen constant. -func newVarbitChunk(enc varbitValueEncoding) *varbitChunk { - if ChunkLen < varbitMinLength || ChunkLen > varbitMaxLength { - panic(fmt.Errorf( - "invalid chunk length of %d bytes, need at least %d bytes and at most %d bytes", - ChunkLen, varbitMinLength, varbitMaxLength, - )) - } - if enc > varbitDirectEncoding { - panic(fmt.Errorf("unknown varbit value encoding: %v", enc)) - } - c := make(varbitChunk, ChunkLen) - c.setValueEncoding(enc) - return &c -} - -// Add implements chunk. -func (c *varbitChunk) Add(s model.SamplePair) ([]Chunk, error) { - offset := c.nextSampleOffset() - switch { - case c.closed(): - return addToOverflowChunk(c, s) - case offset > varbitNextSampleBitOffsetThreshold: - return c.addLastSample(s), nil - case offset == varbitFirstSampleBitOffset: - return c.addFirstSample(s), nil - case offset == varbitSecondSampleBitOffset: - return c.addSecondSample(s) - } - return c.addLaterSample(s, offset) -} - -// Clone implements chunk. -func (c varbitChunk) Clone() Chunk { - clone := make(varbitChunk, len(c)) - copy(clone, c) - return &clone -} - -// NewIterator implements chunk. -func (c varbitChunk) NewIterator() Iterator { - return newVarbitChunkIterator(c) -} - -// Marshal implements chunk. -func (c varbitChunk) Marshal(w io.Writer) error { - n, err := w.Write(c) - if err != nil { - return err - } - if n != cap(c) { - return fmt.Errorf("wanted to write %d bytes, wrote %d", cap(c), n) - } - return nil -} - -// MarshalToBuf implements chunk. -func (c varbitChunk) MarshalToBuf(buf []byte) error { - n := copy(buf, c) - if n != len(c) { - return fmt.Errorf("wanted to copy %d bytes to buffer, copied %d", len(c), n) - } - return nil -} - -// Unmarshal implements chunk. -func (c varbitChunk) Unmarshal(r io.Reader) error { - _, err := io.ReadFull(r, c) - return err -} - -// UnmarshalFromBuf implements chunk. -func (c varbitChunk) UnmarshalFromBuf(buf []byte) error { - if copied := copy(c, buf); copied != cap(c) { - return fmt.Errorf("insufficient bytes copied from buffer during unmarshaling, want %d, got %d", cap(c), copied) - } - return nil -} - -// Encoding implements chunk. -func (c varbitChunk) Encoding() Encoding { return Varbit } - -// Utilization implements chunk. -func (c varbitChunk) Utilization() float64 { - // 15 bytes is the length of the chunk footer. - return math.Min(float64(c.nextSampleOffset()/8+15)/float64(cap(c)), 1) -} - -// Len implements chunk. Runs in O(n). -func (c varbitChunk) Len() int { - it := c.NewIterator() - i := 0 - for ; it.Scan(); i++ { - } - return i -} - -// FirstTime implements chunk. -func (c varbitChunk) FirstTime() model.Time { - return model.Time( - binary.BigEndian.Uint64( - c[varbitFirstTimeOffset:], - ), - ) -} - -func (c varbitChunk) firstValue() model.SampleValue { - return model.SampleValue( - math.Float64frombits( - binary.BigEndian.Uint64( - c[varbitFirstValueOffset:], - ), - ), - ) -} - -func (c varbitChunk) lastTime() model.Time { - return model.Time( - binary.BigEndian.Uint64( - c[varbitLastTimeOffset:], - ), - ) -} - -func (c varbitChunk) lastValue() model.SampleValue { - return model.SampleValue( - math.Float64frombits( - binary.BigEndian.Uint64( - c[varbitLastValueOffset:], - ), - ), - ) -} - -func (c varbitChunk) firstTimeDelta() model.Time { - // Only the first 3 bytes are actually the timestamp, so get rid of the - // last one by bitshifting. - return model.Time(c[varbitFirstTimeDeltaOffset+2]) | - model.Time(c[varbitFirstTimeDeltaOffset+1])<<8 | - model.Time(c[varbitFirstTimeDeltaOffset])<<16 -} - -// firstValueDelta returns an undefined result if the encoding type is not 1. -func (c varbitChunk) firstValueDelta() int32 { - return int32(binary.BigEndian.Uint32(c[varbitFirstValueDeltaOffset:])) -} - -// lastTimeDelta returns an undefined result if the chunk is closed already. -func (c varbitChunk) lastTimeDelta() model.Time { - return model.Time(c[varbitLastTimeDeltaOffset+2]) | - model.Time(c[varbitLastTimeDeltaOffset+1])<<8 | - model.Time(c[varbitLastTimeDeltaOffset])<<16 -} - -// setLastTimeDelta must not be called if the chunk is closed already. It most -// not be called with a time that doesn't fit into 24bit, either. -func (c varbitChunk) setLastTimeDelta(dT model.Time) { - if dT > varbitMaxTimeDelta { - panic("Δt overflows 24 bit") - } - c[varbitLastTimeDeltaOffset] = byte(dT >> 16) - c[varbitLastTimeDeltaOffset+1] = byte(dT >> 8) - c[varbitLastTimeDeltaOffset+2] = byte(dT) -} - -// lastValueDelta returns an undefined result if the chunk is closed already. -func (c varbitChunk) lastValueDelta() int32 { - return int32(binary.BigEndian.Uint32(c[varbitLastValueDeltaOffset:])) -} - -// setLastValueDelta must not be called if the chunk is closed already. -func (c varbitChunk) setLastValueDelta(dV int32) { - binary.BigEndian.PutUint32(c[varbitLastValueDeltaOffset:], uint32(dV)) -} - -func (c varbitChunk) nextSampleOffset() uint16 { - return binary.BigEndian.Uint16(c[varbitNextSampleBitOffsetOffset:]) -} - -func (c varbitChunk) setNextSampleOffset(offset uint16) { - binary.BigEndian.PutUint16(c[varbitNextSampleBitOffsetOffset:], offset) -} - -func (c varbitChunk) valueEncoding() varbitValueEncoding { - return varbitValueEncoding(c[varbitFlagOffset] & 0x03) -} - -func (c varbitChunk) setValueEncoding(enc varbitValueEncoding) { - if enc > varbitDirectEncoding { - panic("invalid varbit value encoding") - } - c[varbitFlagOffset] &^= 0x03 // Clear. - c[varbitFlagOffset] |= byte(enc) // Set. -} - -func (c varbitChunk) closed() bool { - return c[varbitFlagOffset] > 0x7F // Most significant bit set. -} - -func (c varbitChunk) zeroDDTRepeats() (repeats uint64, offset uint16) { - offset = binary.BigEndian.Uint16(c[varbitCountOffsetBitOffset:]) - if offset == 0 { - return 0, 0 - } - return c.readBitPattern(offset, 7) + 1, offset -} - -func (c varbitChunk) setZeroDDTRepeats(repeats uint64, offset uint16) { - switch repeats { - case 0: - // Just clear the offset. - binary.BigEndian.PutUint16(c[varbitCountOffsetBitOffset:], 0) - return - case 1: - // First time we set a repeat here, so set the offset. But only - // if we haven't reached the footer yet. (If that's the case, we - // would overwrite ourselves below, and we don't need the offset - // later anyway because no more samples will be added to this - // chunk.) - if offset+7 <= varbitNextSampleBitOffsetThreshold { - binary.BigEndian.PutUint16(c[varbitCountOffsetBitOffset:], offset) - } - default: - // For a change, we are writing somewhere where we have written - // before. We need to clear the bits first. - posIn1stByte := offset % 8 - c[offset/8] &^= bitMask[7][posIn1stByte] - if posIn1stByte > 1 { - c[offset/8+1] &^= bitMask[posIn1stByte-1][0] - } - } - c.addBitPattern(offset, repeats-1, 7) -} - -func (c varbitChunk) setLastSample(s model.SamplePair) { - binary.BigEndian.PutUint64( - c[varbitLastTimeOffset:], - uint64(s.Timestamp), - ) - binary.BigEndian.PutUint64( - c[varbitLastValueOffset:], - math.Float64bits(float64(s.Value)), - ) -} - -// addFirstSample is a helper method only used by c.add(). It adds timestamp and -// value as base time and value. -func (c *varbitChunk) addFirstSample(s model.SamplePair) []Chunk { - binary.BigEndian.PutUint64( - (*c)[varbitFirstTimeOffset:], - uint64(s.Timestamp), - ) - binary.BigEndian.PutUint64( - (*c)[varbitFirstValueOffset:], - math.Float64bits(float64(s.Value)), - ) - c.setLastSample(s) // To simplify handling of single-sample chunks. - c.setNextSampleOffset(varbitSecondSampleBitOffset) - return []Chunk{c} -} - -// addSecondSample is a helper method only used by c.add(). It calculates the -// first time delta from the provided sample and adds it to the chunk together -// with the provided sample as the last sample. -func (c *varbitChunk) addSecondSample(s model.SamplePair) ([]Chunk, error) { - firstTimeDelta := s.Timestamp - c.FirstTime() - if firstTimeDelta < 0 { - return nil, fmt.Errorf("first Δt is less than zero: %v", firstTimeDelta) - } - if firstTimeDelta > varbitMaxTimeDelta { - // A time delta too great. Still, we can add it as a last sample - // before overflowing. - return c.addLastSample(s), nil - } - (*c)[varbitFirstTimeDeltaOffset] = byte(firstTimeDelta >> 16) - (*c)[varbitFirstTimeDeltaOffset+1] = byte(firstTimeDelta >> 8) - (*c)[varbitFirstTimeDeltaOffset+2] = byte(firstTimeDelta) - - // Also set firstTimeDelta as the last time delta to be able to use the - // normal methods for adding later samples. - c.setLastTimeDelta(firstTimeDelta) - - c.setLastSample(s) - c.setNextSampleOffset(varbitThirdSampleBitOffset) - return []Chunk{c}, nil -} - -// addLastSample is a helper method only used by c.add() and in other helper -// methods called by c.add(). It simply sets the given sample as the last sample -// in the heador and declares the chunk closed. In other words, addLastSample -// adds the very last sample added to this chunk ever, while setLastSample sets -// the sample most recently added to the chunk so that it can be used for the -// calculations required to add the next sample. -func (c *varbitChunk) addLastSample(s model.SamplePair) []Chunk { - c.setLastSample(s) - (*c)[varbitFlagOffset] |= 0x80 - return []Chunk{c} -} - -// addLaterSample is a helper method only used by c.add(). It adds a third or -// later sample. -func (c *varbitChunk) addLaterSample(s model.SamplePair, offset uint16) ([]Chunk, error) { - var ( - lastTime = c.lastTime() - lastTimeDelta = c.lastTimeDelta() - newTimeDelta = s.Timestamp - lastTime - lastValue = c.lastValue() - encoding = c.valueEncoding() - ) - - if newTimeDelta < 0 { - return nil, fmt.Errorf("Δt is less than zero: %v", newTimeDelta) - } - if offset == varbitThirdSampleBitOffset { - offset, encoding = c.prepForThirdSample(lastValue, s.Value, encoding) - } - if newTimeDelta > varbitMaxTimeDelta { - // A time delta too great. Still, we can add it as a last sample - // before overflowing. - return c.addLastSample(s), nil - } - - // Analyze worst case, does it fit? If not, set new sample as the last. - if int(offset)+varbitWorstCaseBitsPerSample[encoding] > ChunkLen*8 { - return c.addLastSample(s), nil - } - - // Transcoding/overflow decisions first. - if encoding == varbitZeroEncoding && s.Value != lastValue { - // Cannot go on with zero encoding. - if offset > ChunkLen*4 { - // Chunk already half full. Don't transcode, overflow instead. - return addToOverflowChunk(c, s) - } - if isInt32(s.Value - lastValue) { - // Trying int encoding looks promising. - return transcodeAndAdd(newVarbitChunk(varbitIntDoubleDeltaEncoding), c, s) - } - return transcodeAndAdd(newVarbitChunk(varbitXOREncoding), c, s) - } - if encoding == varbitIntDoubleDeltaEncoding && !isInt32(s.Value-lastValue) { - // Cannot go on with int encoding. - if offset > ChunkLen*4 { - // Chunk already half full. Don't transcode, overflow instead. - return addToOverflowChunk(c, s) - } - return transcodeAndAdd(newVarbitChunk(varbitXOREncoding), c, s) - } - - offset, overflow := c.addDDTime(offset, lastTimeDelta, newTimeDelta) - if overflow { - return c.addLastSample(s), nil - } - switch encoding { - case varbitZeroEncoding: - // Nothing to do. - case varbitIntDoubleDeltaEncoding: - offset = c.addDDValue(offset, lastValue, s.Value) - case varbitXOREncoding: - offset = c.addXORValue(offset, lastValue, s.Value) - case varbitDirectEncoding: - offset = c.addBitPattern(offset, math.Float64bits(float64(s.Value)), 64) - default: - return nil, fmt.Errorf("unknown Varbit value encoding: %v", encoding) - } - - c.setNextSampleOffset(offset) - c.setLastSample(s) - return []Chunk{c}, nil -} - -func (c varbitChunk) prepForThirdSample( - lastValue, newValue model.SampleValue, encoding varbitValueEncoding, -) (uint16, varbitValueEncoding) { - var ( - offset = varbitThirdSampleBitOffset - firstValue = c.firstValue() - firstValueDelta = lastValue - firstValue - firstXOR = math.Float64bits(float64(firstValue)) ^ math.Float64bits(float64(lastValue)) - _, firstSignificantBits = countBits(firstXOR) - secondXOR = math.Float64bits(float64(lastValue)) ^ math.Float64bits(float64(newValue)) - _, secondSignificantBits = countBits(secondXOR) - ) - // Now pick an initial encoding and prepare things accordingly. - // However, never pick an encoding "below" the one initially set. - switch { - case encoding == varbitZeroEncoding && lastValue == firstValue && lastValue == newValue: - // Stay at zero encoding. - // No value to be set. - // No offset change required. - case encoding <= varbitIntDoubleDeltaEncoding && isInt32(firstValueDelta): - encoding = varbitIntDoubleDeltaEncoding - binary.BigEndian.PutUint32( - c[varbitFirstValueDeltaOffset:], - uint32(int32(firstValueDelta)), - ) - c.setLastValueDelta(int32(firstValueDelta)) - offset += 32 - case encoding == varbitDirectEncoding || firstSignificantBits+secondSignificantBits > 100: - // Heuristics based on three samples only is a bit weak, - // but if we need 50+13 = 63 bits per sample already - // now, we might be better off going for direct encoding. - encoding = varbitDirectEncoding - // Put bit pattern directly where otherwise the delta would have gone. - binary.BigEndian.PutUint64( - c[varbitFirstValueDeltaOffset:], - math.Float64bits(float64(lastValue)), - ) - offset += 64 - default: - encoding = varbitXOREncoding - offset = c.addXORValue(offset, firstValue, lastValue) - } - c.setValueEncoding(encoding) - c.setNextSampleOffset(offset) - return offset, encoding -} - -// addDDTime requires that lastTimeDelta and newTimeDelta are positive and don't overflow 24bit. -func (c varbitChunk) addDDTime(offset uint16, lastTimeDelta, newTimeDelta model.Time) (newOffset uint16, overflow bool) { - timeDD := newTimeDelta - lastTimeDelta - - if !isSignedIntN(int64(timeDD), 23) { - return offset, true - } - - c.setLastTimeDelta(newTimeDelta) - repeats, repeatsOffset := c.zeroDDTRepeats() - - if timeDD == 0 { - if repeats == 0 || repeats == 128 { - // First zeroDDT, or counter full, prepare new counter. - offset = c.addZeroBit(offset) - repeatsOffset = offset - offset += 7 - repeats = 0 - } - c.setZeroDDTRepeats(repeats+1, repeatsOffset) - return offset, false - } - - // No zero repeat. If we had any before, clear the DDT offset. - c.setZeroDDTRepeats(0, repeatsOffset) - - switch { - case isSignedIntN(int64(timeDD), 6): - offset = c.addOneBitsWithTrailingZero(offset, 1) - offset = c.addSignedInt(offset, int64(timeDD), 6) - case isSignedIntN(int64(timeDD), 17): - offset = c.addOneBitsWithTrailingZero(offset, 2) - offset = c.addSignedInt(offset, int64(timeDD), 17) - case isSignedIntN(int64(timeDD), 23): - offset = c.addOneBits(offset, 3) - offset = c.addSignedInt(offset, int64(timeDD), 23) - default: - panic("unexpected required bits for ΔΔt") - } - return offset, false -} - -// addDDValue requires that newValue-lastValue can be represented with an int32. -func (c varbitChunk) addDDValue(offset uint16, lastValue, newValue model.SampleValue) uint16 { - newValueDelta := int64(newValue - lastValue) - lastValueDelta := c.lastValueDelta() - valueDD := newValueDelta - int64(lastValueDelta) - c.setLastValueDelta(int32(newValueDelta)) - - switch { - case valueDD == 0: - return c.addZeroBit(offset) - case isSignedIntN(valueDD, 6): - offset = c.addOneBitsWithTrailingZero(offset, 1) - return c.addSignedInt(offset, valueDD, 6) - case isSignedIntN(valueDD, 13): - offset = c.addOneBitsWithTrailingZero(offset, 2) - return c.addSignedInt(offset, valueDD, 13) - case isSignedIntN(valueDD, 20): - offset = c.addOneBitsWithTrailingZero(offset, 3) - return c.addSignedInt(offset, valueDD, 20) - case isSignedIntN(valueDD, 33): - offset = c.addOneBits(offset, 4) - return c.addSignedInt(offset, valueDD, 33) - default: - panic("unexpected required bits for ΔΔv") - } -} - -func (c varbitChunk) addXORValue(offset uint16, lastValue, newValue model.SampleValue) uint16 { - lastPattern := math.Float64bits(float64(lastValue)) - newPattern := math.Float64bits(float64(newValue)) - xor := lastPattern ^ newPattern - if xor == 0 { - return c.addZeroBit(offset) - } - - lastLeadingBits := c[varbitLastLeadingZerosCountOffset] - lastSignificantBits := c[varbitLastSignificantBitsCountOffset] - newLeadingBits, newSignificantBits := countBits(xor) - - // Short entry if the new significant bits fit into the same box as the - // last significant bits. However, should the new significant bits be - // shorter by 10 or more, go for a long entry instead, as we will - // probably save more (11 bit one-time overhead, potentially more to - // save later). - if newLeadingBits >= lastLeadingBits && - newLeadingBits+newSignificantBits <= lastLeadingBits+lastSignificantBits && - lastSignificantBits-newSignificantBits < 10 { - offset = c.addOneBitsWithTrailingZero(offset, 1) - return c.addBitPattern( - offset, - xor>>(64-lastLeadingBits-lastSignificantBits), - uint16(lastSignificantBits), - ) - } - - // Long entry. - c[varbitLastLeadingZerosCountOffset] = newLeadingBits - c[varbitLastSignificantBitsCountOffset] = newSignificantBits - offset = c.addOneBits(offset, 2) - offset = c.addBitPattern(offset, uint64(newLeadingBits), 5) - offset = c.addBitPattern(offset, uint64(newSignificantBits-1), 6) // Note -1! - return c.addBitPattern( - offset, - xor>>(64-newLeadingBits-newSignificantBits), - uint16(newSignificantBits), - ) -} - -func (c varbitChunk) addZeroBit(offset uint16) uint16 { - if offset < varbitNextSampleBitOffsetThreshold { - // Writing a zero to a never touched area is a no-op. - // Just increase the offset. - return offset + 1 - } - newByte := c[offset/8] &^ bitMask[1][offset%8] - c[offset/8] = newByte - // TODO(beorn7): The two lines above could be written as - // c[offset/8] &^= bitMask[1][offset%8] - // However, that tickles a compiler bug with GOARCH=386. - // See https://github.com/prometheus/prometheus/issues/1509 - return offset + 1 -} - -func (c varbitChunk) addOneBits(offset uint16, n uint16) uint16 { - if n > 7 { - panic("unexpected number of control bits") - } - b := 8 - offset%8 - if b > n { - b = n - } - c[offset/8] |= bitMask[b][offset%8] - offset += b - b = n - b - if b > 0 { - c[offset/8] |= bitMask[b][0] - offset += b - } - return offset -} -func (c varbitChunk) addOneBitsWithTrailingZero(offset uint16, n uint16) uint16 { - offset = c.addOneBits(offset, n) - return c.addZeroBit(offset) -} - -// addSignedInt adds i as a signed integer with n bits. It requires i to be -// representable as such. (Check with isSignedIntN first.) -func (c varbitChunk) addSignedInt(offset uint16, i int64, n uint16) uint16 { - if i < 0 && n < 64 { - i += 1 << n - } - return c.addBitPattern(offset, uint64(i), n) -} - -// addBitPattern adds the last n bits of the given pattern. Other bits in the -// pattern must be 0. -func (c varbitChunk) addBitPattern(offset uint16, pattern uint64, n uint16) uint16 { - var ( - byteOffset = offset / 8 - bitsToWrite = 8 - offset%8 - newOffset = offset + n - ) - - // Clean up the parts of the footer we will write into. (But not more as - // we are still using the value related part of the footer when we have - // already overwritten timestamp related parts.) - if newOffset > varbitNextSampleBitOffsetThreshold { - pos := offset - if pos < varbitNextSampleBitOffsetThreshold { - pos = varbitNextSampleBitOffsetThreshold - } - for pos < newOffset { - posInByte := pos % 8 - bitsToClear := newOffset - pos - if bitsToClear > 8-posInByte { - bitsToClear = 8 - posInByte - } - c[pos/8] &^= bitMask[bitsToClear][posInByte] - pos += bitsToClear - } - } - - for n > 0 { - if n <= bitsToWrite { - c[byteOffset] |= byte(pattern << (bitsToWrite - n)) - break - } - c[byteOffset] |= byte(pattern >> (n - bitsToWrite)) - n -= bitsToWrite - bitsToWrite = 8 - byteOffset++ - } - return newOffset -} - -// readBitPattern reads n bits at the given offset and returns them as the last -// n bits in a uint64. -func (c varbitChunk) readBitPattern(offset, n uint16) uint64 { - var ( - result uint64 - byteOffset = offset / 8 - bitOffset = offset % 8 - trailingBits, bitsToRead uint16 - ) - - for n > 0 { - trailingBits = 0 - bitsToRead = 8 - bitOffset - if bitsToRead > n { - trailingBits = bitsToRead - n - bitsToRead = n - } - result <<= bitsToRead - result |= uint64( - (c[byteOffset] & bitMask[bitsToRead][bitOffset]) >> trailingBits, - ) - n -= bitsToRead - byteOffset++ - bitOffset = 0 - } - return result -} - -type varbitChunkIterator struct { - c varbitChunk - // pos is the bit position within the chunk for the next sample to be - // decoded when scan() is called (i.e. it is _not_ the bit position of - // the sample currently returned by value()). The symbolic values - // varbitFirstSampleBitOffset and varbitSecondSampleBitOffset are also - // used for pos. len is the offset of the first bit in the chunk that is - // not part of the payload. If pos==len, then the iterator is positioned - // behind the last sample in the payload. However, the next call of - // scan() still has to check if the chunk is closed, in which case there - // is one more sample, saved in the header. To mark the iterator as - // having scanned that last sample, too, pos is set to len+1. - pos, len uint16 - t, dT model.Time - repeats byte // Repeats of ΔΔt=0. - v model.SampleValue - dV int64 // Only used for int value encoding. - leading, significant uint16 - enc varbitValueEncoding - lastError error - rewound bool - nextT model.Time // Only for rewound state. - nextV model.SampleValue // Only for rewound state. -} - -func newVarbitChunkIterator(c varbitChunk) *varbitChunkIterator { - return &varbitChunkIterator{ - c: c, - len: c.nextSampleOffset(), - t: model.Earliest, - enc: c.valueEncoding(), - significant: 1, - } -} - -// lastTimestamp implements Iterator. -func (it *varbitChunkIterator) LastTimestamp() (model.Time, error) { - if it.len == varbitFirstSampleBitOffset { - // No samples in the chunk yet. - return model.Earliest, it.lastError - } - return it.c.lastTime(), it.lastError -} - -// contains implements Iterator. -func (it *varbitChunkIterator) Contains(t model.Time) (bool, error) { - last, err := it.LastTimestamp() - if err != nil { - it.lastError = err - return false, err - } - return !t.Before(it.c.FirstTime()) && - !t.After(last), it.lastError -} - -// scan implements Iterator. -func (it *varbitChunkIterator) Scan() bool { - if it.lastError != nil { - return false - } - if it.rewound { - it.t = it.nextT - it.v = it.nextV - it.rewound = false - return true - } - if it.pos > it.len { - return false - } - if it.pos == it.len && it.repeats == 0 { - it.pos = it.len + 1 - if !it.c.closed() { - return false - } - it.t = it.c.lastTime() - it.v = it.c.lastValue() - return it.lastError == nil - } - if it.pos == varbitFirstSampleBitOffset { - it.t = it.c.FirstTime() - it.v = it.c.firstValue() - it.pos = varbitSecondSampleBitOffset - return it.lastError == nil - } - if it.pos == varbitSecondSampleBitOffset { - if it.len == varbitThirdSampleBitOffset && !it.c.closed() { - // Special case: Chunk has only two samples. - it.t = it.c.lastTime() - it.v = it.c.lastValue() - it.pos = it.len + 1 - return it.lastError == nil - } - it.dT = it.c.firstTimeDelta() - it.t += it.dT - // Value depends on encoding. - switch it.enc { - case varbitZeroEncoding: - it.pos = varbitThirdSampleBitOffset - case varbitIntDoubleDeltaEncoding: - it.dV = int64(it.c.firstValueDelta()) - it.v += model.SampleValue(it.dV) - it.pos = varbitThirdSampleBitOffset + 32 - case varbitXOREncoding: - it.pos = varbitThirdSampleBitOffset - it.readXOR() - case varbitDirectEncoding: - it.v = model.SampleValue(math.Float64frombits( - binary.BigEndian.Uint64(it.c[varbitThirdSampleBitOffset/8:]), - )) - it.pos = varbitThirdSampleBitOffset + 64 - default: - it.lastError = fmt.Errorf("unknown varbit value encoding: %v", it.enc) - } - return it.lastError == nil - } - // 3rd sample or later does not have special cases anymore. - it.readDDT() - switch it.enc { - case varbitZeroEncoding: - // Do nothing. - case varbitIntDoubleDeltaEncoding: - it.readDDV() - case varbitXOREncoding: - it.readXOR() - case varbitDirectEncoding: - it.v = model.SampleValue(math.Float64frombits(it.readBitPattern(64))) - return it.lastError == nil - default: - it.lastError = fmt.Errorf("unknown varbit value encoding: %v", it.enc) - return false - } - return it.lastError == nil -} - -// findAtOrBefore implements Iterator. -func (it *varbitChunkIterator) FindAtOrBefore(t model.Time) bool { - if it.len == 0 || t.Before(it.c.FirstTime()) { - return false - } - last := it.c.lastTime() - if !t.Before(last) { - it.t = last - it.v = it.c.lastValue() - it.pos = it.len + 1 - return true - } - if t == it.t { - return it.lastError == nil - } - if t.Before(it.t) || it.rewound { - it.reset() - } - - var ( - prevT = model.Earliest - prevV model.SampleValue - ) - for it.Scan() && t.After(it.t) { - prevT = it.t - prevV = it.v - // TODO(beorn7): If we are in a repeat, we could iterate forward - // much faster. - } - if t == it.t { - return it.lastError == nil - } - it.rewind(prevT, prevV) - return it.lastError == nil -} - -// findAtOrAfter implements Iterator. -func (it *varbitChunkIterator) FindAtOrAfter(t model.Time) bool { - if it.len == 0 || t.After(it.c.lastTime()) { - return false - } - first := it.c.FirstTime() - if !t.After(first) { - it.reset() - return it.Scan() - } - if t == it.t { - return it.lastError == nil - } - if t.Before(it.t) { - it.reset() - } - for it.Scan() && t.After(it.t) { - // TODO(beorn7): If we are in a repeat, we could iterate forward - // much faster. - } - return it.lastError == nil -} - -// value implements Iterator. -func (it *varbitChunkIterator) Value() model.SamplePair { - return model.SamplePair{ - Timestamp: it.t, - Value: it.v, - } -} - -// err implements Iterator. -func (it *varbitChunkIterator) Err() error { - return it.lastError -} - -func (it *varbitChunkIterator) readDDT() { - if it.repeats > 0 { - it.repeats-- - } else { - switch it.readControlBits(3) { - case 0: - it.repeats = byte(it.readBitPattern(7)) - case 1: - it.dT += model.Time(it.readSignedInt(6)) - case 2: - it.dT += model.Time(it.readSignedInt(17)) - case 3: - it.dT += model.Time(it.readSignedInt(23)) - default: - panic("unexpected number of control bits") - } - } - it.t += it.dT -} - -func (it *varbitChunkIterator) readDDV() { - switch it.readControlBits(4) { - case 0: - // Do nothing. - case 1: - it.dV += it.readSignedInt(6) - case 2: - it.dV += it.readSignedInt(13) - case 3: - it.dV += it.readSignedInt(20) - case 4: - it.dV += it.readSignedInt(33) - default: - panic("unexpected number of control bits") - } - it.v += model.SampleValue(it.dV) -} - -func (it *varbitChunkIterator) readXOR() { - switch it.readControlBits(2) { - case 0: - return - case 1: - // Do nothing right now. All done below. - case 2: - it.leading = uint16(it.readBitPattern(5)) - it.significant = uint16(it.readBitPattern(6)) + 1 - default: - panic("unexpected number of control bits") - } - pattern := math.Float64bits(float64(it.v)) - pattern ^= it.readBitPattern(it.significant) << (64 - it.significant - it.leading) - it.v = model.SampleValue(math.Float64frombits(pattern)) -} - -// readControlBits reads successive 1-bits and stops after reading the first -// 0-bit. It also stops once it has read max bits. It returns the number of read -// 1-bits. -func (it *varbitChunkIterator) readControlBits(max uint16) uint16 { - var count uint16 - for count < max && int(it.pos/8) < len(it.c) { - b := it.c[it.pos/8] & bitMask[1][it.pos%8] - it.pos++ - if b == 0 { - return count - } - count++ - } - if int(it.pos/8) >= len(it.c) { - it.lastError = errChunkBoundsExceeded - } - return count -} - -func (it *varbitChunkIterator) readBitPattern(n uint16) uint64 { - if len(it.c)*8 < int(it.pos)+int(n) { - it.lastError = errChunkBoundsExceeded - return 0 - } - u := it.c.readBitPattern(it.pos, n) - it.pos += n - return u -} - -func (it *varbitChunkIterator) readSignedInt(n uint16) int64 { - u := it.readBitPattern(n) - if n < 64 && u >= 1<<(n-1) { - u -= 1 << n - } - return int64(u) -} - -// reset puts the chunk iterator into the state it had upon creation. -func (it *varbitChunkIterator) reset() { - it.pos = 0 - it.t = model.Earliest - it.dT = 0 - it.repeats = 0 - it.v = 0 - it.dV = 0 - it.leading = 0 - it.significant = 1 - it.rewound = false -} - -// rewind "rewinds" the chunk iterator by one step. Since one cannot simply -// rewind a Varbit chunk, the old values have to be provided by the -// caller. Rewinding an already rewound chunk panics. After a call of scan or -// reset, a chunk can be rewound again. -func (it *varbitChunkIterator) rewind(t model.Time, v model.SampleValue) { - if it.rewound { - panic("cannot rewind varbit chunk twice") - } - it.rewound = true - it.nextT = it.t - it.nextV = it.v - it.t = t - it.v = v -} diff --git a/storage/local/chunk/varbit_helpers.go b/storage/local/chunk/varbit_helpers.go deleted file mode 100644 index cc637a992b..0000000000 --- a/storage/local/chunk/varbit_helpers.go +++ /dev/null @@ -1,75 +0,0 @@ -// Copyright 2016 The Prometheus Authors -// Licensed under the Apache License, Version 2.0 (the "License"); -// you may not use this file except in compliance with the License. -// You may obtain a copy of the License at -// -// http://www.apache.org/licenses/LICENSE-2.0 -// -// Unless required by applicable law or agreed to in writing, software -// distributed under the License is distributed on an "AS IS" BASIS, -// WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. -// See the License for the specific language governing permissions and -// limitations under the License. - -package chunk - -import "github.com/prometheus/common/model" - -var ( - // bit masks for consecutive bits in a byte at various offsets. - bitMask = [][]byte{ - {0x00, 0x00, 0x00, 0x00, 0x00, 0x00, 0x00, 0x00}, // 0 bit - {0x80, 0x40, 0x20, 0x10, 0x08, 0x04, 0x02, 0x01}, // 1 bit - {0xC0, 0x60, 0x30, 0x18, 0x0C, 0x06, 0x03, 0x01}, // 2 bit - {0xE0, 0x70, 0x38, 0x1C, 0x0E, 0x07, 0x03, 0x01}, // 3 bit - {0xF0, 0x78, 0x3C, 0x1E, 0x0F, 0x07, 0x03, 0x01}, // 4 bit - {0xF8, 0x7C, 0x3E, 0x1F, 0x0F, 0x07, 0x03, 0x01}, // 5 bit - {0xFC, 0x7E, 0x3F, 0x1F, 0x0F, 0x07, 0x03, 0x01}, // 6 bit - {0xFE, 0x7F, 0x3F, 0x1F, 0x0F, 0x07, 0x03, 0x01}, // 7 bit - {0xFF, 0x7F, 0x3F, 0x1F, 0x0F, 0x07, 0x03, 0x01}, // 8 bit - } -) - -// isInt32 returns true if v can be represented as an int32. -func isInt32(v model.SampleValue) bool { - return model.SampleValue(int32(v)) == v -} - -// countBits returs the number of leading zero bits and the number of -// significant bits after that in the given bit pattern. The maximum number of -// leading zeros is 31 (so that it can be represented by a 5bit number). Leading -// zeros beyond that are considered part of the significant bits. -func countBits(pattern uint64) (leading, significant byte) { - // TODO(beorn7): This would probably be faster with ugly endless switch - // statements. - if pattern == 0 { - return - } - for pattern < 1<<63 { - leading++ - pattern <<= 1 - } - for pattern > 0 { - significant++ - pattern <<= 1 - } - if leading > 31 { // 5 bit limit. - significant += leading - 31 - leading = 31 - } - return -} - -// isSignedIntN returns if n can be represented as a signed int with the given -// bit length. -func isSignedIntN(i int64, n byte) bool { - upper := int64(1) << (n - 1) - if i >= upper { - return false - } - lower := upper - (1 << n) - if i < lower { - return false - } - return true -} diff --git a/storage/local/chunk/varbit_test.go b/storage/local/chunk/varbit_test.go deleted file mode 100644 index e0d0308220..0000000000 --- a/storage/local/chunk/varbit_test.go +++ /dev/null @@ -1,52 +0,0 @@ -// Copyright 2016 The Prometheus Authors -// Licensed under the Apache License, Version 2.0 (the "License"); -// you may not use this file except in compliance with the License. -// You may obtain a copy of the License at -// -// http://www.apache.org/licenses/LICENSE-2.0 -// -// Unless required by applicable law or agreed to in writing, software -// distributed under the License is distributed on an "AS IS" BASIS, -// WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. -// See the License for the specific language governing permissions and -// limitations under the License. - -package chunk - -import "testing" - -func TestCountBits(t *testing.T) { - for i := byte(0); i < 56; i++ { - for j := byte(0); j <= 8; j++ { - for k := byte(0); k < 8; k++ { - p := uint64(bitMask[j][k]) << i - gotLeading, gotSignificant := countBits(p) - wantLeading := 56 - i + k - wantSignificant := j - if j+k > 8 { - wantSignificant -= j + k - 8 - } - if wantLeading > 31 { - wantSignificant += wantLeading - 31 - wantLeading = 31 - } - if p == 0 { - wantLeading = 0 - wantSignificant = 0 - } - if wantLeading != gotLeading { - t.Errorf( - "unexpected leading bit count for i=%d, j=%d, k=%d; want %d, got %d", - i, j, k, wantLeading, gotLeading, - ) - } - if wantSignificant != gotSignificant { - t.Errorf( - "unexpected significant bit count for i=%d, j=%d, k=%d; want %d, got %d", - i, j, k, wantSignificant, gotSignificant, - ) - } - } - } - } -} diff --git a/storage/local/codable/codable.go b/storage/local/codable/codable.go deleted file mode 100644 index 37c273d1fd..0000000000 --- a/storage/local/codable/codable.go +++ /dev/null @@ -1,447 +0,0 @@ -// Copyright 2014 The Prometheus Authors -// Licensed under the Apache License, Version 2.0 (the "License"); -// you may not use this file except in compliance with the License. -// You may obtain a copy of the License at -// -// http://www.apache.org/licenses/LICENSE-2.0 -// -// Unless required by applicable law or agreed to in writing, software -// distributed under the License is distributed on an "AS IS" BASIS, -// WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. -// See the License for the specific language governing permissions and -// limitations under the License. - -// Package codable provides types that implement encoding.BinaryMarshaler and -// encoding.BinaryUnmarshaler and functions that help to encode and decode -// primitives. The Prometheus storage backend uses them to persist objects to -// files and to save objects in LevelDB. -// -// The encodings used in this package are designed in a way that objects can be -// unmarshaled from a continuous byte stream, i.e. the information when to stop -// reading is determined by the format. No separate termination information is -// needed. -// -// Strings are encoded as the length of their bytes as a varint followed by -// their bytes. -// -// Slices are encoded as their length as a varint followed by their elements. -// -// Maps are encoded as the number of mappings as a varint, followed by the -// mappings, each of which consists of the key followed by the value. -package codable - -import ( - "bytes" - "encoding/binary" - "fmt" - "io" - "sync" - - "github.com/prometheus/common/model" -) - -// A byteReader is an io.ByteReader that also implements the vanilla io.Reader -// interface. -type byteReader interface { - io.Reader - io.ByteReader -} - -// bufPool is a pool for staging buffers. Using a pool allows concurrency-safe -// reuse of buffers -var bufPool sync.Pool - -// getBuf returns a buffer from the pool. The length of the returned slice is l. -func getBuf(l int) []byte { - x := bufPool.Get() - if x == nil { - return make([]byte, l) - } - buf := x.([]byte) - if cap(buf) < l { - return make([]byte, l) - } - return buf[:l] -} - -// putBuf returns a buffer to the pool. -func putBuf(buf []byte) { - bufPool.Put(buf) -} - -// EncodeVarint encodes an int64 as a varint and writes it to an io.Writer. -// It returns the number of bytes written. -// This is a GC-friendly implementation that takes the required staging buffer -// from a buffer pool. -func EncodeVarint(w io.Writer, i int64) (int, error) { - buf := getBuf(binary.MaxVarintLen64) - defer putBuf(buf) - - bytesWritten := binary.PutVarint(buf, i) - _, err := w.Write(buf[:bytesWritten]) - return bytesWritten, err -} - -// EncodeUvarint encodes an uint64 as a varint and writes it to an io.Writer. -// It returns the number of bytes written. -// This is a GC-friendly implementation that takes the required staging buffer -// from a buffer pool. -func EncodeUvarint(w io.Writer, i uint64) (int, error) { - buf := getBuf(binary.MaxVarintLen64) - defer putBuf(buf) - - bytesWritten := binary.PutUvarint(buf, i) - _, err := w.Write(buf[:bytesWritten]) - return bytesWritten, err -} - -// EncodeUint64 writes an uint64 to an io.Writer in big-endian byte-order. -// This is a GC-friendly implementation that takes the required staging buffer -// from a buffer pool. -func EncodeUint64(w io.Writer, u uint64) error { - buf := getBuf(8) - defer putBuf(buf) - - binary.BigEndian.PutUint64(buf, u) - _, err := w.Write(buf) - return err -} - -// DecodeUint64 reads an uint64 from an io.Reader in big-endian byte-order. -// This is a GC-friendly implementation that takes the required staging buffer -// from a buffer pool. -func DecodeUint64(r io.Reader) (uint64, error) { - buf := getBuf(8) - defer putBuf(buf) - - if _, err := io.ReadFull(r, buf); err != nil { - return 0, err - } - return binary.BigEndian.Uint64(buf), nil -} - -// encodeString writes the varint encoded length followed by the bytes of s to -// b. -func encodeString(b *bytes.Buffer, s string) error { - if _, err := EncodeVarint(b, int64(len(s))); err != nil { - return err - } - if _, err := b.WriteString(s); err != nil { - return err - } - return nil -} - -// decodeString decodes a string encoded by encodeString. -func decodeString(b byteReader) (string, error) { - length, err := binary.ReadVarint(b) - if err != nil { - return "", err - } - - buf := getBuf(int(length)) - defer putBuf(buf) - - if _, err := io.ReadFull(b, buf); err != nil { - return "", err - } - return string(buf), nil -} - -// A Metric is a model.Metric that implements -// encoding.BinaryMarshaler and encoding.BinaryUnmarshaler. -type Metric model.Metric - -// MarshalBinary implements encoding.BinaryMarshaler. -func (m Metric) MarshalBinary() ([]byte, error) { - buf := &bytes.Buffer{} - if _, err := EncodeVarint(buf, int64(len(m))); err != nil { - return nil, err - } - for l, v := range m { - if err := encodeString(buf, string(l)); err != nil { - return nil, err - } - if err := encodeString(buf, string(v)); err != nil { - return nil, err - } - } - return buf.Bytes(), nil -} - -// UnmarshalBinary implements encoding.BinaryUnmarshaler. It can be used with the -// zero value of Metric. -func (m *Metric) UnmarshalBinary(buf []byte) error { - return m.UnmarshalFromReader(bytes.NewReader(buf)) -} - -// UnmarshalFromReader unmarshals a Metric from a reader that implements -// both, io.Reader and io.ByteReader. It can be used with the zero value of -// Metric. -func (m *Metric) UnmarshalFromReader(r byteReader) error { - numLabelPairs, err := binary.ReadVarint(r) - if err != nil { - return err - } - *m = make(Metric, numLabelPairs) - - for ; numLabelPairs > 0; numLabelPairs-- { - ln, err := decodeString(r) - if err != nil { - return err - } - lv, err := decodeString(r) - if err != nil { - return err - } - (*m)[model.LabelName(ln)] = model.LabelValue(lv) - } - return nil -} - -// A Fingerprint is a model.Fingerprint that implements -// encoding.BinaryMarshaler and encoding.BinaryUnmarshaler. The implementation -// depends on model.Fingerprint to be convertible to uint64. It encodes -// the fingerprint as a big-endian uint64. -type Fingerprint model.Fingerprint - -// MarshalBinary implements encoding.BinaryMarshaler. -func (fp Fingerprint) MarshalBinary() ([]byte, error) { - b := make([]byte, 8) - binary.BigEndian.PutUint64(b, uint64(fp)) - return b, nil -} - -// UnmarshalBinary implements encoding.BinaryUnmarshaler. -func (fp *Fingerprint) UnmarshalBinary(buf []byte) error { - *fp = Fingerprint(binary.BigEndian.Uint64(buf)) - return nil -} - -// FingerprintSet is a map[model.Fingerprint]struct{} that -// implements encoding.BinaryMarshaler and encoding.BinaryUnmarshaler. Its -// binary form is identical to that of Fingerprints. -type FingerprintSet map[model.Fingerprint]struct{} - -// MarshalBinary implements encoding.BinaryMarshaler. -func (fps FingerprintSet) MarshalBinary() ([]byte, error) { - b := make([]byte, binary.MaxVarintLen64+len(fps)*8) - lenBytes := binary.PutVarint(b, int64(len(fps))) - offset := lenBytes - - for fp := range fps { - binary.BigEndian.PutUint64(b[offset:], uint64(fp)) - offset += 8 - } - return b[:len(fps)*8+lenBytes], nil -} - -// UnmarshalBinary implements encoding.BinaryUnmarshaler. -func (fps *FingerprintSet) UnmarshalBinary(buf []byte) error { - numFPs, offset := binary.Varint(buf) - if offset <= 0 { - return fmt.Errorf("could not decode length of Fingerprints, varint decoding returned %d", offset) - } - *fps = make(FingerprintSet, numFPs) - - for i := 0; i < int(numFPs); i++ { - (*fps)[model.Fingerprint(binary.BigEndian.Uint64(buf[offset+i*8:]))] = struct{}{} - } - return nil -} - -// Fingerprints is a model.Fingerprints that implements -// encoding.BinaryMarshaler and encoding.BinaryUnmarshaler. Its binary form is -// identical to that of FingerprintSet. -type Fingerprints model.Fingerprints - -// MarshalBinary implements encoding.BinaryMarshaler. -func (fps Fingerprints) MarshalBinary() ([]byte, error) { - b := make([]byte, binary.MaxVarintLen64+len(fps)*8) - lenBytes := binary.PutVarint(b, int64(len(fps))) - - for i, fp := range fps { - binary.BigEndian.PutUint64(b[i*8+lenBytes:], uint64(fp)) - } - return b[:len(fps)*8+lenBytes], nil -} - -// UnmarshalBinary implements encoding.BinaryUnmarshaler. -func (fps *Fingerprints) UnmarshalBinary(buf []byte) error { - numFPs, offset := binary.Varint(buf) - if offset <= 0 { - return fmt.Errorf("could not decode length of Fingerprints, varint decoding returned %d", offset) - } - *fps = make(Fingerprints, numFPs) - - for i := range *fps { - (*fps)[i] = model.Fingerprint(binary.BigEndian.Uint64(buf[offset+i*8:])) - } - return nil -} - -// LabelPair is a model.LabelPair that implements -// encoding.BinaryMarshaler and encoding.BinaryUnmarshaler. -type LabelPair model.LabelPair - -// MarshalBinary implements encoding.BinaryMarshaler. -func (lp LabelPair) MarshalBinary() ([]byte, error) { - buf := &bytes.Buffer{} - if err := encodeString(buf, string(lp.Name)); err != nil { - return nil, err - } - if err := encodeString(buf, string(lp.Value)); err != nil { - return nil, err - } - return buf.Bytes(), nil -} - -// UnmarshalBinary implements encoding.BinaryUnmarshaler. -func (lp *LabelPair) UnmarshalBinary(buf []byte) error { - r := bytes.NewReader(buf) - n, err := decodeString(r) - if err != nil { - return err - } - v, err := decodeString(r) - if err != nil { - return err - } - lp.Name = model.LabelName(n) - lp.Value = model.LabelValue(v) - return nil -} - -// LabelName is a model.LabelName that implements -// encoding.BinaryMarshaler and encoding.BinaryUnmarshaler. -type LabelName model.LabelName - -// MarshalBinary implements encoding.BinaryMarshaler. -func (l LabelName) MarshalBinary() ([]byte, error) { - buf := &bytes.Buffer{} - if err := encodeString(buf, string(l)); err != nil { - return nil, err - } - return buf.Bytes(), nil -} - -// UnmarshalBinary implements encoding.BinaryUnmarshaler. -func (l *LabelName) UnmarshalBinary(buf []byte) error { - r := bytes.NewReader(buf) - n, err := decodeString(r) - if err != nil { - return err - } - *l = LabelName(n) - return nil -} - -// LabelValueSet is a map[model.LabelValue]struct{} that implements -// encoding.BinaryMarshaler and encoding.BinaryUnmarshaler. Its binary form is -// identical to that of LabelValues. -type LabelValueSet map[model.LabelValue]struct{} - -// MarshalBinary implements encoding.BinaryMarshaler. -func (vs LabelValueSet) MarshalBinary() ([]byte, error) { - buf := &bytes.Buffer{} - if _, err := EncodeVarint(buf, int64(len(vs))); err != nil { - return nil, err - } - for v := range vs { - if err := encodeString(buf, string(v)); err != nil { - return nil, err - } - } - return buf.Bytes(), nil -} - -// UnmarshalBinary implements encoding.BinaryUnmarshaler. -func (vs *LabelValueSet) UnmarshalBinary(buf []byte) error { - r := bytes.NewReader(buf) - numValues, err := binary.ReadVarint(r) - if err != nil { - return err - } - *vs = make(LabelValueSet, numValues) - - for i := int64(0); i < numValues; i++ { - v, err := decodeString(r) - if err != nil { - return err - } - (*vs)[model.LabelValue(v)] = struct{}{} - } - return nil -} - -// LabelValues is a model.LabelValues that implements -// encoding.BinaryMarshaler and encoding.BinaryUnmarshaler. Its binary form is -// identical to that of LabelValueSet. -type LabelValues model.LabelValues - -// MarshalBinary implements encoding.BinaryMarshaler. -func (vs LabelValues) MarshalBinary() ([]byte, error) { - buf := &bytes.Buffer{} - if _, err := EncodeVarint(buf, int64(len(vs))); err != nil { - return nil, err - } - for _, v := range vs { - if err := encodeString(buf, string(v)); err != nil { - return nil, err - } - } - return buf.Bytes(), nil -} - -// UnmarshalBinary implements encoding.BinaryUnmarshaler. -func (vs *LabelValues) UnmarshalBinary(buf []byte) error { - r := bytes.NewReader(buf) - numValues, err := binary.ReadVarint(r) - if err != nil { - return err - } - *vs = make(LabelValues, numValues) - - for i := range *vs { - v, err := decodeString(r) - if err != nil { - return err - } - (*vs)[i] = model.LabelValue(v) - } - return nil -} - -// TimeRange is used to define a time range and implements -// encoding.BinaryMarshaler and encoding.BinaryUnmarshaler. -type TimeRange struct { - First, Last model.Time -} - -// MarshalBinary implements encoding.BinaryMarshaler. -func (tr TimeRange) MarshalBinary() ([]byte, error) { - buf := &bytes.Buffer{} - if _, err := EncodeVarint(buf, int64(tr.First)); err != nil { - return nil, err - } - if _, err := EncodeVarint(buf, int64(tr.Last)); err != nil { - return nil, err - } - return buf.Bytes(), nil -} - -// UnmarshalBinary implements encoding.BinaryUnmarshaler. -func (tr *TimeRange) UnmarshalBinary(buf []byte) error { - r := bytes.NewReader(buf) - first, err := binary.ReadVarint(r) - if err != nil { - return err - } - last, err := binary.ReadVarint(r) - if err != nil { - return err - } - tr.First = model.Time(first) - tr.Last = model.Time(last) - return nil -} diff --git a/storage/local/codable/codable_test.go b/storage/local/codable/codable_test.go deleted file mode 100644 index d5c9cdf83b..0000000000 --- a/storage/local/codable/codable_test.go +++ /dev/null @@ -1,165 +0,0 @@ -// Copyright 2014 The Prometheus Authors -// Licensed under the Apache License, Version 2.0 (the "License"); -// you may not use this file except in compliance with the License. -// You may obtain a copy of the License at -// -// http://www.apache.org/licenses/LICENSE-2.0 -// -// Unless required by applicable law or agreed to in writing, software -// distributed under the License is distributed on an "AS IS" BASIS, -// WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. -// See the License for the specific language governing permissions and -// limitations under the License. - -package codable - -import ( - "bytes" - "encoding" - "reflect" - "testing" -) - -func newFingerprint(fp int64) *Fingerprint { - cfp := Fingerprint(fp) - return &cfp -} - -func newLabelName(ln string) *LabelName { - cln := LabelName(ln) - return &cln -} - -func TestUint64(t *testing.T) { - var b bytes.Buffer - const n = uint64(422010471112345) - if err := EncodeUint64(&b, n); err != nil { - t.Fatal(err) - } - got, err := DecodeUint64(&b) - if err != nil { - t.Fatal(err) - } - if got != n { - t.Errorf("want %d, got %d", n, got) - } -} - -var scenarios = []struct { - in encoding.BinaryMarshaler - out encoding.BinaryUnmarshaler - equal func(in, out interface{}) bool -}{ - { - in: &Metric{ - "label_1": "value_2", - "label_2": "value_2", - "label_3": "value_3", - }, - out: &Metric{}, - }, { - in: newFingerprint(12345), - out: newFingerprint(0), - }, { - in: &Fingerprints{1, 2, 56, 1234}, - out: &Fingerprints{}, - }, { - in: &Fingerprints{1, 2, 56, 1234}, - out: &FingerprintSet{}, - equal: func(in, out interface{}) bool { - inSet := FingerprintSet{} - for _, fp := range *(in.(*Fingerprints)) { - inSet[fp] = struct{}{} - } - return reflect.DeepEqual(inSet, *(out.(*FingerprintSet))) - }, - }, { - in: &FingerprintSet{ - 1: struct{}{}, - 2: struct{}{}, - 56: struct{}{}, - 1234: struct{}{}, - }, - out: &FingerprintSet{}, - }, { - in: &FingerprintSet{ - 1: struct{}{}, - 2: struct{}{}, - 56: struct{}{}, - 1234: struct{}{}, - }, - out: &Fingerprints{}, - equal: func(in, out interface{}) bool { - outSet := FingerprintSet{} - for _, fp := range *(out.(*Fingerprints)) { - outSet[fp] = struct{}{} - } - return reflect.DeepEqual(outSet, *(in.(*FingerprintSet))) - }, - }, { - in: &LabelPair{ - Name: "label_name", - Value: "label_value", - }, - out: &LabelPair{}, - }, { - in: newLabelName("label_name"), - out: newLabelName(""), - }, { - in: &LabelValues{"value_1", "value_2", "value_3"}, - out: &LabelValues{}, - }, { - in: &LabelValues{"value_1", "value_2", "value_3"}, - out: &LabelValueSet{}, - equal: func(in, out interface{}) bool { - inSet := LabelValueSet{} - for _, lv := range *(in.(*LabelValues)) { - inSet[lv] = struct{}{} - } - return reflect.DeepEqual(inSet, *(out.(*LabelValueSet))) - }, - }, { - in: &LabelValueSet{ - "value_1": struct{}{}, - "value_2": struct{}{}, - "value_3": struct{}{}, - }, - out: &LabelValueSet{}, - }, { - in: &LabelValueSet{ - "value_1": struct{}{}, - "value_2": struct{}{}, - "value_3": struct{}{}, - }, - out: &LabelValues{}, - equal: func(in, out interface{}) bool { - outSet := LabelValueSet{} - for _, lv := range *(out.(*LabelValues)) { - outSet[lv] = struct{}{} - } - return reflect.DeepEqual(outSet, *(in.(*LabelValueSet))) - }, - }, { - in: &TimeRange{42, 2001}, - out: &TimeRange{}, - }, -} - -func TestCodec(t *testing.T) { - for i, s := range scenarios { - encoded, err := s.in.MarshalBinary() - if err != nil { - t.Fatal(err) - } - if err := s.out.UnmarshalBinary(encoded); err != nil { - t.Fatal(err) - } - equal := s.equal - if equal == nil { - equal = reflect.DeepEqual - } - if !equal(s.in, s.out) { - t.Errorf("%d. Got: %v; want %v; encoded bytes are: %v", i, s.out, s.in, encoded) - } - } -} diff --git a/storage/local/crashrecovery.go b/storage/local/crashrecovery.go deleted file mode 100644 index 3b29da55e3..0000000000 --- a/storage/local/crashrecovery.go +++ /dev/null @@ -1,548 +0,0 @@ -// Copyright 2015 The Prometheus Authors -// Licensed under the Apache License, Version 2.0 (the "License"); -// you may not use this file except in compliance with the License. -// You may obtain a copy of the License at -// -// http://www.apache.org/licenses/LICENSE-2.0 -// -// Unless required by applicable law or agreed to in writing, software -// distributed under the License is distributed on an "AS IS" BASIS, -// WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. -// See the License for the specific language governing permissions and -// limitations under the License. - -package local - -import ( - "errors" - "fmt" - "io" - "os" - "path/filepath" - "strings" - "sync/atomic" - - "github.com/prometheus/common/log" - "github.com/prometheus/common/model" - - "github.com/prometheus/prometheus/storage/local/chunk" - "github.com/prometheus/prometheus/storage/local/codable" - "github.com/prometheus/prometheus/storage/local/index" -) - -// recoverFromCrash is called by loadSeriesMapAndHeads if the persistence -// appears to be dirty after the loading (either because the loading resulted in -// an error or because the persistence was dirty from the start). Not goroutine -// safe. Only call before anything else is running (except index processing -// queue as started by newPersistence). -func (p *persistence) recoverFromCrash(fingerprintToSeries map[model.Fingerprint]*memorySeries) error { - // TODO(beorn): We need proper tests for the crash recovery. - log.Warn("Starting crash recovery. Prometheus is inoperational until complete.") - log.Warn("To avoid crash recovery in the future, shut down Prometheus with SIGTERM or a HTTP POST to /-/quit.") - - fpsSeen := map[model.Fingerprint]struct{}{} - count := 0 - seriesDirNameFmt := fmt.Sprintf("%%0%dx", seriesDirNameLen) - - // Delete the fingerprint mapping file as it might be stale or - // corrupt. We'll rebuild the mappings as we go. - if err := os.RemoveAll(p.mappingsFileName()); err != nil { - return fmt.Errorf("couldn't remove old fingerprint mapping file %s: %s", p.mappingsFileName(), err) - } - // The mappings to rebuild. - fpm := fpMappings{} - - log.Info("Scanning files.") - for i := 0; i < 1<<(seriesDirNameLen*4); i++ { - dirname := filepath.Join(p.basePath, fmt.Sprintf(seriesDirNameFmt, i)) - dir, err := os.Open(dirname) - if os.IsNotExist(err) { - continue - } - if err != nil { - return err - } - for fis := []os.FileInfo{}; err != io.EOF; fis, err = dir.Readdir(1024) { - if err != nil { - dir.Close() - return err - } - for _, fi := range fis { - fp, ok := p.sanitizeSeries(dirname, fi, fingerprintToSeries, fpm) - if ok { - fpsSeen[fp] = struct{}{} - } - count++ - if count%10000 == 0 { - log.Infof("%d files scanned.", count) - } - } - } - dir.Close() - } - log.Infof("File scan complete. %d series found.", len(fpsSeen)) - - log.Info("Checking for series without series file.") - for fp, s := range fingerprintToSeries { - if _, seen := fpsSeen[fp]; !seen { - // fp exists in fingerprintToSeries, but has no representation on disk. - if s.persistWatermark == len(s.chunkDescs) { - // Oops, everything including the head chunk was - // already persisted, but nothing on disk. - // Thus, we lost that series completely. Clean - // up the remnants. - delete(fingerprintToSeries, fp) - if err := p.purgeArchivedMetric(fp); err != nil { - // Purging the archived metric didn't work, so try - // to unindex it, just in case it's in the indexes. - p.unindexMetric(fp, s.metric) - } - log.Warnf("Lost series detected: fingerprint %v, metric %v.", fp, s.metric) - continue - } - // If we are here, the only chunks we have are the chunks in the checkpoint. - // Adjust things accordingly. - if s.persistWatermark > 0 || s.chunkDescsOffset != 0 { - minLostChunks := s.persistWatermark + s.chunkDescsOffset - if minLostChunks <= 0 { - log.Warnf( - "Possible loss of chunks for fingerprint %v, metric %v.", - fp, s.metric, - ) - } else { - log.Warnf( - "Lost at least %d chunks for fingerprint %v, metric %v.", - minLostChunks, fp, s.metric, - ) - } - s.chunkDescs = append( - make([]*chunk.Desc, 0, len(s.chunkDescs)-s.persistWatermark), - s.chunkDescs[s.persistWatermark:]..., - ) - chunk.NumMemDescs.Sub(float64(s.persistWatermark)) - s.persistWatermark = 0 - s.chunkDescsOffset = 0 - } - maybeAddMapping(fp, s.metric, fpm) - fpsSeen[fp] = struct{}{} // Add so that fpsSeen is complete. - } - } - log.Info("Check for series without series file complete.") - - if err := p.cleanUpArchiveIndexes(fingerprintToSeries, fpsSeen, fpm); err != nil { - return err - } - if err := p.rebuildLabelIndexes(fingerprintToSeries); err != nil { - return err - } - // Finally rewrite the mappings file if there are any mappings. - if len(fpm) > 0 { - if err := p.checkpointFPMappings(fpm); err != nil { - return err - } - } - - p.dirtyMtx.Lock() - // Only declare storage clean if it didn't become dirty during crash recovery. - if !p.becameDirty { - p.dirty = false - } - p.dirtyMtx.Unlock() - - log.Warn("Crash recovery complete.") - return nil -} - -// sanitizeSeries sanitizes a series based on its series file as defined by the -// provided directory and FileInfo. The method returns the fingerprint as -// derived from the directory and file name, and whether the provided file has -// been sanitized. A file that failed to be sanitized is moved into the -// "orphaned" sub-directory, if possible. -// -// The following steps are performed: -// -// - A file whose name doesn't comply with the naming scheme of a series file is -// simply moved into the orphaned directory. -// -// - If the size of the series file isn't a multiple of the chunk size, -// extraneous bytes are truncated. If the truncation fails, the file is -// moved into the orphaned directory. -// -// - A file that is empty (after truncation) is deleted. -// -// - A series that is not archived (i.e. it is in the fingerprintToSeries map) -// is checked for consistency of its various parameters (like persist -// watermark, offset of chunkDescs etc.). In particular, overlap between an -// in-memory head chunk with the most recent persisted chunk is -// checked. Inconsistencies are rectified. -// -// - A series that is archived (i.e. it is not in the fingerprintToSeries map) -// is checked for its presence in the index of archived series. If it cannot -// be found there, it is moved into the orphaned directory. -func (p *persistence) sanitizeSeries( - dirname string, fi os.FileInfo, - fingerprintToSeries map[model.Fingerprint]*memorySeries, - fpm fpMappings, -) (model.Fingerprint, bool) { - var ( - fp model.Fingerprint - err error - filename = filepath.Join(dirname, fi.Name()) - s *memorySeries - ) - - purge := func() { - if fp != 0 { - var metric model.Metric - if s != nil { - metric = s.metric - } - if err = p.quarantineSeriesFile( - fp, errors.New("purge during crash recovery"), metric, - ); err == nil { - return - } - log. - With("file", filename). - With("error", err). - Error("Failed to move lost series file to orphaned directory.") - } - // If we are here, we are either purging an incorrectly named - // file, or quarantining has failed. So simply delete the file. - if err = os.Remove(filename); err != nil { - log. - With("file", filename). - With("error", err). - Error("Failed to delete lost series file.") - } - } - - if len(fi.Name()) != fpLen-seriesDirNameLen+len(seriesFileSuffix) || - !strings.HasSuffix(fi.Name(), seriesFileSuffix) { - log.Warnf("Unexpected series file name %s.", filename) - purge() - return fp, false - } - if fp, err = model.FingerprintFromString(filepath.Base(dirname) + fi.Name()[:fpLen-seriesDirNameLen]); err != nil { - log.Warnf("Error parsing file name %s: %s", filename, err) - purge() - return fp, false - } - - bytesToTrim := fi.Size() % int64(chunkLenWithHeader) - chunksInFile := int(fi.Size()) / chunkLenWithHeader - modTime := fi.ModTime() - if bytesToTrim != 0 { - log.Warnf( - "Truncating file %s to exactly %d chunks, trimming %d extraneous bytes.", - filename, chunksInFile, bytesToTrim, - ) - f, err := os.OpenFile(filename, os.O_WRONLY, 0640) - if err != nil { - log.Errorf("Could not open file %s: %s", filename, err) - purge() - return fp, false - } - if err := f.Truncate(fi.Size() - bytesToTrim); err != nil { - log.Errorf("Failed to truncate file %s: %s", filename, err) - purge() - return fp, false - } - } - if chunksInFile == 0 { - log.Warnf("No chunks left in file %s.", filename) - purge() - return fp, false - } - - s, ok := fingerprintToSeries[fp] - if ok { // This series is supposed to not be archived. - if s == nil { - panic("fingerprint mapped to nil pointer") - } - maybeAddMapping(fp, s.metric, fpm) - if !p.pedanticChecks && - bytesToTrim == 0 && - s.chunkDescsOffset != -1 && - chunksInFile == s.chunkDescsOffset+s.persistWatermark && - modTime.Equal(s.modTime) { - // Everything is consistent. We are good. - return fp, true - } - // If we are here, we cannot be sure the series file is - // consistent with the checkpoint, so we have to take a closer - // look. - if s.headChunkClosed { - // This is the easy case as we have all chunks on - // disk. Treat this series as a freshly unarchived one - // by loading the chunkDescs and setting all parameters - // based on the loaded chunkDescs. - cds, err := p.loadChunkDescs(fp, 0) - if err != nil { - log.Errorf( - "Failed to load chunk descriptors for metric %v, fingerprint %v: %s", - s.metric, fp, err, - ) - purge() - return fp, false - } - log.Warnf( - "Treating recovered metric %v, fingerprint %v, as freshly unarchived, with %d chunks in series file.", - s.metric, fp, len(cds), - ) - s.chunkDescs = cds - s.chunkDescsOffset = 0 - s.savedFirstTime = cds[0].FirstTime() - s.lastTime, err = cds[len(cds)-1].LastTime() - if err != nil { - log.Errorf( - "Failed to determine time of the last sample for metric %v, fingerprint %v: %s", - s.metric, fp, err, - ) - purge() - return fp, false - } - s.persistWatermark = len(cds) - s.modTime = modTime - return fp, true - } - // This is the tricky one: We have chunks from heads.db, but - // some of those chunks might already be in the series - // file. Strategy: Take the last time of the most recent chunk - // in the series file. Then find the oldest chunk among those - // from heads.db that has a first time later or equal to the - // last time from the series file. Throw away the older chunks - // from heads.db and stitch the parts together. - - // First, throw away the chunkDescs without chunks. - s.chunkDescs = s.chunkDescs[s.persistWatermark:] - chunk.NumMemDescs.Sub(float64(s.persistWatermark)) - cds, err := p.loadChunkDescs(fp, 0) - if err != nil { - log.Errorf( - "Failed to load chunk descriptors for metric %v, fingerprint %v: %s", - s.metric, fp, err, - ) - purge() - return fp, false - } - s.persistWatermark = len(cds) - s.chunkDescsOffset = 0 - s.savedFirstTime = cds[0].FirstTime() - s.modTime = modTime - - lastTime, err := cds[len(cds)-1].LastTime() - if err != nil { - log.Errorf( - "Failed to determine time of the last sample for metric %v, fingerprint %v: %s", - s.metric, fp, err, - ) - purge() - return fp, false - } - keepIdx := -1 - for i, cd := range s.chunkDescs { - if cd.FirstTime() >= lastTime { - keepIdx = i - break - } - } - if keepIdx == -1 { - log.Warnf( - "Recovered metric %v, fingerprint %v: all %d chunks recovered from series file.", - s.metric, fp, chunksInFile, - ) - chunk.NumMemDescs.Sub(float64(len(s.chunkDescs))) - atomic.AddInt64(&chunk.NumMemChunks, int64(-len(s.chunkDescs))) - s.chunkDescs = cds - s.headChunkClosed = true - return fp, true - } - log.Warnf( - "Recovered metric %v, fingerprint %v: recovered %d chunks from series file, recovered %d chunks from checkpoint.", - s.metric, fp, chunksInFile, len(s.chunkDescs)-keepIdx, - ) - chunk.NumMemDescs.Sub(float64(keepIdx)) - atomic.AddInt64(&chunk.NumMemChunks, int64(-keepIdx)) - if keepIdx == len(s.chunkDescs) { - // No chunks from series file left, head chunk is evicted, so declare it closed. - s.headChunkClosed = true - } - s.chunkDescs = append(cds, s.chunkDescs[keepIdx:]...) - return fp, true - } - // This series is supposed to be archived. - metric, err := p.archivedMetric(fp) - if err != nil { - log.Errorf( - "Fingerprint %v assumed archived but couldn't be looked up in archived index: %s", - fp, err, - ) - purge() - return fp, false - } - if metric == nil { - log.Warnf( - "Fingerprint %v assumed archived but couldn't be found in archived index.", - fp, - ) - purge() - return fp, false - } - // This series looks like a properly archived one. - maybeAddMapping(fp, metric, fpm) - return fp, true -} - -func (p *persistence) cleanUpArchiveIndexes( - fpToSeries map[model.Fingerprint]*memorySeries, - fpsSeen map[model.Fingerprint]struct{}, - fpm fpMappings, -) error { - log.Info("Cleaning up archive indexes.") - var fp codable.Fingerprint - var m codable.Metric - count := 0 - if err := p.archivedFingerprintToMetrics.ForEach(func(kv index.KeyValueAccessor) error { - count++ - if count%10000 == 0 { - log.Infof("%d archived metrics checked.", count) - } - if err := kv.Key(&fp); err != nil { - return err - } - _, fpSeen := fpsSeen[model.Fingerprint(fp)] - inMemory := false - if fpSeen { - _, inMemory = fpToSeries[model.Fingerprint(fp)] - } - if !fpSeen || inMemory { - if inMemory { - log.Warnf("Archive clean-up: Fingerprint %v is not archived. Purging from archive indexes.", model.Fingerprint(fp)) - } - if !fpSeen { - log.Warnf("Archive clean-up: Fingerprint %v is unknown. Purging from archive indexes.", model.Fingerprint(fp)) - } - // It's fine if the fp is not in the archive indexes. - if _, err := p.archivedFingerprintToMetrics.Delete(fp); err != nil { - return err - } - // Delete from timerange index, too. - _, err := p.archivedFingerprintToTimeRange.Delete(fp) - return err - } - // fp is legitimately archived. Now we need the metric to check for a mapped fingerprint. - if err := kv.Value(&m); err != nil { - return err - } - maybeAddMapping(model.Fingerprint(fp), model.Metric(m), fpm) - // Make sure it is in timerange index, too. - has, err := p.archivedFingerprintToTimeRange.Has(fp) - if err != nil { - return err - } - if has { - return nil // All good. - } - log.Warnf("Archive clean-up: Fingerprint %v is not in time-range index. Unarchiving it for recovery.") - // Again, it's fine if fp is not in the archive index. - if _, err := p.archivedFingerprintToMetrics.Delete(fp); err != nil { - return err - } - cds, err := p.loadChunkDescs(model.Fingerprint(fp), 0) - if err != nil { - return err - } - series, err := newMemorySeries(model.Metric(m), cds, p.seriesFileModTime(model.Fingerprint(fp))) - if err != nil { - return err - } - fpToSeries[model.Fingerprint(fp)] = series - return nil - }); err != nil { - return err - } - count = 0 - if err := p.archivedFingerprintToTimeRange.ForEach(func(kv index.KeyValueAccessor) error { - count++ - if count%10000 == 0 { - log.Infof("%d archived time ranges checked.", count) - } - if err := kv.Key(&fp); err != nil { - return err - } - has, err := p.archivedFingerprintToMetrics.Has(fp) - if err != nil { - return err - } - if has { - return nil // All good. - } - log.Warnf("Archive clean-up: Purging unknown fingerprint %v in time-range index.", fp) - deleted, err := p.archivedFingerprintToTimeRange.Delete(fp) - if err != nil { - return err - } - if !deleted { - log.Errorf("Fingerprint %v to be deleted from archivedFingerprintToTimeRange not found. This should never happen.", fp) - } - return nil - }); err != nil { - return err - } - log.Info("Clean-up of archive indexes complete.") - return nil -} - -func (p *persistence) rebuildLabelIndexes( - fpToSeries map[model.Fingerprint]*memorySeries, -) error { - count := 0 - log.Info("Rebuilding label indexes.") - log.Info("Indexing metrics in memory.") - for fp, s := range fpToSeries { - p.indexMetric(fp, s.metric) - count++ - if count%10000 == 0 { - log.Infof("%d metrics queued for indexing.", count) - } - } - log.Info("Indexing archived metrics.") - var fp codable.Fingerprint - var m codable.Metric - if err := p.archivedFingerprintToMetrics.ForEach(func(kv index.KeyValueAccessor) error { - if err := kv.Key(&fp); err != nil { - return err - } - if err := kv.Value(&m); err != nil { - return err - } - p.indexMetric(model.Fingerprint(fp), model.Metric(m)) - count++ - if count%10000 == 0 { - log.Infof("%d metrics queued for indexing.", count) - } - return nil - }); err != nil { - return err - } - log.Info("All requests for rebuilding the label indexes queued. (Actual processing may lag behind.)") - return nil -} - -// maybeAddMapping adds a fingerprint mapping to fpm if the FastFingerprint of m is different from fp. -func maybeAddMapping(fp model.Fingerprint, m model.Metric, fpm fpMappings) { - if rawFP := m.FastFingerprint(); rawFP != fp { - log.Warnf( - "Metric %v with fingerprint %v is mapped from raw fingerprint %v.", - m, fp, rawFP, - ) - if mappedFPs, ok := fpm[rawFP]; ok { - mappedFPs[metricToUniqueString(m)] = fp - } else { - fpm[rawFP] = map[string]model.Fingerprint{ - metricToUniqueString(m): fp, - } - } - } -} diff --git a/storage/local/fixtures/b0/04b821ca50ba26.db b/storage/local/fixtures/b0/04b821ca50ba26.db deleted file mode 100644 index a9be8736432a15b8e92cad9e5451ceff7f745195..0000000000000000000000000000000000000000 GIT binary patch literal 0 HcmV?d00001 literal 62460 zcmeI5d2Ccw6vk&-loEs#tDzyFEJ8{OQZ`u(n?x)^!(xDHRf-6*Xj%nAv0#y4K-5^H z5Ks|}5P^iIA|Mcgl?W+|r2&Ek5fBin5G?zMRBYejJ9m-^lL0G+o{vA6ujk!)Gwqie+YP^_p|{0ZcgMti}BpDTa5)tueEyFPd79g_cx z_Z)oA;CwshB=~i2lyCHYg2xAs39kQ}9LYIvG$xYz&5j8k|9|D+`MlZp5Xo~ZZ|n14 z2?)LzzQY5;m1uvA4@KPcAwa+Zy|p!5yYV3%6G^D5^Mt-uBxjCfjGj5nU_GQkpA{M+ z858DNVV-_#heBVgsw!dLPv}d8863%>Fk?brD@;Ytqal*PVaD9rq0m>?iUvKiRzic0 zalvc){y8HB1Z&Rk^nmbZbX1HFMcmvWKnOYu#`q(wIkyP^40m)w9lQopAcu3nIjFSu zs@?}44k~WWg_i{cDNBo-A)tsGF$4%W;2^+CAcwP2u~9h*cj_dFA#_|WAZR{ymj{HX zsEUsaDB>%D>x7H7bp-MrWXL5H-x-lO-WB8mXQ*o#qu5AW|lPEOr&{Z#UI z!z-GnC0^EiwaW#~*~w=$Um1Ev^ORAiGzTV>YMwmfSIw7lk7}N@=&MKhgZ$U-LB|tTj(_ zTKtEa546bD{B+0nG?(|2kYwnTPQ}d>c-86UG(N*)t{hc&FTG~PLhD&WVce$&L z=5=wcGRRE2v~K} z{!aaR@YYt)Lz+K*kbofe+D{G$DB{Ef0)bHZt1cpeto-Z7BKtK@_Mnhn8q#<}WOf}Xd0YbuP0YORJFAfMO5(w-a zG=$0f*1Y!uO+XMtLlaD(fpor-q z5C{YU!MkCqu__1<%3c=`EKWG#fPf-_V6jtR)p_kEJ}KyO3lL_v5)gDgaMA$*MU15( z&=6<{HVmQhApt@0JHI&~phzIFOWNv^*zp$ET?}E}`vQV7-A*|mporHg3&j|?TFFxadfFiaA4S|M0L$F~8)0YSck_MMK zAfQMfuuIzNlGyPU*If)DeyD(;F#4PW0*aU(8UhW0hG4@G@@onR($}1KKtPc|u-GZE z>b!Pab|i+7yk0;M7=6(J0Yz*L8UhW0hG4@Gj*Jr!1=&{|5Ktr#EOrX4I#1LZo2ng1lk8(gj5nF?XKtrG* z*f4~}mjwg^m)3AVK#@RTm$cO-vEwbSyBI>}Z&ffFiaA4S|M0L$F~8zD)vx{PDFN5Ktr# z*d=XsN$hxw>n?^cKSMx}+^CKN0*aU(8UhW0hG4@Gl9~tzjuh2(KtPc|u-GZE>b!Pa zb|i*SxLZJwH9gh=0Yz*L8UhW0hG4@G(q{??V%y#8fPf-_z%FU4OJc`cTz4^qKt}<= zvSV=$2q-Ig7RA(S>35X|56kOKmW*cvnh8UhW$h9OMh7GLGaENucd=r2=i`swZ@)YH zIy3V--@OfBPfS_<@F?4|+O6BEj@RSbI4w4PzVDB>*b}FO#gF}uZdMl!9V*XV5GDG* z_{@Q?8K`e%O#_00-unrEuyLAQ!*wT(L@NE=Axt3dLo=x4rh#fun>dgln8lN zun1>Nh-Za(`pumRey`SYLVTX!w+JyfoKqpj1ix1ZM_xxnIDlLk+7F%bB1^NZ+R>Qe*)>mNP++|bX^XA9ahWT3po zr~zu=PwlO}4x0C+{lR4dLFOm>bV6X0HzEUpABWT8F~&D*C<2P$-)T{?5)=~bv^4~R zqy_sG5F8HQB?Frb006Zs_8Ya2ut zR38$3epj97R}OtAy781-bl#=CqR+P3E&An-wW81T*e?2|)NP`F8~Tmt8JX3hJ=t4C zPn)?}^vS}{MNe5;Df+~_Yei44_)PTi4WEjhw6#L?vHc&3&N=d-=p$#|7yaDTGSLmU zuMnNx`5n=RdzFfQChbkp^&^&xo;Yr)=(^k`qQ~dGCi>vQMWUa6qe%3D_XUKG8n!z9royFV|wHYHp1L+^8 zZGA;ob?hU0-~+uyZ%*wcI&D}_(Hk=pMfZQUyXXxwyNK>LFF|zW(oUlLuDDzDx{3~> zecRjamO~)OZ(FB;pza~C$v_B72s#Si0<}MZKSyh;l}Y@{P~KwHz~mY<$00kb9f81o zj*fl(mkez3b_oIkpU^mf6BXBqtv*f7>j;M!seoO(s$M07!py_!sDACpCNCcV!K8Z# zJnNu+blePVt@%<&vvbE12;!T5&_H05-wHqg5TfK(IWJkzEMl4w5WJ2C1%lO;KWZSb z$<+lA00aO*ZI~FoFeD(9ZX^)6UO1|Oz$SozMiX(P;YEYeMOo}VgjD%2=Kp@E9ep~~AO#lIHF3Q>*FIh}n#1lseA!{drz&ZDX1_GO0T_gk~ z1SAAA5<=sA0zrB5Neu)x0R%Lfh#L(r8k8=|N=FG{SswyH`r%U=2yF7)A|W6lAR(BM z5E73N2;8s#s)4{JfPgj^Wo?d^EG91EiKB#2^%j9(`jFEa2yAk7kr0p&kPysB2xEp5 z2&~`FXdti&AfVAi+-P{wpmb4II!Xwh^8|vj^0OKUZ1UV9As`_jA()X6N-78hspA?o z5ZD9|(B`77&GC}O#6>)Dln^>kAQ05Xo!3BMldFq_fP{dAU`9fyY)c@>-E=_%flX6@ zpr*QHmP;Ki2wxpg1T%|ZXKW!5Bu>4mfY8-pTVkpW;#(S=5NN~TCn|sn#ApR<=S%`Y zRl;?h5ZE*&A@KLag=fSUo+w?+{Do+RIJ>+nfndze8yW~~@@#WL;K$fF*f{>>6Ek}O z3)?r91cdax1Om@|tH1xhbW{I?8k<}v00B1@vHjJCiD3p35ZsFh1SNeO8VGCx2*^T+ z0sf5=!yMpEu6SVC(^ClqiFe+ifbfw2rXrhaj4K~E6{%SGi3-Ge1#D+LfuQQEb~+)j z$?(7>11=fN=#sqRO-zQ0b$0ny0>PNP_8JIma(Ms*00BTy8zzRY9S8{N1q1?5;$0dD zYyt=-b_$+#e0L*m9msSO5Zn(E2ucpbYap=6W04S$5ReeeNC?yG2n3y%+@pcOCV+r0 zX;zoy6>s9aQ$nzo5eO>NI%y!V$>l*pKte!5Fe4$94I&U^oVr&7flUCx#7@Dpj_+<( zI--P-dYV9RtSmtTflXc-NC-#>NC;*mgxZw^g2E^6(?DPoKtPu?t4s2VH*wx6A>=+q zAc()#MFW9NE)Nm{5&{x}841DLL?Bqbwwne5n*f4|oq}f_-`%WqL3JHUR{5Nwc~nuXq#ZT>P diff --git a/storage/local/fixtures/b0/37de1e884e5469.db b/storage/local/fixtures/b0/37de1e884e5469.db deleted file mode 100644 index 3445e010911beb8018c3f9e1adb801eb5abc62b6..0000000000000000000000000000000000000000 GIT binary patch literal 0 HcmV?d00001 literal 51009 zcmeI5du)?c7{>c`7^(WA`XX;FocDH0ZNEEl*z>aGgCK2U?67GK^P3!BqS_| zamGj%MVWC%Fo6sN9A@euLdLy8GRH+gM5sb=T!hMPw&&gZo`z<%#ffRp!yoI@r|0|H zzUQ3Z`+Of^SQC0Jd2FO*v$a{XT|HipYvr)9>GOSmyhWZkC2aiM|Lj(E;gBJ6?7S$^ z|HWqxJZGT3oiz!3-{y34pC@p9;Fv)F-{^48+1!|L{7A-_{q_ z5(qxc+NpqWCC+l#*yK$g1O$FTwzi1I=3kOA;fO{;&{P*a+FDM8bLMcy$Oj8CSY8q# zM+J*;#)LR3#NoI0Qt*4VloR6f1iwXy!Qs3VVodORg>d9NBElIQV$7|*6#VYuh>$aj z6D(xRO~P6-i9oR7!Y&1bzvAo;8=EwEs*8cZ&znwCGsC47fxz~UUbsi-^X0c?$Uu3E zPy^J!pW0h`9W?Dr>!ZsAg7nY!=!C!~Z$t(HKM$wHV~lUsPy`ggztf^(B`73VDXRzs ziSza004qXTf3=O;K9TSScQph1%m#W`_-emsmLY+VcMT|!iW0lEh=hv zUG5d!=yFB${9czuUmJK)bYa>#(O1Tw6+LJ2Y0=(%kLcM8e-(YH^tk9*D~^i3xcac@ z{OSXu&+n)c{rZ6)L^qsri_X2YTlCphJ4L_NzE<>^Zd*man!H8y??b*5JuSUjv?p_u z=&92;iauHJrRd3vD@C7pf3@gI6<>%xzV36;6E|0gKDPH0(OE}67JcN*hoWD+S|&Xg5*xh^{Y*uI-g6`l&$~qPL}u6aB=5 zv7*19JWBNNnIlDSSvW%Uuy>ynU9)1i=%H&|qQ9*kEPC+HL87Y<3=sXer@!d0FZC5Y z@QyyBtJ?P#J>a39qBkb@5S=o#yXf`lPSO2d=qh^M^iHDt&PfnmxwwPqKFjVEy{4j_ zXy5j>x?~Xu@>_0< z&(UKa|0M&Pyj_BTz%OVVz=?|M#1@~Xrgeme7^#3&yP{qtgM##f>QVi*Bb&T@00fin zA@Hn&_R(=Ou%+fpA41A$Fb zfS{(jn15v~gs&bbf|*6I(l!wYoKvnUAau4{HZggla6(Ye@dFh=1!A=VmSZ}Bpeo_I zP6%w8k`VZN;=(iH3s00TX8u96Vw_dpnLsdl`wa~QHhH!=A@Fl-9Bdr_@`;(ffQ9Xw zN&-UaZUTX4uC1T{zjRZ-LXAzX6M%r5irD^Y!^AKH2?*{51cKt;b`1nJ0R&_r!~p+B ziD3@#CRaSLSvkoB0_WX#DIh%Jzp2Qk8so~l>A!NLBk>azKm}s10+u75Kv4Bf8=Vl? zWO(3`0hbJBbV*+ECMH7#H>-RzfnaoQTMYy@xjX;@fB+z<4HLuH4g`eMd;)>Td5;DH zn*f4|oq}f_-`$8?2Qu9R1oy)Pg5rJg8VGFiSR@1_1SAAA5<*TLfuQ4}`!o>P1Q5_A z&FYf8;!T`)N(iW+a5NfdqoIQ}=5iun8cT*eQ6{@!icz zN0bnfPZJ1^l_h8(u*pjU2>}TK3BinnP`jK!Q1J8v8VGCx29@Ic! zla~e(0ullef*A=RX%c~;KHjN;z$SozE@@VmMxjaY+NC-#>W+a5Fr38YUL62%6 Uun8cT*eQ6{@!iczM+$g<1Ej{ExBvhE diff --git a/storage/local/heads.go b/storage/local/heads.go deleted file mode 100644 index 14a3175926..0000000000 --- a/storage/local/heads.go +++ /dev/null @@ -1,251 +0,0 @@ -// Copyright 2016 The Prometheus Authors -// Licensed under the Apache License, Version 2.0 (the "License"); -// you may not use this file except in compliance with the License. -// You may obtain a copy of the License at -// -// http://www.apache.org/licenses/LICENSE-2.0 -// -// Unless required by applicable law or agreed to in writing, software -// distributed under the License is distributed on an "AS IS" BASIS, -// WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. -// See the License for the specific language governing permissions and -// limitations under the License. - -package local - -import ( - "bufio" - "encoding/binary" - "fmt" - "io" - "os" - "time" - - "github.com/prometheus/common/model" - - "github.com/prometheus/prometheus/storage/local/chunk" - "github.com/prometheus/prometheus/storage/local/codable" -) - -const ( - headsFileName = "heads.db" - headsTempFileName = "heads.db.tmp" - headsFormatVersion = 2 - headsFormatLegacyVersion = 1 // Can read, but will never write. - headsMagicString = "PrometheusHeads" -) - -// headsScanner is a scanner to read time series with their heads from a -// heads.db file. It follows a similar semantics as the bufio.Scanner. -// It is not safe to use a headsScanner concurrently. -type headsScanner struct { - f *os.File - r *bufio.Reader - fp model.Fingerprint // Read after each scan() call that has returned true. - series *memorySeries // Read after each scan() call that has returned true. - version int64 // Read after newHeadsScanner has returned. - seriesTotal uint64 // Read after newHeadsScanner has returned. - seriesCurrent uint64 - chunksToPersistTotal int64 // Read after scan() has returned false. - err error // Read after scan() has returned false. -} - -func newHeadsScanner(filename string) *headsScanner { - hs := &headsScanner{} - defer func() { - if hs.f != nil && hs.err != nil { - hs.f.Close() - } - }() - - if hs.f, hs.err = os.Open(filename); hs.err != nil { - return hs - } - hs.r = bufio.NewReaderSize(hs.f, fileBufSize) - - buf := make([]byte, len(headsMagicString)) - if _, hs.err = io.ReadFull(hs.r, buf); hs.err != nil { - return hs - } - magic := string(buf) - if magic != headsMagicString { - hs.err = fmt.Errorf( - "unexpected magic string, want %q, got %q", - headsMagicString, magic, - ) - return hs - } - hs.version, hs.err = binary.ReadVarint(hs.r) - if (hs.version != headsFormatVersion && hs.version != headsFormatLegacyVersion) || hs.err != nil { - hs.err = fmt.Errorf( - "unknown or unreadable heads format version, want %d, got %d, error: %s", - headsFormatVersion, hs.version, hs.err, - ) - return hs - } - if hs.seriesTotal, hs.err = codable.DecodeUint64(hs.r); hs.err != nil { - return hs - } - return hs -} - -// scan works like bufio.Scanner.Scan. -func (hs *headsScanner) scan() bool { - if hs.seriesCurrent == hs.seriesTotal || hs.err != nil { - return false - } - - var ( - seriesFlags byte - fpAsInt uint64 - metric codable.Metric - persistWatermark int64 - modTimeNano int64 - modTime time.Time - chunkDescsOffset int64 - savedFirstTime int64 - numChunkDescs int64 - firstTime int64 - lastTime int64 - encoding byte - ch chunk.Chunk - lastTimeHead model.Time - ) - if seriesFlags, hs.err = hs.r.ReadByte(); hs.err != nil { - return false - } - headChunkPersisted := seriesFlags&flagHeadChunkPersisted != 0 - if fpAsInt, hs.err = codable.DecodeUint64(hs.r); hs.err != nil { - return false - } - hs.fp = model.Fingerprint(fpAsInt) - - if hs.err = metric.UnmarshalFromReader(hs.r); hs.err != nil { - return false - } - if hs.version != headsFormatLegacyVersion { - // persistWatermark only present in v2. - persistWatermark, hs.err = binary.ReadVarint(hs.r) - if hs.err != nil { - return false - } - modTimeNano, hs.err = binary.ReadVarint(hs.r) - if hs.err != nil { - return false - } - if modTimeNano != -1 { - modTime = time.Unix(0, modTimeNano) - } - } - if chunkDescsOffset, hs.err = binary.ReadVarint(hs.r); hs.err != nil { - return false - } - if savedFirstTime, hs.err = binary.ReadVarint(hs.r); hs.err != nil { - return false - } - - if numChunkDescs, hs.err = binary.ReadVarint(hs.r); hs.err != nil { - return false - } - chunkDescs := make([]*chunk.Desc, numChunkDescs) - if hs.version == headsFormatLegacyVersion { - if headChunkPersisted { - persistWatermark = numChunkDescs - } else { - persistWatermark = numChunkDescs - 1 - } - } - headChunkClosed := true // Initial assumption. - for i := int64(0); i < numChunkDescs; i++ { - if i < persistWatermark { - if firstTime, hs.err = binary.ReadVarint(hs.r); hs.err != nil { - return false - } - if lastTime, hs.err = binary.ReadVarint(hs.r); hs.err != nil { - return false - } - chunkDescs[i] = &chunk.Desc{ - ChunkFirstTime: model.Time(firstTime), - ChunkLastTime: model.Time(lastTime), - } - chunk.NumMemDescs.Inc() - } else { - // Non-persisted chunk. - // If there are non-persisted chunks at all, we consider - // the head chunk not to be closed yet. - headChunkClosed = false - if encoding, hs.err = hs.r.ReadByte(); hs.err != nil { - return false - } - if ch, hs.err = chunk.NewForEncoding(chunk.Encoding(encoding)); hs.err != nil { - return false - } - if hs.err = ch.Unmarshal(hs.r); hs.err != nil { - return false - } - cd := chunk.NewDesc(ch, ch.FirstTime()) - if i < numChunkDescs-1 { - // This is NOT the head chunk. So it's a chunk - // to be persisted, and we need to populate lastTime. - hs.chunksToPersistTotal++ - cd.MaybePopulateLastTime() - } - chunkDescs[i] = cd - } - } - - if lastTimeHead, hs.err = chunkDescs[len(chunkDescs)-1].LastTime(); hs.err != nil { - return false - } - - hs.series = &memorySeries{ - metric: model.Metric(metric), - chunkDescs: chunkDescs, - persistWatermark: int(persistWatermark), - modTime: modTime, - chunkDescsOffset: int(chunkDescsOffset), - savedFirstTime: model.Time(savedFirstTime), - lastTime: lastTimeHead, - headChunkClosed: headChunkClosed, - } - hs.seriesCurrent++ - return true -} - -// close closes the underlying file if required. -func (hs *headsScanner) close() { - if hs.f != nil { - hs.f.Close() - } -} - -// DumpHeads writes the metadata of the provided heads file in a human-readable -// form. -func DumpHeads(filename string, out io.Writer) error { - hs := newHeadsScanner(filename) - defer hs.close() - - if hs.err == nil { - fmt.Fprintf( - out, - ">>> Dumping %d series from heads file %q with format version %d. <<<\n", - hs.seriesTotal, filename, hs.version, - ) - } - for hs.scan() { - s := hs.series - fmt.Fprintf( - out, - "FP=%v\tMETRIC=%s\tlen(chunkDescs)=%d\tpersistWatermark=%d\tchunkDescOffset=%d\tsavedFirstTime=%v\tlastTime=%v\theadChunkClosed=%t\n", - hs.fp, s.metric, len(s.chunkDescs), s.persistWatermark, s.chunkDescsOffset, s.savedFirstTime, s.lastTime, s.headChunkClosed, - ) - } - if hs.err == nil { - fmt.Fprintf( - out, - ">>> Dump complete. %d chunks to persist. <<<\n", - hs.chunksToPersistTotal, - ) - } - return hs.err -} diff --git a/storage/local/index/index.go b/storage/local/index/index.go deleted file mode 100644 index e189004cc7..0000000000 --- a/storage/local/index/index.go +++ /dev/null @@ -1,296 +0,0 @@ -// Copyright 2014 The Prometheus Authors -// Licensed under the Apache License, Version 2.0 (the "License"); -// you may not use this file except in compliance with the License. -// You may obtain a copy of the License at -// -// http://www.apache.org/licenses/LICENSE-2.0 -// -// Unless required by applicable law or agreed to in writing, software -// distributed under the License is distributed on an "AS IS" BASIS, -// WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. -// See the License for the specific language governing permissions and -// limitations under the License. - -// Package index provides a number of indexes backed by persistent key-value -// stores. The only supported implementation of a key-value store is currently -// goleveldb, but other implementations can easily be added. -package index - -import ( - "os" - "path" - "path/filepath" - - "github.com/prometheus/common/model" - - "github.com/prometheus/prometheus/storage/local/codable" -) - -const ( - fingerprintToMetricDir = "archived_fingerprint_to_metric" - fingerprintTimeRangeDir = "archived_fingerprint_to_timerange" - labelNameToLabelValuesDir = "labelname_to_labelvalues" - labelPairToFingerprintsDir = "labelpair_to_fingerprints" -) - -// LevelDB cache sizes, changeable via flags. -var ( - FingerprintMetricCacheSize = 10 * 1024 * 1024 - FingerprintTimeRangeCacheSize = 5 * 1024 * 1024 - LabelNameLabelValuesCacheSize = 10 * 1024 * 1024 - LabelPairFingerprintsCacheSize = 20 * 1024 * 1024 -) - -// FingerprintMetricMapping is an in-memory map of fingerprints to metrics. -type FingerprintMetricMapping map[model.Fingerprint]model.Metric - -// FingerprintMetricIndex models a database mapping fingerprints to metrics. -type FingerprintMetricIndex struct { - KeyValueStore -} - -// IndexBatch indexes a batch of mappings from fingerprints to metrics. -// -// This method is goroutine-safe, but note that no specific order of execution -// can be guaranteed (especially critical if IndexBatch and UnindexBatch are -// called concurrently for the same fingerprint). -func (i *FingerprintMetricIndex) IndexBatch(mapping FingerprintMetricMapping) error { - b := i.NewBatch() - - for fp, m := range mapping { - if err := b.Put(codable.Fingerprint(fp), codable.Metric(m)); err != nil { - return err - } - } - - return i.Commit(b) -} - -// UnindexBatch unindexes a batch of mappings from fingerprints to metrics. -// -// This method is goroutine-safe, but note that no specific order of execution -// can be guaranteed (especially critical if IndexBatch and UnindexBatch are -// called concurrently for the same fingerprint). -func (i *FingerprintMetricIndex) UnindexBatch(mapping FingerprintMetricMapping) error { - b := i.NewBatch() - - for fp := range mapping { - if err := b.Delete(codable.Fingerprint(fp)); err != nil { - return err - } - } - - return i.Commit(b) -} - -// Lookup looks up a metric by fingerprint. Looking up a non-existing -// fingerprint is not an error. In that case, (nil, false, nil) is returned. -// -// This method is goroutine-safe. -func (i *FingerprintMetricIndex) Lookup(fp model.Fingerprint) (metric model.Metric, ok bool, err error) { - ok, err = i.Get(codable.Fingerprint(fp), (*codable.Metric)(&metric)) - return -} - -// NewFingerprintMetricIndex returns a LevelDB-backed FingerprintMetricIndex -// ready to use. -func NewFingerprintMetricIndex(basePath string) (*FingerprintMetricIndex, error) { - fingerprintToMetricDB, err := NewLevelDB(LevelDBOptions{ - Path: filepath.Join(basePath, fingerprintToMetricDir), - CacheSizeBytes: FingerprintMetricCacheSize, - }) - if err != nil { - return nil, err - } - return &FingerprintMetricIndex{ - KeyValueStore: fingerprintToMetricDB, - }, nil -} - -// LabelNameLabelValuesMapping is an in-memory map of label names to -// label values. -type LabelNameLabelValuesMapping map[model.LabelName]codable.LabelValueSet - -// LabelNameLabelValuesIndex is a KeyValueStore that maps existing label names -// to all label values stored for that label name. -type LabelNameLabelValuesIndex struct { - KeyValueStore -} - -// IndexBatch adds a batch of label name to label values mappings to the -// index. A mapping of a label name to an empty slice of label values results in -// a deletion of that mapping from the index. -// -// While this method is fundamentally goroutine-safe, note that the order of -// execution for multiple batches executed concurrently is undefined. -func (i *LabelNameLabelValuesIndex) IndexBatch(b LabelNameLabelValuesMapping) error { - batch := i.NewBatch() - - for name, values := range b { - if len(values) == 0 { - if err := batch.Delete(codable.LabelName(name)); err != nil { - return err - } - } else { - if err := batch.Put(codable.LabelName(name), values); err != nil { - return err - } - } - } - - return i.Commit(batch) -} - -// Lookup looks up all label values for a given label name and returns them as -// model.LabelValues (which is a slice). Looking up a non-existing label -// name is not an error. In that case, (nil, false, nil) is returned. -// -// This method is goroutine-safe. -func (i *LabelNameLabelValuesIndex) Lookup(l model.LabelName) (values model.LabelValues, ok bool, err error) { - ok, err = i.Get(codable.LabelName(l), (*codable.LabelValues)(&values)) - return -} - -// LookupSet looks up all label values for a given label name and returns them -// as a set. Looking up a non-existing label name is not an error. In that case, -// (nil, false, nil) is returned. -// -// This method is goroutine-safe. -func (i *LabelNameLabelValuesIndex) LookupSet(l model.LabelName) (values map[model.LabelValue]struct{}, ok bool, err error) { - ok, err = i.Get(codable.LabelName(l), (*codable.LabelValueSet)(&values)) - if values == nil { - values = map[model.LabelValue]struct{}{} - } - return -} - -// NewLabelNameLabelValuesIndex returns a LevelDB-backed -// LabelNameLabelValuesIndex ready to use. -func NewLabelNameLabelValuesIndex(basePath string) (*LabelNameLabelValuesIndex, error) { - labelNameToLabelValuesDB, err := NewLevelDB(LevelDBOptions{ - Path: filepath.Join(basePath, labelNameToLabelValuesDir), - CacheSizeBytes: LabelNameLabelValuesCacheSize, - }) - if err != nil { - return nil, err - } - return &LabelNameLabelValuesIndex{ - KeyValueStore: labelNameToLabelValuesDB, - }, nil -} - -// DeleteLabelNameLabelValuesIndex deletes the LevelDB-backed -// LabelNameLabelValuesIndex. Use only for a not yet opened index. -func DeleteLabelNameLabelValuesIndex(basePath string) error { - return os.RemoveAll(path.Join(basePath, labelNameToLabelValuesDir)) -} - -// LabelPairFingerprintsMapping is an in-memory map of label pairs to -// fingerprints. -type LabelPairFingerprintsMapping map[model.LabelPair]codable.FingerprintSet - -// LabelPairFingerprintIndex is a KeyValueStore that maps existing label pairs -// to the fingerprints of all metrics containing those label pairs. -type LabelPairFingerprintIndex struct { - KeyValueStore -} - -// IndexBatch indexes a batch of mappings from label pairs to fingerprints. A -// mapping to an empty slice of fingerprints results in deletion of that mapping -// from the index. -// -// While this method is fundamentally goroutine-safe, note that the order of -// execution for multiple batches executed concurrently is undefined. -func (i *LabelPairFingerprintIndex) IndexBatch(m LabelPairFingerprintsMapping) (err error) { - batch := i.NewBatch() - - for pair, fps := range m { - if len(fps) == 0 { - err = batch.Delete(codable.LabelPair(pair)) - } else { - err = batch.Put(codable.LabelPair(pair), fps) - } - - if err != nil { - return err - } - } - - return i.Commit(batch) -} - -// Lookup looks up all fingerprints for a given label pair. Looking up a -// non-existing label pair is not an error. In that case, (nil, false, nil) is -// returned. -// -// This method is goroutine-safe. -func (i *LabelPairFingerprintIndex) Lookup(p model.LabelPair) (fps model.Fingerprints, ok bool, err error) { - ok, err = i.Get((codable.LabelPair)(p), (*codable.Fingerprints)(&fps)) - return -} - -// LookupSet looks up all fingerprints for a given label pair. Looking up a -// non-existing label pair is not an error. In that case, (nil, false, nil) is -// returned. -// -// This method is goroutine-safe. -func (i *LabelPairFingerprintIndex) LookupSet(p model.LabelPair) (fps map[model.Fingerprint]struct{}, ok bool, err error) { - ok, err = i.Get((codable.LabelPair)(p), (*codable.FingerprintSet)(&fps)) - if fps == nil { - fps = map[model.Fingerprint]struct{}{} - } - return -} - -// NewLabelPairFingerprintIndex returns a LevelDB-backed -// LabelPairFingerprintIndex ready to use. -func NewLabelPairFingerprintIndex(basePath string) (*LabelPairFingerprintIndex, error) { - labelPairToFingerprintsDB, err := NewLevelDB(LevelDBOptions{ - Path: filepath.Join(basePath, labelPairToFingerprintsDir), - CacheSizeBytes: LabelPairFingerprintsCacheSize, - }) - if err != nil { - return nil, err - } - return &LabelPairFingerprintIndex{ - KeyValueStore: labelPairToFingerprintsDB, - }, nil -} - -// DeleteLabelPairFingerprintIndex deletes the LevelDB-backed -// LabelPairFingerprintIndex. Use only for a not yet opened index. -func DeleteLabelPairFingerprintIndex(basePath string) error { - return os.RemoveAll(path.Join(basePath, labelPairToFingerprintsDir)) -} - -// FingerprintTimeRangeIndex models a database tracking the time ranges -// of metrics by their fingerprints. -type FingerprintTimeRangeIndex struct { - KeyValueStore -} - -// Lookup returns the time range for the given fingerprint. Looking up a -// non-existing fingerprint is not an error. In that case, (0, 0, false, nil) is -// returned. -// -// This method is goroutine-safe. -func (i *FingerprintTimeRangeIndex) Lookup(fp model.Fingerprint) (firstTime, lastTime model.Time, ok bool, err error) { - var tr codable.TimeRange - ok, err = i.Get(codable.Fingerprint(fp), &tr) - return tr.First, tr.Last, ok, err -} - -// NewFingerprintTimeRangeIndex returns a LevelDB-backed -// FingerprintTimeRangeIndex ready to use. -func NewFingerprintTimeRangeIndex(basePath string) (*FingerprintTimeRangeIndex, error) { - fingerprintTimeRangeDB, err := NewLevelDB(LevelDBOptions{ - Path: filepath.Join(basePath, fingerprintTimeRangeDir), - CacheSizeBytes: FingerprintTimeRangeCacheSize, - }) - if err != nil { - return nil, err - } - return &FingerprintTimeRangeIndex{ - KeyValueStore: fingerprintTimeRangeDB, - }, nil -} diff --git a/storage/local/index/interface.go b/storage/local/index/interface.go deleted file mode 100644 index 40080c7f35..0000000000 --- a/storage/local/index/interface.go +++ /dev/null @@ -1,61 +0,0 @@ -// Copyright 2014 The Prometheus Authors -// Licensed under the Apache License, Version 2.0 (the "License"); -// you may not use this file except in compliance with the License. -// You may obtain a copy of the License at -// -// http://www.apache.org/licenses/LICENSE-2.0 -// -// Unless required by applicable law or agreed to in writing, software -// distributed under the License is distributed on an "AS IS" BASIS, -// WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. -// See the License for the specific language governing permissions and -// limitations under the License. - -package index - -import "encoding" - -// KeyValueStore persists key/value pairs. Implementations must be fundamentally -// goroutine-safe. However, it is the caller's responsibility that keys and -// values can be safely marshaled and unmarshaled (via the MarshalBinary and -// UnmarshalBinary methods of the keys and values). For example, if you call the -// Put method of a KeyValueStore implementation, but the key or the value are -// modified concurrently while being marshaled into its binary representation, -// you obviously have a problem. Methods of KeyValueStore return only after -// (un)marshaling is complete. -type KeyValueStore interface { - Put(key, value encoding.BinaryMarshaler) error - // Get unmarshals the result into value. It returns false if no entry - // could be found for key. If value is nil, Get behaves like Has. - Get(key encoding.BinaryMarshaler, value encoding.BinaryUnmarshaler) (bool, error) - Has(key encoding.BinaryMarshaler) (bool, error) - // Delete returns (false, nil) if key does not exist. - Delete(key encoding.BinaryMarshaler) (bool, error) - - NewBatch() Batch - Commit(b Batch) error - - // ForEach iterates through the complete KeyValueStore and calls the - // supplied function for each mapping. - ForEach(func(kv KeyValueAccessor) error) error - - Close() error -} - -// KeyValueAccessor allows access to the key and value of an entry in a -// KeyValueStore. -type KeyValueAccessor interface { - Key(encoding.BinaryUnmarshaler) error - Value(encoding.BinaryUnmarshaler) error -} - -// Batch allows KeyValueStore mutations to be pooled and committed together. An -// implementation does not have to be goroutine-safe. Never modify a Batch -// concurrently or commit the same batch multiple times concurrently. Marshaling -// of keys and values is guaranteed to be complete when the Put or Delete methods -// have returned. -type Batch interface { - Put(key, value encoding.BinaryMarshaler) error - Delete(key encoding.BinaryMarshaler) error - Reset() -} diff --git a/storage/local/index/leveldb.go b/storage/local/index/leveldb.go deleted file mode 100644 index c4c46421cd..0000000000 --- a/storage/local/index/leveldb.go +++ /dev/null @@ -1,210 +0,0 @@ -// Copyright 2014 The Prometheus Authors -// Licensed under the Apache License, Version 2.0 (the "License"); -// you may not use this file except in compliance with the License. -// You may obtain a copy of the License at -// -// http://www.apache.org/licenses/LICENSE-2.0 -// -// Unless required by applicable law or agreed to in writing, software -// distributed under the License is distributed on an "AS IS" BASIS, -// WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. -// See the License for the specific language governing permissions and -// limitations under the License. - -package index - -import ( - "encoding" - - "github.com/syndtr/goleveldb/leveldb" - leveldb_filter "github.com/syndtr/goleveldb/leveldb/filter" - leveldb_iterator "github.com/syndtr/goleveldb/leveldb/iterator" - leveldb_opt "github.com/syndtr/goleveldb/leveldb/opt" - leveldb_util "github.com/syndtr/goleveldb/leveldb/util" -) - -var ( - keyspace = &leveldb_util.Range{ - Start: nil, - Limit: nil, - } - - iteratorOpts = &leveldb_opt.ReadOptions{ - DontFillCache: true, - } -) - -// LevelDB is a LevelDB-backed sorted KeyValueStore. -type LevelDB struct { - storage *leveldb.DB - readOpts *leveldb_opt.ReadOptions - writeOpts *leveldb_opt.WriteOptions -} - -// LevelDBOptions provides options for a LevelDB. -type LevelDBOptions struct { - Path string // Base path to store files. - CacheSizeBytes int -} - -// NewLevelDB returns a newly allocated LevelDB-backed KeyValueStore ready to -// use. -func NewLevelDB(o LevelDBOptions) (KeyValueStore, error) { - options := &leveldb_opt.Options{ - BlockCacheCapacity: o.CacheSizeBytes, - Filter: leveldb_filter.NewBloomFilter(10), - } - - storage, err := leveldb.OpenFile(o.Path, options) - if err != nil { - return nil, err - } - - return &LevelDB{ - storage: storage, - readOpts: &leveldb_opt.ReadOptions{}, - writeOpts: &leveldb_opt.WriteOptions{}, - }, nil -} - -// NewBatch implements KeyValueStore. -func (l *LevelDB) NewBatch() Batch { - return &LevelDBBatch{ - batch: &leveldb.Batch{}, - } -} - -// Close implements KeyValueStore. -func (l *LevelDB) Close() error { - return l.storage.Close() -} - -// Get implements KeyValueStore. -func (l *LevelDB) Get(key encoding.BinaryMarshaler, value encoding.BinaryUnmarshaler) (bool, error) { - k, err := key.MarshalBinary() - if err != nil { - return false, err - } - raw, err := l.storage.Get(k, l.readOpts) - if err == leveldb.ErrNotFound { - return false, nil - } - if err != nil { - return false, err - } - if value == nil { - return true, nil - } - return true, value.UnmarshalBinary(raw) -} - -// Has implements KeyValueStore. -func (l *LevelDB) Has(key encoding.BinaryMarshaler) (has bool, err error) { - return l.Get(key, nil) -} - -// Delete implements KeyValueStore. -func (l *LevelDB) Delete(key encoding.BinaryMarshaler) (bool, error) { - k, err := key.MarshalBinary() - if err != nil { - return false, err - } - // Note that Delete returns nil if k does not exist. So we have to test - // for existence with Has first. - if has, err := l.storage.Has(k, l.readOpts); !has || err != nil { - return false, err - } - if err = l.storage.Delete(k, l.writeOpts); err != nil { - return false, err - } - return true, nil -} - -// Put implements KeyValueStore. -func (l *LevelDB) Put(key, value encoding.BinaryMarshaler) error { - k, err := key.MarshalBinary() - if err != nil { - return err - } - v, err := value.MarshalBinary() - if err != nil { - return err - } - return l.storage.Put(k, v, l.writeOpts) -} - -// Commit implements KeyValueStore. -func (l *LevelDB) Commit(b Batch) error { - return l.storage.Write(b.(*LevelDBBatch).batch, l.writeOpts) -} - -// ForEach implements KeyValueStore. -func (l *LevelDB) ForEach(cb func(kv KeyValueAccessor) error) error { - snap, err := l.storage.GetSnapshot() - if err != nil { - return err - } - defer snap.Release() - - iter := snap.NewIterator(keyspace, iteratorOpts) - - kv := &levelDBKeyValueAccessor{it: iter} - - for valid := iter.First(); valid; valid = iter.Next() { - if err = iter.Error(); err != nil { - return err - } - - if err := cb(kv); err != nil { - return err - } - } - return nil -} - -// LevelDBBatch is a Batch implementation for LevelDB. -type LevelDBBatch struct { - batch *leveldb.Batch -} - -// Put implements Batch. -func (b *LevelDBBatch) Put(key, value encoding.BinaryMarshaler) error { - k, err := key.MarshalBinary() - if err != nil { - return err - } - v, err := value.MarshalBinary() - if err != nil { - return err - } - b.batch.Put(k, v) - return nil -} - -// Delete implements Batch. -func (b *LevelDBBatch) Delete(key encoding.BinaryMarshaler) error { - k, err := key.MarshalBinary() - if err != nil { - return err - } - b.batch.Delete(k) - return nil -} - -// Reset implements Batch. -func (b *LevelDBBatch) Reset() { - b.batch.Reset() -} - -// levelDBKeyValueAccessor implements KeyValueAccessor. -type levelDBKeyValueAccessor struct { - it leveldb_iterator.Iterator -} - -func (i *levelDBKeyValueAccessor) Key(key encoding.BinaryUnmarshaler) error { - return key.UnmarshalBinary(i.it.Key()) -} - -func (i *levelDBKeyValueAccessor) Value(value encoding.BinaryUnmarshaler) error { - return value.UnmarshalBinary(i.it.Value()) -} diff --git a/storage/local/instrumentation.go b/storage/local/instrumentation.go deleted file mode 100644 index 479e138215..0000000000 --- a/storage/local/instrumentation.go +++ /dev/null @@ -1,46 +0,0 @@ -// Copyright 2014 The Prometheus Authors -// Licensed under the Apache License, Version 2.0 (the "License"); -// you may not use this file except in compliance with the License. -// You may obtain a copy of the License at -// -// http://www.apache.org/licenses/LICENSE-2.0 -// -// Unless required by applicable law or agreed to in writing, software -// distributed under the License is distributed on an "AS IS" BASIS, -// WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. -// See the License for the specific language governing permissions and -// limitations under the License. - -package local - -const ( - namespace = "prometheus" - subsystem = "local_storage" - - opTypeLabel = "type" - - // Op-types for seriesOps. - create = "create" - archive = "archive" - unarchive = "unarchive" - memoryPurge = "purge_from_memory" - archivePurge = "purge_from_archive" - requestedPurge = "purge_on_request" - memoryMaintenance = "maintenance_in_memory" - archiveMaintenance = "maintenance_in_archive" - completedQurantine = "quarantine_completed" - droppedQuarantine = "quarantine_dropped" - failedQuarantine = "quarantine_failed" - - seriesLocationLabel = "location" - - // Maintenance types for maintainSeriesDuration. - maintainInMemory = "memory" - maintainArchived = "archived" - - discardReasonLabel = "reason" - - // Reasons to discard samples. - outOfOrderTimestamp = "timestamp_out_of_order" - duplicateSample = "multiple_values_for_timestamp" -) diff --git a/storage/local/interface.go b/storage/local/interface.go index 3f1fda7130..851f0c062f 100644 --- a/storage/local/interface.go +++ b/storage/local/interface.go @@ -14,6 +14,7 @@ package local import ( + "errors" "time" "github.com/prometheus/common/model" @@ -21,8 +22,18 @@ import ( "github.com/prometheus/prometheus/storage" "github.com/prometheus/prometheus/storage/metric" + "github.com/prometheus/prometheus/util/testutil" ) +var ( + ErrOutOfOrderSample = errors.New("out of order sample") + ErrDuplicateSampleForTimestamp = errors.New("duplicate sample for timestamp") +) + +func NewTestStorage(t testutil.T, enc byte) (Storage, testutil.Closer) { + return nil, nil +} + // Storage ingests and manages samples, along with various indexes. All methods // are goroutine-safe. Storage implements storage.SampleAppender. type Storage interface { diff --git a/storage/local/locker.go b/storage/local/locker.go deleted file mode 100644 index 85effcdbec..0000000000 --- a/storage/local/locker.go +++ /dev/null @@ -1,79 +0,0 @@ -// Copyright 2016 The Prometheus Authors -// Licensed under the Apache License, Version 2.0 (the "License"); -// you may not use this file except in compliance with the License. -// You may obtain a copy of the License at -// -// http://www.apache.org/licenses/LICENSE-2.0 -// -// Unless required by applicable law or agreed to in writing, software -// distributed under the License is distributed on an "AS IS" BASIS, -// WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. -// See the License for the specific language governing permissions and -// limitations under the License. - -package local - -import ( - "sync" - "unsafe" - - "github.com/prometheus/common/model" -) - -const ( - cacheLineSize = 64 -) - -// Avoid false sharing when using array of mutexes. -type paddedMutex struct { - sync.Mutex - pad [cacheLineSize - unsafe.Sizeof(sync.Mutex{})]byte -} - -// fingerprintLocker allows locking individual fingerprints. To limit the number -// of mutexes needed for that, only a fixed number of mutexes are -// allocated. Fingerprints to be locked are assigned to those pre-allocated -// mutexes by their value. Collisions are not detected. If two fingerprints get -// assigned to the same mutex, only one of them can be locked at the same -// time. As long as the number of pre-allocated mutexes is much larger than the -// number of goroutines requiring a fingerprint lock concurrently, the loss in -// efficiency is small. However, a goroutine must never lock more than one -// fingerprint at the same time. (In that case a collision would try to acquire -// the same mutex twice). -type fingerprintLocker struct { - fpMtxs []paddedMutex - numFpMtxs uint -} - -// newFingerprintLocker returns a new fingerprintLocker ready for use. At least -// 1024 preallocated mutexes are used, even if preallocatedMutexes is lower. -func newFingerprintLocker(preallocatedMutexes int) *fingerprintLocker { - if preallocatedMutexes < 1024 { - preallocatedMutexes = 1024 - } - return &fingerprintLocker{ - make([]paddedMutex, preallocatedMutexes), - uint(preallocatedMutexes), - } -} - -// Lock locks the given fingerprint. -func (l *fingerprintLocker) Lock(fp model.Fingerprint) { - l.fpMtxs[hashFP(fp)%l.numFpMtxs].Lock() -} - -// Unlock unlocks the given fingerprint. -func (l *fingerprintLocker) Unlock(fp model.Fingerprint) { - l.fpMtxs[hashFP(fp)%l.numFpMtxs].Unlock() -} - -// hashFP simply moves entropy from the most significant 48 bits of the -// fingerprint into the least significant 16 bits (by XORing) so that a simple -// MOD on the result can be used to pick a mutex while still making use of -// changes in more significant bits of the fingerprint. (The fast fingerprinting -// function we use is prone to only change a few bits for similar metrics. We -// really want to make use of every change in the fingerprint to vary mutex -// selection.) -func hashFP(fp model.Fingerprint) uint { - return uint(fp ^ (fp >> 32) ^ (fp >> 16)) -} diff --git a/storage/local/locker_test.go b/storage/local/locker_test.go deleted file mode 100644 index 8450afcbbf..0000000000 --- a/storage/local/locker_test.go +++ /dev/null @@ -1,58 +0,0 @@ -// Copyright 2016 The Prometheus Authors -// Licensed under the Apache License, Version 2.0 (the "License"); -// you may not use this file except in compliance with the License. -// You may obtain a copy of the License at -// -// http://www.apache.org/licenses/LICENSE-2.0 -// -// Unless required by applicable law or agreed to in writing, software -// distributed under the License is distributed on an "AS IS" BASIS, -// WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. -// See the License for the specific language governing permissions and -// limitations under the License. - -package local - -import ( - "sync" - "testing" - - "github.com/prometheus/common/model" -) - -func BenchmarkFingerprintLockerParallel(b *testing.B) { - numGoroutines := 10 - numFingerprints := 10 - numLockOps := b.N - locker := newFingerprintLocker(100) - - wg := sync.WaitGroup{} - b.ResetTimer() - for i := 0; i < numGoroutines; i++ { - wg.Add(1) - go func(i int) { - for j := 0; j < numLockOps; j++ { - fp1 := model.Fingerprint(j % numFingerprints) - fp2 := model.Fingerprint(j%numFingerprints + numFingerprints) - locker.Lock(fp1) - locker.Lock(fp2) - locker.Unlock(fp2) - locker.Unlock(fp1) - } - wg.Done() - }(i) - } - wg.Wait() -} - -func BenchmarkFingerprintLockerSerial(b *testing.B) { - numFingerprints := 10 - locker := newFingerprintLocker(100) - - b.ResetTimer() - for i := 0; i < b.N; i++ { - fp := model.Fingerprint(i % numFingerprints) - locker.Lock(fp) - locker.Unlock(fp) - } -} diff --git a/storage/local/mapper.go b/storage/local/mapper.go deleted file mode 100644 index cbd58d6904..0000000000 --- a/storage/local/mapper.go +++ /dev/null @@ -1,218 +0,0 @@ -// Copyright 2016 The Prometheus Authors -// Licensed under the Apache License, Version 2.0 (the "License"); -// you may not use this file except in compliance with the License. -// You may obtain a copy of the License at -// -// http://www.apache.org/licenses/LICENSE-2.0 -// -// Unless required by applicable law or agreed to in writing, software -// distributed under the License is distributed on an "AS IS" BASIS, -// WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. -// See the License for the specific language governing permissions and -// limitations under the License. - -package local - -import ( - "fmt" - "sort" - "strings" - "sync" - "sync/atomic" - - "github.com/prometheus/client_golang/prometheus" - "github.com/prometheus/common/log" - - "github.com/prometheus/common/model" -) - -const maxMappedFP = 1 << 20 // About 1M fingerprints reserved for mapping. - -var separatorString = string([]byte{model.SeparatorByte}) - -// fpMappings maps original fingerprints to a map of string representations of -// metrics to the truly unique fingerprint. -type fpMappings map[model.Fingerprint]map[string]model.Fingerprint - -// fpMapper is used to map fingerprints in order to work around fingerprint -// collisions. -type fpMapper struct { - // highestMappedFP has to be aligned for atomic operations. - highestMappedFP model.Fingerprint - - mtx sync.RWMutex // Protects mappings. - mappings fpMappings - - fpToSeries *seriesMap - p *persistence - - mappingsCounter prometheus.Counter -} - -// newFPMapper loads the collision map from the persistence and -// returns an fpMapper ready to use. -func newFPMapper(fpToSeries *seriesMap, p *persistence) (*fpMapper, error) { - m := &fpMapper{ - fpToSeries: fpToSeries, - p: p, - mappingsCounter: prometheus.NewCounter(prometheus.CounterOpts{ - Namespace: namespace, - Subsystem: subsystem, - Name: "fingerprint_mappings_total", - Help: "The total number of fingerprints being mapped to avoid collisions.", - }), - } - mappings, nextFP, err := p.loadFPMappings() - if err != nil { - return nil, err - } - m.mappings = mappings - m.mappingsCounter.Set(float64(len(m.mappings))) - m.highestMappedFP = nextFP - return m, nil -} - -// checkpoint persists the current mappings. The caller has to ensure that the -// provided mappings are not changed concurrently. This method is only called -// upon shutdown, when no samples are ingested anymore. -func (m *fpMapper) checkpoint() error { - return m.p.checkpointFPMappings(m.mappings) -} - -// mapFP takes a raw fingerprint (as returned by Metrics.FastFingerprint) and -// returns a truly unique fingerprint. The caller must have locked the raw -// fingerprint. -// -// If an error is encountered, it is returned together with the unchanged raw -// fingerprint. -func (m *fpMapper) mapFP(fp model.Fingerprint, metric model.Metric) model.Fingerprint { - // First check if we are in the reserved FP space, in which case this is - // automatically a collision that has to be mapped. - if fp <= maxMappedFP { - return m.maybeAddMapping(fp, metric) - } - - // Then check the most likely case: This fp belongs to a series that is - // already in memory. - s, ok := m.fpToSeries.get(fp) - if ok { - // FP exists in memory, but is it for the same metric? - if metric.Equal(s.metric) { - // Yupp. We are done. - return fp - } - // Collision detected! - return m.maybeAddMapping(fp, metric) - } - // Metric is not in memory. Before doing the expensive archive lookup, - // check if we have a mapping for this metric in place already. - m.mtx.RLock() - mappedFPs, fpAlreadyMapped := m.mappings[fp] - m.mtx.RUnlock() - if fpAlreadyMapped { - // We indeed have mapped fp historically. - ms := metricToUniqueString(metric) - // fp is locked by the caller, so no further locking of - // 'collisions' required (it is specific to fp). - mappedFP, ok := mappedFPs[ms] - if ok { - // Historical mapping found, return the mapped FP. - return mappedFP - } - } - // If we are here, FP does not exist in memory and is either not mapped - // at all, or existing mappings for FP are not for m. Check if we have - // something for FP in the archive. - archivedMetric, err := m.p.archivedMetric(fp) - if err != nil || archivedMetric == nil { - // Either the archive lookup has returend an error, or fp does - // not exist in the archive. In the former case, the storage has - // been marked as dirty already. We just carry on for as long as - // it goes, assuming that fp does not exist. In either case, - // since now we know (or assume) now that fp does not exist, - // neither in memory nor in archive, we can safely keep it - // unmapped. - return fp - } - // FP exists in archive, but is it for the same metric? - if metric.Equal(archivedMetric) { - // Yupp. We are done. - return fp - } - // Collision detected! - return m.maybeAddMapping(fp, metric) -} - -// maybeAddMapping is only used internally. It takes a detected collision and -// adds it to the collisions map if not yet there. In any case, it returns the -// truly unique fingerprint for the colliding metric. -func (m *fpMapper) maybeAddMapping( - fp model.Fingerprint, - collidingMetric model.Metric, -) model.Fingerprint { - ms := metricToUniqueString(collidingMetric) - m.mtx.RLock() - mappedFPs, ok := m.mappings[fp] - m.mtx.RUnlock() - if ok { - // fp is locked by the caller, so no further locking required. - mappedFP, ok := mappedFPs[ms] - if ok { - return mappedFP // Existing mapping. - } - // A new mapping has to be created. - mappedFP = m.nextMappedFP() - mappedFPs[ms] = mappedFP - log.Infof( - "Collision detected for fingerprint %v, metric %v, mapping to new fingerprint %v.", - fp, collidingMetric, mappedFP, - ) - return mappedFP - } - // This is the first collision for fp. - mappedFP := m.nextMappedFP() - mappedFPs = map[string]model.Fingerprint{ms: mappedFP} - m.mtx.Lock() - m.mappings[fp] = mappedFPs - m.mappingsCounter.Inc() - m.mtx.Unlock() - log.Infof( - "Collision detected for fingerprint %v, metric %v, mapping to new fingerprint %v.", - fp, collidingMetric, mappedFP, - ) - return mappedFP -} - -func (m *fpMapper) nextMappedFP() model.Fingerprint { - mappedFP := model.Fingerprint(atomic.AddUint64((*uint64)(&m.highestMappedFP), 1)) - if mappedFP > maxMappedFP { - panic(fmt.Errorf("more than %v fingerprints mapped in collision detection", maxMappedFP)) - } - return mappedFP -} - -// Describe implements prometheus.Collector. -func (m *fpMapper) Describe(ch chan<- *prometheus.Desc) { - ch <- m.mappingsCounter.Desc() -} - -// Collect implements prometheus.Collector. -func (m *fpMapper) Collect(ch chan<- prometheus.Metric) { - ch <- m.mappingsCounter -} - -// metricToUniqueString turns a metric into a string in a reproducible and -// unique way, i.e. the same metric will always create the same string, and -// different metrics will always create different strings. In a way, it is the -// "ideal" fingerprint function, only that it is more expensive than the -// FastFingerprint function, and its result is not suitable as a key for maps -// and indexes as it might become really large, causing a lot of hashing effort -// in maps and a lot of storage overhead in indexes. -func metricToUniqueString(m model.Metric) string { - parts := make([]string, 0, len(m)) - for ln, lv := range m { - parts = append(parts, string(ln)+separatorString+string(lv)) - } - sort.Strings(parts) - return strings.Join(parts, separatorString) -} diff --git a/storage/local/mapper_test.go b/storage/local/mapper_test.go deleted file mode 100644 index 4d036ebb2d..0000000000 --- a/storage/local/mapper_test.go +++ /dev/null @@ -1,292 +0,0 @@ -// Copyright 2016 The Prometheus Authors -// Licensed under the Apache License, Version 2.0 (the "License"); -// you may not use this file except in compliance with the License. -// You may obtain a copy of the License at -// -// http://www.apache.org/licenses/LICENSE-2.0 -// -// Unless required by applicable law or agreed to in writing, software -// distributed under the License is distributed on an "AS IS" BASIS, -// WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. -// See the License for the specific language governing permissions and -// limitations under the License. - -package local - -import ( - "testing" - - "github.com/prometheus/common/model" -) - -var ( - // cm11, cm12, cm13 are colliding with fp1. - // cm21, cm22 are colliding with fp2. - // cm31, cm32 are colliding with fp3, which is below maxMappedFP. - // Note that fingerprints are set and not actually calculated. - // The collision detection is independent from the actually used - // fingerprinting algorithm. - fp1 = model.Fingerprint(maxMappedFP + 1) - fp2 = model.Fingerprint(maxMappedFP + 2) - fp3 = model.Fingerprint(1) - cm11 = model.Metric{ - "foo": "bar", - "dings": "bumms", - } - cm12 = model.Metric{ - "bar": "foo", - } - cm13 = model.Metric{ - "foo": "bar", - } - cm21 = model.Metric{ - "foo": "bumms", - "dings": "bar", - } - cm22 = model.Metric{ - "dings": "foo", - "bar": "bumms", - } - cm31 = model.Metric{ - "bumms": "dings", - } - cm32 = model.Metric{ - "bumms": "dings", - "bar": "foo", - } -) - -func TestFPMapper(t *testing.T) { - sm := newSeriesMap() - - p, closer := newTestPersistence(t, 1) - defer closer.Close() - - mapper, err := newFPMapper(sm, p) - - // Everything is empty, resolving a FP should do nothing. - gotFP := mapper.mapFP(fp1, cm11) - if wantFP := fp1; gotFP != wantFP { - t.Errorf("got fingerprint %v, want fingerprint %v", gotFP, wantFP) - } - gotFP = mapper.mapFP(fp1, cm12) - if wantFP := fp1; gotFP != wantFP { - t.Errorf("got fingerprint %v, want fingerprint %v", gotFP, wantFP) - } - - // cm11 is in sm. Adding cm11 should do nothing. Mapping cm12 should resolve - // the collision. - sm.put(fp1, &memorySeries{metric: cm11}) - gotFP = mapper.mapFP(fp1, cm11) - if wantFP := fp1; gotFP != wantFP { - t.Errorf("got fingerprint %v, want fingerprint %v", gotFP, wantFP) - } - gotFP = mapper.mapFP(fp1, cm12) - if wantFP := model.Fingerprint(1); gotFP != wantFP { - t.Errorf("got fingerprint %v, want fingerprint %v", gotFP, wantFP) - } - - // The mapped cm12 is added to sm, too. That should not change the outcome. - sm.put(model.Fingerprint(1), &memorySeries{metric: cm12}) - gotFP = mapper.mapFP(fp1, cm11) - if wantFP := fp1; gotFP != wantFP { - t.Errorf("got fingerprint %v, want fingerprint %v", gotFP, wantFP) - } - gotFP = mapper.mapFP(fp1, cm12) - if wantFP := model.Fingerprint(1); gotFP != wantFP { - t.Errorf("got fingerprint %v, want fingerprint %v", gotFP, wantFP) - } - - // Now map cm13, should reproducibly result in the next mapped FP. - gotFP = mapper.mapFP(fp1, cm13) - if wantFP := model.Fingerprint(2); gotFP != wantFP { - t.Errorf("got fingerprint %v, want fingerprint %v", gotFP, wantFP) - } - gotFP = mapper.mapFP(fp1, cm13) - if wantFP := model.Fingerprint(2); gotFP != wantFP { - t.Errorf("got fingerprint %v, want fingerprint %v", gotFP, wantFP) - } - - // Add cm13 to sm. Should not change anything. - sm.put(model.Fingerprint(2), &memorySeries{metric: cm13}) - gotFP = mapper.mapFP(fp1, cm11) - if wantFP := fp1; gotFP != wantFP { - t.Errorf("got fingerprint %v, want fingerprint %v", gotFP, wantFP) - } - gotFP = mapper.mapFP(fp1, cm12) - if wantFP := model.Fingerprint(1); gotFP != wantFP { - t.Errorf("got fingerprint %v, want fingerprint %v", gotFP, wantFP) - } - gotFP = mapper.mapFP(fp1, cm13) - if wantFP := model.Fingerprint(2); gotFP != wantFP { - t.Errorf("got fingerprint %v, want fingerprint %v", gotFP, wantFP) - } - - // Now add cm21 and cm22 in the same way, checking the mapped FPs. - gotFP = mapper.mapFP(fp2, cm21) - if wantFP := fp2; gotFP != wantFP { - t.Errorf("got fingerprint %v, want fingerprint %v", gotFP, wantFP) - } - sm.put(fp2, &memorySeries{metric: cm21}) - gotFP = mapper.mapFP(fp2, cm21) - if wantFP := fp2; gotFP != wantFP { - t.Errorf("got fingerprint %v, want fingerprint %v", gotFP, wantFP) - } - gotFP = mapper.mapFP(fp2, cm22) - if wantFP := model.Fingerprint(3); gotFP != wantFP { - t.Errorf("got fingerprint %v, want fingerprint %v", gotFP, wantFP) - } - sm.put(model.Fingerprint(3), &memorySeries{metric: cm22}) - gotFP = mapper.mapFP(fp2, cm21) - if wantFP := fp2; gotFP != wantFP { - t.Errorf("got fingerprint %v, want fingerprint %v", gotFP, wantFP) - } - gotFP = mapper.mapFP(fp2, cm22) - if wantFP := model.Fingerprint(3); gotFP != wantFP { - t.Errorf("got fingerprint %v, want fingerprint %v", gotFP, wantFP) - } - - // Map cm31, resulting in a mapping straight away. - gotFP = mapper.mapFP(fp3, cm31) - if wantFP := model.Fingerprint(4); gotFP != wantFP { - t.Errorf("got fingerprint %v, want fingerprint %v", gotFP, wantFP) - } - sm.put(model.Fingerprint(4), &memorySeries{metric: cm31}) - - // Map cm32, which is now mapped for two reasons... - gotFP = mapper.mapFP(fp3, cm32) - if wantFP := model.Fingerprint(5); gotFP != wantFP { - t.Errorf("got fingerprint %v, want fingerprint %v", gotFP, wantFP) - } - sm.put(model.Fingerprint(5), &memorySeries{metric: cm32}) - - // Now check ALL the mappings, just to be sure. - gotFP = mapper.mapFP(fp1, cm11) - if wantFP := fp1; gotFP != wantFP { - t.Errorf("got fingerprint %v, want fingerprint %v", gotFP, wantFP) - } - gotFP = mapper.mapFP(fp1, cm12) - if wantFP := model.Fingerprint(1); gotFP != wantFP { - t.Errorf("got fingerprint %v, want fingerprint %v", gotFP, wantFP) - } - gotFP = mapper.mapFP(fp1, cm13) - if wantFP := model.Fingerprint(2); gotFP != wantFP { - t.Errorf("got fingerprint %v, want fingerprint %v", gotFP, wantFP) - } - gotFP = mapper.mapFP(fp2, cm21) - if wantFP := fp2; gotFP != wantFP { - t.Errorf("got fingerprint %v, want fingerprint %v", gotFP, wantFP) - } - gotFP = mapper.mapFP(fp2, cm22) - if wantFP := model.Fingerprint(3); gotFP != wantFP { - t.Errorf("got fingerprint %v, want fingerprint %v", gotFP, wantFP) - } - gotFP = mapper.mapFP(fp3, cm31) - if wantFP := model.Fingerprint(4); gotFP != wantFP { - t.Errorf("got fingerprint %v, want fingerprint %v", gotFP, wantFP) - } - gotFP = mapper.mapFP(fp3, cm32) - if wantFP := model.Fingerprint(5); gotFP != wantFP { - t.Errorf("got fingerprint %v, want fingerprint %v", gotFP, wantFP) - } - - // Remove all the fingerprints from sm, which should change nothing, as - // the existing mappings stay and should be detected. - sm.del(fp1) - sm.del(fp2) - sm.del(fp3) - gotFP = mapper.mapFP(fp1, cm11) - if wantFP := fp1; gotFP != wantFP { - t.Errorf("got fingerprint %v, want fingerprint %v", gotFP, wantFP) - } - gotFP = mapper.mapFP(fp1, cm12) - if wantFP := model.Fingerprint(1); gotFP != wantFP { - t.Errorf("got fingerprint %v, want fingerprint %v", gotFP, wantFP) - } - gotFP = mapper.mapFP(fp1, cm13) - if wantFP := model.Fingerprint(2); gotFP != wantFP { - t.Errorf("got fingerprint %v, want fingerprint %v", gotFP, wantFP) - } - gotFP = mapper.mapFP(fp2, cm21) - if wantFP := fp2; gotFP != wantFP { - t.Errorf("got fingerprint %v, want fingerprint %v", gotFP, wantFP) - } - gotFP = mapper.mapFP(fp2, cm22) - if wantFP := model.Fingerprint(3); gotFP != wantFP { - t.Errorf("got fingerprint %v, want fingerprint %v", gotFP, wantFP) - } - gotFP = mapper.mapFP(fp3, cm31) - if wantFP := model.Fingerprint(4); gotFP != wantFP { - t.Errorf("got fingerprint %v, want fingerprint %v", gotFP, wantFP) - } - gotFP = mapper.mapFP(fp3, cm32) - if wantFP := model.Fingerprint(5); gotFP != wantFP { - t.Errorf("got fingerprint %v, want fingerprint %v", gotFP, wantFP) - } - err = mapper.checkpoint() - if err != nil { - t.Fatal(err) - } - - // Load the mapper anew from disk and then check all the mappings again - // to make sure all changes have made it to disk. - mapper, err = newFPMapper(sm, p) - if err != nil { - t.Fatal(err) - } - - gotFP = mapper.mapFP(fp1, cm11) - if wantFP := fp1; gotFP != wantFP { - t.Errorf("got fingerprint %v, want fingerprint %v", gotFP, wantFP) - } - gotFP = mapper.mapFP(fp1, cm12) - if wantFP := model.Fingerprint(1); gotFP != wantFP { - t.Errorf("got fingerprint %v, want fingerprint %v", gotFP, wantFP) - } - gotFP = mapper.mapFP(fp1, cm13) - if wantFP := model.Fingerprint(2); gotFP != wantFP { - t.Errorf("got fingerprint %v, want fingerprint %v", gotFP, wantFP) - } - gotFP = mapper.mapFP(fp2, cm21) - if wantFP := fp2; gotFP != wantFP { - t.Errorf("got fingerprint %v, want fingerprint %v", gotFP, wantFP) - } - gotFP = mapper.mapFP(fp2, cm22) - if wantFP := model.Fingerprint(3); gotFP != wantFP { - t.Errorf("got fingerprint %v, want fingerprint %v", gotFP, wantFP) - } - gotFP = mapper.mapFP(fp3, cm31) - if wantFP := model.Fingerprint(4); gotFP != wantFP { - t.Errorf("got fingerprint %v, want fingerprint %v", gotFP, wantFP) - } - gotFP = mapper.mapFP(fp3, cm32) - if wantFP := model.Fingerprint(5); gotFP != wantFP { - t.Errorf("got fingerprint %v, want fingerprint %v", gotFP, wantFP) - } - - // To make sure that the mapping layer is not queried if the FP is found - // in sm but the mapping layer is queried before going to the archive, - // now put fp1 with cm12 in sm and fp2 with cm22 into archive (which - // will never happen in practice as only mapped FPs are put into sm and - // the archive). - sm.put(fp1, &memorySeries{metric: cm12}) - p.archiveMetric(fp2, cm22, 0, 0) - gotFP = mapper.mapFP(fp1, cm12) - if wantFP := fp1; gotFP != wantFP { // No mapping happened. - t.Errorf("got fingerprint %v, want fingerprint %v", gotFP, wantFP) - } - gotFP = mapper.mapFP(fp2, cm22) - if wantFP := model.Fingerprint(3); gotFP != wantFP { // Old mapping still applied. - t.Errorf("got fingerprint %v, want fingerprint %v", gotFP, wantFP) - } - - // If we now map cm21, we should get a mapping as the collision with the - // archived metric is detected. Again, this is a pathological situation - // that must never happen in real operations. It's just staged here to - // test the expected behavior. - gotFP = mapper.mapFP(fp2, cm21) - if wantFP := model.Fingerprint(6); gotFP != wantFP { - t.Errorf("got fingerprint %v, want fingerprint %v", gotFP, wantFP) - } - -} diff --git a/storage/local/noop_storage.go b/storage/local/noop_storage.go deleted file mode 100644 index 70b5a32f19..0000000000 --- a/storage/local/noop_storage.go +++ /dev/null @@ -1,100 +0,0 @@ -// Copyright 2016 The Prometheus Authors -// Licensed under the Apache License, Version 2.0 (the "License"); -// you may not use this file except in compliance with the License. -// You may obtain a copy of the License at -// -// http://www.apache.org/licenses/LICENSE-2.0 -// -// Unless required by applicable law or agreed to in writing, software -// distributed under the License is distributed on an "AS IS" BASIS, -// WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. -// See the License for the specific language governing permissions and -// limitations under the License. - -package local - -import ( - "time" - - "github.com/prometheus/common/model" - "golang.org/x/net/context" - - "github.com/prometheus/prometheus/storage/metric" -) - -// NoopStorage is a dummy storage for use when Prometheus's local storage is -// disabled. It throws away any appended samples and returns empty results. -type NoopStorage struct{} - -// Start implements Storage. -func (s *NoopStorage) Start() (err error) { - return nil -} - -// Stop implements Storage. -func (s *NoopStorage) Stop() error { - return nil -} - -// WaitForIndexing implements Storage. -func (s *NoopStorage) WaitForIndexing() { -} - -// Querier implements Storage. -func (s *NoopStorage) Querier() (Querier, error) { - return &NoopQuerier{}, nil -} - -// NoopQuerier is a dummy Querier for use when Prometheus's local storage is -// disabled. It is returned by the NoopStorage Querier method and always returns -// empty results. -type NoopQuerier struct{} - -// Close implements Querier. -func (s *NoopQuerier) Close() error { - return nil -} - -// LastSampleForLabelMatchers implements Querier. -func (s *NoopQuerier) LastSampleForLabelMatchers(ctx context.Context, cutoff model.Time, matcherSets ...metric.LabelMatchers) (model.Vector, error) { - return nil, nil -} - -// QueryRange implements Querier -func (s *NoopQuerier) QueryRange(ctx context.Context, from, through model.Time, matchers ...*metric.LabelMatcher) ([]SeriesIterator, error) { - return nil, nil -} - -// QueryInstant implements Querier. -func (s *NoopQuerier) QueryInstant(ctx context.Context, ts model.Time, stalenessDelta time.Duration, matchers ...*metric.LabelMatcher) ([]SeriesIterator, error) { - return nil, nil -} - -// MetricsForLabelMatchers implements Querier. -func (s *NoopQuerier) MetricsForLabelMatchers( - ctx context.Context, - from, through model.Time, - matcherSets ...metric.LabelMatchers, -) ([]metric.Metric, error) { - return nil, nil -} - -// LabelValuesForLabelName implements Querier. -func (s *NoopQuerier) LabelValuesForLabelName(ctx context.Context, labelName model.LabelName) (model.LabelValues, error) { - return nil, nil -} - -// DropMetricsForLabelMatchers implements Storage. -func (s *NoopStorage) DropMetricsForLabelMatchers(ctx context.Context, matchers ...*metric.LabelMatcher) (int, error) { - return 0, nil -} - -// Append implements Storage. -func (s *NoopStorage) Append(sample *model.Sample) error { - return nil -} - -// NeedsThrottling implements Storage. -func (s *NoopStorage) NeedsThrottling() bool { - return false -} diff --git a/storage/local/persistence.go b/storage/local/persistence.go deleted file mode 100644 index 0c8c83ff10..0000000000 --- a/storage/local/persistence.go +++ /dev/null @@ -1,1578 +0,0 @@ -// Copyright 2014 The Prometheus Authors -// Licensed under the Apache License, Version 2.0 (the "License"); -// you may not use this file except in compliance with the License. -// You may obtain a copy of the License at -// -// http://www.apache.org/licenses/LICENSE-2.0 -// -// Unless required by applicable law or agreed to in writing, software -// distributed under the License is distributed on an "AS IS" BASIS, -// WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. -// See the License for the specific language governing permissions and -// limitations under the License. - -package local - -import ( - "bufio" - "encoding/binary" - "fmt" - "io" - "io/ioutil" - "os" - "path/filepath" - "strconv" - "strings" - "sync" - "sync/atomic" - "time" - - "github.com/prometheus/client_golang/prometheus" - "github.com/prometheus/common/log" - "github.com/prometheus/common/model" - - "github.com/prometheus/prometheus/storage/local/chunk" - "github.com/prometheus/prometheus/storage/local/codable" - "github.com/prometheus/prometheus/storage/local/index" - "github.com/prometheus/prometheus/util/flock" -) - -const ( - // Version of the storage as it can be found in the version file. - // Increment to protect against incompatible changes. - Version = 1 - versionFileName = "VERSION" - - seriesFileSuffix = ".db" - seriesTempFileSuffix = ".db.tmp" - seriesDirNameLen = 2 // How many bytes of the fingerprint in dir name. - hintFileSuffix = ".hint" - - mappingsFileName = "mappings.db" - mappingsTempFileName = "mappings.db.tmp" - mappingsFormatVersion = 1 - mappingsMagicString = "PrometheusMappings" - - dirtyFileName = "DIRTY" - - fileBufSize = 1 << 16 // 64kiB. - - chunkHeaderLen = 17 - chunkHeaderTypeOffset = 0 - chunkHeaderFirstTimeOffset = 1 - chunkHeaderLastTimeOffset = 9 - chunkLenWithHeader = chunk.ChunkLen + chunkHeaderLen - chunkMaxBatchSize = 62 // Max no. of chunks to load or write in - // one batch. Note that 62 is the largest number of chunks that fit - // into 64kiB on disk because chunkHeaderLen is added to each 1k chunk. - - indexingMaxBatchSize = 1024 * 1024 - indexingBatchTimeout = 500 * time.Millisecond // Commit batch when idle for that long. - indexingQueueCapacity = 1024 * 16 -) - -var fpLen = len(model.Fingerprint(0).String()) // Length of a fingerprint as string. - -const ( - flagHeadChunkPersisted byte = 1 << iota - // Add more flags here like: - // flagFoo - // flagBar -) - -type indexingOpType byte - -const ( - add indexingOpType = iota - remove -) - -type indexingOp struct { - fingerprint model.Fingerprint - metric model.Metric - opType indexingOpType -} - -// A Persistence is used by a Storage implementation to store samples -// persistently across restarts. The methods are only goroutine-safe if -// explicitly marked as such below. The chunk-related methods persistChunks, -// dropChunks, loadChunks, and loadChunkDescs can be called concurrently with -// each other if each call refers to a different fingerprint. -type persistence struct { - basePath string - - archivedFingerprintToMetrics *index.FingerprintMetricIndex - archivedFingerprintToTimeRange *index.FingerprintTimeRangeIndex - labelPairToFingerprints *index.LabelPairFingerprintIndex - labelNameToLabelValues *index.LabelNameLabelValuesIndex - - indexingQueue chan indexingOp - indexingStopped chan struct{} - indexingFlush chan chan int - - indexingQueueLength prometheus.Gauge - indexingQueueCapacity prometheus.Metric - indexingBatchSizes prometheus.Summary - indexingBatchDuration prometheus.Summary - checkpointDuration prometheus.Gauge - dirtyCounter prometheus.Counter - startedDirty prometheus.Gauge - - dirtyMtx sync.Mutex // Protects dirty and becameDirty. - dirty bool // true if persistence was started in dirty state. - becameDirty bool // true if an inconsistency came up during runtime. - pedanticChecks bool // true if crash recovery should check each series. - dirtyFileName string // The file used for locking and to mark dirty state. - fLock flock.Releaser // The file lock to protect against concurrent usage. - - shouldSync syncStrategy - - minShrinkRatio float64 // How much a series file has to shrink to justify dropping chunks. - - bufPool sync.Pool -} - -// newPersistence returns a newly allocated persistence backed by local disk storage, ready to use. -func newPersistence( - basePath string, - dirty, pedanticChecks bool, - shouldSync syncStrategy, - minShrinkRatio float64, -) (*persistence, error) { - dirtyPath := filepath.Join(basePath, dirtyFileName) - versionPath := filepath.Join(basePath, versionFileName) - - if versionData, err := ioutil.ReadFile(versionPath); err == nil { - if persistedVersion, err := strconv.Atoi(strings.TrimSpace(string(versionData))); err != nil { - return nil, fmt.Errorf("cannot parse content of %s: %s", versionPath, versionData) - } else if persistedVersion != Version { - return nil, fmt.Errorf("found storage version %d on disk, need version %d - please wipe storage or run a version of Prometheus compatible with storage version %d", persistedVersion, Version, persistedVersion) - } - } else if os.IsNotExist(err) { - // No version file found. Let's create the directory (in case - // it's not there yet) and then check if it is actually - // empty. If not, we have found an old storage directory without - // version file, so we have to bail out. - if err := os.MkdirAll(basePath, 0700); err != nil { - if abspath, e := filepath.Abs(basePath); e == nil { - return nil, fmt.Errorf("cannot create persistent directory %s: %s", abspath, err) - } - return nil, fmt.Errorf("cannot create persistent directory %s: %s", basePath, err) - } - fis, err := ioutil.ReadDir(basePath) - if err != nil { - return nil, err - } - filesPresent := len(fis) - for i := range fis { - switch { - case fis[i].Name() == "lost+found" && fis[i].IsDir(): - filesPresent-- - case strings.HasPrefix(fis[i].Name(), "."): - filesPresent-- - } - } - if filesPresent > 0 { - return nil, fmt.Errorf("found existing files in storage path that do not look like storage files compatible with this version of Prometheus; please delete the files in the storage path or choose a different storage path") - } - // Finally we can write our own version into a new version file. - file, err := os.Create(versionPath) - if err != nil { - return nil, err - } - defer file.Close() - if _, err := fmt.Fprintf(file, "%d\n", Version); err != nil { - return nil, err - } - } else { - return nil, err - } - - fLock, dirtyfileExisted, err := flock.New(dirtyPath) - if err != nil { - log.Errorf("Could not lock %s, Prometheus already running?", dirtyPath) - return nil, err - } - if dirtyfileExisted { - dirty = true - } - - archivedFingerprintToMetrics, err := index.NewFingerprintMetricIndex(basePath) - if err != nil { - return nil, err - } - archivedFingerprintToTimeRange, err := index.NewFingerprintTimeRangeIndex(basePath) - if err != nil { - return nil, err - } - - p := &persistence{ - basePath: basePath, - - archivedFingerprintToMetrics: archivedFingerprintToMetrics, - archivedFingerprintToTimeRange: archivedFingerprintToTimeRange, - - indexingQueue: make(chan indexingOp, indexingQueueCapacity), - indexingStopped: make(chan struct{}), - indexingFlush: make(chan chan int), - - indexingQueueLength: prometheus.NewGauge(prometheus.GaugeOpts{ - Namespace: namespace, - Subsystem: subsystem, - Name: "indexing_queue_length", - Help: "The number of metrics waiting to be indexed.", - }), - indexingQueueCapacity: prometheus.MustNewConstMetric( - prometheus.NewDesc( - prometheus.BuildFQName(namespace, subsystem, "indexing_queue_capacity"), - "The capacity of the indexing queue.", - nil, nil, - ), - prometheus.GaugeValue, - float64(indexingQueueCapacity), - ), - indexingBatchSizes: prometheus.NewSummary( - prometheus.SummaryOpts{ - Namespace: namespace, - Subsystem: subsystem, - Name: "indexing_batch_sizes", - Help: "Quantiles for indexing batch sizes (number of metrics per batch).", - }, - ), - indexingBatchDuration: prometheus.NewSummary( - prometheus.SummaryOpts{ - Namespace: namespace, - Subsystem: subsystem, - Name: "indexing_batch_duration_seconds", - Help: "Quantiles for batch indexing duration in seconds.", - }, - ), - checkpointDuration: prometheus.NewGauge(prometheus.GaugeOpts{ - Namespace: namespace, - Subsystem: subsystem, - Name: "checkpoint_duration_seconds", - Help: "The duration in seconds it took to checkpoint in-memory metrics and head chunks.", - }), - dirtyCounter: prometheus.NewCounter(prometheus.CounterOpts{ - Namespace: namespace, - Subsystem: subsystem, - Name: "inconsistencies_total", - Help: "A counter incremented each time an inconsistency in the local storage is detected. If this is greater zero, restart the server as soon as possible.", - }), - startedDirty: prometheus.NewGauge(prometheus.GaugeOpts{ - Namespace: namespace, - Subsystem: subsystem, - Name: "started_dirty", - Help: "Whether the local storage was found to be dirty (and crash recovery occurred) during Prometheus startup.", - }), - dirty: dirty, - pedanticChecks: pedanticChecks, - dirtyFileName: dirtyPath, - fLock: fLock, - shouldSync: shouldSync, - // Create buffers of length 3*chunkLenWithHeader by default because that is still reasonably small - // and at the same time enough for many uses. The contract is to never return buffer smaller than - // that to the pool so that callers can rely on a minimum buffer size. - bufPool: sync.Pool{New: func() interface{} { return make([]byte, 0, 3*chunkLenWithHeader) }}, - } - - if p.dirty { - // Blow away the label indexes. We'll rebuild them later. - if err := index.DeleteLabelPairFingerprintIndex(basePath); err != nil { - return nil, err - } - if err := index.DeleteLabelNameLabelValuesIndex(basePath); err != nil { - return nil, err - } - } - labelPairToFingerprints, err := index.NewLabelPairFingerprintIndex(basePath) - if err != nil { - return nil, err - } - labelNameToLabelValues, err := index.NewLabelNameLabelValuesIndex(basePath) - if err != nil { - return nil, err - } - p.labelPairToFingerprints = labelPairToFingerprints - p.labelNameToLabelValues = labelNameToLabelValues - - return p, nil -} - -func (p *persistence) run() { - p.processIndexingQueue() -} - -// Describe implements prometheus.Collector. -func (p *persistence) Describe(ch chan<- *prometheus.Desc) { - ch <- p.indexingQueueLength.Desc() - ch <- p.indexingQueueCapacity.Desc() - p.indexingBatchSizes.Describe(ch) - p.indexingBatchDuration.Describe(ch) - ch <- p.checkpointDuration.Desc() - ch <- p.dirtyCounter.Desc() - ch <- p.startedDirty.Desc() -} - -// Collect implements prometheus.Collector. -func (p *persistence) Collect(ch chan<- prometheus.Metric) { - p.indexingQueueLength.Set(float64(len(p.indexingQueue))) - - ch <- p.indexingQueueLength - ch <- p.indexingQueueCapacity - p.indexingBatchSizes.Collect(ch) - p.indexingBatchDuration.Collect(ch) - ch <- p.checkpointDuration - ch <- p.dirtyCounter - ch <- p.startedDirty -} - -// isDirty returns the dirty flag in a goroutine-safe way. -func (p *persistence) isDirty() bool { - p.dirtyMtx.Lock() - defer p.dirtyMtx.Unlock() - return p.dirty -} - -// setDirty flags the storage as dirty in a goroutine-safe way. The provided -// error will be logged as a reason the first time the storage is flagged as dirty. -func (p *persistence) setDirty(err error) { - p.dirtyCounter.Inc() - p.dirtyMtx.Lock() - defer p.dirtyMtx.Unlock() - if p.becameDirty { - return - } - p.dirty = true - p.becameDirty = true - log.With("error", err).Error("The storage is now inconsistent. Restart Prometheus ASAP to initiate recovery.") -} - -// fingerprintsForLabelPair returns the fingerprints for the given label -// pair. This method is goroutine-safe but take into account that metrics queued -// for indexing with IndexMetric might not have made it into the index -// yet. (Same applies correspondingly to UnindexMetric.) -func (p *persistence) fingerprintsForLabelPair(lp model.LabelPair) model.Fingerprints { - fps, _, err := p.labelPairToFingerprints.Lookup(lp) - if err != nil { - p.setDirty(fmt.Errorf("error in method fingerprintsForLabelPair(%v): %s", lp, err)) - return nil - } - return fps -} - -// labelValuesForLabelName returns the label values for the given label -// name. This method is goroutine-safe but take into account that metrics queued -// for indexing with IndexMetric might not have made it into the index -// yet. (Same applies correspondingly to UnindexMetric.) -func (p *persistence) labelValuesForLabelName(ln model.LabelName) (model.LabelValues, error) { - lvs, _, err := p.labelNameToLabelValues.Lookup(ln) - if err != nil { - p.setDirty(fmt.Errorf("error in method labelValuesForLabelName(%v): %s", ln, err)) - return nil, err - } - return lvs, nil -} - -// persistChunks persists a number of consecutive chunks of a series. It is the -// caller's responsibility to not modify the chunks concurrently and to not -// persist or drop anything for the same fingerprint concurrently. It returns -// the (zero-based) index of the first persisted chunk within the series -// file. In case of an error, the returned index is -1 (to avoid the -// misconception that the chunk was written at position 0). -// -// Returning an error signals problems with the series file. In this case, the -// caller should quarantine the series. -func (p *persistence) persistChunks(fp model.Fingerprint, chunks []chunk.Chunk) (index int, err error) { - f, err := p.openChunkFileForWriting(fp) - if err != nil { - return -1, err - } - defer p.closeChunkFile(f) - - if err := p.writeChunks(f, chunks); err != nil { - return -1, err - } - - // Determine index within the file. - offset, err := f.Seek(0, os.SEEK_CUR) - if err != nil { - return -1, err - } - index, err = chunkIndexForOffset(offset) - if err != nil { - return -1, err - } - - return index - len(chunks), err -} - -// loadChunks loads a group of chunks of a timeseries by their index. The chunk -// with the earliest time will have index 0, the following ones will have -// incrementally larger indexes. The indexOffset denotes the offset to be added to -// each index in indexes. It is the caller's responsibility to not persist or -// drop anything for the same fingerprint concurrently. -func (p *persistence) loadChunks(fp model.Fingerprint, indexes []int, indexOffset int) ([]chunk.Chunk, error) { - f, err := p.openChunkFileForReading(fp) - if err != nil { - return nil, err - } - defer f.Close() - - chunks := make([]chunk.Chunk, 0, len(indexes)) - buf := p.bufPool.Get().([]byte) - defer func() { - // buf may change below. An unwrapped 'defer p.bufPool.Put(buf)' - // would only put back the original buf. - p.bufPool.Put(buf) - }() - - for i := 0; i < len(indexes); i++ { - // This loads chunks in batches. A batch is a streak of - // consecutive chunks, read from disk in one go. - batchSize := 1 - if _, err := f.Seek(offsetForChunkIndex(indexes[i]+indexOffset), os.SEEK_SET); err != nil { - return nil, err - } - - for ; batchSize < chunkMaxBatchSize && - i+1 < len(indexes) && - indexes[i]+1 == indexes[i+1]; i, batchSize = i+1, batchSize+1 { - } - readSize := batchSize * chunkLenWithHeader - if cap(buf) < readSize { - buf = make([]byte, readSize) - } - buf = buf[:readSize] - - if _, err := io.ReadFull(f, buf); err != nil { - return nil, err - } - for c := 0; c < batchSize; c++ { - chunk, err := chunk.NewForEncoding(chunk.Encoding(buf[c*chunkLenWithHeader+chunkHeaderTypeOffset])) - if err != nil { - return nil, err - } - if err := chunk.UnmarshalFromBuf(buf[c*chunkLenWithHeader+chunkHeaderLen:]); err != nil { - return nil, err - } - chunks = append(chunks, chunk) - } - } - chunk.Ops.WithLabelValues(chunk.Load).Add(float64(len(chunks))) - atomic.AddInt64(&chunk.NumMemChunks, int64(len(chunks))) - return chunks, nil -} - -// loadChunkDescs loads the chunk.Descs for a series from disk. offsetFromEnd is -// the number of chunk.Descs to skip from the end of the series file. It is the -// caller's responsibility to not persist or drop anything for the same -// fingerprint concurrently. -func (p *persistence) loadChunkDescs(fp model.Fingerprint, offsetFromEnd int) ([]*chunk.Desc, error) { - f, err := p.openChunkFileForReading(fp) - if os.IsNotExist(err) { - return nil, nil - } - if err != nil { - return nil, err - } - defer f.Close() - - fi, err := f.Stat() - if err != nil { - return nil, err - } - if fi.Size()%int64(chunkLenWithHeader) != 0 { - // The returned error will bubble up and lead to quarantining of the whole series. - return nil, fmt.Errorf( - "size of series file for fingerprint %v is %d, which is not a multiple of the chunk length %d", - fp, fi.Size(), chunkLenWithHeader, - ) - } - - numChunks := int(fi.Size())/chunkLenWithHeader - offsetFromEnd - cds := make([]*chunk.Desc, numChunks) - chunkTimesBuf := make([]byte, 16) - for i := 0; i < numChunks; i++ { - _, err := f.Seek(offsetForChunkIndex(i)+chunkHeaderFirstTimeOffset, os.SEEK_SET) - if err != nil { - return nil, err - } - - _, err = io.ReadAtLeast(f, chunkTimesBuf, 16) - if err != nil { - return nil, err - } - cds[i] = &chunk.Desc{ - ChunkFirstTime: model.Time(binary.LittleEndian.Uint64(chunkTimesBuf)), - ChunkLastTime: model.Time(binary.LittleEndian.Uint64(chunkTimesBuf[8:])), - } - } - chunk.DescOps.WithLabelValues(chunk.Load).Add(float64(len(cds))) - chunk.NumMemDescs.Add(float64(len(cds))) - return cds, nil -} - -// checkpointSeriesMapAndHeads persists the fingerprint to memory-series mapping -// and all non persisted chunks. Do not call concurrently with -// loadSeriesMapAndHeads. This method will only write heads format v2, but -// loadSeriesMapAndHeads can also understand v1. -// -// Description of the file format (for both, v1 and v2): -// -// (1) Magic string (const headsMagicString). -// -// (2) Varint-encoded format version (const headsFormatVersion). -// -// (3) Number of series in checkpoint as big-endian uint64. -// -// (4) Repeated once per series: -// -// (4.1) A flag byte, see flag constants above. (Present but unused in v2.) -// -// (4.2) The fingerprint as big-endian uint64. -// -// (4.3) The metric as defined by codable.Metric. -// -// (4.4) The varint-encoded persistWatermark. (Missing in v1.) -// -// (4.5) The modification time of the series file as nanoseconds elapsed since -// January 1, 1970 UTC. -1 if the modification time is unknown or no series file -// exists yet. (Missing in v1.) -// -// (4.6) The varint-encoded chunkDescsOffset. -// -// (4.6) The varint-encoded savedFirstTime. -// -// (4.7) The varint-encoded number of chunk descriptors. -// -// (4.8) Repeated once per chunk descriptor, oldest to most recent, either -// variant 4.8.1 (if index < persistWatermark) or variant 4.8.2 (if index >= -// persistWatermark). In v1, everything is variant 4.8.1 except for a -// non-persisted head-chunk (determined by the flags). -// -// (4.8.1.1) The varint-encoded first time. -// (4.8.1.2) The varint-encoded last time. -// -// (4.8.2.1) A byte defining the chunk type. -// (4.8.2.2) The chunk itself, marshaled with the Marshal() method. -// -func (p *persistence) checkpointSeriesMapAndHeads(fingerprintToSeries *seriesMap, fpLocker *fingerprintLocker) (err error) { - log.Info("Checkpointing in-memory metrics and chunks...") - begin := time.Now() - f, err := os.OpenFile(p.headsTempFileName(), os.O_WRONLY|os.O_TRUNC|os.O_CREATE, 0640) - if err != nil { - return err - } - - defer func() { - syncErr := f.Sync() - closeErr := f.Close() - if err != nil { - return - } - err = syncErr - if err != nil { - return - } - err = closeErr - if err != nil { - return - } - err = os.Rename(p.headsTempFileName(), p.headsFileName()) - duration := time.Since(begin) - p.checkpointDuration.Set(duration.Seconds()) - log.Infof("Done checkpointing in-memory metrics and chunks in %v.", duration) - }() - - w := bufio.NewWriterSize(f, fileBufSize) - - if _, err = w.WriteString(headsMagicString); err != nil { - return err - } - var numberOfSeriesOffset int - if numberOfSeriesOffset, err = codable.EncodeVarint(w, headsFormatVersion); err != nil { - return err - } - numberOfSeriesOffset += len(headsMagicString) - numberOfSeriesInHeader := uint64(fingerprintToSeries.length()) - // We have to write the number of series as uint64 because we might need - // to overwrite it later, and a varint might change byte width then. - if err = codable.EncodeUint64(w, numberOfSeriesInHeader); err != nil { - return err - } - - iter := fingerprintToSeries.iter() - defer func() { - // Consume the iterator in any case to not leak goroutines. - for range iter { - } - }() - - var realNumberOfSeries uint64 - for m := range iter { - func() { // Wrapped in function to use defer for unlocking the fp. - fpLocker.Lock(m.fp) - defer fpLocker.Unlock(m.fp) - - if len(m.series.chunkDescs) == 0 { - // This series was completely purged or archived in the meantime. Ignore. - return - } - realNumberOfSeries++ - // seriesFlags left empty in v2. - if err = w.WriteByte(0); err != nil { - return - } - if err = codable.EncodeUint64(w, uint64(m.fp)); err != nil { - return - } - var buf []byte - buf, err = codable.Metric(m.series.metric).MarshalBinary() - if err != nil { - return - } - if _, err = w.Write(buf); err != nil { - return - } - if _, err = codable.EncodeVarint(w, int64(m.series.persistWatermark)); err != nil { - return - } - if m.series.modTime.IsZero() { - if _, err = codable.EncodeVarint(w, -1); err != nil { - return - } - } else { - if _, err = codable.EncodeVarint(w, m.series.modTime.UnixNano()); err != nil { - return - } - } - if _, err = codable.EncodeVarint(w, int64(m.series.chunkDescsOffset)); err != nil { - return - } - if _, err = codable.EncodeVarint(w, int64(m.series.savedFirstTime)); err != nil { - return - } - if _, err = codable.EncodeVarint(w, int64(len(m.series.chunkDescs))); err != nil { - return - } - for i, chunkDesc := range m.series.chunkDescs { - if i < m.series.persistWatermark { - if _, err = codable.EncodeVarint(w, int64(chunkDesc.FirstTime())); err != nil { - return - } - lt, err := chunkDesc.LastTime() - if err != nil { - return - } - if _, err = codable.EncodeVarint(w, int64(lt)); err != nil { - return - } - } else { - // This is a non-persisted chunk. Fully marshal it. - if err = w.WriteByte(byte(chunkDesc.C.Encoding())); err != nil { - return - } - if err = chunkDesc.C.Marshal(w); err != nil { - return - } - } - } - // Series is checkpointed now, so declare it clean. In case the entire - // checkpoint fails later on, this is fine, as the storage's series - // maintenance will mark these series newly dirty again, continuously - // increasing the total number of dirty series as seen by the storage. - // This has the effect of triggering a new checkpoint attempt even - // earlier than if we hadn't incorrectly set "dirty" to "false" here - // already. - m.series.dirty = false - }() - if err != nil { - return err - } - } - if err = w.Flush(); err != nil { - return err - } - if realNumberOfSeries != numberOfSeriesInHeader { - // The number of series has changed in the meantime. - // Rewrite it in the header. - if _, err = f.Seek(int64(numberOfSeriesOffset), os.SEEK_SET); err != nil { - return err - } - if err = codable.EncodeUint64(f, realNumberOfSeries); err != nil { - return err - } - } - return err -} - -// loadSeriesMapAndHeads loads the fingerprint to memory-series mapping and all -// the chunks contained in the checkpoint (and thus not yet persisted to series -// files). The method is capable of loading the checkpoint format v1 and v2. If -// recoverable corruption is detected, or if the dirty flag was set from the -// beginning, crash recovery is run, which might take a while. If an -// unrecoverable error is encountered, it is returned. Call this method during -// start-up while nothing else is running in storage land. This method is -// utterly goroutine-unsafe. -func (p *persistence) loadSeriesMapAndHeads() (sm *seriesMap, chunksToPersist int64, err error) { - fingerprintToSeries := make(map[model.Fingerprint]*memorySeries) - sm = &seriesMap{m: fingerprintToSeries} - - defer func() { - if p.dirty { - log.Warn("Persistence layer appears dirty.") - p.startedDirty.Set(1) - err = p.recoverFromCrash(fingerprintToSeries) - if err != nil { - sm = nil - } - } else { - p.startedDirty.Set(0) - } - }() - - hs := newHeadsScanner(p.headsFileName()) - defer hs.close() - for hs.scan() { - fingerprintToSeries[hs.fp] = hs.series - } - if os.IsNotExist(hs.err) { - return sm, 0, nil - } - if hs.err != nil { - p.dirty = true - log. - With("file", p.headsFileName()). - With("error", hs.err). - Error("Error reading heads file.") - return sm, 0, hs.err - } - return sm, hs.chunksToPersistTotal, nil -} - -// dropAndPersistChunks deletes all chunks from a series file whose last sample -// time is before beforeTime, and then appends the provided chunks, leaving out -// those whose last sample time is before beforeTime. It returns the timestamp -// of the first sample in the oldest chunk _not_ dropped, the offset within the -// series file of the first chunk persisted (out of the provided chunks), the -// number of deleted chunks, and true if all chunks of the series have been -// deleted (in which case the returned timestamp will be 0 and must be ignored). -// It is the caller's responsibility to make sure nothing is persisted or loaded -// for the same fingerprint concurrently. -// -// Returning an error signals problems with the series file. In this case, the -// caller should quarantine the series. -func (p *persistence) dropAndPersistChunks( - fp model.Fingerprint, beforeTime model.Time, chunks []chunk.Chunk, -) ( - firstTimeNotDropped model.Time, - offset int, - numDropped int, - allDropped bool, - err error, -) { - // Style note: With the many return values, it was decided to use naked - // returns in this method. They make the method more readable, but - // please handle with care! - if len(chunks) > 0 { - // We have chunks to persist. First check if those are already - // too old. If that's the case, the chunks in the series file - // are all too old, too. - i := 0 - for ; i < len(chunks); i++ { - var lt model.Time - lt, err = chunks[i].NewIterator().LastTimestamp() - if err != nil { - return - } - if !lt.Before(beforeTime) { - break - } - } - if i < len(chunks) { - firstTimeNotDropped = chunks[i].FirstTime() - } - if i > 0 || firstTimeNotDropped.Before(beforeTime) { - // Series file has to go. - if numDropped, err = p.deleteSeriesFile(fp); err != nil { - return - } - numDropped += i - if i == len(chunks) { - allDropped = true - return - } - // Now simply persist what has to be persisted to a new file. - _, err = p.persistChunks(fp, chunks[i:]) - return - } - } - - // If we are here, we have to check the series file itself. - f, err := p.openChunkFileForReading(fp) - if os.IsNotExist(err) { - // No series file. Only need to create new file with chunks to - // persist, if there are any. - if len(chunks) == 0 { - allDropped = true - err = nil // Do not report not-exist err. - return - } - offset, err = p.persistChunks(fp, chunks) - return - } - if err != nil { - return - } - defer f.Close() - - headerBuf := make([]byte, chunkHeaderLen) - var firstTimeInFile model.Time - // Find the first chunk in the file that should be kept. - for ; ; numDropped++ { - _, err = f.Seek(offsetForChunkIndex(numDropped), os.SEEK_SET) - if err != nil { - return - } - _, err = io.ReadFull(f, headerBuf) - if err == io.EOF { - // Close the file before trying to delete it. This is necessary on Windows - // (this will cause the defer f.Close to fail, but the error is silently ignored) - f.Close() - // We ran into the end of the file without finding any chunks that should - // be kept. Remove the whole file. - if numDropped, err = p.deleteSeriesFile(fp); err != nil { - return - } - if len(chunks) == 0 { - allDropped = true - return - } - offset, err = p.persistChunks(fp, chunks) - return - } - if err != nil { - return - } - if numDropped == 0 { - firstTimeInFile = model.Time( - binary.LittleEndian.Uint64(headerBuf[chunkHeaderFirstTimeOffset:]), - ) - } - lastTime := model.Time( - binary.LittleEndian.Uint64(headerBuf[chunkHeaderLastTimeOffset:]), - ) - if !lastTime.Before(beforeTime) { - break - } - } - - // We've found the first chunk that should be kept. - // First check if the shrink ratio is good enough to perform the the - // actual drop or leave it for next time if it is not worth the effort. - fi, err := f.Stat() - if err != nil { - return - } - totalChunks := int(fi.Size())/chunkLenWithHeader + len(chunks) - if numDropped == 0 || float64(numDropped)/float64(totalChunks) < p.minShrinkRatio { - // Nothing to drop. Just adjust the return values and append the chunks (if any). - numDropped = 0 - firstTimeNotDropped = firstTimeInFile - if len(chunks) > 0 { - offset, err = p.persistChunks(fp, chunks) - } - return - } - // If we are here, we have to drop some chunks for real. So we need to - // record firstTimeNotDropped from the last read header, seek backwards - // to the beginning of its header, and start copying everything from - // there into a new file. Then append the chunks to the new file. - firstTimeNotDropped = model.Time( - binary.LittleEndian.Uint64(headerBuf[chunkHeaderFirstTimeOffset:]), - ) - chunk.Ops.WithLabelValues(chunk.Drop).Add(float64(numDropped)) - _, err = f.Seek(-chunkHeaderLen, os.SEEK_CUR) - if err != nil { - return - } - - temp, err := os.OpenFile(p.tempFileNameForFingerprint(fp), os.O_WRONLY|os.O_CREATE, 0640) - if err != nil { - return - } - defer func() { - // Close the file before trying to rename to it. This is necessary on Windows - // (this will cause the defer f.Close to fail, but the error is silently ignored) - f.Close() - p.closeChunkFile(temp) - if err == nil { - err = os.Rename(p.tempFileNameForFingerprint(fp), p.fileNameForFingerprint(fp)) - } - }() - - written, err := io.Copy(temp, f) - if err != nil { - return - } - offset = int(written / chunkLenWithHeader) - - if len(chunks) > 0 { - if err = p.writeChunks(temp, chunks); err != nil { - return - } - } - return -} - -// deleteSeriesFile deletes a series file belonging to the provided -// fingerprint. It returns the number of chunks that were contained in the -// deleted file. -func (p *persistence) deleteSeriesFile(fp model.Fingerprint) (int, error) { - fname := p.fileNameForFingerprint(fp) - fi, err := os.Stat(fname) - if os.IsNotExist(err) { - // Great. The file is already gone. - return 0, nil - } - if err != nil { - return -1, err - } - numChunks := int(fi.Size() / chunkLenWithHeader) - if err := os.Remove(fname); err != nil { - return -1, err - } - chunk.Ops.WithLabelValues(chunk.Drop).Add(float64(numChunks)) - return numChunks, nil -} - -// quarantineSeriesFile moves a series file to the orphaned directory. It also -// writes a hint file with the provided quarantine reason and, if series is -// non-nil, the string representation of the metric. -func (p *persistence) quarantineSeriesFile(fp model.Fingerprint, quarantineReason error, metric model.Metric) error { - var ( - oldName = p.fileNameForFingerprint(fp) - orphanedDir = filepath.Join(p.basePath, "orphaned", filepath.Base(filepath.Dir(oldName))) - newName = filepath.Join(orphanedDir, filepath.Base(oldName)) - hintName = newName[:len(newName)-len(seriesFileSuffix)] + hintFileSuffix - ) - - renameErr := os.MkdirAll(orphanedDir, 0700) - if renameErr != nil { - return renameErr - } - renameErr = os.Rename(oldName, newName) - if os.IsNotExist(renameErr) { - // Source file dosn't exist. That's normal. - renameErr = nil - } - // Write hint file even if the rename ended in an error. At least try... - // And ignore errors writing the hint file. It's best effort. - if f, err := os.Create(hintName); err == nil { - if metric != nil { - f.WriteString(metric.String() + "\n") - } else { - f.WriteString("[UNKNOWN METRIC]\n") - } - if quarantineReason != nil { - f.WriteString(quarantineReason.Error() + "\n") - } else { - f.WriteString("[UNKNOWN REASON]\n") - } - f.Close() - } - return renameErr -} - -// seriesFileModTime returns the modification time of the series file belonging -// to the provided fingerprint. In case of an error, the zero value of time.Time -// is returned. -func (p *persistence) seriesFileModTime(fp model.Fingerprint) time.Time { - var modTime time.Time - if fi, err := os.Stat(p.fileNameForFingerprint(fp)); err == nil { - return fi.ModTime() - } - return modTime -} - -// indexMetric queues the given metric for addition to the indexes needed by -// fingerprintsForLabelPair, labelValuesForLabelName, and -// fingerprintsModifiedBefore. If the queue is full, this method blocks until -// the metric can be queued. This method is goroutine-safe. -func (p *persistence) indexMetric(fp model.Fingerprint, m model.Metric) { - p.indexingQueue <- indexingOp{fp, m, add} -} - -// unindexMetric queues references to the given metric for removal from the -// indexes used for fingerprintsForLabelPair, labelValuesForLabelName, and -// fingerprintsModifiedBefore. The index of fingerprints to archived metrics is -// not affected by this removal. (In fact, never call this method for an -// archived metric. To purge an archived metric, call purgeArchivedMetric.) -// If the queue is full, this method blocks until the metric can be queued. This -// method is goroutine-safe. -func (p *persistence) unindexMetric(fp model.Fingerprint, m model.Metric) { - p.indexingQueue <- indexingOp{fp, m, remove} -} - -// waitForIndexing waits until all items in the indexing queue are processed. If -// queue processing is currently on hold (to gather more ops for batching), this -// method will trigger an immediate start of processing. This method is -// goroutine-safe. -func (p *persistence) waitForIndexing() { - wait := make(chan int) - for { - p.indexingFlush <- wait - if <-wait == 0 { - break - } - } -} - -// archiveMetric persists the mapping of the given fingerprint to the given -// metric, together with the first and last timestamp of the series belonging to -// the metric. The caller must have locked the fingerprint. -func (p *persistence) archiveMetric( - fp model.Fingerprint, m model.Metric, first, last model.Time, -) { - if err := p.archivedFingerprintToMetrics.Put(codable.Fingerprint(fp), codable.Metric(m)); err != nil { - p.setDirty(fmt.Errorf("error in method archiveMetric inserting fingerprint %v into FingerprintToMetrics: %s", fp, err)) - return - } - if err := p.archivedFingerprintToTimeRange.Put(codable.Fingerprint(fp), codable.TimeRange{First: first, Last: last}); err != nil { - p.setDirty(fmt.Errorf("error in method archiveMetric inserting fingerprint %v into FingerprintToTimeRange: %s", fp, err)) - } -} - -// hasArchivedMetric returns whether the archived metric for the given -// fingerprint exists and if yes, what the first and last timestamp in the -// corresponding series is. This method is goroutine-safe. -func (p *persistence) hasArchivedMetric(fp model.Fingerprint) ( - hasMetric bool, firstTime, lastTime model.Time, -) { - firstTime, lastTime, hasMetric, err := p.archivedFingerprintToTimeRange.Lookup(fp) - if err != nil { - p.setDirty(fmt.Errorf("error in method hasArchivedMetric(%v): %s", fp, err)) - hasMetric = false - } - return hasMetric, firstTime, lastTime -} - -// updateArchivedTimeRange updates an archived time range. The caller must make -// sure that the fingerprint is currently archived (the time range will -// otherwise be added without the corresponding metric in the archive). -func (p *persistence) updateArchivedTimeRange( - fp model.Fingerprint, first, last model.Time, -) error { - return p.archivedFingerprintToTimeRange.Put(codable.Fingerprint(fp), codable.TimeRange{First: first, Last: last}) -} - -// fingerprintsModifiedBefore returns the fingerprints of archived timeseries -// that have live samples before the provided timestamp. This method is -// goroutine-safe. -func (p *persistence) fingerprintsModifiedBefore(beforeTime model.Time) ([]model.Fingerprint, error) { - var fp codable.Fingerprint - var tr codable.TimeRange - fps := []model.Fingerprint{} - err := p.archivedFingerprintToTimeRange.ForEach(func(kv index.KeyValueAccessor) error { - if err := kv.Value(&tr); err != nil { - return err - } - if tr.First.Before(beforeTime) { - if err := kv.Key(&fp); err != nil { - return err - } - fps = append(fps, model.Fingerprint(fp)) - } - return nil - }) - return fps, err -} - -// archivedMetric retrieves the archived metric with the given fingerprint. This -// method is goroutine-safe. -func (p *persistence) archivedMetric(fp model.Fingerprint) (model.Metric, error) { - metric, _, err := p.archivedFingerprintToMetrics.Lookup(fp) - if err != nil { - p.setDirty(fmt.Errorf("error in method archivedMetric(%v): %s", fp, err)) - return nil, err - } - return metric, nil -} - -// purgeArchivedMetric deletes an archived fingerprint and its corresponding -// metric entirely. It also queues the metric for un-indexing (no need to call -// unindexMetric for the deleted metric.) It does not touch the series file, -// though. The caller must have locked the fingerprint. -func (p *persistence) purgeArchivedMetric(fp model.Fingerprint) (err error) { - defer func() { - if err != nil { - p.setDirty(fmt.Errorf("error in method purgeArchivedMetric(%v): %s", fp, err)) - } - }() - - metric, err := p.archivedMetric(fp) - if err != nil || metric == nil { - return err - } - deleted, err := p.archivedFingerprintToMetrics.Delete(codable.Fingerprint(fp)) - if err != nil { - return err - } - if !deleted { - log.Errorf("Tried to delete non-archived fingerprint %s from archivedFingerprintToMetrics index. This should never happen.", fp) - } - deleted, err = p.archivedFingerprintToTimeRange.Delete(codable.Fingerprint(fp)) - if err != nil { - return err - } - if !deleted { - log.Errorf("Tried to delete non-archived fingerprint %s from archivedFingerprintToTimeRange index. This should never happen.", fp) - } - p.unindexMetric(fp, metric) - return nil -} - -// unarchiveMetric deletes an archived fingerprint and its metric, but (in -// contrast to purgeArchivedMetric) does not un-index the metric. If a metric -// was actually deleted, the method returns true and the first time and last -// time of the deleted metric. The caller must have locked the fingerprint. -func (p *persistence) unarchiveMetric(fp model.Fingerprint) (deletedAnything bool, err error) { - // An error returned here will bubble up and lead to quarantining of the - // series, so no setDirty required. - deleted, err := p.archivedFingerprintToMetrics.Delete(codable.Fingerprint(fp)) - if err != nil || !deleted { - return false, err - } - deleted, err = p.archivedFingerprintToTimeRange.Delete(codable.Fingerprint(fp)) - if err != nil { - return false, err - } - if !deleted { - log.Errorf("Tried to delete non-archived fingerprint %s from archivedFingerprintToTimeRange index. This should never happen.", fp) - } - return true, nil -} - -// close flushes the indexing queue and other buffered data and releases any -// held resources. It also removes the dirty marker file if successful and if -// the persistence is currently not marked as dirty. -func (p *persistence) close() error { - close(p.indexingQueue) - <-p.indexingStopped - - var lastError, dirtyFileRemoveError error - if err := p.archivedFingerprintToMetrics.Close(); err != nil { - lastError = err - log.Error("Error closing archivedFingerprintToMetric index DB: ", err) - } - if err := p.archivedFingerprintToTimeRange.Close(); err != nil { - lastError = err - log.Error("Error closing archivedFingerprintToTimeRange index DB: ", err) - } - if err := p.labelPairToFingerprints.Close(); err != nil { - lastError = err - log.Error("Error closing labelPairToFingerprints index DB: ", err) - } - if err := p.labelNameToLabelValues.Close(); err != nil { - lastError = err - log.Error("Error closing labelNameToLabelValues index DB: ", err) - } - if lastError == nil && !p.isDirty() { - dirtyFileRemoveError = os.Remove(p.dirtyFileName) - } - if err := p.fLock.Release(); err != nil { - lastError = err - log.Error("Error releasing file lock: ", err) - } - if dirtyFileRemoveError != nil { - // On Windows, removing the dirty file before unlocking is not - // possible. So remove it here if it failed above. - lastError = os.Remove(p.dirtyFileName) - } - return lastError -} - -func (p *persistence) dirNameForFingerprint(fp model.Fingerprint) string { - fpStr := fp.String() - return filepath.Join(p.basePath, fpStr[0:seriesDirNameLen]) -} - -func (p *persistence) fileNameForFingerprint(fp model.Fingerprint) string { - fpStr := fp.String() - return filepath.Join(p.basePath, fpStr[0:seriesDirNameLen], fpStr[seriesDirNameLen:]+seriesFileSuffix) -} - -func (p *persistence) tempFileNameForFingerprint(fp model.Fingerprint) string { - fpStr := fp.String() - return filepath.Join(p.basePath, fpStr[0:seriesDirNameLen], fpStr[seriesDirNameLen:]+seriesTempFileSuffix) -} - -func (p *persistence) openChunkFileForWriting(fp model.Fingerprint) (*os.File, error) { - if err := os.MkdirAll(p.dirNameForFingerprint(fp), 0700); err != nil { - return nil, err - } - return os.OpenFile(p.fileNameForFingerprint(fp), os.O_WRONLY|os.O_APPEND|os.O_CREATE, 0640) - // NOTE: Although the file was opened for append, - // f.Seek(0, os.SEEK_CUR) - // would now return '0, nil', so we cannot check for a consistent file length right now. - // However, the chunkIndexForOffset function is doing that check, so a wrong file length - // would still be detected. -} - -// closeChunkFile first syncs the provided file if mandated so by the sync -// strategy. Then it closes the file. Errors are logged. -func (p *persistence) closeChunkFile(f *os.File) { - if p.shouldSync() { - if err := f.Sync(); err != nil { - log.Error("Error syncing file:", err) - } - } - if err := f.Close(); err != nil { - log.Error("Error closing chunk file:", err) - } -} - -func (p *persistence) openChunkFileForReading(fp model.Fingerprint) (*os.File, error) { - return os.Open(p.fileNameForFingerprint(fp)) -} - -func (p *persistence) headsFileName() string { - return filepath.Join(p.basePath, headsFileName) -} - -func (p *persistence) headsTempFileName() string { - return filepath.Join(p.basePath, headsTempFileName) -} - -func (p *persistence) mappingsFileName() string { - return filepath.Join(p.basePath, mappingsFileName) -} - -func (p *persistence) mappingsTempFileName() string { - return filepath.Join(p.basePath, mappingsTempFileName) -} - -func (p *persistence) processIndexingQueue() { - batchSize := 0 - nameToValues := index.LabelNameLabelValuesMapping{} - pairToFPs := index.LabelPairFingerprintsMapping{} - batchTimeout := time.NewTimer(indexingBatchTimeout) - defer batchTimeout.Stop() - - commitBatch := func() { - p.indexingBatchSizes.Observe(float64(batchSize)) - defer func(begin time.Time) { - p.indexingBatchDuration.Observe(time.Since(begin).Seconds()) - }(time.Now()) - - if err := p.labelPairToFingerprints.IndexBatch(pairToFPs); err != nil { - log.Error("Error indexing label pair to fingerprints batch: ", err) - } - if err := p.labelNameToLabelValues.IndexBatch(nameToValues); err != nil { - log.Error("Error indexing label name to label values batch: ", err) - } - batchSize = 0 - nameToValues = index.LabelNameLabelValuesMapping{} - pairToFPs = index.LabelPairFingerprintsMapping{} - batchTimeout.Reset(indexingBatchTimeout) - } - - var flush chan chan int -loop: - for { - // Only process flush requests if the queue is currently empty. - if len(p.indexingQueue) == 0 { - flush = p.indexingFlush - } else { - flush = nil - } - select { - case <-batchTimeout.C: - // Only commit if we have something to commit _and_ - // nothing is waiting in the queue to be picked up. That - // prevents a death spiral if the LookupSet calls below - // are slow for some reason. - if batchSize > 0 && len(p.indexingQueue) == 0 { - commitBatch() - } else { - batchTimeout.Reset(indexingBatchTimeout) - } - case r := <-flush: - if batchSize > 0 { - commitBatch() - } - r <- len(p.indexingQueue) - case op, ok := <-p.indexingQueue: - if !ok { - if batchSize > 0 { - commitBatch() - } - break loop - } - - batchSize++ - for ln, lv := range op.metric { - lp := model.LabelPair{Name: ln, Value: lv} - baseFPs, ok := pairToFPs[lp] - if !ok { - var err error - baseFPs, _, err = p.labelPairToFingerprints.LookupSet(lp) - if err != nil { - log.Errorf("Error looking up label pair %v: %s", lp, err) - continue - } - pairToFPs[lp] = baseFPs - } - baseValues, ok := nameToValues[ln] - if !ok { - var err error - baseValues, _, err = p.labelNameToLabelValues.LookupSet(ln) - if err != nil { - log.Errorf("Error looking up label name %v: %s", ln, err) - continue - } - nameToValues[ln] = baseValues - } - switch op.opType { - case add: - baseFPs[op.fingerprint] = struct{}{} - baseValues[lv] = struct{}{} - case remove: - delete(baseFPs, op.fingerprint) - if len(baseFPs) == 0 { - delete(baseValues, lv) - } - default: - panic("unknown op type") - } - } - - if batchSize >= indexingMaxBatchSize { - commitBatch() - } - } - } - close(p.indexingStopped) -} - -// checkpointFPMappings persists the fingerprint mappings. The caller has to -// ensure that the provided mappings are not changed concurrently. This method -// is only called upon shutdown or during crash recovery, when no samples are -// ingested. -// -// Description of the file format, v1: -// -// (1) Magic string (const mappingsMagicString). -// -// (2) Uvarint-encoded format version (const mappingsFormatVersion). -// -// (3) Uvarint-encoded number of mappings in fpMappings. -// -// (4) Repeated once per mapping: -// -// (4.1) The raw fingerprint as big-endian uint64. -// -// (4.2) The uvarint-encoded number of sub-mappings for the raw fingerprint. -// -// (4.3) Repeated once per sub-mapping: -// -// (4.3.1) The uvarint-encoded length of the unique metric string. -// (4.3.2) The unique metric string. -// (4.3.3) The mapped fingerprint as big-endian uint64. -func (p *persistence) checkpointFPMappings(fpm fpMappings) (err error) { - log.Info("Checkpointing fingerprint mappings...") - begin := time.Now() - f, err := os.OpenFile(p.mappingsTempFileName(), os.O_WRONLY|os.O_TRUNC|os.O_CREATE, 0640) - if err != nil { - return - } - - defer func() { - syncErr := f.Sync() - closeErr := f.Close() - if err != nil { - return - } - err = syncErr - if err != nil { - return - } - err = closeErr - if err != nil { - return - } - err = os.Rename(p.mappingsTempFileName(), p.mappingsFileName()) - duration := time.Since(begin) - log.Infof("Done checkpointing fingerprint mappings in %v.", duration) - }() - - w := bufio.NewWriterSize(f, fileBufSize) - - if _, err = w.WriteString(mappingsMagicString); err != nil { - return - } - if _, err = codable.EncodeUvarint(w, mappingsFormatVersion); err != nil { - return - } - if _, err = codable.EncodeUvarint(w, uint64(len(fpm))); err != nil { - return - } - - for fp, mappings := range fpm { - if err = codable.EncodeUint64(w, uint64(fp)); err != nil { - return - } - if _, err = codable.EncodeUvarint(w, uint64(len(mappings))); err != nil { - return - } - for ms, mappedFP := range mappings { - if _, err = codable.EncodeUvarint(w, uint64(len(ms))); err != nil { - return - } - if _, err = w.WriteString(ms); err != nil { - return - } - if err = codable.EncodeUint64(w, uint64(mappedFP)); err != nil { - return - } - } - } - err = w.Flush() - return -} - -// loadFPMappings loads the fingerprint mappings. It also returns the highest -// mapped fingerprint and any error encountered. If p.mappingsFileName is not -// found, the method returns (fpMappings{}, 0, nil). Do not call concurrently -// with checkpointFPMappings. -func (p *persistence) loadFPMappings() (fpMappings, model.Fingerprint, error) { - fpm := fpMappings{} - var highestMappedFP model.Fingerprint - - f, err := os.Open(p.mappingsFileName()) - if os.IsNotExist(err) { - return fpm, 0, nil - } - if err != nil { - return nil, 0, err - } - defer f.Close() - r := bufio.NewReaderSize(f, fileBufSize) - - buf := make([]byte, len(mappingsMagicString)) - if _, err := io.ReadFull(r, buf); err != nil { - return nil, 0, err - } - magic := string(buf) - if magic != mappingsMagicString { - return nil, 0, fmt.Errorf( - "unexpected magic string, want %q, got %q", - mappingsMagicString, magic, - ) - } - version, err := binary.ReadUvarint(r) - if version != mappingsFormatVersion || err != nil { - return nil, 0, fmt.Errorf("unknown fingerprint mappings format version, want %d", mappingsFormatVersion) - } - numRawFPs, err := binary.ReadUvarint(r) - if err != nil { - return nil, 0, err - } - for ; numRawFPs > 0; numRawFPs-- { - rawFP, err := codable.DecodeUint64(r) - if err != nil { - return nil, 0, err - } - numMappings, err := binary.ReadUvarint(r) - if err != nil { - return nil, 0, err - } - mappings := make(map[string]model.Fingerprint, numMappings) - for ; numMappings > 0; numMappings-- { - lenMS, err := binary.ReadUvarint(r) - if err != nil { - return nil, 0, err - } - buf := make([]byte, lenMS) - if _, err := io.ReadFull(r, buf); err != nil { - return nil, 0, err - } - fp, err := codable.DecodeUint64(r) - if err != nil { - return nil, 0, err - } - mappedFP := model.Fingerprint(fp) - if mappedFP > highestMappedFP { - highestMappedFP = mappedFP - } - mappings[string(buf)] = mappedFP - } - fpm[model.Fingerprint(rawFP)] = mappings - } - return fpm, highestMappedFP, nil -} - -func (p *persistence) writeChunks(w io.Writer, chunks []chunk.Chunk) error { - b := p.bufPool.Get().([]byte) - defer func() { - // buf may change below. An unwrapped 'defer p.bufPool.Put(buf)' - // would only put back the original buf. - p.bufPool.Put(b) - }() - - for batchSize := chunkMaxBatchSize; len(chunks) > 0; chunks = chunks[batchSize:] { - if batchSize > len(chunks) { - batchSize = len(chunks) - } - writeSize := batchSize * chunkLenWithHeader - if cap(b) < writeSize { - b = make([]byte, writeSize) - } - b = b[:writeSize] - - for i, chunk := range chunks[:batchSize] { - if err := writeChunkHeader(b[i*chunkLenWithHeader:], chunk); err != nil { - return err - } - if err := chunk.MarshalToBuf(b[i*chunkLenWithHeader+chunkHeaderLen:]); err != nil { - return err - } - } - if _, err := w.Write(b); err != nil { - return err - } - } - return nil -} - -func offsetForChunkIndex(i int) int64 { - return int64(i * chunkLenWithHeader) -} - -func chunkIndexForOffset(offset int64) (int, error) { - if int(offset)%chunkLenWithHeader != 0 { - return -1, fmt.Errorf( - "offset %d is not a multiple of on-disk chunk length %d", - offset, chunkLenWithHeader, - ) - } - return int(offset) / chunkLenWithHeader, nil -} - -func writeChunkHeader(header []byte, c chunk.Chunk) error { - header[chunkHeaderTypeOffset] = byte(c.Encoding()) - binary.LittleEndian.PutUint64( - header[chunkHeaderFirstTimeOffset:], - uint64(c.FirstTime()), - ) - lt, err := c.NewIterator().LastTimestamp() - if err != nil { - return err - } - binary.LittleEndian.PutUint64( - header[chunkHeaderLastTimeOffset:], - uint64(lt), - ) - return nil -} diff --git a/storage/local/persistence_test.go b/storage/local/persistence_test.go deleted file mode 100644 index 2b37857668..0000000000 --- a/storage/local/persistence_test.go +++ /dev/null @@ -1,1256 +0,0 @@ -// Copyright 2014 The Prometheus Authors -// Licensed under the Apache License, Version 2.0 (the "License"); -// you may not use this file except in compliance with the License. -// You may obtain a copy of the License at -// -// http://www.apache.org/licenses/LICENSE-2.0 -// -// Unless required by applicable law or agreed to in writing, software -// distributed under the License is distributed on an "AS IS" BASIS, -// WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. -// See the License for the specific language governing permissions and -// limitations under the License. - -package local - -import ( - "bufio" - "errors" - "os" - "path/filepath" - "reflect" - "sync" - "testing" - "time" - - "github.com/prometheus/common/model" - - "github.com/prometheus/prometheus/storage/local/chunk" - "github.com/prometheus/prometheus/storage/local/codable" - "github.com/prometheus/prometheus/storage/local/index" - "github.com/prometheus/prometheus/util/testutil" -) - -var ( - m1 = model.Metric{"label": "value1"} - m2 = model.Metric{"label": "value2"} - m3 = model.Metric{"label": "value3"} - m4 = model.Metric{"label": "value4"} - m5 = model.Metric{"label": "value5"} -) - -func newTestPersistence(t *testing.T, encoding chunk.Encoding) (*persistence, testutil.Closer) { - chunk.DefaultEncoding = encoding - dir := testutil.NewTemporaryDirectory("test_persistence", t) - p, err := newPersistence(dir.Path(), false, false, func() bool { return false }, 0.1) - if err != nil { - dir.Close() - t.Fatal(err) - } - go p.run() - return p, testutil.NewCallbackCloser(func() { - p.close() - dir.Close() - }) -} - -func buildTestChunks(t *testing.T, encoding chunk.Encoding) map[model.Fingerprint][]chunk.Chunk { - fps := model.Fingerprints{ - m1.FastFingerprint(), - m2.FastFingerprint(), - m3.FastFingerprint(), - } - fpToChunks := map[model.Fingerprint][]chunk.Chunk{} - - for _, fp := range fps { - fpToChunks[fp] = make([]chunk.Chunk, 0, 10) - for i := 0; i < 10; i++ { - ch, err := chunk.NewForEncoding(encoding) - if err != nil { - t.Fatal(err) - } - chs, err := ch.Add(model.SamplePair{ - Timestamp: model.Time(i), - Value: model.SampleValue(fp), - }) - if err != nil { - t.Fatal(err) - } - fpToChunks[fp] = append(fpToChunks[fp], chs[0]) - } - } - return fpToChunks -} - -func chunksEqual(c1, c2 chunk.Chunk) bool { - it1 := c1.NewIterator() - it2 := c2.NewIterator() - for it1.Scan() && it2.Scan() { - if !(it1.Value() == it2.Value()) { - return false - } - } - return it1.Err() == nil && it2.Err() == nil -} - -func testPersistLoadDropChunks(t *testing.T, encoding chunk.Encoding) { - p, closer := newTestPersistence(t, encoding) - defer closer.Close() - - fpToChunks := buildTestChunks(t, encoding) - - for fp, chunks := range fpToChunks { - firstTimeNotDropped, offset, numDropped, allDropped, err := - p.dropAndPersistChunks(fp, model.Earliest, chunks) - if err != nil { - t.Fatal(err) - } - if got, want := firstTimeNotDropped, model.Time(0); got != want { - t.Errorf("Want firstTimeNotDropped %v, got %v.", got, want) - } - if got, want := offset, 0; got != want { - t.Errorf("Want offset %v, got %v.", got, want) - } - if got, want := numDropped, 0; got != want { - t.Errorf("Want numDropped %v, got %v.", got, want) - } - if allDropped { - t.Error("All dropped.") - } - } - - for fp, expectedChunks := range fpToChunks { - indexes := make([]int, 0, len(expectedChunks)) - for i := range expectedChunks { - indexes = append(indexes, i) - } - actualChunks, err := p.loadChunks(fp, indexes, 0) - if err != nil { - t.Fatal(err) - } - for _, i := range indexes { - if !chunksEqual(expectedChunks[i], actualChunks[i]) { - t.Errorf("%d. Chunks not equal.", i) - } - } - // Load all chunk descs. - actualChunkDescs, err := p.loadChunkDescs(fp, 0) - if len(actualChunkDescs) != 10 { - t.Errorf("Got %d chunkDescs, want %d.", len(actualChunkDescs), 10) - } - for i, cd := range actualChunkDescs { - lastTime, err := cd.LastTime() - if err != nil { - t.Fatal(err) - } - if cd.FirstTime() != model.Time(i) || lastTime != model.Time(i) { - t.Errorf( - "Want ts=%v, got firstTime=%v, lastTime=%v.", - i, cd.FirstTime(), lastTime, - ) - } - - } - // Load chunk descs partially. - actualChunkDescs, err = p.loadChunkDescs(fp, 5) - if err != nil { - t.Fatal(err) - } - if len(actualChunkDescs) != 5 { - t.Errorf("Got %d chunkDescs, want %d.", len(actualChunkDescs), 5) - } - for i, cd := range actualChunkDescs { - lastTime, err := cd.LastTime() - if err != nil { - t.Fatal(err) - } - if cd.FirstTime() != model.Time(i) || lastTime != model.Time(i) { - t.Errorf( - "Want ts=%v, got firstTime=%v, lastTime=%v.", - i, cd.FirstTime(), lastTime, - ) - } - - } - } - // Drop half of the chunks. - for fp, expectedChunks := range fpToChunks { - firstTime, offset, numDropped, allDropped, err := p.dropAndPersistChunks(fp, 5, nil) - if err != nil { - t.Fatal(err) - } - if offset != 5 { - t.Errorf("want offset 5, got %d", offset) - } - if firstTime != 5 { - t.Errorf("want first time 5, got %d", firstTime) - } - if numDropped != 5 { - t.Errorf("want 5 dropped chunks, got %v", numDropped) - } - if allDropped { - t.Error("all chunks dropped") - } - indexes := make([]int, 5) - for i := range indexes { - indexes[i] = i - } - actualChunks, err := p.loadChunks(fp, indexes, 0) - if err != nil { - t.Fatal(err) - } - for _, i := range indexes { - if !chunksEqual(expectedChunks[i+5], actualChunks[i]) { - t.Errorf("%d. Chunks not equal.", i) - } - } - } - // Drop all the chunks. - for fp := range fpToChunks { - firstTime, offset, numDropped, allDropped, err := p.dropAndPersistChunks(fp, 100, nil) - if firstTime != 0 { - t.Errorf("want first time 0, got %d", firstTime) - } - if err != nil { - t.Fatal(err) - } - if offset != 0 { - t.Errorf("want offset 0, got %d", offset) - } - if numDropped != 5 { - t.Errorf("want 5 dropped chunks, got %v", numDropped) - } - if !allDropped { - t.Error("not all chunks dropped") - } - } - // Re-add first two of the chunks. - for fp, chunks := range fpToChunks { - firstTimeNotDropped, offset, numDropped, allDropped, err := - p.dropAndPersistChunks(fp, model.Earliest, chunks[:2]) - if err != nil { - t.Fatal(err) - } - if got, want := firstTimeNotDropped, model.Time(0); got != want { - t.Errorf("Want firstTimeNotDropped %v, got %v.", got, want) - } - if got, want := offset, 0; got != want { - t.Errorf("Want offset %v, got %v.", got, want) - } - if got, want := numDropped, 0; got != want { - t.Errorf("Want numDropped %v, got %v.", got, want) - } - if allDropped { - t.Error("All dropped.") - } - } - // Drop the first of the chunks while adding two more. - for fp, chunks := range fpToChunks { - firstTime, offset, numDropped, allDropped, err := p.dropAndPersistChunks(fp, 1, chunks[2:4]) - if err != nil { - t.Fatal(err) - } - if offset != 1 { - t.Errorf("want offset 1, got %d", offset) - } - if firstTime != 1 { - t.Errorf("want first time 1, got %d", firstTime) - } - if numDropped != 1 { - t.Errorf("want 1 dropped chunk, got %v", numDropped) - } - if allDropped { - t.Error("all chunks dropped") - } - wantChunks := chunks[1:4] - indexes := make([]int, len(wantChunks)) - for i := range indexes { - indexes[i] = i - } - gotChunks, err := p.loadChunks(fp, indexes, 0) - if err != nil { - t.Fatal(err) - } - for i, wantChunk := range wantChunks { - if !chunksEqual(wantChunk, gotChunks[i]) { - t.Errorf("%d. Chunks not equal.", i) - } - } - } - // Drop all the chunks while adding two more. - for fp, chunks := range fpToChunks { - firstTime, offset, numDropped, allDropped, err := p.dropAndPersistChunks(fp, 4, chunks[4:6]) - if err != nil { - t.Fatal(err) - } - if offset != 0 { - t.Errorf("want offset 0, got %d", offset) - } - if firstTime != 4 { - t.Errorf("want first time 4, got %d", firstTime) - } - if numDropped != 3 { - t.Errorf("want 3 dropped chunks, got %v", numDropped) - } - if allDropped { - t.Error("all chunks dropped") - } - wantChunks := chunks[4:6] - indexes := make([]int, len(wantChunks)) - for i := range indexes { - indexes[i] = i - } - gotChunks, err := p.loadChunks(fp, indexes, 0) - if err != nil { - t.Fatal(err) - } - for i, wantChunk := range wantChunks { - if !chunksEqual(wantChunk, gotChunks[i]) { - t.Errorf("%d. Chunks not equal.", i) - } - } - } - // While adding two more, drop all but one of the added ones. - for fp, chunks := range fpToChunks { - firstTime, offset, numDropped, allDropped, err := p.dropAndPersistChunks(fp, 7, chunks[6:8]) - if err != nil { - t.Fatal(err) - } - if offset != 0 { - t.Errorf("want offset 0, got %d", offset) - } - if firstTime != 7 { - t.Errorf("want first time 7, got %d", firstTime) - } - if numDropped != 3 { - t.Errorf("want 3 dropped chunks, got %v", numDropped) - } - if allDropped { - t.Error("all chunks dropped") - } - wantChunks := chunks[7:8] - indexes := make([]int, len(wantChunks)) - for i := range indexes { - indexes[i] = i - } - gotChunks, err := p.loadChunks(fp, indexes, 0) - if err != nil { - t.Fatal(err) - } - for i, wantChunk := range wantChunks { - if !chunksEqual(wantChunk, gotChunks[i]) { - t.Errorf("%d. Chunks not equal.", i) - } - } - } - // While adding two more, drop all chunks including the added ones. - for fp, chunks := range fpToChunks { - firstTime, offset, numDropped, allDropped, err := p.dropAndPersistChunks(fp, 10, chunks[8:]) - if err != nil { - t.Fatal(err) - } - if offset != 0 { - t.Errorf("want offset 0, got %d", offset) - } - if firstTime != 0 { - t.Errorf("want first time 0, got %d", firstTime) - } - if numDropped != 3 { - t.Errorf("want 3 dropped chunks, got %v", numDropped) - } - if !allDropped { - t.Error("not all chunks dropped") - } - } - // Now set minShrinkRatio to 0.25 and play with it. - p.minShrinkRatio = 0.25 - // Re-add 8 chunks. - for fp, chunks := range fpToChunks { - firstTimeNotDropped, offset, numDropped, allDropped, err := - p.dropAndPersistChunks(fp, model.Earliest, chunks[:8]) - if err != nil { - t.Fatal(err) - } - if got, want := firstTimeNotDropped, model.Time(0); got != want { - t.Errorf("Want firstTimeNotDropped %v, got %v.", got, want) - } - if got, want := offset, 0; got != want { - t.Errorf("Want offset %v, got %v.", got, want) - } - if got, want := numDropped, 0; got != want { - t.Errorf("Want numDropped %v, got %v.", got, want) - } - if allDropped { - t.Error("All dropped.") - } - } - // Drop only the first chunk should not happen, but persistence should still work. - for fp, chunks := range fpToChunks { - firstTime, offset, numDropped, allDropped, err := p.dropAndPersistChunks(fp, 1, chunks[8:9]) - if err != nil { - t.Fatal(err) - } - if offset != 8 { - t.Errorf("want offset 8, got %d", offset) - } - if firstTime != 0 { - t.Errorf("want first time 0, got %d", firstTime) - } - if numDropped != 0 { - t.Errorf("want 0 dropped chunk, got %v", numDropped) - } - if allDropped { - t.Error("all chunks dropped") - } - } - // Drop only the first two chunks should not happen, either. - for fp := range fpToChunks { - firstTime, offset, numDropped, allDropped, err := p.dropAndPersistChunks(fp, 2, nil) - if err != nil { - t.Fatal(err) - } - if offset != 0 { - t.Errorf("want offset 0, got %d", offset) - } - if firstTime != 0 { - t.Errorf("want first time 0, got %d", firstTime) - } - if numDropped != 0 { - t.Errorf("want 0 dropped chunk, got %v", numDropped) - } - if allDropped { - t.Error("all chunks dropped") - } - } - // Drop the first three chunks should finally work. - for fp, chunks := range fpToChunks { - firstTime, offset, numDropped, allDropped, err := p.dropAndPersistChunks(fp, 3, chunks[9:]) - if err != nil { - t.Fatal(err) - } - if offset != 6 { - t.Errorf("want offset 6, got %d", offset) - } - if firstTime != 3 { - t.Errorf("want first time 3, got %d", firstTime) - } - if numDropped != 3 { - t.Errorf("want 3 dropped chunk, got %v", numDropped) - } - if allDropped { - t.Error("all chunks dropped") - } - } -} - -func TestPersistLoadDropChunksType0(t *testing.T) { - testPersistLoadDropChunks(t, 0) -} - -func TestPersistLoadDropChunksType1(t *testing.T) { - testPersistLoadDropChunks(t, 1) -} - -func testCheckpointAndLoadSeriesMapAndHeads(t *testing.T, encoding chunk.Encoding) { - p, closer := newTestPersistence(t, encoding) - defer closer.Close() - - fpLocker := newFingerprintLocker(10) - sm := newSeriesMap() - s1, _ := newMemorySeries(m1, nil, time.Time{}) - s2, _ := newMemorySeries(m2, nil, time.Time{}) - s3, _ := newMemorySeries(m3, nil, time.Time{}) - s4, _ := newMemorySeries(m4, nil, time.Time{}) - s5, _ := newMemorySeries(m5, nil, time.Time{}) - s1.add(model.SamplePair{Timestamp: 1, Value: 3.14}) - s3.add(model.SamplePair{Timestamp: 2, Value: 2.7}) - s3.headChunkClosed = true - s3.persistWatermark = 1 - for i := 0; i < 10000; i++ { - s4.add(model.SamplePair{ - Timestamp: model.Time(i), - Value: model.SampleValue(i) / 2, - }) - s5.add(model.SamplePair{ - Timestamp: model.Time(i), - Value: model.SampleValue(i * i), - }) - } - s5.persistWatermark = 3 - chunkCountS4 := len(s4.chunkDescs) - chunkCountS5 := len(s5.chunkDescs) - sm.put(m1.FastFingerprint(), s1) - sm.put(m2.FastFingerprint(), s2) - sm.put(m3.FastFingerprint(), s3) - sm.put(m4.FastFingerprint(), s4) - sm.put(m5.FastFingerprint(), s5) - - if err := p.checkpointSeriesMapAndHeads(sm, fpLocker); err != nil { - t.Fatal(err) - } - - loadedSM, _, err := p.loadSeriesMapAndHeads() - if err != nil { - t.Fatal(err) - } - if loadedSM.length() != 4 { - t.Errorf("want 4 series in map, got %d", loadedSM.length()) - } - if loadedS1, ok := loadedSM.get(m1.FastFingerprint()); ok { - if !reflect.DeepEqual(loadedS1.metric, m1) { - t.Errorf("want metric %v, got %v", m1, loadedS1.metric) - } - if !reflect.DeepEqual(loadedS1.head().C, s1.head().C) { - t.Error("head chunks differ") - } - if loadedS1.chunkDescsOffset != 0 { - t.Errorf("want chunkDescsOffset 0, got %d", loadedS1.chunkDescsOffset) - } - if loadedS1.headChunkClosed { - t.Error("headChunkClosed is true") - } - if loadedS1.head().ChunkFirstTime != 1 { - t.Errorf("want ChunkFirstTime in head chunk to be 1, got %d", loadedS1.head().ChunkFirstTime) - } - if loadedS1.head().ChunkLastTime != model.Earliest { - t.Error("want ChunkLastTime in head chunk to be unset") - } - } else { - t.Errorf("couldn't find %v in loaded map", m1) - } - if loadedS3, ok := loadedSM.get(m3.FastFingerprint()); ok { - if !reflect.DeepEqual(loadedS3.metric, m3) { - t.Errorf("want metric %v, got %v", m3, loadedS3.metric) - } - if loadedS3.head().C != nil { - t.Error("head chunk not evicted") - } - if loadedS3.chunkDescsOffset != 0 { - t.Errorf("want chunkDescsOffset 0, got %d", loadedS3.chunkDescsOffset) - } - if !loadedS3.headChunkClosed { - t.Error("headChunkClosed is false") - } - if loadedS3.head().ChunkFirstTime != 2 { - t.Errorf("want ChunkFirstTime in head chunk to be 2, got %d", loadedS3.head().ChunkFirstTime) - } - if loadedS3.head().ChunkLastTime != 2 { - t.Errorf("want ChunkLastTime in head chunk to be 2, got %d", loadedS3.head().ChunkLastTime) - } - } else { - t.Errorf("couldn't find %v in loaded map", m3) - } - if loadedS4, ok := loadedSM.get(m4.FastFingerprint()); ok { - if !reflect.DeepEqual(loadedS4.metric, m4) { - t.Errorf("want metric %v, got %v", m4, loadedS4.metric) - } - if got, want := len(loadedS4.chunkDescs), chunkCountS4; got != want { - t.Errorf("got %d chunkDescs, want %d", got, want) - } - if got, want := loadedS4.persistWatermark, 0; got != want { - t.Errorf("got persistWatermark %d, want %d", got, want) - } - if loadedS4.chunkDescs[2].IsEvicted() { - t.Error("3rd chunk evicted") - } - if loadedS4.chunkDescs[3].IsEvicted() { - t.Error("4th chunk evicted") - } - if loadedS4.chunkDescsOffset != 0 { - t.Errorf("want chunkDescsOffset 0, got %d", loadedS4.chunkDescsOffset) - } - if loadedS4.headChunkClosed { - t.Error("headChunkClosed is true") - } - for i, cd := range loadedS4.chunkDescs { - if cd.ChunkFirstTime != cd.C.FirstTime() { - t.Errorf( - "chunk.Desc[%d]: ChunkFirstTime not consistent with chunk, want %d, got %d", - i, cd.C.FirstTime(), cd.ChunkFirstTime, - ) - } - if i == len(loadedS4.chunkDescs)-1 { - // Head chunk. - if cd.ChunkLastTime != model.Earliest { - t.Error("want ChunkLastTime in head chunk to be unset") - } - continue - } - lastTime, err := cd.C.NewIterator().LastTimestamp() - if err != nil { - t.Fatal(err) - } - if cd.ChunkLastTime != lastTime { - t.Errorf( - "chunk.Desc[%d]: ChunkLastTime not consistent with chunk, want %d, got %d", - i, lastTime, cd.ChunkLastTime, - ) - } - } - } else { - t.Errorf("couldn't find %v in loaded map", m4) - } - if loadedS5, ok := loadedSM.get(m5.FastFingerprint()); ok { - if !reflect.DeepEqual(loadedS5.metric, m5) { - t.Errorf("want metric %v, got %v", m5, loadedS5.metric) - } - if got, want := len(loadedS5.chunkDescs), chunkCountS5; got != want { - t.Errorf("got %d chunkDescs, want %d", got, want) - } - if got, want := loadedS5.persistWatermark, 3; got != want { - t.Errorf("got persistWatermark %d, want %d", got, want) - } - if !loadedS5.chunkDescs[2].IsEvicted() { - t.Error("3rd chunk not evicted") - } - if loadedS5.chunkDescs[3].IsEvicted() { - t.Error("4th chunk evicted") - } - if loadedS5.chunkDescsOffset != 0 { - t.Errorf("want chunkDescsOffset 0, got %d", loadedS5.chunkDescsOffset) - } - if loadedS5.headChunkClosed { - t.Error("headChunkClosed is true") - } - for i, cd := range loadedS5.chunkDescs { - if i < 3 { - // Evicted chunks. - if cd.ChunkFirstTime == model.Earliest { - t.Errorf("chunk.Desc[%d]: ChunkLastTime not set", i) - } - continue - } - if cd.ChunkFirstTime != cd.C.FirstTime() { - t.Errorf( - "chunk.Desc[%d]: ChunkFirstTime not consistent with chunk, want %d, got %d", - i, cd.C.FirstTime(), cd.ChunkFirstTime, - ) - } - if i == len(loadedS5.chunkDescs)-1 { - // Head chunk. - if cd.ChunkLastTime != model.Earliest { - t.Error("want ChunkLastTime in head chunk to be unset") - } - continue - } - lastTime, err := cd.C.NewIterator().LastTimestamp() - if err != nil { - t.Fatal(err) - } - if cd.ChunkLastTime != lastTime { - t.Errorf( - "chunk.Desc[%d]: ChunkLastTime not consistent with chunk, want %d, got %d", - i, cd.ChunkLastTime, lastTime, - ) - } - } - } else { - t.Errorf("couldn't find %v in loaded map", m5) - } -} - -func TestCheckpointAndLoadSeriesMapAndHeadsChunkType0(t *testing.T) { - testCheckpointAndLoadSeriesMapAndHeads(t, 0) -} - -func TestCheckpointAndLoadSeriesMapAndHeadsChunkType1(t *testing.T) { - testCheckpointAndLoadSeriesMapAndHeads(t, 1) -} - -func TestCheckpointAndLoadSeriesMapAndHeadsChunkType2(t *testing.T) { - testCheckpointAndLoadSeriesMapAndHeads(t, 2) -} - -func TestCheckpointAndLoadFPMappings(t *testing.T) { - p, closer := newTestPersistence(t, 1) - defer closer.Close() - - in := fpMappings{ - 1: map[string]model.Fingerprint{ - "foo": 1, - "bar": 2, - }, - 3: map[string]model.Fingerprint{ - "baz": 4, - }, - } - - if err := p.checkpointFPMappings(in); err != nil { - t.Fatal(err) - } - - out, fp, err := p.loadFPMappings() - if err != nil { - t.Fatal(err) - } - if got, want := fp, model.Fingerprint(4); got != want { - t.Errorf("got highest FP %v, want %v", got, want) - } - if !reflect.DeepEqual(in, out) { - t.Errorf("got collision map %v, want %v", out, in) - } -} - -func testFingerprintsModifiedBefore(t *testing.T, encoding chunk.Encoding) { - p, closer := newTestPersistence(t, encoding) - defer closer.Close() - - m1 := model.Metric{"n1": "v1"} - m2 := model.Metric{"n2": "v2"} - m3 := model.Metric{"n1": "v2"} - p.archiveMetric(1, m1, 2, 4) - p.archiveMetric(2, m2, 1, 6) - p.archiveMetric(3, m3, 5, 5) - - expectedFPs := map[model.Time][]model.Fingerprint{ - 0: {}, - 1: {}, - 2: {2}, - 3: {1, 2}, - 4: {1, 2}, - 5: {1, 2}, - 6: {1, 2, 3}, - } - - for ts, want := range expectedFPs { - got, err := p.fingerprintsModifiedBefore(ts) - if err != nil { - t.Fatal(err) - } - if !reflect.DeepEqual(want, got) { - t.Errorf("timestamp: %v, want FPs %v, got %v", ts, want, got) - } - } - - unarchived, err := p.unarchiveMetric(1) - if err != nil { - t.Fatal(err) - } - if !unarchived { - t.Error("expected actual unarchival") - } - unarchived, err = p.unarchiveMetric(1) - if err != nil { - t.Fatal(err) - } - if unarchived { - t.Error("expected no unarchival") - } - - expectedFPs = map[model.Time][]model.Fingerprint{ - 0: {}, - 1: {}, - 2: {2}, - 3: {2}, - 4: {2}, - 5: {2}, - 6: {2, 3}, - } - - for ts, want := range expectedFPs { - got, err := p.fingerprintsModifiedBefore(ts) - if err != nil { - t.Fatal(err) - } - if !reflect.DeepEqual(want, got) { - t.Errorf("timestamp: %v, want FPs %v, got %v", ts, want, got) - } - } -} - -func TestFingerprintsModifiedBeforeChunkType0(t *testing.T) { - testFingerprintsModifiedBefore(t, 0) -} - -func TestFingerprintsModifiedBeforeChunkType1(t *testing.T) { - testFingerprintsModifiedBefore(t, 1) -} - -func TestFingerprintsModifiedBeforeChunkType2(t *testing.T) { - testFingerprintsModifiedBefore(t, 2) -} - -func testDropArchivedMetric(t *testing.T, encoding chunk.Encoding) { - p, closer := newTestPersistence(t, encoding) - defer closer.Close() - - m1 := model.Metric{"n1": "v1"} - m2 := model.Metric{"n2": "v2"} - p.archiveMetric(1, m1, 2, 4) - p.archiveMetric(2, m2, 1, 6) - p.indexMetric(1, m1) - p.indexMetric(2, m2) - p.waitForIndexing() - - outFPs := p.fingerprintsForLabelPair(model.LabelPair{Name: "n1", Value: "v1"}) - want := model.Fingerprints{1} - if !reflect.DeepEqual(outFPs, want) { - t.Errorf("want %#v, got %#v", want, outFPs) - } - outFPs = p.fingerprintsForLabelPair(model.LabelPair{Name: "n2", Value: "v2"}) - want = model.Fingerprints{2} - if !reflect.DeepEqual(outFPs, want) { - t.Errorf("want %#v, got %#v", want, outFPs) - } - if archived, _, _ := p.hasArchivedMetric(1); !archived { - t.Error("want FP 1 archived") - } - if archived, _, _ := p.hasArchivedMetric(2); !archived { - t.Error("want FP 2 archived") - } - - if err := p.purgeArchivedMetric(1); err != nil { - t.Fatal(err) - } - if err := p.purgeArchivedMetric(3); err != nil { - // Purging something that has not beet archived is not an error. - t.Fatal(err) - } - p.waitForIndexing() - - outFPs = p.fingerprintsForLabelPair(model.LabelPair{Name: "n1", Value: "v1"}) - want = nil - if !reflect.DeepEqual(outFPs, want) { - t.Errorf("want %#v, got %#v", want, outFPs) - } - outFPs = p.fingerprintsForLabelPair(model.LabelPair{Name: "n2", Value: "v2"}) - want = model.Fingerprints{2} - if !reflect.DeepEqual(outFPs, want) { - t.Errorf("want %#v, got %#v", want, outFPs) - } - if archived, _, _ := p.hasArchivedMetric(1); archived { - t.Error("want FP 1 not archived") - } - if archived, _, _ := p.hasArchivedMetric(2); !archived { - t.Error("want FP 2 archived") - } -} - -func TestDropArchivedMetricChunkType0(t *testing.T) { - testDropArchivedMetric(t, 0) -} - -func TestDropArchivedMetricChunkType1(t *testing.T) { - testDropArchivedMetric(t, 1) -} - -func TestDropArchivedMetricChunkType2(t *testing.T) { - testDropArchivedMetric(t, 2) -} - -type incrementalBatch struct { - fpToMetric index.FingerprintMetricMapping - expectedLnToLvs index.LabelNameLabelValuesMapping - expectedLpToFps index.LabelPairFingerprintsMapping -} - -func testIndexing(t *testing.T, encoding chunk.Encoding) { - batches := []incrementalBatch{ - { - fpToMetric: index.FingerprintMetricMapping{ - 0: { - model.MetricNameLabel: "metric_0", - "label_1": "value_1", - }, - 1: { - model.MetricNameLabel: "metric_0", - "label_2": "value_2", - "label_3": "value_3", - }, - 2: { - model.MetricNameLabel: "metric_1", - "label_1": "value_2", - }, - }, - expectedLnToLvs: index.LabelNameLabelValuesMapping{ - model.MetricNameLabel: codable.LabelValueSet{ - "metric_0": struct{}{}, - "metric_1": struct{}{}, - }, - "label_1": codable.LabelValueSet{ - "value_1": struct{}{}, - "value_2": struct{}{}, - }, - "label_2": codable.LabelValueSet{ - "value_2": struct{}{}, - }, - "label_3": codable.LabelValueSet{ - "value_3": struct{}{}, - }, - }, - expectedLpToFps: index.LabelPairFingerprintsMapping{ - model.LabelPair{ - Name: model.MetricNameLabel, - Value: "metric_0", - }: codable.FingerprintSet{0: struct{}{}, 1: struct{}{}}, - model.LabelPair{ - Name: model.MetricNameLabel, - Value: "metric_1", - }: codable.FingerprintSet{2: struct{}{}}, - model.LabelPair{ - Name: "label_1", - Value: "value_1", - }: codable.FingerprintSet{0: struct{}{}}, - model.LabelPair{ - Name: "label_1", - Value: "value_2", - }: codable.FingerprintSet{2: struct{}{}}, - model.LabelPair{ - Name: "label_2", - Value: "value_2", - }: codable.FingerprintSet{1: struct{}{}}, - model.LabelPair{ - Name: "label_3", - Value: "value_3", - }: codable.FingerprintSet{1: struct{}{}}, - }, - }, { - fpToMetric: index.FingerprintMetricMapping{ - 3: { - model.MetricNameLabel: "metric_0", - "label_1": "value_3", - }, - 4: { - model.MetricNameLabel: "metric_2", - "label_2": "value_2", - "label_3": "value_1", - }, - 5: { - model.MetricNameLabel: "metric_1", - "label_1": "value_3", - }, - }, - expectedLnToLvs: index.LabelNameLabelValuesMapping{ - model.MetricNameLabel: codable.LabelValueSet{ - "metric_0": struct{}{}, - "metric_1": struct{}{}, - "metric_2": struct{}{}, - }, - "label_1": codable.LabelValueSet{ - "value_1": struct{}{}, - "value_2": struct{}{}, - "value_3": struct{}{}, - }, - "label_2": codable.LabelValueSet{ - "value_2": struct{}{}, - }, - "label_3": codable.LabelValueSet{ - "value_1": struct{}{}, - "value_3": struct{}{}, - }, - }, - expectedLpToFps: index.LabelPairFingerprintsMapping{ - model.LabelPair{ - Name: model.MetricNameLabel, - Value: "metric_0", - }: codable.FingerprintSet{0: struct{}{}, 1: struct{}{}, 3: struct{}{}}, - model.LabelPair{ - Name: model.MetricNameLabel, - Value: "metric_1", - }: codable.FingerprintSet{2: struct{}{}, 5: struct{}{}}, - model.LabelPair{ - Name: model.MetricNameLabel, - Value: "metric_2", - }: codable.FingerprintSet{4: struct{}{}}, - model.LabelPair{ - Name: "label_1", - Value: "value_1", - }: codable.FingerprintSet{0: struct{}{}}, - model.LabelPair{ - Name: "label_1", - Value: "value_2", - }: codable.FingerprintSet{2: struct{}{}}, - model.LabelPair{ - Name: "label_1", - Value: "value_3", - }: codable.FingerprintSet{3: struct{}{}, 5: struct{}{}}, - model.LabelPair{ - Name: "label_2", - Value: "value_2", - }: codable.FingerprintSet{1: struct{}{}, 4: struct{}{}}, - model.LabelPair{ - Name: "label_3", - Value: "value_1", - }: codable.FingerprintSet{4: struct{}{}}, - model.LabelPair{ - Name: "label_3", - Value: "value_3", - }: codable.FingerprintSet{1: struct{}{}}, - }, - }, - } - - p, closer := newTestPersistence(t, encoding) - defer closer.Close() - - indexedFpsToMetrics := index.FingerprintMetricMapping{} - for i, b := range batches { - for fp, m := range b.fpToMetric { - p.indexMetric(fp, m) - p.archiveMetric(fp, m, 1, 2) - indexedFpsToMetrics[fp] = m - } - verifyIndexedState(i, t, b, indexedFpsToMetrics, p) - } - - for i := len(batches) - 1; i >= 0; i-- { - b := batches[i] - verifyIndexedState(i, t, batches[i], indexedFpsToMetrics, p) - for fp, m := range b.fpToMetric { - p.unindexMetric(fp, m) - unarchived, err := p.unarchiveMetric(fp) - if err != nil { - t.Fatal(err) - } - if !unarchived { - t.Errorf("%d. metric not unarchived", i) - } - delete(indexedFpsToMetrics, fp) - } - } -} - -func TestIndexingChunkType0(t *testing.T) { - testIndexing(t, 0) -} - -func TestIndexingChunkType1(t *testing.T) { - testIndexing(t, 1) -} - -func TestIndexingChunkType2(t *testing.T) { - testIndexing(t, 2) -} - -func verifyIndexedState(i int, t *testing.T, b incrementalBatch, indexedFpsToMetrics index.FingerprintMetricMapping, p *persistence) { - p.waitForIndexing() - for fp, m := range indexedFpsToMetrics { - // Compare archived metrics with input metrics. - mOut, err := p.archivedMetric(fp) - if err != nil { - t.Fatal(err) - } - if !mOut.Equal(m) { - t.Errorf("%d. %v: Got: %s; want %s", i, fp, mOut, m) - } - - // Check that archived metrics are in membership index. - has, first, last := p.hasArchivedMetric(fp) - if !has { - t.Errorf("%d. fingerprint %v not found", i, fp) - } - if first != 1 || last != 2 { - t.Errorf( - "%d. %v: Got first: %d, last %d; want first: %d, last %d", - i, fp, first, last, 1, 2, - ) - } - } - - // Compare label name -> label values mappings. - for ln, lvs := range b.expectedLnToLvs { - outLvs, err := p.labelValuesForLabelName(ln) - if err != nil { - t.Fatal(err) - } - - outSet := codable.LabelValueSet{} - for _, lv := range outLvs { - outSet[lv] = struct{}{} - } - - if !reflect.DeepEqual(lvs, outSet) { - t.Errorf("%d. label values don't match. Got: %v; want %v", i, outSet, lvs) - } - } - - // Compare label pair -> fingerprints mappings. - for lp, fps := range b.expectedLpToFps { - outFPs := p.fingerprintsForLabelPair(lp) - - outSet := codable.FingerprintSet{} - for _, fp := range outFPs { - outSet[fp] = struct{}{} - } - - if !reflect.DeepEqual(fps, outSet) { - t.Errorf("%d. %v: fingerprints don't match. Got: %v; want %v", i, lp, outSet, fps) - } - } -} - -func TestQuranatineSeriesFile(t *testing.T) { - p, closer := newTestPersistence(t, 1) - defer closer.Close() - - verify := func(fp model.Fingerprint, seriesFileShouldExist bool, contentHintFile ...string) { - var ( - fpStr = fp.String() - originalFile = p.fileNameForFingerprint(fp) - quarantinedFile = filepath.Join(p.basePath, "orphaned", fpStr[0:seriesDirNameLen], fpStr[seriesDirNameLen:]+seriesFileSuffix) - hintFile = filepath.Join(p.basePath, "orphaned", fpStr[0:seriesDirNameLen], fpStr[seriesDirNameLen:]+hintFileSuffix) - ) - if _, err := os.Stat(originalFile); !os.IsNotExist(err) { - t.Errorf("Expected file %q to not exist.", originalFile) - } - if _, err := os.Stat(quarantinedFile); (os.IsNotExist(err) && seriesFileShouldExist) || (err == nil && !seriesFileShouldExist) { - t.Errorf("Unexpected state of quarantined file %q. Expected it to exist: %t. os.Stat returned: %s.", quarantinedFile, seriesFileShouldExist, err) - } - f, err := os.Open(hintFile) - if err != nil { - t.Errorf("Could not open hint file %q: %s", hintFile, err) - return - } - defer f.Close() - - scanner := bufio.NewScanner(f) - for _, want := range contentHintFile { - if !scanner.Scan() { - t.Errorf("Unexpected end of hint file %q.", hintFile) - return - } - got := scanner.Text() - if want != got { - t.Errorf("Want hint line %q, got %q.", want, got) - } - } - if scanner.Scan() { - t.Errorf("Unexpected spurious content in hint file %q: %q", hintFile, scanner.Text()) - } - } - - if err := p.quarantineSeriesFile(0, nil, nil); err != nil { - t.Error(err) - } - verify(0, false, "[UNKNOWN METRIC]", "[UNKNOWN REASON]") - - if err := p.quarantineSeriesFile( - 1, errors.New("file does not exist"), - nil, - ); err != nil { - t.Error(err) - } - verify(1, false, "[UNKNOWN METRIC]", "file does not exist") - - if err := p.quarantineSeriesFile( - 2, errors.New("file does not exist"), - model.Metric{"foo": "bar", "dings": "bums"}, - ); err != nil { - t.Error(err) - } - verify(2, false, `{dings="bums", foo="bar"}`, "file does not exist") - - if err := p.quarantineSeriesFile( - 3, nil, - model.Metric{"foo": "bar", "dings": "bums"}, - ); err != nil { - t.Error(err) - } - verify(3, false, `{dings="bums", foo="bar"}`, "[UNKNOWN REASON]") - - err := os.Mkdir(filepath.Join(p.basePath, "00"), os.ModePerm) - if err != nil { - t.Fatal(err) - } - f, err := os.Create(p.fileNameForFingerprint(4)) - if err != nil { - t.Fatal(err) - } - f.Close() - - if err := p.quarantineSeriesFile( - 4, errors.New("file exists"), - model.Metric{"sound": "cloud"}, - ); err != nil { - t.Error(err) - } - verify(4, true, `{sound="cloud"}`, "file exists") - - if err := p.quarantineSeriesFile(4, nil, nil); err != nil { - t.Error(err) - } - // Overwrites hint file but leaves series file intact. - verify(4, true, "[UNKNOWN METRIC]", "[UNKNOWN REASON]") - - if err := p.quarantineSeriesFile( - 4, errors.New("file exists"), - model.Metric{"sound": "cloud"}, - ); err != nil { - t.Error(err) - } - // Overwrites everything. - verify(4, true, `{sound="cloud"}`, "file exists") -} - -var fpStrings = []string{ - "b004b821ca50ba26", - "b037c21e884e4fc5", - "b037de1e884e5469", -} - -func BenchmarkLoadChunksSequentially(b *testing.B) { - p := persistence{ - basePath: "fixtures", - bufPool: sync.Pool{New: func() interface{} { return make([]byte, 0, 3*chunkLenWithHeader) }}, - } - sequentialIndexes := make([]int, 47) - for i := range sequentialIndexes { - sequentialIndexes[i] = i - } - - var fp model.Fingerprint - for i := 0; i < b.N; i++ { - for _, s := range fpStrings { - fp, _ = model.FingerprintFromString(s) - cds, err := p.loadChunks(fp, sequentialIndexes, 0) - if err != nil { - b.Error(err) - } - if len(cds) == 0 { - b.Error("could not read any chunks") - } - } - } -} - -func BenchmarkLoadChunksRandomly(b *testing.B) { - p := persistence{ - basePath: "fixtures", - bufPool: sync.Pool{New: func() interface{} { return make([]byte, 0, 3*chunkLenWithHeader) }}, - } - randomIndexes := []int{1, 5, 6, 8, 11, 14, 18, 23, 29, 33, 42, 46} - - var fp model.Fingerprint - for i := 0; i < b.N; i++ { - for _, s := range fpStrings { - fp, _ = model.FingerprintFromString(s) - cds, err := p.loadChunks(fp, randomIndexes, 0) - if err != nil { - b.Error(err) - } - if len(cds) == 0 { - b.Error("could not read any chunks") - } - } - } -} - -func BenchmarkLoadChunkDescs(b *testing.B) { - p := persistence{ - basePath: "fixtures", - } - - var fp model.Fingerprint - for i := 0; i < b.N; i++ { - for _, s := range fpStrings { - fp, _ = model.FingerprintFromString(s) - cds, err := p.loadChunkDescs(fp, 0) - if err != nil { - b.Error(err) - } - if len(cds) == 0 { - b.Error("could not read any chunk descs") - } - } - } -} diff --git a/storage/local/series.go b/storage/local/series.go deleted file mode 100644 index 4a97c3c5c3..0000000000 --- a/storage/local/series.go +++ /dev/null @@ -1,737 +0,0 @@ -// Copyright 2014 The Prometheus Authors -// Licensed under the Apache License, Version 2.0 (the "License"); -// you may not use this file except in compliance with the License. -// You may obtain a copy of the License at -// -// http://www.apache.org/licenses/LICENSE-2.0 -// -// Unless required by applicable law or agreed to in writing, software -// distributed under the License is distributed on an "AS IS" BASIS, -// WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. -// See the License for the specific language governing permissions and -// limitations under the License. - -package local - -import ( - "fmt" - "sort" - "sync" - "time" - - "github.com/prometheus/common/model" - - "github.com/prometheus/prometheus/storage/local/chunk" - "github.com/prometheus/prometheus/storage/metric" -) - -const ( - headChunkTimeout = time.Hour // Close head chunk if not touched for that long. -) - -// fingerprintSeriesPair pairs a fingerprint with a memorySeries pointer. -type fingerprintSeriesPair struct { - fp model.Fingerprint - series *memorySeries -} - -// seriesMap maps fingerprints to memory series. All its methods are -// goroutine-safe. A SeriesMap is effectively is a goroutine-safe version of -// map[model.Fingerprint]*memorySeries. -type seriesMap struct { - mtx sync.RWMutex - m map[model.Fingerprint]*memorySeries -} - -// newSeriesMap returns a newly allocated empty seriesMap. To create a seriesMap -// based on a prefilled map, use an explicit initializer. -func newSeriesMap() *seriesMap { - return &seriesMap{m: make(map[model.Fingerprint]*memorySeries)} -} - -// length returns the number of mappings in the seriesMap. -func (sm *seriesMap) length() int { - sm.mtx.RLock() - defer sm.mtx.RUnlock() - - return len(sm.m) -} - -// get returns a memorySeries for a fingerprint. Return values have the same -// semantics as the native Go map. -func (sm *seriesMap) get(fp model.Fingerprint) (s *memorySeries, ok bool) { - sm.mtx.RLock() - s, ok = sm.m[fp] - // Note that the RUnlock is not done via defer for performance reasons. - // TODO(beorn7): Once https://github.com/golang/go/issues/14939 is - // fixed, revert to the usual defer idiom. - sm.mtx.RUnlock() - return -} - -// put adds a mapping to the seriesMap. It panics if s == nil. -func (sm *seriesMap) put(fp model.Fingerprint, s *memorySeries) { - sm.mtx.Lock() - defer sm.mtx.Unlock() - - if s == nil { - panic("tried to add nil pointer to seriesMap") - } - sm.m[fp] = s -} - -// del removes a mapping from the series Map. -func (sm *seriesMap) del(fp model.Fingerprint) { - sm.mtx.Lock() - defer sm.mtx.Unlock() - - delete(sm.m, fp) -} - -// iter returns a channel that produces all mappings in the seriesMap. The -// channel will be closed once all fingerprints have been received. Not -// consuming all fingerprints from the channel will leak a goroutine. The -// semantics of concurrent modification of seriesMap is the similar as the one -// for iterating over a map with a 'range' clause. However, if the next element -// in iteration order is removed after the current element has been received -// from the channel, it will still be produced by the channel. -func (sm *seriesMap) iter() <-chan fingerprintSeriesPair { - ch := make(chan fingerprintSeriesPair) - go func() { - sm.mtx.RLock() - for fp, s := range sm.m { - sm.mtx.RUnlock() - ch <- fingerprintSeriesPair{fp, s} - sm.mtx.RLock() - } - sm.mtx.RUnlock() - close(ch) - }() - return ch -} - -// fpIter returns a channel that produces all fingerprints in the seriesMap. The -// channel will be closed once all fingerprints have been received. Not -// consuming all fingerprints from the channel will leak a goroutine. The -// semantics of concurrent modification of seriesMap is the similar as the one -// for iterating over a map with a 'range' clause. However, if the next element -// in iteration order is removed after the current element has been received -// from the channel, it will still be produced by the channel. -func (sm *seriesMap) fpIter() <-chan model.Fingerprint { - ch := make(chan model.Fingerprint) - go func() { - sm.mtx.RLock() - for fp := range sm.m { - sm.mtx.RUnlock() - ch <- fp - sm.mtx.RLock() - } - sm.mtx.RUnlock() - close(ch) - }() - return ch -} - -type memorySeries struct { - metric model.Metric - // Sorted by start time, overlapping chunk ranges are forbidden. - chunkDescs []*chunk.Desc - // The index (within chunkDescs above) of the first chunk.Desc that - // points to a non-persisted chunk. If all chunks are persisted, then - // persistWatermark == len(chunkDescs). - persistWatermark int - // The modification time of the series file. The zero value of time.Time - // is used to mark an unknown modification time. - modTime time.Time - // The chunkDescs in memory might not have all the chunkDescs for the - // chunks that are persisted to disk. The missing chunkDescs are all - // contiguous and at the tail end. chunkDescsOffset is the index of the - // chunk on disk that corresponds to the first chunk.Desc in memory. If - // it is 0, the chunkDescs are all loaded. A value of -1 denotes a - // special case: There are chunks on disk, but the offset to the - // chunkDescs in memory is unknown. Also, in this special case, there is - // no overlap between chunks on disk and chunks in memory (implying that - // upon first persisting of a chunk in memory, the offset has to be - // set). - chunkDescsOffset int - // The savedFirstTime field is used as a fallback when the - // chunkDescsOffset is not 0. It can be used to save the FirstTime of the - // first chunk before its chunk desc is evicted. In doubt, this field is - // just set to the oldest possible timestamp. - savedFirstTime model.Time - // The timestamp of the last sample in this series. Needed for fast - // access for federation and to ensure timestamp monotonicity during - // ingestion. - lastTime model.Time - // The last ingested sample value. Needed for fast access for - // federation. - lastSampleValue model.SampleValue - // Whether lastSampleValue has been set already. - lastSampleValueSet bool - // Whether the current head chunk has already been finished. If true, - // the current head chunk must not be modified anymore. - headChunkClosed bool - // Whether the current head chunk is used by an iterator. In that case, - // a non-closed head chunk has to be cloned before more samples are - // appended. - headChunkUsedByIterator bool - // Whether the series is inconsistent with the last checkpoint in a way - // that would require a disk seek during crash recovery. - dirty bool -} - -// newMemorySeries returns a pointer to a newly allocated memorySeries for the -// given metric. chunkDescs and modTime in the new series are set according to -// the provided parameters. chunkDescs can be nil or empty if this is a -// genuinely new time series (i.e. not one that is being unarchived). In that -// case, headChunkClosed is set to false, and firstTime and lastTime are both -// set to model.Earliest. The zero value for modTime can be used if the -// modification time of the series file is unknown (e.g. if this is a genuinely -// new series). -func newMemorySeries(m model.Metric, chunkDescs []*chunk.Desc, modTime time.Time) (*memorySeries, error) { - var err error - firstTime := model.Earliest - lastTime := model.Earliest - if len(chunkDescs) > 0 { - firstTime = chunkDescs[0].FirstTime() - if lastTime, err = chunkDescs[len(chunkDescs)-1].LastTime(); err != nil { - return nil, err - } - } - return &memorySeries{ - metric: m, - chunkDescs: chunkDescs, - headChunkClosed: len(chunkDescs) > 0, - savedFirstTime: firstTime, - lastTime: lastTime, - persistWatermark: len(chunkDescs), - modTime: modTime, - }, nil -} - -// add adds a sample pair to the series. It returns the number of newly -// completed chunks (which are now eligible for persistence). -// -// The caller must have locked the fingerprint of the series. -func (s *memorySeries) add(v model.SamplePair) (int, error) { - if len(s.chunkDescs) == 0 || s.headChunkClosed { - newHead := chunk.NewDesc(chunk.New(), v.Timestamp) - s.chunkDescs = append(s.chunkDescs, newHead) - s.headChunkClosed = false - } else if s.headChunkUsedByIterator && s.head().RefCount() > 1 { - // We only need to clone the head chunk if the current head - // chunk was used in an iterator at all and if the refCount is - // still greater than the 1 we always have because the head - // chunk is not yet persisted. The latter is just an - // approximation. We will still clone unnecessarily if an older - // iterator using a previous version of the head chunk is still - // around and keep the head chunk pinned. We needed to track - // pins by version of the head chunk, which is probably not - // worth the effort. - chunk.Ops.WithLabelValues(chunk.Clone).Inc() - // No locking needed here because a non-persisted head chunk can - // not get evicted concurrently. - s.head().C = s.head().C.Clone() - s.headChunkUsedByIterator = false - } - - chunks, err := s.head().Add(v) - if err != nil { - return 0, err - } - s.head().C = chunks[0] - - for _, c := range chunks[1:] { - s.chunkDescs = append(s.chunkDescs, chunk.NewDesc(c, c.FirstTime())) - } - - // Populate lastTime of now-closed chunks. - for _, cd := range s.chunkDescs[len(s.chunkDescs)-len(chunks) : len(s.chunkDescs)-1] { - cd.MaybePopulateLastTime() - } - - s.lastTime = v.Timestamp - s.lastSampleValue = v.Value - s.lastSampleValueSet = true - return len(chunks) - 1, nil -} - -// maybeCloseHeadChunk closes the head chunk if it has not been touched for the -// duration of headChunkTimeout. It returns whether the head chunk was closed. -// If the head chunk is already closed, the method is a no-op and returns false. -// -// The caller must have locked the fingerprint of the series. -func (s *memorySeries) maybeCloseHeadChunk() bool { - if s.headChunkClosed { - return false - } - if time.Now().Sub(s.lastTime.Time()) > headChunkTimeout { - s.headChunkClosed = true - // Since we cannot modify the head chunk from now on, we - // don't need to bother with cloning anymore. - s.headChunkUsedByIterator = false - s.head().MaybePopulateLastTime() - return true - } - return false -} - -// evictChunkDescs evicts chunkDescs if the chunk is evicted. -// iOldestNotEvicted is the index within the current chunkDescs of the oldest -// chunk that is not evicted. -func (s *memorySeries) evictChunkDescs(iOldestNotEvicted int) { - lenToKeep := len(s.chunkDescs) - iOldestNotEvicted - if lenToKeep < len(s.chunkDescs) { - s.savedFirstTime = s.firstTime() - lenEvicted := len(s.chunkDescs) - lenToKeep - s.chunkDescsOffset += lenEvicted - s.persistWatermark -= lenEvicted - chunk.DescOps.WithLabelValues(chunk.Evict).Add(float64(lenEvicted)) - chunk.NumMemDescs.Sub(float64(lenEvicted)) - s.chunkDescs = append( - make([]*chunk.Desc, 0, lenToKeep), - s.chunkDescs[lenEvicted:]..., - ) - s.dirty = true - } -} - -// dropChunks removes chunkDescs older than t. The caller must have locked the -// fingerprint of the series. -func (s *memorySeries) dropChunks(t model.Time) error { - keepIdx := len(s.chunkDescs) - for i, cd := range s.chunkDescs { - lt, err := cd.LastTime() - if err != nil { - return err - } - if !lt.Before(t) { - keepIdx = i - break - } - } - if keepIdx == len(s.chunkDescs) && !s.headChunkClosed { - // Never drop an open head chunk. - keepIdx-- - } - if keepIdx <= 0 { - // Nothing to drop. - return nil - } - s.chunkDescs = append( - make([]*chunk.Desc, 0, len(s.chunkDescs)-keepIdx), - s.chunkDescs[keepIdx:]..., - ) - s.persistWatermark -= keepIdx - if s.persistWatermark < 0 { - panic("dropped unpersisted chunks from memory") - } - if s.chunkDescsOffset != -1 { - s.chunkDescsOffset += keepIdx - } - chunk.NumMemDescs.Sub(float64(keepIdx)) - s.dirty = true - return nil -} - -// preloadChunks is an internal helper method. -func (s *memorySeries) preloadChunks( - indexes []int, fp model.Fingerprint, mss *MemorySeriesStorage, -) (SeriesIterator, error) { - loadIndexes := []int{} - pinnedChunkDescs := make([]*chunk.Desc, 0, len(indexes)) - for _, idx := range indexes { - cd := s.chunkDescs[idx] - pinnedChunkDescs = append(pinnedChunkDescs, cd) - cd.Pin(mss.evictRequests) // Have to pin everything first to prevent immediate eviction on chunk loading. - if cd.IsEvicted() { - loadIndexes = append(loadIndexes, idx) - } - } - chunk.Ops.WithLabelValues(chunk.Pin).Add(float64(len(pinnedChunkDescs))) - - if len(loadIndexes) > 0 { - if s.chunkDescsOffset == -1 { - panic("requested loading chunks from persistence in a situation where we must not have persisted data for chunk descriptors in memory") - } - chunks, err := mss.loadChunks(fp, loadIndexes, s.chunkDescsOffset) - if err != nil { - // Unpin the chunks since we won't return them as pinned chunks now. - for _, cd := range pinnedChunkDescs { - cd.Unpin(mss.evictRequests) - } - chunk.Ops.WithLabelValues(chunk.Unpin).Add(float64(len(pinnedChunkDescs))) - return nopIter, err - } - for i, c := range chunks { - s.chunkDescs[loadIndexes[i]].SetChunk(c) - } - } - - if !s.headChunkClosed && indexes[len(indexes)-1] == len(s.chunkDescs)-1 { - s.headChunkUsedByIterator = true - } - - curriedQuarantineSeries := func(err error) { - mss.quarantineSeries(fp, s.metric, err) - } - - iter := &boundedIterator{ - it: s.newIterator(pinnedChunkDescs, curriedQuarantineSeries, mss.evictRequests), - start: model.Now().Add(-mss.dropAfter), - } - - return iter, nil -} - -// newIterator returns a new SeriesIterator for the provided chunkDescs (which -// must be pinned). -// -// The caller must have locked the fingerprint of the memorySeries. -func (s *memorySeries) newIterator( - pinnedChunkDescs []*chunk.Desc, - quarantine func(error), - evictRequests chan<- chunk.EvictRequest, -) SeriesIterator { - chunks := make([]chunk.Chunk, 0, len(pinnedChunkDescs)) - for _, cd := range pinnedChunkDescs { - // It's OK to directly access cd.c here (without locking) as the - // series FP is locked and the chunk is pinned. - chunks = append(chunks, cd.C) - } - return &memorySeriesIterator{ - chunks: chunks, - chunkIts: make([]chunk.Iterator, len(chunks)), - quarantine: quarantine, - metric: s.metric, - pinnedChunkDescs: pinnedChunkDescs, - evictRequests: evictRequests, - } -} - -// preloadChunksForInstant preloads chunks for the latest value in the given -// range. If the last sample saved in the memorySeries itself is the latest -// value in the given range, it will in fact preload zero chunks and just take -// that value. -func (s *memorySeries) preloadChunksForInstant( - fp model.Fingerprint, - from model.Time, through model.Time, - mss *MemorySeriesStorage, -) (SeriesIterator, error) { - // If we have a lastSamplePair in the series, and thas last samplePair - // is in the interval, just take it in a singleSampleSeriesIterator. No - // need to pin or load anything. - lastSample := s.lastSamplePair() - if !through.Before(lastSample.Timestamp) && - !from.After(lastSample.Timestamp) && - lastSample != model.ZeroSamplePair { - iter := &boundedIterator{ - it: &singleSampleSeriesIterator{ - samplePair: lastSample, - metric: s.metric, - }, - start: model.Now().Add(-mss.dropAfter), - } - return iter, nil - } - // If we are here, we are out of luck and have to delegate to the more - // expensive method. - return s.preloadChunksForRange(fp, from, through, mss) -} - -// preloadChunksForRange loads chunks for the given range from the persistence. -// The caller must have locked the fingerprint of the series. -func (s *memorySeries) preloadChunksForRange( - fp model.Fingerprint, - from model.Time, through model.Time, - mss *MemorySeriesStorage, -) (SeriesIterator, error) { - firstChunkDescTime := model.Latest - if len(s.chunkDescs) > 0 { - firstChunkDescTime = s.chunkDescs[0].FirstTime() - } - if s.chunkDescsOffset != 0 && from.Before(firstChunkDescTime) { - cds, err := mss.loadChunkDescs(fp, s.persistWatermark) - if err != nil { - return nopIter, err - } - if s.chunkDescsOffset != -1 && len(cds) != s.chunkDescsOffset { - return nopIter, fmt.Errorf( - "unexpected number of chunk descs loaded for fingerprint %v: expected %d, got %d", - fp, s.chunkDescsOffset, len(cds), - ) - } - s.chunkDescs = append(cds, s.chunkDescs...) - s.chunkDescsOffset = 0 - s.persistWatermark += len(cds) - if len(s.chunkDescs) > 0 { - firstChunkDescTime = s.chunkDescs[0].FirstTime() - } - } - - if len(s.chunkDescs) == 0 || through.Before(firstChunkDescTime) { - return nopIter, nil - } - - // Find first chunk with start time after "from". - fromIdx := sort.Search(len(s.chunkDescs), func(i int) bool { - return s.chunkDescs[i].FirstTime().After(from) - }) - // Find first chunk with start time after "through". - throughIdx := sort.Search(len(s.chunkDescs), func(i int) bool { - return s.chunkDescs[i].FirstTime().After(through) - }) - if fromIdx == len(s.chunkDescs) { - // Even the last chunk starts before "from". Find out if the - // series ends before "from" and we don't need to do anything. - lt, err := s.chunkDescs[len(s.chunkDescs)-1].LastTime() - if err != nil { - return nopIter, err - } - if lt.Before(from) { - return nopIter, nil - } - } - if fromIdx > 0 { - fromIdx-- - } - if throughIdx == len(s.chunkDescs) { - throughIdx-- - } - if fromIdx > throughIdx { - // Guard against nonsensical result. The caller will quarantine the series with a meaningful log entry. - return nopIter, fmt.Errorf("fromIdx=%d is greater than throughIdx=%d, likely caused by data corruption", fromIdx, throughIdx) - } - - pinIndexes := make([]int, 0, throughIdx-fromIdx+1) - for i := fromIdx; i <= throughIdx; i++ { - pinIndexes = append(pinIndexes, i) - } - return s.preloadChunks(pinIndexes, fp, mss) -} - -// head returns a pointer to the head chunk descriptor. The caller must have -// locked the fingerprint of the memorySeries. This method will panic if this -// series has no chunk descriptors. -func (s *memorySeries) head() *chunk.Desc { - return s.chunkDescs[len(s.chunkDescs)-1] -} - -// firstTime returns the timestamp of the first sample in the series. -// -// The caller must have locked the fingerprint of the memorySeries. -func (s *memorySeries) firstTime() model.Time { - if s.chunkDescsOffset == 0 && len(s.chunkDescs) > 0 { - return s.chunkDescs[0].FirstTime() - } - return s.savedFirstTime -} - -// lastSamplePair returns the last ingested SamplePair. It returns -// model.ZeroSamplePair if this memorySeries has never received a sample (via the add -// method), which is the case for freshly unarchived series or newly created -// ones and also for all series after a server restart. However, in that case, -// series will most likely be considered stale anyway. -// -// The caller must have locked the fingerprint of the memorySeries. -func (s *memorySeries) lastSamplePair() model.SamplePair { - if !s.lastSampleValueSet { - return model.ZeroSamplePair - } - return model.SamplePair{ - Timestamp: s.lastTime, - Value: s.lastSampleValue, - } -} - -// chunksToPersist returns a slice of chunkDescs eligible for persistence. It's -// the caller's responsibility to actually persist the returned chunks -// afterwards. The method sets the persistWatermark and the dirty flag -// accordingly. -// -// The caller must have locked the fingerprint of the series. -func (s *memorySeries) chunksToPersist() []*chunk.Desc { - newWatermark := len(s.chunkDescs) - if !s.headChunkClosed { - newWatermark-- - } - if newWatermark == s.persistWatermark { - return nil - } - cds := s.chunkDescs[s.persistWatermark:newWatermark] - s.dirty = true - s.persistWatermark = newWatermark - return cds -} - -// memorySeriesIterator implements SeriesIterator. -type memorySeriesIterator struct { - // Last chunk.Iterator used by ValueAtOrBeforeTime. - chunkIt chunk.Iterator - // Caches chunkIterators. - chunkIts []chunk.Iterator - // The actual sample chunks. - chunks []chunk.Chunk - // Call to quarantine the series this iterator belongs to. - quarantine func(error) - // The metric corresponding to the iterator. - metric model.Metric - // Chunks that were pinned for this iterator. - pinnedChunkDescs []*chunk.Desc - // Where to send evict requests when unpinning pinned chunks. - evictRequests chan<- chunk.EvictRequest -} - -// ValueAtOrBeforeTime implements SeriesIterator. -func (it *memorySeriesIterator) ValueAtOrBeforeTime(t model.Time) model.SamplePair { - // The most common case. We are iterating through a chunk. - if it.chunkIt != nil { - containsT, err := it.chunkIt.Contains(t) - if err != nil { - it.quarantine(err) - return model.ZeroSamplePair - } - if containsT { - if it.chunkIt.FindAtOrBefore(t) { - return it.chunkIt.Value() - } - if it.chunkIt.Err() != nil { - it.quarantine(it.chunkIt.Err()) - } - return model.ZeroSamplePair - } - } - - if len(it.chunks) == 0 { - return model.ZeroSamplePair - } - - // Find the last chunk where FirstTime() is before or equal to t. - l := len(it.chunks) - 1 - i := sort.Search(len(it.chunks), func(i int) bool { - return !it.chunks[l-i].FirstTime().After(t) - }) - if i == len(it.chunks) { - // Even the first chunk starts after t. - return model.ZeroSamplePair - } - it.chunkIt = it.chunkIterator(l - i) - if it.chunkIt.FindAtOrBefore(t) { - return it.chunkIt.Value() - } - if it.chunkIt.Err() != nil { - it.quarantine(it.chunkIt.Err()) - } - return model.ZeroSamplePair -} - -// RangeValues implements SeriesIterator. -func (it *memorySeriesIterator) RangeValues(in metric.Interval) []model.SamplePair { - // Find the first chunk for which the first sample is within the interval. - i := sort.Search(len(it.chunks), func(i int) bool { - return !it.chunks[i].FirstTime().Before(in.OldestInclusive) - }) - // Only now check the last timestamp of the previous chunk (which is - // fairly expensive). - if i > 0 { - lt, err := it.chunkIterator(i - 1).LastTimestamp() - if err != nil { - it.quarantine(err) - return nil - } - if !lt.Before(in.OldestInclusive) { - i-- - } - } - - values := []model.SamplePair{} - for j, c := range it.chunks[i:] { - if c.FirstTime().After(in.NewestInclusive) { - break - } - chValues, err := chunk.RangeValues(it.chunkIterator(i+j), in) - if err != nil { - it.quarantine(err) - return nil - } - values = append(values, chValues...) - } - return values -} - -func (it *memorySeriesIterator) Metric() metric.Metric { - return metric.Metric{Metric: it.metric} -} - -// chunkIterator returns the chunk.Iterator for the chunk at position i (and -// creates it if needed). -func (it *memorySeriesIterator) chunkIterator(i int) chunk.Iterator { - chunkIt := it.chunkIts[i] - if chunkIt == nil { - chunkIt = it.chunks[i].NewIterator() - it.chunkIts[i] = chunkIt - } - return chunkIt -} - -func (it *memorySeriesIterator) Close() { - for _, cd := range it.pinnedChunkDescs { - cd.Unpin(it.evictRequests) - } - chunk.Ops.WithLabelValues(chunk.Unpin).Add(float64(len(it.pinnedChunkDescs))) -} - -// singleSampleSeriesIterator implements Series Iterator. It is a "shortcut -// iterator" that returns a single sample only. The sample is saved in the -// iterator itself, so no chunks need to be pinned. -type singleSampleSeriesIterator struct { - samplePair model.SamplePair - metric model.Metric -} - -// ValueAtTime implements SeriesIterator. -func (it *singleSampleSeriesIterator) ValueAtOrBeforeTime(t model.Time) model.SamplePair { - if it.samplePair.Timestamp.After(t) { - return model.ZeroSamplePair - } - return it.samplePair -} - -// RangeValues implements SeriesIterator. -func (it *singleSampleSeriesIterator) RangeValues(in metric.Interval) []model.SamplePair { - if it.samplePair.Timestamp.After(in.NewestInclusive) || - it.samplePair.Timestamp.Before(in.OldestInclusive) { - return []model.SamplePair{} - } - return []model.SamplePair{it.samplePair} -} - -func (it *singleSampleSeriesIterator) Metric() metric.Metric { - return metric.Metric{Metric: it.metric} -} - -// Close implements SeriesIterator. -func (it *singleSampleSeriesIterator) Close() {} - -// nopSeriesIterator implements Series Iterator. It never returns any values. -type nopSeriesIterator struct{} - -// ValueAtTime implements SeriesIterator. -func (i nopSeriesIterator) ValueAtOrBeforeTime(t model.Time) model.SamplePair { - return model.ZeroSamplePair -} - -// RangeValues implements SeriesIterator. -func (i nopSeriesIterator) RangeValues(in metric.Interval) []model.SamplePair { - return []model.SamplePair{} -} - -// Metric implements SeriesIterator. -func (i nopSeriesIterator) Metric() metric.Metric { - return metric.Metric{} -} - -// Close implements SeriesIterator. -func (i nopSeriesIterator) Close() {} - -var nopIter nopSeriesIterator // A nopSeriesIterator for convenience. Can be shared. diff --git a/storage/local/series_test.go b/storage/local/series_test.go deleted file mode 100644 index 86467c2702..0000000000 --- a/storage/local/series_test.go +++ /dev/null @@ -1,63 +0,0 @@ -// Copyright 2016 The Prometheus Authors -// Licensed under the Apache License, Version 2.0 (the "License"); -// you may not use this file except in compliance with the License. -// You may obtain a copy of the License at -// -// http://www.apache.org/licenses/LICENSE-2.0 -// -// Unless required by applicable law or agreed to in writing, software -// distributed under the License is distributed on an "AS IS" BASIS, -// WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. -// See the License for the specific language governing permissions and -// limitations under the License. - -package local - -import ( - "testing" - "time" - - "github.com/prometheus/common/model" -) - -func TestDropChunks(t *testing.T) { - s, err := newMemorySeries(nil, nil, time.Time{}) - if err != nil { - t.Fatal(err) - } - - s.add(model.SamplePair{ - Timestamp: 100, - Value: 42, - }) - s.add(model.SamplePair{ - Timestamp: 110, - Value: 4711, - }) - - err = s.dropChunks(110) - if err != nil { - t.Fatal(err) - } - if len(s.chunkDescs) == 0 { - t.Fatal("chunk dropped too early") - } - - err = s.dropChunks(115) - if err != nil { - t.Fatal(err) - } - if len(s.chunkDescs) == 0 { - t.Fatal("open head chunk dropped") - } - - s.headChunkClosed = true - s.persistWatermark = 1 - err = s.dropChunks(115) - if err != nil { - t.Fatal(err) - } - if len(s.chunkDescs) != 0 { - t.Error("did not drop closed head chunk") - } -} diff --git a/storage/local/storage.go b/storage/local/storage.go deleted file mode 100644 index 8c9f54a4c7..0000000000 --- a/storage/local/storage.go +++ /dev/null @@ -1,1774 +0,0 @@ -// Copyright 2014 The Prometheus Authors -// Licensed under the Apache License, Version 2.0 (the "License"); -// you may not use this file except in compliance with the License. -// You may obtain a copy of the License at -// -// http://www.apache.org/licenses/LICENSE-2.0 -// -// Unless required by applicable law or agreed to in writing, software -// distributed under the License is distributed on an "AS IS" BASIS, -// WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. -// See the License for the specific language governing permissions and -// limitations under the License. - -// Package local contains the local time series storage used by Prometheus. -package local - -import ( - "container/list" - "errors" - "fmt" - "math" - "sort" - "sync" - "sync/atomic" - "time" - - "github.com/prometheus/client_golang/prometheus" - "github.com/prometheus/common/log" - "github.com/prometheus/common/model" - "golang.org/x/net/context" - - "github.com/prometheus/prometheus/storage/local/chunk" - "github.com/prometheus/prometheus/storage/metric" -) - -const ( - evictRequestsCap = 1024 - quarantineRequestsCap = 1024 - - // See waitForNextFP. - fpMaxSweepTime = 6 * time.Hour - fpMaxWaitDuration = 10 * time.Second - - // See waitForNextFP. - maxEvictInterval = time.Minute - - // Constants to control the hysteresis of entering and leaving "rushed - // mode". In rushed mode, the dirty series count is ignored for - // checkpointing, series are maintained as frequently as possible, and - // series files are not synced if the adaptive sync strategy is used. - persintenceUrgencyScoreForEnteringRushedMode = 0.8 - persintenceUrgencyScoreForLeavingRushedMode = 0.7 - - // This factor times -storage.local.memory-chunks is the number of - // memory chunks we tolerate before throttling the storage. It is also a - // basis for calculating the persistenceUrgencyScore. - toleranceFactorMemChunks = 1.1 - // This factor times -storage.local.max-chunks-to-persist is the minimum - // required number of chunks waiting for persistence before the number - // of chunks in memory may influence the persistenceUrgencyScore. (In - // other words: if there are no chunks to persist, it doesn't help chunk - // eviction if we speed up persistence.) - factorMinChunksToPersist = 0.2 - - // Threshold for when to stop using LabelMatchers to retrieve and - // intersect fingerprints. The rationale here is that looking up more - // fingerprints has diminishing returns if we already have narrowed down - // the possible fingerprints significantly. It is then easier to simply - // lookup the metrics for all the fingerprints and directly compare them - // to the matchers. Since a fingerprint lookup for an Equal matcher is - // much less expensive, there is a lower threshold for that case. - // TODO(beorn7): These numbers need to be tweaked, probably a bit lower. - // 5x higher numbers have resulted in slightly worse performance in a - // real-life production scenario. - fpEqualMatchThreshold = 1000 - fpOtherMatchThreshold = 10000 -) - -var ( - numChunksToPersistDesc = prometheus.NewDesc( - prometheus.BuildFQName(namespace, subsystem, "chunks_to_persist"), - "The current number of chunks waiting for persistence.", - nil, nil, - ) - maxChunksToPersistDesc = prometheus.NewDesc( - prometheus.BuildFQName(namespace, subsystem, "max_chunks_to_persist"), - "The maximum number of chunks that can be waiting for persistence before sample ingestion will stop.", - nil, nil, - ) -) - -type quarantineRequest struct { - fp model.Fingerprint - metric model.Metric - reason error -} - -// SyncStrategy is an enum to select a sync strategy for series files. -type SyncStrategy int - -// String implements flag.Value. -func (ss SyncStrategy) String() string { - switch ss { - case Adaptive: - return "adaptive" - case Always: - return "always" - case Never: - return "never" - } - return "" -} - -// Set implements flag.Value. -func (ss *SyncStrategy) Set(s string) error { - switch s { - case "adaptive": - *ss = Adaptive - case "always": - *ss = Always - case "never": - *ss = Never - default: - return fmt.Errorf("invalid sync strategy: %s", s) - } - return nil -} - -// Possible values for SyncStrategy. -const ( - _ SyncStrategy = iota - Never - Always - Adaptive -) - -// A syncStrategy is a function that returns whether series files should be -// synced or not. It does not need to be goroutine safe. -type syncStrategy func() bool - -// A MemorySeriesStorage manages series in memory over time, while also -// interfacing with a persistence layer to make time series data persistent -// across restarts and evictable from memory. -type MemorySeriesStorage struct { - // archiveHighWatermark and numChunksToPersist have to be aligned for atomic operations. - archiveHighWatermark model.Time // No archived series has samples after this time. - numChunksToPersist int64 // The number of chunks waiting for persistence. - maxChunksToPersist int // If numChunksToPersist reaches this threshold, ingestion will be throttled. - rushed bool // Whether the storage is in rushed mode. - rushedMtx sync.Mutex // Protects entering and exiting rushed mode. - throttled chan struct{} // This chan is sent to whenever NeedsThrottling() returns true (for logging). - - fpLocker *fingerprintLocker - fpToSeries *seriesMap - - options *MemorySeriesStorageOptions - - loopStopping, loopStopped chan struct{} - logThrottlingStopped chan struct{} - maxMemoryChunks int - dropAfter time.Duration - checkpointInterval time.Duration - checkpointDirtySeriesLimit int - - persistence *persistence - mapper *fpMapper - - evictList *list.List - evictRequests chan chunk.EvictRequest - evictStopping, evictStopped chan struct{} - - quarantineRequests chan quarantineRequest - quarantineStopping, quarantineStopped chan struct{} - - persistErrors prometheus.Counter - numSeries prometheus.Gauge - seriesOps *prometheus.CounterVec - ingestedSamplesCount prometheus.Counter - discardedSamplesCount *prometheus.CounterVec - nonExistentSeriesMatchesCount prometheus.Counter - maintainSeriesDuration *prometheus.SummaryVec - persistenceUrgencyScore prometheus.Gauge - rushedMode prometheus.Gauge -} - -// MemorySeriesStorageOptions contains options needed by -// NewMemorySeriesStorage. It is not safe to leave any of those at their zero -// values. -type MemorySeriesStorageOptions struct { - MemoryChunks int // How many chunks to keep in memory. - MaxChunksToPersist int // Max number of chunks waiting to be persisted. - PersistenceStoragePath string // Location of persistence files. - PersistenceRetentionPeriod time.Duration // Chunks at least that old are dropped. - CheckpointInterval time.Duration // How often to checkpoint the series map and head chunks. - CheckpointDirtySeriesLimit int // How many dirty series will trigger an early checkpoint. - Dirty bool // Force the storage to consider itself dirty on startup. - PedanticChecks bool // If dirty, perform crash-recovery checks on each series file. - SyncStrategy SyncStrategy // Which sync strategy to apply to series files. - MinShrinkRatio float64 // Minimum ratio a series file has to shrink during truncation. - NumMutexes int // Number of mutexes used for stochastic fingerprint locking. -} - -// NewMemorySeriesStorage returns a newly allocated Storage. Storage.Serve still -// has to be called to start the storage. -func NewMemorySeriesStorage(o *MemorySeriesStorageOptions) *MemorySeriesStorage { - s := &MemorySeriesStorage{ - fpLocker: newFingerprintLocker(o.NumMutexes), - - options: o, - - loopStopping: make(chan struct{}), - loopStopped: make(chan struct{}), - logThrottlingStopped: make(chan struct{}), - throttled: make(chan struct{}, 1), - maxMemoryChunks: o.MemoryChunks, - dropAfter: o.PersistenceRetentionPeriod, - checkpointInterval: o.CheckpointInterval, - checkpointDirtySeriesLimit: o.CheckpointDirtySeriesLimit, - archiveHighWatermark: model.Now().Add(-headChunkTimeout), - - maxChunksToPersist: o.MaxChunksToPersist, - - evictList: list.New(), - evictRequests: make(chan chunk.EvictRequest, evictRequestsCap), - evictStopping: make(chan struct{}), - evictStopped: make(chan struct{}), - - quarantineRequests: make(chan quarantineRequest, quarantineRequestsCap), - quarantineStopping: make(chan struct{}), - quarantineStopped: make(chan struct{}), - - persistErrors: prometheus.NewCounter(prometheus.CounterOpts{ - Namespace: namespace, - Subsystem: subsystem, - Name: "persist_errors_total", - Help: "The total number of errors while persisting chunks.", - }), - numSeries: prometheus.NewGauge(prometheus.GaugeOpts{ - Namespace: namespace, - Subsystem: subsystem, - Name: "memory_series", - Help: "The current number of series in memory.", - }), - seriesOps: prometheus.NewCounterVec( - prometheus.CounterOpts{ - Namespace: namespace, - Subsystem: subsystem, - Name: "series_ops_total", - Help: "The total number of series operations by their type.", - }, - []string{opTypeLabel}, - ), - ingestedSamplesCount: prometheus.NewCounter(prometheus.CounterOpts{ - Namespace: namespace, - Subsystem: subsystem, - Name: "ingested_samples_total", - Help: "The total number of samples ingested.", - }), - discardedSamplesCount: prometheus.NewCounterVec( - prometheus.CounterOpts{ - Namespace: namespace, - Subsystem: subsystem, - Name: "out_of_order_samples_total", - Help: "The total number of samples that were discarded because their timestamps were at or before the last received sample for a series.", - }, - []string{discardReasonLabel}, - ), - nonExistentSeriesMatchesCount: prometheus.NewCounter(prometheus.CounterOpts{ - Namespace: namespace, - Subsystem: subsystem, - Name: "non_existent_series_matches_total", - Help: "How often a non-existent series was referred to during label matching or chunk preloading. This is an indication of outdated label indexes.", - }), - maintainSeriesDuration: prometheus.NewSummaryVec( - prometheus.SummaryOpts{ - Namespace: namespace, - Subsystem: subsystem, - Name: "maintain_series_duration_seconds", - Help: "The duration in seconds it took to perform maintenance on a series.", - }, - []string{seriesLocationLabel}, - ), - persistenceUrgencyScore: prometheus.NewGauge(prometheus.GaugeOpts{ - Namespace: namespace, - Subsystem: subsystem, - Name: "persistence_urgency_score", - Help: "A score of urgency to persist chunks, 0 is least urgent, 1 most.", - }), - rushedMode: prometheus.NewGauge(prometheus.GaugeOpts{ - Namespace: namespace, - Subsystem: subsystem, - Name: "rushed_mode", - Help: "1 if the storage is in rushed mode, 0 otherwise. In rushed mode, the system behaves as if the persistence_urgency_score is 1.", - }), - } - - // Initialize metric vectors. - // TODO(beorn7): Rework once we have a utility function for it in client_golang. - s.discardedSamplesCount.WithLabelValues(outOfOrderTimestamp) - s.discardedSamplesCount.WithLabelValues(duplicateSample) - s.maintainSeriesDuration.WithLabelValues(maintainInMemory) - s.maintainSeriesDuration.WithLabelValues(maintainArchived) - s.seriesOps.WithLabelValues(create) - s.seriesOps.WithLabelValues(archive) - s.seriesOps.WithLabelValues(unarchive) - s.seriesOps.WithLabelValues(memoryPurge) - s.seriesOps.WithLabelValues(archivePurge) - s.seriesOps.WithLabelValues(requestedPurge) - s.seriesOps.WithLabelValues(memoryMaintenance) - s.seriesOps.WithLabelValues(archiveMaintenance) - s.seriesOps.WithLabelValues(completedQurantine) - s.seriesOps.WithLabelValues(droppedQuarantine) - s.seriesOps.WithLabelValues(failedQuarantine) - - return s -} - -// Start implements Storage. -func (s *MemorySeriesStorage) Start() (err error) { - var syncStrategy syncStrategy - switch s.options.SyncStrategy { - case Never: - syncStrategy = func() bool { return false } - case Always: - syncStrategy = func() bool { return true } - case Adaptive: - syncStrategy = func() bool { return s.calculatePersistenceUrgencyScore() < 1 } - default: - panic("unknown sync strategy") - } - - var p *persistence - p, err = newPersistence( - s.options.PersistenceStoragePath, - s.options.Dirty, s.options.PedanticChecks, - syncStrategy, - s.options.MinShrinkRatio, - ) - if err != nil { - return err - } - s.persistence = p - // Persistence must start running before loadSeriesMapAndHeads() is called. - go s.persistence.run() - - defer func() { - if err != nil { - if e := p.close(); e != nil { - log.Errorln("Error closing persistence:", e) - } - } - }() - - log.Info("Loading series map and head chunks...") - s.fpToSeries, s.numChunksToPersist, err = p.loadSeriesMapAndHeads() - if err != nil { - return err - } - log.Infof("%d series loaded.", s.fpToSeries.length()) - s.numSeries.Set(float64(s.fpToSeries.length())) - - s.mapper, err = newFPMapper(s.fpToSeries, p) - if err != nil { - return err - } - - go s.handleEvictList() - go s.handleQuarantine() - go s.logThrottling() - go s.loop() - - return nil -} - -// Stop implements Storage. -func (s *MemorySeriesStorage) Stop() error { - log.Info("Stopping local storage...") - - log.Info("Stopping maintenance loop...") - close(s.loopStopping) - <-s.loopStopped - - log.Info("Stopping series quarantining...") - close(s.quarantineStopping) - <-s.quarantineStopped - - log.Info("Stopping chunk eviction...") - close(s.evictStopping) - <-s.evictStopped - - // One final checkpoint of the series map and the head chunks. - if err := s.persistence.checkpointSeriesMapAndHeads(s.fpToSeries, s.fpLocker); err != nil { - return err - } - if err := s.mapper.checkpoint(); err != nil { - return err - } - - if err := s.persistence.close(); err != nil { - return err - } - log.Info("Local storage stopped.") - return nil -} - -type memorySeriesStorageQuerier struct { - *MemorySeriesStorage -} - -func (memorySeriesStorageQuerier) Close() error { - return nil -} - -// Querier implements the storage interface. -func (s *MemorySeriesStorage) Querier() (Querier, error) { - return memorySeriesStorageQuerier{s}, nil -} - -// WaitForIndexing implements Storage. -func (s *MemorySeriesStorage) WaitForIndexing() { - s.persistence.waitForIndexing() -} - -// LastSampleForLabelMatchers implements Storage. -func (s *MemorySeriesStorage) LastSampleForLabelMatchers(_ context.Context, cutoff model.Time, matcherSets ...metric.LabelMatchers) (model.Vector, error) { - mergedFPs := map[model.Fingerprint]struct{}{} - for _, matchers := range matcherSets { - fps, err := s.fpsForLabelMatchers(cutoff, model.Latest, matchers...) - if err != nil { - return nil, err - } - for fp := range fps { - mergedFPs[fp] = struct{}{} - } - } - - res := make(model.Vector, 0, len(mergedFPs)) - for fp := range mergedFPs { - s.fpLocker.Lock(fp) - - series, ok := s.fpToSeries.get(fp) - if !ok { - // A series could have disappeared between resolving label matchers and here. - s.fpLocker.Unlock(fp) - continue - } - sp := series.lastSamplePair() - res = append(res, &model.Sample{ - Metric: series.metric, - Value: sp.Value, - Timestamp: sp.Timestamp, - }) - s.fpLocker.Unlock(fp) - } - return res, nil -} - -// boundedIterator wraps a SeriesIterator and does not allow fetching -// data from earlier than the configured start time. -type boundedIterator struct { - it SeriesIterator - start model.Time -} - -// ValueAtOrBeforeTime implements the SeriesIterator interface. -func (bit *boundedIterator) ValueAtOrBeforeTime(ts model.Time) model.SamplePair { - if ts < bit.start { - return model.ZeroSamplePair - } - return bit.it.ValueAtOrBeforeTime(ts) -} - -// RangeValues implements the SeriesIterator interface. -func (bit *boundedIterator) RangeValues(interval metric.Interval) []model.SamplePair { - if interval.NewestInclusive < bit.start { - return []model.SamplePair{} - } - if interval.OldestInclusive < bit.start { - interval.OldestInclusive = bit.start - } - return bit.it.RangeValues(interval) -} - -// Metric implements SeriesIterator. -func (bit *boundedIterator) Metric() metric.Metric { - return bit.it.Metric() -} - -// Close implements SeriesIterator. -func (bit *boundedIterator) Close() { - bit.it.Close() -} - -// QueryRange implements Storage. -func (s *MemorySeriesStorage) QueryRange(_ context.Context, from, through model.Time, matchers ...*metric.LabelMatcher) ([]SeriesIterator, error) { - if through.Before(from) { - // In that case, nothing will match. - return nil, nil - } - fpSeriesPairs, err := s.seriesForLabelMatchers(from, through, matchers...) - if err != nil { - return nil, err - } - iterators := make([]SeriesIterator, 0, len(fpSeriesPairs)) - for _, pair := range fpSeriesPairs { - it := s.preloadChunksForRange(pair, from, through) - iterators = append(iterators, it) - } - return iterators, nil -} - -// QueryInstant implements Storage. -func (s *MemorySeriesStorage) QueryInstant(_ context.Context, ts model.Time, stalenessDelta time.Duration, matchers ...*metric.LabelMatcher) ([]SeriesIterator, error) { - if stalenessDelta < 0 { - panic("negative staleness delta") - } - from := ts.Add(-stalenessDelta) - through := ts - - fpSeriesPairs, err := s.seriesForLabelMatchers(from, through, matchers...) - if err != nil { - return nil, err - } - iterators := make([]SeriesIterator, 0, len(fpSeriesPairs)) - for _, pair := range fpSeriesPairs { - it := s.preloadChunksForInstant(pair, from, through) - iterators = append(iterators, it) - } - return iterators, nil -} - -// fingerprintsForLabelPair returns the fingerprints with the given -// LabelPair. If intersectWith is non-nil, the method will only return -// fingerprints that are also contained in intersectsWith. If mergeWith is -// non-nil, the found fingerprints are added to the given map. The returned map -// is the same as the given one. -func (s *MemorySeriesStorage) fingerprintsForLabelPair( - pair model.LabelPair, - mergeWith map[model.Fingerprint]struct{}, - intersectWith map[model.Fingerprint]struct{}, -) map[model.Fingerprint]struct{} { - if mergeWith == nil { - mergeWith = map[model.Fingerprint]struct{}{} - } - for _, fp := range s.persistence.fingerprintsForLabelPair(pair) { - if intersectWith == nil { - mergeWith[fp] = struct{}{} - continue - } - if _, ok := intersectWith[fp]; ok { - mergeWith[fp] = struct{}{} - } - } - return mergeWith -} - -// MetricsForLabelMatchers implements Storage. -func (s *MemorySeriesStorage) MetricsForLabelMatchers( - _ context.Context, - from, through model.Time, - matcherSets ...metric.LabelMatchers, -) ([]metric.Metric, error) { - fpToMetric := map[model.Fingerprint]metric.Metric{} - for _, matchers := range matcherSets { - metrics, err := s.metricsForLabelMatchers(from, through, matchers...) - if err != nil { - return nil, err - } - for fp, m := range metrics { - fpToMetric[fp] = m - } - } - - metrics := make([]metric.Metric, 0, len(fpToMetric)) - for _, m := range fpToMetric { - metrics = append(metrics, m) - } - return metrics, nil -} - -// candidateFPsForLabelMatchers returns candidate FPs for given matchers and remaining matchers to be checked. -func (s *MemorySeriesStorage) candidateFPsForLabelMatchers( - matchers ...*metric.LabelMatcher, -) (map[model.Fingerprint]struct{}, []*metric.LabelMatcher, error) { - sort.Sort(metric.LabelMatchers(matchers)) - - if len(matchers) == 0 || matchers[0].MatchesEmptyString() { - // No matchers at all or even the best matcher matches the empty string. - return nil, nil, nil - } - - var ( - matcherIdx int - candidateFPs map[model.Fingerprint]struct{} - ) - - // Equal matchers. - for ; matcherIdx < len(matchers) && (candidateFPs == nil || len(candidateFPs) > fpEqualMatchThreshold); matcherIdx++ { - m := matchers[matcherIdx] - if m.Type != metric.Equal || m.MatchesEmptyString() { - break - } - candidateFPs = s.fingerprintsForLabelPair( - model.LabelPair{ - Name: m.Name, - Value: m.Value, - }, - nil, - candidateFPs, - ) - if len(candidateFPs) == 0 { - return nil, nil, nil - } - } - - // Other matchers. - for ; matcherIdx < len(matchers) && (candidateFPs == nil || len(candidateFPs) > fpOtherMatchThreshold); matcherIdx++ { - m := matchers[matcherIdx] - if m.MatchesEmptyString() { - break - } - - lvs, err := s.LabelValuesForLabelName(context.TODO(), m.Name) - if err != nil { - return nil, nil, err - } - lvs = m.Filter(lvs) - if len(lvs) == 0 { - return nil, nil, nil - } - fps := map[model.Fingerprint]struct{}{} - for _, lv := range lvs { - s.fingerprintsForLabelPair( - model.LabelPair{ - Name: m.Name, - Value: lv, - }, - fps, - candidateFPs, - ) - } - candidateFPs = fps - if len(candidateFPs) == 0 { - return nil, nil, nil - } - } - return candidateFPs, matchers[matcherIdx:], nil -} - -func (s *MemorySeriesStorage) seriesForLabelMatchers( - from, through model.Time, - matchers ...*metric.LabelMatcher, -) ([]fingerprintSeriesPair, error) { - candidateFPs, matchersToCheck, err := s.candidateFPsForLabelMatchers(matchers...) - if err != nil { - return nil, err - } - - result := []fingerprintSeriesPair{} -FPLoop: - for fp := range candidateFPs { - s.fpLocker.Lock(fp) - series := s.seriesForRange(fp, from, through) - s.fpLocker.Unlock(fp) - - if series == nil { - continue FPLoop - } - - for _, m := range matchersToCheck { - if !m.Match(series.metric[m.Name]) { - continue FPLoop - } - } - result = append(result, fingerprintSeriesPair{fp, series}) - } - return result, nil -} - -func (s *MemorySeriesStorage) fpsForLabelMatchers( - from, through model.Time, - matchers ...*metric.LabelMatcher, -) (map[model.Fingerprint]struct{}, error) { - candidateFPs, matchersToCheck, err := s.candidateFPsForLabelMatchers(matchers...) - if err != nil { - return nil, err - } - -FPLoop: - for fp := range candidateFPs { - s.fpLocker.Lock(fp) - met, _, ok := s.metricForRange(fp, from, through) - s.fpLocker.Unlock(fp) - - if !ok { - delete(candidateFPs, fp) - continue FPLoop - } - - for _, m := range matchersToCheck { - if !m.Match(met[m.Name]) { - delete(candidateFPs, fp) - continue FPLoop - } - } - } - return candidateFPs, nil -} - -func (s *MemorySeriesStorage) metricsForLabelMatchers( - from, through model.Time, - matchers ...*metric.LabelMatcher, -) (map[model.Fingerprint]metric.Metric, error) { - - candidateFPs, matchersToCheck, err := s.candidateFPsForLabelMatchers(matchers...) - if err != nil { - return nil, err - } - - result := map[model.Fingerprint]metric.Metric{} -FPLoop: - for fp := range candidateFPs { - s.fpLocker.Lock(fp) - met, _, ok := s.metricForRange(fp, from, through) - s.fpLocker.Unlock(fp) - - if !ok { - continue FPLoop - } - - for _, m := range matchersToCheck { - if !m.Match(met[m.Name]) { - continue FPLoop - } - } - result[fp] = metric.Metric{Metric: met} - } - return result, nil -} - -// metricForRange returns the metric for the given fingerprint if the -// corresponding time series has samples between 'from' and 'through', together -// with a pointer to the series if it is in memory already. For a series that -// does not have samples between 'from' and 'through', the returned bool is -// false. For an archived series that does contain samples between 'from' and -// 'through', it returns (metric, nil, true). -// -// The caller must have locked the fp. -func (s *MemorySeriesStorage) metricForRange( - fp model.Fingerprint, - from, through model.Time, -) (model.Metric, *memorySeries, bool) { - series, ok := s.fpToSeries.get(fp) - if ok { - if series.lastTime.Before(from) || series.firstTime().After(through) { - return nil, nil, false - } - return series.metric, series, true - } - // From here on, we are only concerned with archived metrics. - // If the high watermark of archived series is before 'from', we are done. - watermark := model.Time(atomic.LoadInt64((*int64)(&s.archiveHighWatermark))) - if watermark < from { - return nil, nil, false - } - if from.After(model.Earliest) || through.Before(model.Latest) { - // The range lookup is relatively cheap, so let's do it first if - // we have a chance the archived metric is not in the range. - has, first, last := s.persistence.hasArchivedMetric(fp) - if !has { - s.nonExistentSeriesMatchesCount.Inc() - return nil, nil, false - } - if first.After(through) || last.Before(from) { - return nil, nil, false - } - } - - metric, err := s.persistence.archivedMetric(fp) - if err != nil { - // archivedMetric has already flagged the storage as dirty in this case. - return nil, nil, false - } - return metric, nil, true -} - -// LabelValuesForLabelName implements Storage. -func (s *MemorySeriesStorage) LabelValuesForLabelName(_ context.Context, labelName model.LabelName) (model.LabelValues, error) { - return s.persistence.labelValuesForLabelName(labelName) -} - -// DropMetricsForLabelMatchers implements Storage. -func (s *MemorySeriesStorage) DropMetricsForLabelMatchers(_ context.Context, matchers ...*metric.LabelMatcher) (int, error) { - fps, err := s.fpsForLabelMatchers(model.Earliest, model.Latest, matchers...) - if err != nil { - return 0, err - } - for fp := range fps { - s.purgeSeries(fp, nil, nil) - } - return len(fps), nil -} - -var ( - // ErrOutOfOrderSample is returned if a sample has a timestamp before the latest - // timestamp in the series it is appended to. - ErrOutOfOrderSample = fmt.Errorf("sample timestamp out of order") - // ErrDuplicateSampleForTimestamp is returned if a sample has the same - // timestamp as the latest sample in the series it is appended to but a - // different value. (Appending an identical sample is a no-op and does - // not cause an error.) - ErrDuplicateSampleForTimestamp = fmt.Errorf("sample with repeated timestamp but different value") -) - -// Append implements Storage. -func (s *MemorySeriesStorage) Append(sample *model.Sample) error { - for ln, lv := range sample.Metric { - if len(lv) == 0 { - delete(sample.Metric, ln) - } - } - rawFP := sample.Metric.FastFingerprint() - s.fpLocker.Lock(rawFP) - fp := s.mapper.mapFP(rawFP, sample.Metric) - defer func() { - s.fpLocker.Unlock(fp) - }() // Func wrapper because fp might change below. - if fp != rawFP { - // Switch locks. - s.fpLocker.Unlock(rawFP) - s.fpLocker.Lock(fp) - } - series, err := s.getOrCreateSeries(fp, sample.Metric) - if err != nil { - return err // getOrCreateSeries took care of quarantining already. - } - - if sample.Timestamp == series.lastTime { - // Don't report "no-op appends", i.e. where timestamp and sample - // value are the same as for the last append, as they are a - // common occurrence when using client-side timestamps - // (e.g. Pushgateway or federation). - if sample.Timestamp == series.lastTime && - series.lastSampleValueSet && - sample.Value.Equal(series.lastSampleValue) { - return nil - } - s.discardedSamplesCount.WithLabelValues(duplicateSample).Inc() - return ErrDuplicateSampleForTimestamp // Caused by the caller. - } - if sample.Timestamp < series.lastTime { - s.discardedSamplesCount.WithLabelValues(outOfOrderTimestamp).Inc() - return ErrOutOfOrderSample // Caused by the caller. - } - completedChunksCount, err := series.add(model.SamplePair{ - Value: sample.Value, - Timestamp: sample.Timestamp, - }) - if err != nil { - s.quarantineSeries(fp, sample.Metric, err) - return err - } - s.ingestedSamplesCount.Inc() - s.incNumChunksToPersist(completedChunksCount) - - return nil -} - -// NeedsThrottling implements Storage. -func (s *MemorySeriesStorage) NeedsThrottling() bool { - if s.getNumChunksToPersist() > s.maxChunksToPersist || - float64(atomic.LoadInt64(&chunk.NumMemChunks)) > float64(s.maxMemoryChunks)*toleranceFactorMemChunks { - select { - case s.throttled <- struct{}{}: - default: // Do nothing, signal already pending. - } - return true - } - return false -} - -// logThrottling handles logging of throttled events and has to be started as a -// goroutine. It stops once s.loopStopping is closed. -// -// Logging strategy: Whenever Throttle() is called and returns true, an signal -// is sent to s.throttled. If that happens for the first time, an Error is -// logged that the storage is now throttled. As long as signals continues to be -// sent via s.throttled at least once per minute, nothing else is logged. Once -// no signal has arrived for a minute, an Info is logged that the storage is not -// throttled anymore. This resets things to the initial state, i.e. once a -// signal arrives again, the Error will be logged again. -func (s *MemorySeriesStorage) logThrottling() { - timer := time.NewTimer(time.Minute) - timer.Stop() - - // Signal exit of the goroutine. Currently only needed by test code. - defer close(s.logThrottlingStopped) - - for { - select { - case <-s.throttled: - if !timer.Reset(time.Minute) { - log. - With("chunksToPersist", s.getNumChunksToPersist()). - With("maxChunksToPersist", s.maxChunksToPersist). - With("memoryChunks", atomic.LoadInt64(&chunk.NumMemChunks)). - With("maxToleratedMemChunks", int(float64(s.maxMemoryChunks)*toleranceFactorMemChunks)). - Error("Storage needs throttling. Scrapes and rule evaluations will be skipped.") - } - case <-timer.C: - log. - With("chunksToPersist", s.getNumChunksToPersist()). - With("maxChunksToPersist", s.maxChunksToPersist). - With("memoryChunks", atomic.LoadInt64(&chunk.NumMemChunks)). - With("maxToleratedMemChunks", int(float64(s.maxMemoryChunks)*toleranceFactorMemChunks)). - Info("Storage does not need throttling anymore.") - case <-s.loopStopping: - return - } - } -} - -func (s *MemorySeriesStorage) getOrCreateSeries(fp model.Fingerprint, m model.Metric) (*memorySeries, error) { - series, ok := s.fpToSeries.get(fp) - if !ok { - var cds []*chunk.Desc - var modTime time.Time - unarchived, err := s.persistence.unarchiveMetric(fp) - if err != nil { - log.Errorf("Error unarchiving fingerprint %v (metric %v): %v", fp, m, err) - return nil, err - } - if unarchived { - s.seriesOps.WithLabelValues(unarchive).Inc() - // We have to load chunk.Descs anyway to do anything with - // the series, so let's do it right now so that we don't - // end up with a series without any chunk.Descs for a - // while (which is confusing as it makes the series - // appear as archived or purged). - cds, err = s.loadChunkDescs(fp, 0) - if err == nil && len(cds) == 0 { - err = fmt.Errorf("unarchived fingerprint %v (metric %v) has no chunks on disk", fp, m) - } - if err != nil { - s.quarantineSeries(fp, m, err) - return nil, err - } - modTime = s.persistence.seriesFileModTime(fp) - } else { - // This was a genuinely new series, so index the metric. - s.persistence.indexMetric(fp, m) - s.seriesOps.WithLabelValues(create).Inc() - } - series, err = newMemorySeries(m, cds, modTime) - if err != nil { - s.quarantineSeries(fp, m, err) - return nil, err - } - s.fpToSeries.put(fp, series) - s.numSeries.Inc() - } - return series, nil -} - -// seriesForRange is a helper method for seriesForLabelMatchers. -// -// The caller must have locked the fp. -func (s *MemorySeriesStorage) seriesForRange( - fp model.Fingerprint, - from model.Time, through model.Time, -) *memorySeries { - metric, series, ok := s.metricForRange(fp, from, through) - if !ok { - return nil - } - if series == nil { - series, _ = s.getOrCreateSeries(fp, metric) - // getOrCreateSeries took care of quarantining already, so ignore the error. - } - return series -} - -func (s *MemorySeriesStorage) preloadChunksForRange( - pair fingerprintSeriesPair, - from model.Time, through model.Time, -) SeriesIterator { - fp, series := pair.fp, pair.series - if series == nil { - return nopIter - } - - s.fpLocker.Lock(fp) - defer s.fpLocker.Unlock(fp) - - iter, err := series.preloadChunksForRange(fp, from, through, s) - if err != nil { - s.quarantineSeries(fp, series.metric, err) - return nopIter - } - return iter -} - -func (s *MemorySeriesStorage) preloadChunksForInstant( - pair fingerprintSeriesPair, - from model.Time, through model.Time, -) SeriesIterator { - fp, series := pair.fp, pair.series - if series == nil { - return nopIter - } - - s.fpLocker.Lock(fp) - defer s.fpLocker.Unlock(fp) - - iter, err := series.preloadChunksForInstant(fp, from, through, s) - if err != nil { - s.quarantineSeries(fp, series.metric, err) - return nopIter - } - return iter -} - -func (s *MemorySeriesStorage) handleEvictList() { - ticker := time.NewTicker(maxEvictInterval) - count := 0 - - for { - // To batch up evictions a bit, this tries evictions at least - // once per evict interval, but earlier if the number of evict - // requests with evict==true that have happened since the last - // evict run is more than maxMemoryChunks/1000. - select { - case req := <-s.evictRequests: - if req.Evict { - req.Desc.EvictListElement = s.evictList.PushBack(req.Desc) - count++ - if count > s.maxMemoryChunks/1000 { - s.maybeEvict() - count = 0 - } - } else { - if req.Desc.EvictListElement != nil { - s.evictList.Remove(req.Desc.EvictListElement) - req.Desc.EvictListElement = nil - } - } - case <-ticker.C: - if s.evictList.Len() > 0 { - s.maybeEvict() - } - case <-s.evictStopping: - // Drain evictRequests forever in a goroutine to not let - // requesters hang. - go func() { - for { - <-s.evictRequests - } - }() - ticker.Stop() - log.Info("Chunk eviction stopped.") - close(s.evictStopped) - return - } - } -} - -// maybeEvict is a local helper method. Must only be called by handleEvictList. -func (s *MemorySeriesStorage) maybeEvict() { - numChunksToEvict := int(atomic.LoadInt64(&chunk.NumMemChunks)) - s.maxMemoryChunks - if numChunksToEvict <= 0 { - return - } - chunkDescsToEvict := make([]*chunk.Desc, numChunksToEvict) - for i := range chunkDescsToEvict { - e := s.evictList.Front() - if e == nil { - break - } - cd := e.Value.(*chunk.Desc) - cd.EvictListElement = nil - chunkDescsToEvict[i] = cd - s.evictList.Remove(e) - } - // Do the actual eviction in a goroutine as we might otherwise deadlock, - // in the following way: A chunk was Unpinned completely and therefore - // scheduled for eviction. At the time we actually try to evict it, - // another goroutine is pinning the chunk. The pinning goroutine has - // currently locked the chunk and tries to send the evict request (to - // remove the chunk from the evict list) to the evictRequests - // channel. The send blocks because evictRequests is full. However, the - // goroutine that is supposed to empty the channel is waiting for the - // Chunk.Desc lock to try to evict the chunk. - go func() { - for _, cd := range chunkDescsToEvict { - if cd == nil { - break - } - cd.MaybeEvict() - // We don't care if the eviction succeeds. If the chunk - // was pinned in the meantime, it will be added to the - // evict list once it gets Unpinned again. - } - }() -} - -// waitForNextFP waits an estimated duration, after which we want to process -// another fingerprint so that we will process all fingerprints in a tenth of -// s.dropAfter assuming that the system is doing nothing else, e.g. if we want -// to drop chunks after 40h, we want to cycle through all fingerprints within -// 4h. The estimation is based on the total number of fingerprints as passed -// in. However, the maximum sweep time is capped at fpMaxSweepTime. Also, the -// method will never wait for longer than fpMaxWaitDuration. -// -// The maxWaitDurationFactor can be used to reduce the waiting time if a faster -// processing is required (for example because unpersisted chunks pile up too -// much). -// -// Normally, the method returns true once the wait duration has passed. However, -// if s.loopStopped is closed, it will return false immediately. -func (s *MemorySeriesStorage) waitForNextFP(numberOfFPs int, maxWaitDurationFactor float64) bool { - d := fpMaxWaitDuration - if numberOfFPs != 0 { - sweepTime := s.dropAfter / 10 - if sweepTime > fpMaxSweepTime { - sweepTime = fpMaxSweepTime - } - calculatedWait := time.Duration(float64(sweepTime) / float64(numberOfFPs) * maxWaitDurationFactor) - if calculatedWait < d { - d = calculatedWait - } - } - if d == 0 { - return true - } - t := time.NewTimer(d) - select { - case <-t.C: - return true - case <-s.loopStopping: - return false - } -} - -// cycleThroughMemoryFingerprints returns a channel that emits fingerprints for -// series in memory in a throttled fashion. It continues to cycle through all -// fingerprints in memory until s.loopStopping is closed. -func (s *MemorySeriesStorage) cycleThroughMemoryFingerprints() chan model.Fingerprint { - memoryFingerprints := make(chan model.Fingerprint) - go func() { - var fpIter <-chan model.Fingerprint - - defer func() { - if fpIter != nil { - for range fpIter { - // Consume the iterator. - } - } - close(memoryFingerprints) - }() - - for { - // Initial wait, also important if there are no FPs yet. - if !s.waitForNextFP(s.fpToSeries.length(), 1) { - return - } - begin := time.Now() - fpIter = s.fpToSeries.fpIter() - count := 0 - for fp := range fpIter { - select { - case memoryFingerprints <- fp: - case <-s.loopStopping: - return - } - // Reduce the wait time according to the urgency score. - s.waitForNextFP(s.fpToSeries.length(), 1-s.calculatePersistenceUrgencyScore()) - count++ - } - if count > 0 { - log.Infof( - "Completed maintenance sweep through %d in-memory fingerprints in %v.", - count, time.Since(begin), - ) - } - } - }() - - return memoryFingerprints -} - -// cycleThroughArchivedFingerprints returns a channel that emits fingerprints -// for archived series in a throttled fashion. It continues to cycle through all -// archived fingerprints until s.loopStopping is closed. -func (s *MemorySeriesStorage) cycleThroughArchivedFingerprints() chan model.Fingerprint { - archivedFingerprints := make(chan model.Fingerprint) - go func() { - defer close(archivedFingerprints) - - for { - archivedFPs, err := s.persistence.fingerprintsModifiedBefore( - model.Now().Add(-s.dropAfter), - ) - if err != nil { - log.Error("Failed to lookup archived fingerprint ranges: ", err) - s.waitForNextFP(0, 1) - continue - } - // Initial wait, also important if there are no FPs yet. - if !s.waitForNextFP(len(archivedFPs), 1) { - return - } - begin := time.Now() - for _, fp := range archivedFPs { - select { - case archivedFingerprints <- fp: - case <-s.loopStopping: - return - } - // Never speed up maintenance of archived FPs. - s.waitForNextFP(len(archivedFPs), 1) - } - if len(archivedFPs) > 0 { - log.Infof( - "Completed maintenance sweep through %d archived fingerprints in %v.", - len(archivedFPs), time.Since(begin), - ) - } - } - }() - return archivedFingerprints -} - -func (s *MemorySeriesStorage) loop() { - checkpointTimer := time.NewTimer(s.checkpointInterval) - - dirtySeriesCount := 0 - - defer func() { - checkpointTimer.Stop() - log.Info("Maintenance loop stopped.") - close(s.loopStopped) - }() - - memoryFingerprints := s.cycleThroughMemoryFingerprints() - archivedFingerprints := s.cycleThroughArchivedFingerprints() - -loop: - for { - select { - case <-s.loopStopping: - break loop - case <-checkpointTimer.C: - err := s.persistence.checkpointSeriesMapAndHeads(s.fpToSeries, s.fpLocker) - if err != nil { - log.Errorln("Error while checkpointing:", err) - } else { - dirtySeriesCount = 0 - } - // If a checkpoint takes longer than checkpointInterval, unluckily timed - // combination with the Reset(0) call below can lead to a case where a - // time is lurking in C leading to repeated checkpointing without break. - select { - case <-checkpointTimer.C: // Get rid of the lurking time. - default: - } - checkpointTimer.Reset(s.checkpointInterval) - case fp := <-memoryFingerprints: - if s.maintainMemorySeries(fp, model.Now().Add(-s.dropAfter)) { - dirtySeriesCount++ - // Check if we have enough "dirty" series so that we need an early checkpoint. - // However, if we are already behind persisting chunks, creating a checkpoint - // would be counterproductive, as it would slow down chunk persisting even more, - // while in a situation like that, where we are clearly lacking speed of disk - // maintenance, the best we can do for crash recovery is to persist chunks as - // quickly as possible. So only checkpoint if the urgency score is < 1. - if dirtySeriesCount >= s.checkpointDirtySeriesLimit && - s.calculatePersistenceUrgencyScore() < 1 { - checkpointTimer.Reset(0) - } - } - case fp := <-archivedFingerprints: - s.maintainArchivedSeries(fp, model.Now().Add(-s.dropAfter)) - } - } - // Wait until both channels are closed. - for range memoryFingerprints { - } - for range archivedFingerprints { - } -} - -// maintainMemorySeries maintains a series that is in memory (i.e. not -// archived). It returns true if the method has changed from clean to dirty -// (i.e. it is inconsistent with the latest checkpoint now so that in case of a -// crash a recovery operation that requires a disk seek needed to be applied). -// -// The method first closes the head chunk if it was not touched for the duration -// of headChunkTimeout. -// -// Then it determines the chunks that need to be purged and the chunks that need -// to be persisted. Depending on the result, it does the following: -// -// - If all chunks of a series need to be purged, the whole series is deleted -// for good and the method returns false. (Detecting non-existence of a series -// file does not require a disk seek.) -// -// - If any chunks need to be purged (but not all of them), it purges those -// chunks from memory and rewrites the series file on disk, leaving out the -// purged chunks and appending all chunks not yet persisted (with the exception -// of a still open head chunk). -// -// - If no chunks on disk need to be purged, but chunks need to be persisted, -// those chunks are simply appended to the existing series file (or the file is -// created if it does not exist yet). -// -// - If no chunks need to be purged and no chunks need to be persisted, nothing -// happens in this step. -// -// Next, the method checks if all chunks in the series are evicted. In that -// case, it archives the series and returns true. -// -// Finally, it evicts chunk.Descs if there are too many. -func (s *MemorySeriesStorage) maintainMemorySeries( - fp model.Fingerprint, beforeTime model.Time, -) (becameDirty bool) { - defer func(begin time.Time) { - s.maintainSeriesDuration.WithLabelValues(maintainInMemory).Observe( - time.Since(begin).Seconds(), - ) - }(time.Now()) - - s.fpLocker.Lock(fp) - defer s.fpLocker.Unlock(fp) - - series, ok := s.fpToSeries.get(fp) - if !ok { - // Series is actually not in memory, perhaps archived or dropped in the meantime. - return false - } - - defer s.seriesOps.WithLabelValues(memoryMaintenance).Inc() - - if series.maybeCloseHeadChunk() { - s.incNumChunksToPersist(1) - } - - seriesWasDirty := series.dirty - - if s.writeMemorySeries(fp, series, beforeTime) { - // Series is gone now, we are done. - return false - } - - iOldestNotEvicted := -1 - for i, cd := range series.chunkDescs { - if !cd.IsEvicted() { - iOldestNotEvicted = i - break - } - } - - // Archive if all chunks are evicted. Also make sure the last sample has - // an age of at least headChunkTimeout (which is very likely anyway). - if iOldestNotEvicted == -1 && model.Now().Sub(series.lastTime) > headChunkTimeout { - s.fpToSeries.del(fp) - s.numSeries.Dec() - s.persistence.archiveMetric(fp, series.metric, series.firstTime(), series.lastTime) - s.seriesOps.WithLabelValues(archive).Inc() - oldWatermark := atomic.LoadInt64((*int64)(&s.archiveHighWatermark)) - if oldWatermark < int64(series.lastTime) { - if !atomic.CompareAndSwapInt64( - (*int64)(&s.archiveHighWatermark), - oldWatermark, int64(series.lastTime), - ) { - panic("s.archiveHighWatermark modified outside of maintainMemorySeries") - } - } - return - } - // If we are here, the series is not archived, so check for Chunk.Desc - // eviction next. - series.evictChunkDescs(iOldestNotEvicted) - - return series.dirty && !seriesWasDirty -} - -// writeMemorySeries (re-)writes a memory series file. While doing so, it drops -// chunks older than beforeTime from both the series file (if it exists) as well -// as from memory. The provided chunksToPersist are appended to the newly -// written series file. If no chunks need to be purged, but chunksToPersist is -// not empty, those chunks are simply appended to the series file. If the series -// contains no chunks after dropping old chunks, it is purged entirely. In that -// case, the method returns true. -// -// The caller must have locked the fp. -func (s *MemorySeriesStorage) writeMemorySeries( - fp model.Fingerprint, series *memorySeries, beforeTime model.Time, -) bool { - var ( - persistErr error - cds = series.chunksToPersist() - ) - - defer func() { - if persistErr != nil { - s.quarantineSeries(fp, series.metric, persistErr) - s.persistErrors.Inc() - } - // The following is done even in case of an error to ensure - // correct counter bookkeeping and to not pin chunks in memory - // that belong to a series that is scheduled for quarantine - // anyway. - for _, cd := range cds { - cd.Unpin(s.evictRequests) - } - s.incNumChunksToPersist(-len(cds)) - chunk.Ops.WithLabelValues(chunk.PersistAndUnpin).Add(float64(len(cds))) - series.modTime = s.persistence.seriesFileModTime(fp) - }() - - // Get the actual chunks from underneath the chunk.Descs. - // No lock required as chunks still to persist cannot be evicted. - chunks := make([]chunk.Chunk, len(cds)) - for i, cd := range cds { - chunks[i] = cd.C - } - - if !series.firstTime().Before(beforeTime) { - // Oldest sample not old enough, just append chunks, if any. - if len(cds) == 0 { - return false - } - var offset int - offset, persistErr = s.persistence.persistChunks(fp, chunks) - if persistErr != nil { - return false - } - if series.chunkDescsOffset == -1 { - // This is the first chunk persisted for a newly created - // series that had prior chunks on disk. Finally, we can - // set the chunkDescsOffset. - series.chunkDescsOffset = offset - } - return false - } - - newFirstTime, offset, numDroppedFromPersistence, allDroppedFromPersistence, persistErr := - s.persistence.dropAndPersistChunks(fp, beforeTime, chunks) - if persistErr != nil { - return false - } - if persistErr = series.dropChunks(beforeTime); persistErr != nil { - return false - } - if len(series.chunkDescs) == 0 && allDroppedFromPersistence { - // All chunks dropped from both memory and persistence. Delete the series for good. - s.fpToSeries.del(fp) - s.numSeries.Dec() - s.seriesOps.WithLabelValues(memoryPurge).Inc() - s.persistence.unindexMetric(fp, series.metric) - return true - } - series.savedFirstTime = newFirstTime - if series.chunkDescsOffset == -1 { - series.chunkDescsOffset = offset - } else { - series.chunkDescsOffset -= numDroppedFromPersistence - if series.chunkDescsOffset < 0 { - persistErr = errors.New("dropped more chunks from persistence than from memory") - series.chunkDescsOffset = -1 - } - } - return false -} - -// maintainArchivedSeries drops chunks older than beforeTime from an archived -// series. If the series contains no chunks after that, it is purged entirely. -func (s *MemorySeriesStorage) maintainArchivedSeries(fp model.Fingerprint, beforeTime model.Time) { - defer func(begin time.Time) { - s.maintainSeriesDuration.WithLabelValues(maintainArchived).Observe( - time.Since(begin).Seconds(), - ) - }(time.Now()) - - s.fpLocker.Lock(fp) - defer s.fpLocker.Unlock(fp) - - has, firstTime, lastTime := s.persistence.hasArchivedMetric(fp) - if !has || !firstTime.Before(beforeTime) { - // Oldest sample not old enough, or metric purged or unarchived in the meantime. - return - } - - defer s.seriesOps.WithLabelValues(archiveMaintenance).Inc() - - newFirstTime, _, _, allDropped, err := s.persistence.dropAndPersistChunks(fp, beforeTime, nil) - if err != nil { - log.Error("Error dropping persisted chunks: ", err) - } - if allDropped { - s.persistence.purgeArchivedMetric(fp) // Ignoring error. Nothing we can do. - s.seriesOps.WithLabelValues(archivePurge).Inc() - return - } - if err := s.persistence.updateArchivedTimeRange(fp, newFirstTime, lastTime); err != nil { - log.Errorf("Error updating archived time range for fingerprint %v: %s", fp, err) - } -} - -// See persistence.loadChunks for detailed explanation. -func (s *MemorySeriesStorage) loadChunks(fp model.Fingerprint, indexes []int, indexOffset int) ([]chunk.Chunk, error) { - return s.persistence.loadChunks(fp, indexes, indexOffset) -} - -// See persistence.loadChunkDescs for detailed explanation. -func (s *MemorySeriesStorage) loadChunkDescs(fp model.Fingerprint, offsetFromEnd int) ([]*chunk.Desc, error) { - return s.persistence.loadChunkDescs(fp, offsetFromEnd) -} - -// getNumChunksToPersist returns numChunksToPersist in a goroutine-safe way. -func (s *MemorySeriesStorage) getNumChunksToPersist() int { - return int(atomic.LoadInt64(&s.numChunksToPersist)) -} - -// incNumChunksToPersist increments numChunksToPersist in a goroutine-safe way. Use a -// negative 'by' to decrement. -func (s *MemorySeriesStorage) incNumChunksToPersist(by int) { - atomic.AddInt64(&s.numChunksToPersist, int64(by)) -} - -// calculatePersistenceUrgencyScore calculates and returns an urgency score for -// the speed of persisting chunks. The score is between 0 and 1, where 0 means -// no urgency at all and 1 means highest urgency. -// -// The score is the maximum of the two following sub-scores: -// -// (1) The first sub-score is the number of chunks waiting for persistence -// divided by the maximum number of chunks allowed to be waiting for -// persistence. -// -// (2) If there are more chunks in memory than allowed AND there are more chunks -// waiting for persistence than factorMinChunksToPersist times -// -storage.local.max-chunks-to-persist, then the second sub-score is the -// fraction the number of memory chunks has reached between -// -storage.local.memory-chunks and toleranceFactorForMemChunks times -// -storage.local.memory-chunks. -// -// Should the score ever hit persintenceUrgencyScoreForEnteringRushedMode, the -// storage locks into "rushed mode", in which the returned score is always -// bumped up to 1 until the non-bumped score is below -// persintenceUrgencyScoreForLeavingRushedMode. -// -// This method is not goroutine-safe, but it is only ever called by the single -// goroutine that is in charge of series maintenance. According to the returned -// score, series maintenance should be sped up. If a score of 1 is returned, -// checkpointing based on dirty-series count should be disabled, and series -// files should not by synced anymore provided the user has specified the -// adaptive sync strategy. -func (s *MemorySeriesStorage) calculatePersistenceUrgencyScore() float64 { - s.rushedMtx.Lock() - defer s.rushedMtx.Unlock() - - var ( - chunksToPersist = float64(s.getNumChunksToPersist()) - maxChunksToPersist = float64(s.maxChunksToPersist) - memChunks = float64(atomic.LoadInt64(&chunk.NumMemChunks)) - maxMemChunks = float64(s.maxMemoryChunks) - ) - score := chunksToPersist / maxChunksToPersist - if chunksToPersist > maxChunksToPersist*factorMinChunksToPersist { - score = math.Max( - score, - (memChunks/maxMemChunks-1)/(toleranceFactorMemChunks-1), - ) - } - if score > 1 { - score = 1 - } - s.persistenceUrgencyScore.Set(score) - - if s.rushed { - // We are already in rushed mode. If the score is still above - // persintenceUrgencyScoreForLeavingRushedMode, return 1 and - // leave things as they are. - if score > persintenceUrgencyScoreForLeavingRushedMode { - return 1 - } - // We are out of rushed mode! - s.rushed = false - s.rushedMode.Set(0) - log. - With("urgencyScore", score). - With("chunksToPersist", int(chunksToPersist)). - With("maxChunksToPersist", int(maxChunksToPersist)). - With("memoryChunks", int(memChunks)). - With("maxMemoryChunks", int(maxMemChunks)). - Info("Storage has left rushed mode.") - return score - } - if score > persintenceUrgencyScoreForEnteringRushedMode { - // Enter rushed mode. - s.rushed = true - s.rushedMode.Set(1) - log. - With("urgencyScore", score). - With("chunksToPersist", int(chunksToPersist)). - With("maxChunksToPersist", int(maxChunksToPersist)). - With("memoryChunks", int(memChunks)). - With("maxMemoryChunks", int(maxMemChunks)). - Warn("Storage has entered rushed mode.") - return 1 - } - return score -} - -// quarantineSeries registers the provided fingerprint for quarantining. It -// always returns immediately. Quarantine requests are processed -// asynchronously. If there are too many requests queued, they are simply -// dropped. -// -// Quarantining means that the series file is moved to the orphaned directory, -// and all its traces are removed from indices. Call this method if an -// unrecoverable error is detected while dealing with a series, and pass in the -// encountered error. It will be saved as a hint in the orphaned directory. -func (s *MemorySeriesStorage) quarantineSeries(fp model.Fingerprint, metric model.Metric, err error) { - req := quarantineRequest{fp: fp, metric: metric, reason: err} - select { - case s.quarantineRequests <- req: - // Request submitted. - default: - log. - With("fingerprint", fp). - With("metric", metric). - With("reason", err). - Warn("Quarantine queue full. Dropped quarantine request.") - s.seriesOps.WithLabelValues(droppedQuarantine).Inc() - } -} - -func (s *MemorySeriesStorage) handleQuarantine() { - for { - select { - case req := <-s.quarantineRequests: - s.purgeSeries(req.fp, req.metric, req.reason) - log. - With("fingerprint", req.fp). - With("metric", req.metric). - With("reason", req.reason). - Warn("Series quarantined.") - case <-s.quarantineStopping: - log.Info("Series quarantining stopped.") - close(s.quarantineStopped) - return - } - } - -} - -// purgeSeries removes all traces of a series. If a non-nil quarantine reason is -// provided, the series file will not be deleted completely, but moved to the -// orphaned directory with the reason and the metric in a hint file. The -// provided metric might be nil if unknown. -func (s *MemorySeriesStorage) purgeSeries(fp model.Fingerprint, m model.Metric, quarantineReason error) { - s.fpLocker.Lock(fp) - - var ( - series *memorySeries - ok bool - ) - - if series, ok = s.fpToSeries.get(fp); ok { - s.fpToSeries.del(fp) - s.numSeries.Dec() - m = series.metric - - // Adjust s.numChunksToPersist and chunk.NumMemChunks down by - // the number of chunks in this series that are not - // persisted yet. Persisted chunks will be deducted from - // chunk.NumMemChunks upon eviction. - numChunksNotYetPersisted := len(series.chunkDescs) - series.persistWatermark - atomic.AddInt64(&chunk.NumMemChunks, int64(-numChunksNotYetPersisted)) - if !series.headChunkClosed { - // Head chunk wasn't counted as waiting for persistence yet. - // (But it was counted as a chunk in memory.) - numChunksNotYetPersisted-- - } - s.incNumChunksToPersist(-numChunksNotYetPersisted) - - } else { - s.persistence.purgeArchivedMetric(fp) // Ignoring error. There is nothing we can do. - } - if m != nil { - // If we know a metric now, unindex it in any case. - // purgeArchivedMetric might have done so already, but we cannot - // be sure. Unindexing in idempotent, though. - s.persistence.unindexMetric(fp, m) - } - // Attempt to delete/quarantine the series file in any case. - if quarantineReason == nil { - // No reason stated, simply delete the file. - if _, err := s.persistence.deleteSeriesFile(fp); err != nil { - log. - With("fingerprint", fp). - With("metric", m). - With("error", err). - Error("Error deleting series file.") - } - s.seriesOps.WithLabelValues(requestedPurge).Inc() - } else { - if err := s.persistence.quarantineSeriesFile(fp, quarantineReason, m); err == nil { - s.seriesOps.WithLabelValues(completedQurantine).Inc() - } else { - s.seriesOps.WithLabelValues(failedQuarantine).Inc() - log. - With("fingerprint", fp). - With("metric", m). - With("reason", quarantineReason). - With("error", err). - Error("Error quarantining series file.") - } - } - - s.fpLocker.Unlock(fp) -} - -// Describe implements prometheus.Collector. -func (s *MemorySeriesStorage) Describe(ch chan<- *prometheus.Desc) { - s.persistence.Describe(ch) - s.mapper.Describe(ch) - - ch <- s.persistErrors.Desc() - ch <- maxChunksToPersistDesc - ch <- numChunksToPersistDesc - ch <- s.numSeries.Desc() - s.seriesOps.Describe(ch) - ch <- s.ingestedSamplesCount.Desc() - s.discardedSamplesCount.Describe(ch) - ch <- s.nonExistentSeriesMatchesCount.Desc() - ch <- chunk.NumMemChunksDesc - s.maintainSeriesDuration.Describe(ch) - ch <- s.persistenceUrgencyScore.Desc() - ch <- s.rushedMode.Desc() -} - -// Collect implements prometheus.Collector. -func (s *MemorySeriesStorage) Collect(ch chan<- prometheus.Metric) { - s.persistence.Collect(ch) - s.mapper.Collect(ch) - - ch <- s.persistErrors - ch <- prometheus.MustNewConstMetric( - maxChunksToPersistDesc, - prometheus.GaugeValue, - float64(s.maxChunksToPersist), - ) - ch <- prometheus.MustNewConstMetric( - numChunksToPersistDesc, - prometheus.GaugeValue, - float64(s.getNumChunksToPersist()), - ) - ch <- s.numSeries - s.seriesOps.Collect(ch) - ch <- s.ingestedSamplesCount - s.discardedSamplesCount.Collect(ch) - ch <- s.nonExistentSeriesMatchesCount - ch <- prometheus.MustNewConstMetric( - chunk.NumMemChunksDesc, - prometheus.GaugeValue, - float64(atomic.LoadInt64(&chunk.NumMemChunks)), - ) - s.maintainSeriesDuration.Collect(ch) - ch <- s.persistenceUrgencyScore - ch <- s.rushedMode -} diff --git a/storage/local/storage_test.go b/storage/local/storage_test.go deleted file mode 100644 index d63b6d3ab1..0000000000 --- a/storage/local/storage_test.go +++ /dev/null @@ -1,2005 +0,0 @@ -// Copyright 2014 The Prometheus Authors -// Licensed under the Apache License, Version 2.0 (the "License"); -// you may not use this file except in compliance with the License. -// You may obtain a copy of the License at -// -// http://www.apache.org/licenses/LICENSE-2.0 -// -// Unless required by applicable law or agreed to in writing, software -// distributed under the License is distributed on an "AS IS" BASIS, -// WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. -// See the License for the specific language governing permissions and -// limitations under the License. - -package local - -import ( - "fmt" - "hash/fnv" - "math" - "math/rand" - "os" - "strconv" - "sync/atomic" - "testing" - "testing/quick" - "time" - - "github.com/prometheus/common/log" - "github.com/prometheus/common/model" - "golang.org/x/net/context" - - "github.com/prometheus/prometheus/storage/local/chunk" - "github.com/prometheus/prometheus/storage/metric" - "github.com/prometheus/prometheus/util/testutil" -) - -func TestMatches(t *testing.T) { - storage, closer := NewTestStorage(t, 2) - defer closer.Close() - - storage.archiveHighWatermark = 90 - samples := make([]*model.Sample, 100) - fingerprints := make(model.Fingerprints, 100) - - for i := range samples { - metric := model.Metric{ - model.MetricNameLabel: model.LabelValue(fmt.Sprintf("test_metric_%d", i)), - "label1": model.LabelValue(fmt.Sprintf("test_%d", i/10)), - "label2": model.LabelValue(fmt.Sprintf("test_%d", (i+5)/10)), - "all": "const", - } - samples[i] = &model.Sample{ - Metric: metric, - Timestamp: model.Time(i), - Value: model.SampleValue(i), - } - fingerprints[i] = metric.FastFingerprint() - } - for _, s := range samples { - storage.Append(s) - } - storage.WaitForIndexing() - - // Archive every tenth metric. - for i, fp := range fingerprints { - if i%10 != 0 { - continue - } - s, ok := storage.fpToSeries.get(fp) - if !ok { - t.Fatal("could not retrieve series for fp", fp) - } - storage.fpLocker.Lock(fp) - storage.persistence.archiveMetric(fp, s.metric, s.firstTime(), s.lastTime) - storage.fpLocker.Unlock(fp) - } - - newMatcher := func(matchType metric.MatchType, name model.LabelName, value model.LabelValue) *metric.LabelMatcher { - lm, err := metric.NewLabelMatcher(matchType, name, value) - if err != nil { - t.Fatalf("error creating label matcher: %s", err) - } - return lm - } - - var matcherTests = []struct { - matchers metric.LabelMatchers - expected model.Fingerprints - }{ - { - matchers: metric.LabelMatchers{newMatcher(metric.Equal, "label1", "x")}, - expected: model.Fingerprints{}, - }, - { - matchers: metric.LabelMatchers{newMatcher(metric.Equal, "label1", "test_0")}, - expected: fingerprints[:10], - }, - { - matchers: metric.LabelMatchers{ - newMatcher(metric.Equal, "label1", "test_0"), - newMatcher(metric.Equal, "label2", "test_1"), - }, - expected: fingerprints[5:10], - }, - { - matchers: metric.LabelMatchers{ - newMatcher(metric.Equal, "all", "const"), - newMatcher(metric.NotEqual, "label1", "x"), - }, - expected: fingerprints, - }, - { - matchers: metric.LabelMatchers{ - newMatcher(metric.Equal, "all", "const"), - newMatcher(metric.NotEqual, "label1", "test_0"), - }, - expected: fingerprints[10:], - }, - { - matchers: metric.LabelMatchers{ - newMatcher(metric.Equal, "all", "const"), - newMatcher(metric.NotEqual, "label1", "test_0"), - newMatcher(metric.NotEqual, "label1", "test_1"), - newMatcher(metric.NotEqual, "label1", "test_2"), - }, - expected: fingerprints[30:], - }, - { - matchers: metric.LabelMatchers{ - newMatcher(metric.Equal, "label1", ""), - }, - expected: fingerprints[:0], - }, - { - matchers: metric.LabelMatchers{ - newMatcher(metric.NotEqual, "label1", "test_0"), - newMatcher(metric.Equal, "label1", ""), - }, - expected: fingerprints[:0], - }, - { - matchers: metric.LabelMatchers{ - newMatcher(metric.NotEqual, "label1", "test_0"), - newMatcher(metric.Equal, "label2", ""), - }, - expected: fingerprints[:0], - }, - { - matchers: metric.LabelMatchers{ - newMatcher(metric.Equal, "all", "const"), - newMatcher(metric.NotEqual, "label1", "test_0"), - newMatcher(metric.Equal, "not_existent", ""), - }, - expected: fingerprints[10:], - }, - { - matchers: metric.LabelMatchers{ - newMatcher(metric.RegexMatch, "label1", `test_[3-5]`), - }, - expected: fingerprints[30:60], - }, - { - matchers: metric.LabelMatchers{ - newMatcher(metric.Equal, "all", "const"), - newMatcher(metric.RegexNoMatch, "label1", `test_[3-5]`), - }, - expected: append(append(model.Fingerprints{}, fingerprints[:30]...), fingerprints[60:]...), - }, - { - matchers: metric.LabelMatchers{ - newMatcher(metric.RegexMatch, "label1", `test_[3-5]`), - newMatcher(metric.RegexMatch, "label2", `test_[4-6]`), - }, - expected: fingerprints[35:60], - }, - { - matchers: metric.LabelMatchers{ - newMatcher(metric.RegexMatch, "label1", `test_[3-5]`), - newMatcher(metric.NotEqual, "label2", `test_4`), - }, - expected: append(append(model.Fingerprints{}, fingerprints[30:35]...), fingerprints[45:60]...), - }, - { - matchers: metric.LabelMatchers{ - newMatcher(metric.Equal, "label1", `nonexistent`), - newMatcher(metric.RegexMatch, "label2", `test`), - }, - expected: model.Fingerprints{}, - }, - { - matchers: metric.LabelMatchers{ - newMatcher(metric.Equal, "label1", `test_0`), - newMatcher(metric.RegexMatch, "label2", `nonexistent`), - }, - expected: model.Fingerprints{}, - }, - } - - for _, mt := range matcherTests { - metrics, err := storage.MetricsForLabelMatchers( - context.Background(), - model.Earliest, model.Latest, - mt.matchers, - ) - if err != nil { - t.Fatal(err) - } - if len(mt.expected) != len(metrics) { - t.Fatalf("expected %d matches for %q, found %d", len(mt.expected), mt.matchers, len(metrics)) - } - for _, m := range metrics { - fp1 := m.Metric.FastFingerprint() - found := false - for _, fp2 := range mt.expected { - if fp1 == fp2 { - found = true - break - } - } - if !found { - t.Errorf("expected fingerprint %s for %q not in result", fp1, mt.matchers) - } - } - // Smoketest for from/through. - metrics, err = storage.MetricsForLabelMatchers( - context.Background(), - model.Earliest, -10000, - mt.matchers, - ) - if err != nil { - t.Fatal(err) - } - if len(metrics) > 0 { - t.Error("expected no matches with 'through' older than any sample") - } - metrics, err = storage.MetricsForLabelMatchers( - context.Background(), - 10000, model.Latest, - mt.matchers, - ) - if err != nil { - t.Fatal(err) - } - if len(metrics) > 0 { - t.Error("expected no matches with 'from' newer than any sample") - } - // Now the tricky one, cut out something from the middle. - var ( - from model.Time = 25 - through model.Time = 75 - ) - metrics, err = storage.MetricsForLabelMatchers( - context.Background(), - from, through, - mt.matchers, - ) - if err != nil { - t.Fatal(err) - } - expected := model.Fingerprints{} - for _, fp := range mt.expected { - i := 0 - for ; fingerprints[i] != fp && i < len(fingerprints); i++ { - } - if i == len(fingerprints) { - t.Fatal("expected fingerprint does not exist") - } - if !model.Time(i).Before(from) && !model.Time(i).After(through) { - expected = append(expected, fp) - } - } - if len(expected) != len(metrics) { - t.Errorf("expected %d range-limited matches for %q, found %d", len(expected), mt.matchers, len(metrics)) - } - for _, m := range metrics { - fp1 := m.Metric.FastFingerprint() - found := false - for _, fp2 := range expected { - if fp1 == fp2 { - found = true - break - } - } - if !found { - t.Errorf("expected fingerprint %s for %q not in range-limited result", fp1, mt.matchers) - } - } - - } -} - -func TestFingerprintsForLabels(t *testing.T) { - storage, closer := NewTestStorage(t, 2) - defer closer.Close() - - samples := make([]*model.Sample, 100) - fingerprints := make(model.Fingerprints, 100) - - for i := range samples { - metric := model.Metric{ - model.MetricNameLabel: model.LabelValue(fmt.Sprintf("test_metric_%d", i)), - "label1": model.LabelValue(fmt.Sprintf("test_%d", i/10)), - "label2": model.LabelValue(fmt.Sprintf("test_%d", (i+5)/10)), - } - samples[i] = &model.Sample{ - Metric: metric, - Timestamp: model.Time(i), - Value: model.SampleValue(i), - } - fingerprints[i] = metric.FastFingerprint() - } - for _, s := range samples { - storage.Append(s) - } - storage.WaitForIndexing() - - var matcherTests = []struct { - pairs []model.LabelPair - expected model.Fingerprints - }{ - { - pairs: []model.LabelPair{{Name: "label1", Value: "x"}}, - expected: fingerprints[:0], - }, - { - pairs: []model.LabelPair{{Name: "label1", Value: "test_0"}}, - expected: fingerprints[:10], - }, - { - pairs: []model.LabelPair{ - {Name: "label1", Value: "test_0"}, - {Name: "label1", Value: "test_1"}, - }, - expected: fingerprints[:0], - }, - { - pairs: []model.LabelPair{ - {Name: "label1", Value: "test_0"}, - {Name: "label2", Value: "test_1"}, - }, - expected: fingerprints[5:10], - }, - { - pairs: []model.LabelPair{ - {Name: "label1", Value: "test_1"}, - {Name: "label2", Value: "test_2"}, - }, - expected: fingerprints[15:20], - }, - } - - for _, mt := range matcherTests { - var resfps map[model.Fingerprint]struct{} - for _, pair := range mt.pairs { - resfps = storage.fingerprintsForLabelPair(pair, nil, resfps) - } - if len(mt.expected) != len(resfps) { - t.Fatalf("expected %d matches for %q, found %d", len(mt.expected), mt.pairs, len(resfps)) - } - for fp1 := range resfps { - found := false - for _, fp2 := range mt.expected { - if fp1 == fp2 { - found = true - break - } - } - if !found { - t.Errorf("expected fingerprint %s for %q not in result", fp1, mt.pairs) - } - } - } -} - -var benchLabelMatchingRes []metric.Metric - -func BenchmarkLabelMatching(b *testing.B) { - s, closer := NewTestStorage(b, 2) - defer closer.Close() - - h := fnv.New64a() - lbl := func(x int) model.LabelValue { - h.Reset() - h.Write([]byte(fmt.Sprintf("%d", x))) - return model.LabelValue(fmt.Sprintf("%d", h.Sum64())) - } - - M := 32 - met := model.Metric{} - for i := 0; i < M; i++ { - met["label_a"] = lbl(i) - for j := 0; j < M; j++ { - met["label_b"] = lbl(j) - for k := 0; k < M; k++ { - met["label_c"] = lbl(k) - for l := 0; l < M; l++ { - met["label_d"] = lbl(l) - s.Append(&model.Sample{ - Metric: met.Clone(), - Timestamp: 0, - Value: 1, - }) - } - } - } - } - s.WaitForIndexing() - - newMatcher := func(matchType metric.MatchType, name model.LabelName, value model.LabelValue) *metric.LabelMatcher { - lm, err := metric.NewLabelMatcher(matchType, name, value) - if err != nil { - b.Fatalf("error creating label matcher: %s", err) - } - return lm - } - - var matcherTests = []metric.LabelMatchers{ - { - newMatcher(metric.Equal, "label_a", lbl(1)), - }, - { - newMatcher(metric.Equal, "label_a", lbl(3)), - newMatcher(metric.Equal, "label_c", lbl(3)), - }, - { - newMatcher(metric.Equal, "label_a", lbl(3)), - newMatcher(metric.Equal, "label_c", lbl(3)), - newMatcher(metric.NotEqual, "label_d", lbl(3)), - }, - { - newMatcher(metric.Equal, "label_a", lbl(3)), - newMatcher(metric.Equal, "label_b", lbl(3)), - newMatcher(metric.Equal, "label_c", lbl(3)), - newMatcher(metric.NotEqual, "label_d", lbl(3)), - }, - { - newMatcher(metric.RegexMatch, "label_a", ".+"), - }, - { - newMatcher(metric.Equal, "label_a", lbl(3)), - newMatcher(metric.RegexMatch, "label_a", ".+"), - }, - { - newMatcher(metric.Equal, "label_a", lbl(1)), - newMatcher(metric.RegexMatch, "label_c", "("+lbl(3)+"|"+lbl(10)+")"), - }, - { - newMatcher(metric.Equal, "label_a", lbl(3)), - newMatcher(metric.Equal, "label_a", lbl(4)), - newMatcher(metric.RegexMatch, "label_c", "("+lbl(3)+"|"+lbl(10)+")"), - }, - } - - b.ReportAllocs() - b.ResetTimer() - - var err error - for i := 0; i < b.N; i++ { - benchLabelMatchingRes = []metric.Metric{} - for _, mt := range matcherTests { - benchLabelMatchingRes, err = s.MetricsForLabelMatchers( - context.Background(), - model.Earliest, model.Latest, - mt, - ) - if err != nil { - b.Fatal(err) - } - } - } - // Stop timer to not count the storage closing. - b.StopTimer() -} - -func BenchmarkQueryRange(b *testing.B) { - now := model.Now() - insertStart := now.Add(-2 * time.Hour) - - s, closer := NewTestStorage(b, 2) - defer closer.Close() - - // Stop maintenance loop to prevent actual purging. - close(s.loopStopping) - <-s.loopStopped - <-s.logThrottlingStopped - // Recreate channel to avoid panic when we really shut down. - s.loopStopping = make(chan struct{}) - - for i := 0; i < 8192; i++ { - s.Append(&model.Sample{ - Metric: model.Metric{"__name__": model.LabelValue(strconv.Itoa(i)), "job": "test"}, - Timestamp: insertStart, - Value: 1, - }) - } - s.WaitForIndexing() - - b.ResetTimer() - b.RunParallel(func(pb *testing.PB) { - lm, _ := metric.NewLabelMatcher(metric.Equal, "job", "test") - for pb.Next() { - s.QueryRange(context.Background(), insertStart, now, lm) - } - }) -} - -func TestQueryRangeThroughBeforeFrom(t *testing.T) { - now := model.Now() - insertStart := now.Add(-2 * time.Hour) - - s, closer := NewTestStorage(t, 2) - defer closer.Close() - - // Stop maintenance loop to prevent actual purging. - close(s.loopStopping) - <-s.loopStopped - <-s.logThrottlingStopped - // Recreate channel to avoid panic when we really shut down. - s.loopStopping = make(chan struct{}) - - for i := 0; i < 8192; i++ { - s.Append(&model.Sample{ - Metric: model.Metric{"__name__": "testmetric", "job": "test"}, - Timestamp: insertStart.Add(time.Duration(i) * time.Second), - Value: model.SampleValue(rand.Float64()), - }) - } - s.WaitForIndexing() - - lm, _ := metric.NewLabelMatcher(metric.Equal, "job", "test") - iters, err := s.QueryRange(context.Background(), now.Add(-30*time.Minute), now.Add(-90*time.Minute), lm) - if err != nil { - t.Error(err) - } - if len(iters) != 0 { - t.Errorf("expected no iters to be returned, got %d", len(iters)) - } -} - -func TestRetentionCutoff(t *testing.T) { - now := model.Now() - insertStart := now.Add(-2 * time.Hour) - - s, closer := NewTestStorage(t, 2) - defer closer.Close() - - // Stop maintenance loop to prevent actual purging. - close(s.loopStopping) - <-s.loopStopped - <-s.logThrottlingStopped - // Recreate channel to avoid panic when we really shut down. - s.loopStopping = make(chan struct{}) - - s.dropAfter = 1 * time.Hour - - for i := 0; i < 120; i++ { - smpl := &model.Sample{ - Metric: model.Metric{"job": "test"}, - Timestamp: insertStart.Add(time.Duration(i) * time.Minute), // 1 minute intervals. - Value: 1, - } - s.Append(smpl) - } - s.WaitForIndexing() - - lm, err := metric.NewLabelMatcher(metric.Equal, "job", "test") - if err != nil { - t.Fatalf("error creating label matcher: %s", err) - } - its, err := s.QueryRange(context.Background(), insertStart, now, lm) - if err != nil { - t.Fatal(err) - } - - if len(its) != 1 { - t.Fatalf("expected one iterator but got %d", len(its)) - } - - val := its[0].ValueAtOrBeforeTime(now.Add(-61 * time.Minute)) - if val.Timestamp != model.Earliest { - t.Errorf("unexpected result for timestamp before retention period") - } - - vals := its[0].RangeValues(metric.Interval{OldestInclusive: insertStart, NewestInclusive: now}) - // We get 59 values here because the model.Now() is slightly later - // than our now. - if len(vals) != 59 { - t.Errorf("expected 59 values but got %d", len(vals)) - } - if expt := now.Add(-1 * time.Hour).Add(time.Minute); vals[0].Timestamp != expt { - t.Errorf("unexpected timestamp for first sample: %v, expected %v", vals[0].Timestamp.Time(), expt.Time()) - } -} - -func TestDropMetrics(t *testing.T) { - now := model.Now() - insertStart := now.Add(-2 * time.Hour) - - s, closer := NewTestStorage(t, 2) - defer closer.Close() - - chunkFileExists := func(fp model.Fingerprint) (bool, error) { - f, err := s.persistence.openChunkFileForReading(fp) - if err == nil { - f.Close() - return true, nil - } - if os.IsNotExist(err) { - return false, nil - } - return false, err - } - - m1 := model.Metric{model.MetricNameLabel: "test", "n1": "v1"} - m2 := model.Metric{model.MetricNameLabel: "test", "n1": "v2"} - m3 := model.Metric{model.MetricNameLabel: "test", "n1": "v3"} - - lm1, err := metric.NewLabelMatcher(metric.Equal, "n1", "v1") - if err != nil { - t.Fatal(err) - } - lmAll, err := metric.NewLabelMatcher(metric.Equal, model.MetricNameLabel, "test") - if err != nil { - t.Fatal(err) - } - - N := 120000 - - for j, m := range []model.Metric{m1, m2, m3} { - for i := 0; i < N; i++ { - smpl := &model.Sample{ - Metric: m, - Timestamp: insertStart.Add(time.Duration(i) * time.Millisecond), // 1 millisecond intervals. - Value: model.SampleValue(j), - } - s.Append(smpl) - } - } - s.WaitForIndexing() - - // Archive m3, but first maintain it so that at least something is written to disk. - fpToBeArchived := m3.FastFingerprint() - s.maintainMemorySeries(fpToBeArchived, 0) - s.fpLocker.Lock(fpToBeArchived) - s.fpToSeries.del(fpToBeArchived) - s.persistence.archiveMetric(fpToBeArchived, m3, 0, insertStart.Add(time.Duration(N-1)*time.Millisecond)) - s.fpLocker.Unlock(fpToBeArchived) - - fps := s.fingerprintsForLabelPair(model.LabelPair{ - Name: model.MetricNameLabel, Value: "test", - }, nil, nil) - if len(fps) != 3 { - t.Errorf("unexpected number of fingerprints: %d", len(fps)) - } - - fpList := model.Fingerprints{m1.FastFingerprint(), m2.FastFingerprint(), fpToBeArchived} - - n, err := s.DropMetricsForLabelMatchers(context.Background(), lm1) - if err != nil { - t.Fatal(err) - } - if n != 1 { - t.Fatalf("expected 1 series to be dropped, got %d", n) - } - s.WaitForIndexing() - - fps2 := s.fingerprintsForLabelPair(model.LabelPair{ - Name: model.MetricNameLabel, Value: "test", - }, nil, nil) - if len(fps2) != 2 { - t.Errorf("unexpected number of fingerprints: %d", len(fps2)) - } - - it := s.preloadChunksForRange(makeFingerprintSeriesPair(s, fpList[0]), model.Earliest, model.Latest) - if vals := it.RangeValues(metric.Interval{OldestInclusive: insertStart, NewestInclusive: now}); len(vals) != 0 { - t.Errorf("unexpected number of samples: %d", len(vals)) - } - - it = s.preloadChunksForRange(makeFingerprintSeriesPair(s, fpList[1]), model.Earliest, model.Latest) - if vals := it.RangeValues(metric.Interval{OldestInclusive: insertStart, NewestInclusive: now}); len(vals) != N { - t.Errorf("unexpected number of samples: %d", len(vals)) - } - exists, err := chunkFileExists(fpList[2]) - if err != nil { - t.Fatal(err) - } - if !exists { - t.Errorf("chunk file does not exist for fp=%v", fpList[2]) - } - - n, err = s.DropMetricsForLabelMatchers(context.Background(), lmAll) - if err != nil { - t.Fatal(err) - } - if n != 2 { - t.Fatalf("expected 2 series to be dropped, got %d", n) - } - s.WaitForIndexing() - - fps3 := s.fingerprintsForLabelPair(model.LabelPair{ - Name: model.MetricNameLabel, Value: "test", - }, nil, nil) - if len(fps3) != 0 { - t.Errorf("unexpected number of fingerprints: %d", len(fps3)) - } - - it = s.preloadChunksForRange(makeFingerprintSeriesPair(s, fpList[0]), model.Earliest, model.Latest) - if vals := it.RangeValues(metric.Interval{OldestInclusive: insertStart, NewestInclusive: now}); len(vals) != 0 { - t.Errorf("unexpected number of samples: %d", len(vals)) - } - - it = s.preloadChunksForRange(makeFingerprintSeriesPair(s, fpList[1]), model.Earliest, model.Latest) - if vals := it.RangeValues(metric.Interval{OldestInclusive: insertStart, NewestInclusive: now}); len(vals) != 0 { - t.Errorf("unexpected number of samples: %d", len(vals)) - } - exists, err = chunkFileExists(fpList[2]) - if err != nil { - t.Fatal(err) - } - if exists { - t.Errorf("chunk file still exists for fp=%v", fpList[2]) - } -} - -func TestQuarantineMetric(t *testing.T) { - now := model.Now() - insertStart := now.Add(-2 * time.Hour) - - s, closer := NewTestStorage(t, 2) - defer closer.Close() - - chunkFileExists := func(fp model.Fingerprint) (bool, error) { - f, err := s.persistence.openChunkFileForReading(fp) - if err == nil { - f.Close() - return true, nil - } - if os.IsNotExist(err) { - return false, nil - } - return false, err - } - - m1 := model.Metric{model.MetricNameLabel: "test", "n1": "v1"} - m2 := model.Metric{model.MetricNameLabel: "test", "n1": "v2"} - m3 := model.Metric{model.MetricNameLabel: "test", "n1": "v3"} - - N := 120000 - - for j, m := range []model.Metric{m1, m2, m3} { - for i := 0; i < N; i++ { - smpl := &model.Sample{ - Metric: m, - Timestamp: insertStart.Add(time.Duration(i) * time.Millisecond), // 1 millisecond intervals. - Value: model.SampleValue(j), - } - s.Append(smpl) - } - } - s.WaitForIndexing() - - // Archive m3, but first maintain it so that at least something is written to disk. - fpToBeArchived := m3.FastFingerprint() - s.maintainMemorySeries(fpToBeArchived, 0) - s.fpLocker.Lock(fpToBeArchived) - s.fpToSeries.del(fpToBeArchived) - s.persistence.archiveMetric(fpToBeArchived, m3, 0, insertStart.Add(time.Duration(N-1)*time.Millisecond)) - s.fpLocker.Unlock(fpToBeArchived) - - // Corrupt the series file for m3. - f, err := os.Create(s.persistence.fileNameForFingerprint(fpToBeArchived)) - if err != nil { - t.Fatal(err) - } - if _, err := f.WriteString("This is clearly not the content of a series file."); err != nil { - t.Fatal(err) - } - if f.Close(); err != nil { - t.Fatal(err) - } - - fps := s.fingerprintsForLabelPair(model.LabelPair{ - Name: model.MetricNameLabel, Value: "test", - }, nil, nil) - if len(fps) != 3 { - t.Errorf("unexpected number of fingerprints: %d", len(fps)) - } - - // This will access the corrupt file and lead to quarantining. - iter := s.preloadChunksForInstant(makeFingerprintSeriesPair(s, fpToBeArchived), now.Add(-2*time.Hour-1*time.Minute), now.Add(-2*time.Hour)) - iter.Close() - time.Sleep(time.Second) // Give time to quarantine. TODO(beorn7): Find a better way to wait. - s.WaitForIndexing() - - fps2 := s.fingerprintsForLabelPair(model.LabelPair{ - Name: model.MetricNameLabel, Value: "test", - }, nil, nil) - if len(fps2) != 2 { - t.Errorf("unexpected number of fingerprints: %d", len(fps2)) - } - - exists, err := chunkFileExists(fpToBeArchived) - if err != nil { - t.Fatal(err) - } - if exists { - t.Errorf("chunk file exists for fp=%v", fpToBeArchived) - } -} - -// TestLoop is just a smoke test for the loop method, if we can switch it on and -// off without disaster. -func TestLoop(t *testing.T) { - if testing.Short() { - t.Skip("Skipping test in short mode.") - } - samples := make(model.Samples, 1000) - for i := range samples { - samples[i] = &model.Sample{ - Timestamp: model.Time(2 * i), - Value: model.SampleValue(float64(i) * 0.2), - } - } - directory := testutil.NewTemporaryDirectory("test_storage", t) - defer directory.Close() - o := &MemorySeriesStorageOptions{ - MemoryChunks: 50, - MaxChunksToPersist: 1000000, - PersistenceRetentionPeriod: 24 * 7 * time.Hour, - PersistenceStoragePath: directory.Path(), - CheckpointInterval: 250 * time.Millisecond, - SyncStrategy: Adaptive, - MinShrinkRatio: 0.1, - } - storage := NewMemorySeriesStorage(o) - if err := storage.Start(); err != nil { - t.Errorf("Error starting storage: %s", err) - } - for _, s := range samples { - storage.Append(s) - } - storage.WaitForIndexing() - series, _ := storage.fpToSeries.get(model.Metric{}.FastFingerprint()) - cdsBefore := len(series.chunkDescs) - time.Sleep(fpMaxWaitDuration + time.Second) // TODO(beorn7): Ugh, need to wait for maintenance to kick in. - cdsAfter := len(series.chunkDescs) - storage.Stop() - if cdsBefore <= cdsAfter { - t.Errorf( - "Number of chunk descriptors should have gone down by now. Got before %d, after %d.", - cdsBefore, cdsAfter, - ) - } -} - -func testChunk(t *testing.T, encoding chunk.Encoding) { - samples := make(model.Samples, 500000) - for i := range samples { - samples[i] = &model.Sample{ - Timestamp: model.Time(i), - Value: model.SampleValue(float64(i) * 0.2), - } - } - s, closer := NewTestStorage(t, encoding) - defer closer.Close() - - for _, sample := range samples { - s.Append(sample) - } - s.WaitForIndexing() - - for m := range s.fpToSeries.iter() { - s.fpLocker.Lock(m.fp) - defer s.fpLocker.Unlock(m.fp) // TODO remove, see below - var values []model.SamplePair - for _, cd := range m.series.chunkDescs { - if cd.IsEvicted() { - continue - } - it := cd.C.NewIterator() - for it.Scan() { - values = append(values, it.Value()) - } - if it.Err() != nil { - t.Error(it.Err()) - } - } - - for i, v := range values { - if samples[i].Timestamp != v.Timestamp { - t.Errorf("%d. Got %v; want %v", i, v.Timestamp, samples[i].Timestamp) - } - if samples[i].Value != v.Value { - t.Errorf("%d. Got %v; want %v", i, v.Value, samples[i].Value) - } - } - //s.fpLocker.Unlock(m.fp) - } - log.Info("test done, closing") -} - -func TestChunkType0(t *testing.T) { - testChunk(t, 0) -} - -func TestChunkType1(t *testing.T) { - testChunk(t, 1) -} - -func TestChunkType2(t *testing.T) { - testChunk(t, 2) -} - -func testValueAtOrBeforeTime(t *testing.T, encoding chunk.Encoding) { - samples := make(model.Samples, 10000) - for i := range samples { - samples[i] = &model.Sample{ - Timestamp: model.Time(2 * i), - Value: model.SampleValue(float64(i) * 0.2), - } - } - s, closer := NewTestStorage(t, encoding) - defer closer.Close() - - for _, sample := range samples { - s.Append(sample) - } - s.WaitForIndexing() - - fp := model.Metric{}.FastFingerprint() - - it := s.preloadChunksForRange(makeFingerprintSeriesPair(s, fp), model.Earliest, model.Latest) - - // #1 Exactly on a sample. - for i, expected := range samples { - actual := it.ValueAtOrBeforeTime(expected.Timestamp) - - if expected.Timestamp != actual.Timestamp { - t.Errorf("1.%d. Got %v; want %v", i, actual.Timestamp, expected.Timestamp) - } - if expected.Value != actual.Value { - t.Errorf("1.%d. Got %v; want %v", i, actual.Value, expected.Value) - } - } - - // #2 Between samples. - for i, expected := range samples { - if i == len(samples)-1 { - continue - } - actual := it.ValueAtOrBeforeTime(expected.Timestamp + 1) - - if expected.Timestamp != actual.Timestamp { - t.Errorf("2.%d. Got %v; want %v", i, actual.Timestamp, expected.Timestamp) - } - if expected.Value != actual.Value { - t.Errorf("2.%d. Got %v; want %v", i, actual.Value, expected.Value) - } - } - - // #3 Corner cases: Just before the first sample, just after the last. - expected := &model.Sample{Timestamp: model.Earliest} - actual := it.ValueAtOrBeforeTime(samples[0].Timestamp - 1) - if expected.Timestamp != actual.Timestamp { - t.Errorf("3.1. Got %v; want %v", actual.Timestamp, expected.Timestamp) - } - if expected.Value != actual.Value { - t.Errorf("3.1. Got %v; want %v", actual.Value, expected.Value) - } - expected = samples[len(samples)-1] - actual = it.ValueAtOrBeforeTime(expected.Timestamp + 1) - if expected.Timestamp != actual.Timestamp { - t.Errorf("3.2. Got %v; want %v", actual.Timestamp, expected.Timestamp) - } - if expected.Value != actual.Value { - t.Errorf("3.2. Got %v; want %v", actual.Value, expected.Value) - } -} - -func TestValueAtTimeChunkType0(t *testing.T) { - testValueAtOrBeforeTime(t, 0) -} - -func TestValueAtTimeChunkType1(t *testing.T) { - testValueAtOrBeforeTime(t, 1) -} - -func TestValueAtTimeChunkType2(t *testing.T) { - testValueAtOrBeforeTime(t, 2) -} - -func benchmarkValueAtOrBeforeTime(b *testing.B, encoding chunk.Encoding) { - samples := make(model.Samples, 10000) - for i := range samples { - samples[i] = &model.Sample{ - Timestamp: model.Time(2 * i), - Value: model.SampleValue(float64(i) * 0.2), - } - } - s, closer := NewTestStorage(b, encoding) - defer closer.Close() - - for _, sample := range samples { - s.Append(sample) - } - s.WaitForIndexing() - - fp := model.Metric{}.FastFingerprint() - - it := s.preloadChunksForRange(makeFingerprintSeriesPair(s, fp), model.Earliest, model.Latest) - - b.ResetTimer() - - for i := 0; i < b.N; i++ { - // #1 Exactly on a sample. - for i, expected := range samples { - actual := it.ValueAtOrBeforeTime(expected.Timestamp) - - if expected.Timestamp != actual.Timestamp { - b.Errorf("1.%d. Got %v; want %v", i, actual.Timestamp, expected.Timestamp) - } - if expected.Value != actual.Value { - b.Errorf("1.%d. Got %v; want %v", i, actual.Value, expected.Value) - } - } - - // #2 Between samples. - for i, expected := range samples { - if i == len(samples)-1 { - continue - } - actual := it.ValueAtOrBeforeTime(expected.Timestamp + 1) - - if expected.Timestamp != actual.Timestamp { - b.Errorf("2.%d. Got %v; want %v", i, actual.Timestamp, expected.Timestamp) - } - if expected.Value != actual.Value { - b.Errorf("2.%d. Got %v; want %v", i, actual.Value, expected.Value) - } - } - - // #3 Corner cases: Just before the first sample, just after the last. - expected := &model.Sample{Timestamp: model.Earliest} - actual := it.ValueAtOrBeforeTime(samples[0].Timestamp - 1) - if expected.Timestamp != actual.Timestamp { - b.Errorf("3.1. Got %v; want %v", actual.Timestamp, expected.Timestamp) - } - if expected.Value != actual.Value { - b.Errorf("3.1. Got %v; want %v", actual.Value, expected.Value) - } - expected = samples[len(samples)-1] - actual = it.ValueAtOrBeforeTime(expected.Timestamp + 1) - if expected.Timestamp != actual.Timestamp { - b.Errorf("3.2. Got %v; want %v", actual.Timestamp, expected.Timestamp) - } - if expected.Value != actual.Value { - b.Errorf("3.2. Got %v; want %v", actual.Value, expected.Value) - } - } -} - -func BenchmarkValueAtOrBeforeTimeChunkType0(b *testing.B) { - benchmarkValueAtOrBeforeTime(b, 0) -} - -func BenchmarkValueAtTimeChunkType1(b *testing.B) { - benchmarkValueAtOrBeforeTime(b, 1) -} - -func BenchmarkValueAtTimeChunkType2(b *testing.B) { - benchmarkValueAtOrBeforeTime(b, 2) -} - -func testRangeValues(t *testing.T, encoding chunk.Encoding) { - samples := make(model.Samples, 10000) - for i := range samples { - samples[i] = &model.Sample{ - Timestamp: model.Time(2 * i), - Value: model.SampleValue(float64(i) * 0.2), - } - } - s, closer := NewTestStorage(t, encoding) - defer closer.Close() - - for _, sample := range samples { - s.Append(sample) - } - s.WaitForIndexing() - - fp := model.Metric{}.FastFingerprint() - - it := s.preloadChunksForRange(makeFingerprintSeriesPair(s, fp), model.Earliest, model.Latest) - - // #1 Zero length interval at sample. - for i, expected := range samples { - actual := it.RangeValues(metric.Interval{ - OldestInclusive: expected.Timestamp, - NewestInclusive: expected.Timestamp, - }) - - if len(actual) != 1 { - t.Fatalf("1.%d. Expected exactly one result, got %d.", i, len(actual)) - } - if expected.Timestamp != actual[0].Timestamp { - t.Errorf("1.%d. Got %v; want %v.", i, actual[0].Timestamp, expected.Timestamp) - } - if expected.Value != actual[0].Value { - t.Errorf("1.%d. Got %v; want %v.", i, actual[0].Value, expected.Value) - } - } - - // #2 Zero length interval off sample. - for i, expected := range samples { - actual := it.RangeValues(metric.Interval{ - OldestInclusive: expected.Timestamp + 1, - NewestInclusive: expected.Timestamp + 1, - }) - - if len(actual) != 0 { - t.Fatalf("2.%d. Expected no result, got %d.", i, len(actual)) - } - } - - // #3 2sec interval around sample. - for i, expected := range samples { - actual := it.RangeValues(metric.Interval{ - OldestInclusive: expected.Timestamp - 1, - NewestInclusive: expected.Timestamp + 1, - }) - - if len(actual) != 1 { - t.Fatalf("3.%d. Expected exactly one result, got %d.", i, len(actual)) - } - if expected.Timestamp != actual[0].Timestamp { - t.Errorf("3.%d. Got %v; want %v.", i, actual[0].Timestamp, expected.Timestamp) - } - if expected.Value != actual[0].Value { - t.Errorf("3.%d. Got %v; want %v.", i, actual[0].Value, expected.Value) - } - } - - // #4 2sec interval sample to sample. - for i, expected1 := range samples { - if i == len(samples)-1 { - continue - } - expected2 := samples[i+1] - actual := it.RangeValues(metric.Interval{ - OldestInclusive: expected1.Timestamp, - NewestInclusive: expected1.Timestamp + 2, - }) - - if len(actual) != 2 { - t.Fatalf("4.%d. Expected exactly 2 results, got %d.", i, len(actual)) - } - if expected1.Timestamp != actual[0].Timestamp { - t.Errorf("4.%d. Got %v for 1st result; want %v.", i, actual[0].Timestamp, expected1.Timestamp) - } - if expected1.Value != actual[0].Value { - t.Errorf("4.%d. Got %v for 1st result; want %v.", i, actual[0].Value, expected1.Value) - } - if expected2.Timestamp != actual[1].Timestamp { - t.Errorf("4.%d. Got %v for 2nd result; want %v.", i, actual[1].Timestamp, expected2.Timestamp) - } - if expected2.Value != actual[1].Value { - t.Errorf("4.%d. Got %v for 2nd result; want %v.", i, actual[1].Value, expected2.Value) - } - } - - // #5 corner cases: Interval ends at first sample, interval starts - // at last sample, interval entirely before/after samples. - expected := samples[0] - actual := it.RangeValues(metric.Interval{ - OldestInclusive: expected.Timestamp - 2, - NewestInclusive: expected.Timestamp, - }) - if len(actual) != 1 { - t.Fatalf("5.1. Expected exactly one result, got %d.", len(actual)) - } - if expected.Timestamp != actual[0].Timestamp { - t.Errorf("5.1. Got %v; want %v.", actual[0].Timestamp, expected.Timestamp) - } - if expected.Value != actual[0].Value { - t.Errorf("5.1. Got %v; want %v.", actual[0].Value, expected.Value) - } - expected = samples[len(samples)-1] - actual = it.RangeValues(metric.Interval{ - OldestInclusive: expected.Timestamp, - NewestInclusive: expected.Timestamp + 2, - }) - if len(actual) != 1 { - t.Fatalf("5.2. Expected exactly one result, got %d.", len(actual)) - } - if expected.Timestamp != actual[0].Timestamp { - t.Errorf("5.2. Got %v; want %v.", actual[0].Timestamp, expected.Timestamp) - } - if expected.Value != actual[0].Value { - t.Errorf("5.2. Got %v; want %v.", actual[0].Value, expected.Value) - } - firstSample := samples[0] - actual = it.RangeValues(metric.Interval{ - OldestInclusive: firstSample.Timestamp - 4, - NewestInclusive: firstSample.Timestamp - 2, - }) - if len(actual) != 0 { - t.Fatalf("5.3. Expected no results, got %d.", len(actual)) - } - lastSample := samples[len(samples)-1] - actual = it.RangeValues(metric.Interval{ - OldestInclusive: lastSample.Timestamp + 2, - NewestInclusive: lastSample.Timestamp + 4, - }) - if len(actual) != 0 { - t.Fatalf("5.3. Expected no results, got %d.", len(actual)) - } -} - -func TestRangeValuesChunkType0(t *testing.T) { - testRangeValues(t, 0) -} - -func TestRangeValuesChunkType1(t *testing.T) { - testRangeValues(t, 1) -} - -func TestRangeValuesChunkType2(t *testing.T) { - testRangeValues(t, 2) -} - -func benchmarkRangeValues(b *testing.B, encoding chunk.Encoding) { - samples := make(model.Samples, 10000) - for i := range samples { - samples[i] = &model.Sample{ - Timestamp: model.Time(2 * i), - Value: model.SampleValue(float64(i) * 0.2), - } - } - s, closer := NewTestStorage(b, encoding) - defer closer.Close() - - for _, sample := range samples { - s.Append(sample) - } - s.WaitForIndexing() - - fp := model.Metric{}.FastFingerprint() - - it := s.preloadChunksForRange(makeFingerprintSeriesPair(s, fp), model.Earliest, model.Latest) - - b.ResetTimer() - - for i := 0; i < b.N; i++ { - for _, sample := range samples { - actual := it.RangeValues(metric.Interval{ - OldestInclusive: sample.Timestamp - 20, - NewestInclusive: sample.Timestamp + 20, - }) - - if len(actual) < 10 { - b.Fatalf("not enough samples found") - } - } - } -} - -func BenchmarkRangeValuesChunkType0(b *testing.B) { - benchmarkRangeValues(b, 0) -} - -func BenchmarkRangeValuesChunkType1(b *testing.B) { - benchmarkRangeValues(b, 1) -} - -func BenchmarkRangeValuesChunkType2(b *testing.B) { - benchmarkRangeValues(b, 2) -} - -func testEvictAndPurgeSeries(t *testing.T, encoding chunk.Encoding) { - samples := make(model.Samples, 10000) - for i := range samples { - samples[i] = &model.Sample{ - Timestamp: model.Time(2 * i), - Value: model.SampleValue(float64(i * i)), - } - } - s, closer := NewTestStorage(t, encoding) - defer closer.Close() - - for _, sample := range samples { - s.Append(sample) - } - s.WaitForIndexing() - - fp := model.Metric{}.FastFingerprint() - - // Drop ~half of the chunks. - s.maintainMemorySeries(fp, 10000) - it := s.preloadChunksForRange(makeFingerprintSeriesPair(s, fp), model.Earliest, model.Latest) - actual := it.RangeValues(metric.Interval{ - OldestInclusive: 0, - NewestInclusive: 100000, - }) - if len(actual) < 4000 { - t.Fatalf("expected more than %d results after purging half of series, got %d", 4000, len(actual)) - } - if actual[0].Timestamp < 6000 || actual[0].Timestamp > 10000 { - t.Errorf("1st timestamp out of expected range: %v", actual[0].Timestamp) - } - want := model.Time(19998) - if actual[len(actual)-1].Timestamp != want { - t.Errorf("2nd timestamp: want %v, got %v", want, actual[1].Timestamp) - } - - // Drop everything. - s.maintainMemorySeries(fp, 100000) - it = s.preloadChunksForRange(makeFingerprintSeriesPair(s, fp), model.Earliest, model.Latest) - actual = it.RangeValues(metric.Interval{ - OldestInclusive: 0, - NewestInclusive: 100000, - }) - if len(actual) != 0 { - t.Fatal("expected zero results after purging the whole series") - } - - // Recreate series. - for _, sample := range samples { - s.Append(sample) - } - s.WaitForIndexing() - - series, ok := s.fpToSeries.get(fp) - if !ok { - t.Fatal("could not find series") - } - - // Persist head chunk so we can safely archive. - series.headChunkClosed = true - s.maintainMemorySeries(fp, model.Earliest) - - // Archive metrics. - s.fpToSeries.del(fp) - lastTime, err := series.head().LastTime() - if err != nil { - t.Fatal(err) - } - s.persistence.archiveMetric(fp, series.metric, series.firstTime(), lastTime) - archived, _, _ := s.persistence.hasArchivedMetric(fp) - if !archived { - t.Fatal("not archived") - } - - // Drop ~half of the chunks of an archived series. - s.maintainArchivedSeries(fp, 10000) - archived, _, _ = s.persistence.hasArchivedMetric(fp) - if !archived { - t.Fatal("archived series purged although only half of the chunks dropped") - } - - // Drop everything. - s.maintainArchivedSeries(fp, 100000) - archived, _, _ = s.persistence.hasArchivedMetric(fp) - if archived { - t.Fatal("archived series not dropped") - } - - // Recreate series. - for _, sample := range samples { - s.Append(sample) - } - s.WaitForIndexing() - - series, ok = s.fpToSeries.get(fp) - if !ok { - t.Fatal("could not find series") - } - - // Persist head chunk so we can safely archive. - series.headChunkClosed = true - s.maintainMemorySeries(fp, model.Earliest) - - // Archive metrics. - s.fpToSeries.del(fp) - lastTime, err = series.head().LastTime() - if err != nil { - t.Fatal(err) - } - s.persistence.archiveMetric(fp, series.metric, series.firstTime(), lastTime) - archived, _, _ = s.persistence.hasArchivedMetric(fp) - if !archived { - t.Fatal("not archived") - } - - // Unarchive metrics. - s.getOrCreateSeries(fp, model.Metric{}) - - series, ok = s.fpToSeries.get(fp) - if !ok { - t.Fatal("could not find series") - } - archived, _, _ = s.persistence.hasArchivedMetric(fp) - if archived { - t.Fatal("archived") - } - - // Set archiveHighWatermark to a low value so that we can see it increase. - s.archiveHighWatermark = 42 - - // This will archive again, but must not drop it completely, despite the - // memorySeries being empty. - s.maintainMemorySeries(fp, 10000) - archived, _, _ = s.persistence.hasArchivedMetric(fp) - if !archived { - t.Fatal("series purged completely") - } - // archiveHighWatermark must have been set by maintainMemorySeries. - if want, got := model.Time(19998), s.archiveHighWatermark; want != got { - t.Errorf("want archiveHighWatermark %v, got %v", want, got) - } -} - -func TestEvictAndPurgeSeriesChunkType0(t *testing.T) { - testEvictAndPurgeSeries(t, 0) -} - -func TestEvictAndPurgeSeriesChunkType1(t *testing.T) { - testEvictAndPurgeSeries(t, 1) -} - -func TestEvictAndPurgeSeriesChunkType2(t *testing.T) { - testEvictAndPurgeSeries(t, 2) -} - -func testEvictAndLoadChunkDescs(t *testing.T, encoding chunk.Encoding) { - samples := make(model.Samples, 10000) - for i := range samples { - samples[i] = &model.Sample{ - Timestamp: model.Time(2 * i), - Value: model.SampleValue(float64(i * i)), - } - } - // Give last sample a timestamp of now so that the head chunk will not - // be closed (which would then archive the time series later as - // everything will get evicted). - samples[len(samples)-1] = &model.Sample{ - Timestamp: model.Now(), - Value: model.SampleValue(3.14), - } - - // Sadly, chunk.NumMemChunks is a global variable. We have to reset it - // explicitly here. - atomic.StoreInt64(&chunk.NumMemChunks, 0) - - s, closer := NewTestStorage(t, encoding) - defer closer.Close() - - // Adjust memory chunks to lower value to see evictions. - s.maxMemoryChunks = 1 - - for _, sample := range samples { - s.Append(sample) - } - s.WaitForIndexing() - - fp := model.Metric{}.FastFingerprint() - - series, ok := s.fpToSeries.get(fp) - if !ok { - t.Fatal("could not find series") - } - - oldLen := len(series.chunkDescs) - // Maintain series without any dropped chunks. - s.maintainMemorySeries(fp, 0) - // Give the evict goroutine an opportunity to run. - time.Sleep(250 * time.Millisecond) - // Maintain series again to trigger chunk.Desc eviction. - s.maintainMemorySeries(fp, 0) - - if oldLen <= len(series.chunkDescs) { - t.Errorf("Expected number of chunkDescs to decrease, old number %d, current number %d.", oldLen, len(series.chunkDescs)) - } - if int64(len(series.chunkDescs)) < atomic.LoadInt64(&chunk.NumMemChunks) { - t.Errorf("NumMemChunks is larger than number of chunk descs, number of chunk descs: %d, NumMemChunks: %d.", len(series.chunkDescs), atomic.LoadInt64(&chunk.NumMemChunks)) - } - - // Load everything back. - it := s.preloadChunksForRange(makeFingerprintSeriesPair(s, fp), 0, 100000) - - if oldLen != len(series.chunkDescs) { - t.Errorf("Expected number of chunkDescs to have reached old value again, old number %d, current number %d.", oldLen, len(series.chunkDescs)) - } - - it.Close() - - // Now maintain series with drops to make sure nothing crazy happens. - s.maintainMemorySeries(fp, 100000) - - if len(series.chunkDescs) != 1 { - t.Errorf("Expected exactly one chunk.Desc left, got %d.", len(series.chunkDescs)) - } -} - -func TestEvictAndLoadChunkDescsType0(t *testing.T) { - testEvictAndLoadChunkDescs(t, 0) -} - -func TestEvictAndLoadChunkDescsType1(t *testing.T) { - testEvictAndLoadChunkDescs(t, 1) -} - -func benchmarkAppend(b *testing.B, encoding chunk.Encoding) { - samples := make(model.Samples, b.N) - for i := range samples { - samples[i] = &model.Sample{ - Metric: model.Metric{ - model.MetricNameLabel: model.LabelValue(fmt.Sprintf("test_metric_%d", i%10)), - "label1": model.LabelValue(fmt.Sprintf("test_metric_%d", i%10)), - "label2": model.LabelValue(fmt.Sprintf("test_metric_%d", i%10)), - }, - Timestamp: model.Time(i), - Value: model.SampleValue(i), - } - } - b.ResetTimer() - s, closer := NewTestStorage(b, encoding) - defer closer.Close() - - for _, sample := range samples { - s.Append(sample) - } -} - -func BenchmarkAppendType0(b *testing.B) { - benchmarkAppend(b, 0) -} - -func BenchmarkAppendType1(b *testing.B) { - benchmarkAppend(b, 1) -} - -func BenchmarkAppendType2(b *testing.B) { - benchmarkAppend(b, 2) -} - -// Append a large number of random samples and then check if we can get them out -// of the storage alright. -func testFuzz(t *testing.T, encoding chunk.Encoding) { - if testing.Short() { - t.Skip("Skipping test in short mode.") - } - - check := func(seed int64) bool { - rand.Seed(seed) - s, c := NewTestStorage(t, encoding) - defer c.Close() - - samples := createRandomSamples("test_fuzz", 10000) - for _, sample := range samples { - s.Append(sample) - } - if !verifyStorageRandom(t, s, samples) { - return false - } - return verifyStorageSequential(t, s, samples) - } - - if err := quick.Check(check, nil); err != nil { - t.Fatal(err) - } -} - -func TestFuzzChunkType0(t *testing.T) { - testFuzz(t, 0) -} - -func TestFuzzChunkType1(t *testing.T) { - testFuzz(t, 1) -} - -func TestFuzzChunkType2(t *testing.T) { - testFuzz(t, 2) -} - -// benchmarkFuzz is the benchmark version of testFuzz. The storage options are -// set such that evictions, checkpoints, and purging will happen concurrently, -// too. This benchmark will have a very long runtime (up to minutes). You can -// use it as an actual benchmark. Run it like this: -// -// go test -cpu 1,2,4,8 -run=NONE -bench BenchmarkFuzzChunkType -benchmem -// -// You can also use it as a test for races. In that case, run it like this (will -// make things even slower): -// -// go test -race -cpu 8 -short -bench BenchmarkFuzzChunkType -func benchmarkFuzz(b *testing.B, encoding chunk.Encoding) { - chunk.DefaultEncoding = encoding - const samplesPerRun = 100000 - rand.Seed(42) - directory := testutil.NewTemporaryDirectory("test_storage", b) - defer directory.Close() - o := &MemorySeriesStorageOptions{ - MemoryChunks: 100, - MaxChunksToPersist: 1000000, - PersistenceRetentionPeriod: time.Hour, - PersistenceStoragePath: directory.Path(), - CheckpointInterval: time.Second, - SyncStrategy: Adaptive, - MinShrinkRatio: 0.1, - } - s := NewMemorySeriesStorage(o) - if err := s.Start(); err != nil { - b.Fatalf("Error starting storage: %s", err) - } - s.Start() - defer s.Stop() - - samples := createRandomSamples("benchmark_fuzz", samplesPerRun*b.N) - - b.ResetTimer() - - for i := 0; i < b.N; i++ { - start := samplesPerRun * i - end := samplesPerRun * (i + 1) - middle := (start + end) / 2 - for _, sample := range samples[start:middle] { - s.Append(sample) - } - verifyStorageRandom(b, s, samples[:middle]) - for _, sample := range samples[middle:end] { - s.Append(sample) - } - verifyStorageRandom(b, s, samples[:end]) - verifyStorageSequential(b, s, samples) - } -} - -func BenchmarkFuzzChunkType0(b *testing.B) { - benchmarkFuzz(b, 0) -} - -func BenchmarkFuzzChunkType1(b *testing.B) { - benchmarkFuzz(b, 1) -} - -func BenchmarkFuzzChunkType2(b *testing.B) { - benchmarkFuzz(b, 2) -} - -func createRandomSamples(metricName string, minLen int) model.Samples { - type valueCreator func() model.SampleValue - type deltaApplier func(model.SampleValue) model.SampleValue - - var ( - maxMetrics = 5 - maxStreakLength = 2000 - maxTimeDelta = 10000 - timestamp = model.Now() - model.Time(maxTimeDelta*minLen) // So that some timestamps are in the future. - generators = []struct { - createValue valueCreator - applyDelta []deltaApplier - }{ - { // "Boolean". - createValue: func() model.SampleValue { - return model.SampleValue(rand.Intn(2)) - }, - applyDelta: []deltaApplier{ - func(_ model.SampleValue) model.SampleValue { - return model.SampleValue(rand.Intn(2)) - }, - }, - }, - { // Integer with int deltas of various byte length. - createValue: func() model.SampleValue { - return model.SampleValue(rand.Int63() - 1<<62) - }, - applyDelta: []deltaApplier{ - func(v model.SampleValue) model.SampleValue { - return model.SampleValue(rand.Intn(1<<8) - 1<<7 + int(v)) - }, - func(v model.SampleValue) model.SampleValue { - return model.SampleValue(rand.Intn(1<<16) - 1<<15 + int(v)) - }, - func(v model.SampleValue) model.SampleValue { - return model.SampleValue(rand.Int63n(1<<32) - 1<<31 + int64(v)) - }, - }, - }, - { // Float with float32 and float64 deltas. - createValue: func() model.SampleValue { - return model.SampleValue(rand.NormFloat64()) - }, - applyDelta: []deltaApplier{ - func(v model.SampleValue) model.SampleValue { - return v + model.SampleValue(float32(rand.NormFloat64())) - }, - func(v model.SampleValue) model.SampleValue { - return v + model.SampleValue(rand.NormFloat64()) - }, - }, - }, - } - timestampIncrementers = []func(baseDelta model.Time) model.Time{ - // Regular increments. - func(delta model.Time) model.Time { - return delta - }, - // Jittered increments. σ is 1/100 of delta, e.g. 10ms for 10s scrape interval. - func(delta model.Time) model.Time { - return delta + model.Time(rand.NormFloat64()*float64(delta)/100) - }, - // Regular increments, but missing a scrape with 10% chance. - func(delta model.Time) model.Time { - i := rand.Intn(100) - if i < 90 { - return delta - } - if i < 99 { - return 2 * delta - } - return 3 * delta - // Ignoring the case with more than two missed scrapes in a row. - }, - } - ) - - // Prefill result with two samples with colliding metrics (to test fingerprint mapping). - result := model.Samples{ - &model.Sample{ - Metric: model.Metric{ - "instance": "ip-10-33-84-73.l05.ams5.s-cloud.net:24483", - "status": "503", - }, - Value: 42, - Timestamp: timestamp, - }, - &model.Sample{ - Metric: model.Metric{ - "instance": "ip-10-33-84-73.l05.ams5.s-cloud.net:24480", - "status": "500", - }, - Value: 2010, - Timestamp: timestamp + 1, - }, - } - - metrics := []model.Metric{} - for n := rand.Intn(maxMetrics); n >= 0; n-- { - metrics = append(metrics, model.Metric{ - model.MetricNameLabel: model.LabelValue(metricName), - model.LabelName(fmt.Sprintf("labelname_%d", n+1)): model.LabelValue(fmt.Sprintf("labelvalue_%d", rand.Int())), - }) - } - - for len(result) < minLen { - var ( - // Pick a metric for this cycle. - metric = metrics[rand.Intn(len(metrics))] - timeDelta = model.Time(rand.Intn(maxTimeDelta) + 1) - generator = generators[rand.Intn(len(generators))] - createValue = generator.createValue - applyDelta = generator.applyDelta[rand.Intn(len(generator.applyDelta))] - incTimestamp = timestampIncrementers[rand.Intn(len(timestampIncrementers))] - ) - - switch rand.Intn(4) { - case 0: // A single sample. - result = append(result, &model.Sample{ - Metric: metric, - Value: createValue(), - Timestamp: timestamp, - }) - timestamp += incTimestamp(timeDelta) - case 1: // A streak of random sample values. - for n := rand.Intn(maxStreakLength); n >= 0; n-- { - result = append(result, &model.Sample{ - Metric: metric, - Value: createValue(), - Timestamp: timestamp, - }) - timestamp += incTimestamp(timeDelta) - } - case 2: // A streak of sample values with incremental changes. - value := createValue() - for n := rand.Intn(maxStreakLength); n >= 0; n-- { - result = append(result, &model.Sample{ - Metric: metric, - Value: value, - Timestamp: timestamp, - }) - timestamp += incTimestamp(timeDelta) - value = applyDelta(value) - } - case 3: // A streak of constant sample values. - value := createValue() - for n := rand.Intn(maxStreakLength); n >= 0; n-- { - result = append(result, &model.Sample{ - Metric: metric, - Value: value, - Timestamp: timestamp, - }) - timestamp += incTimestamp(timeDelta) - } - } - } - - return result -} - -func verifyStorageRandom(t testing.TB, s *MemorySeriesStorage, samples model.Samples) bool { - s.WaitForIndexing() - result := true - for _, i := range rand.Perm(len(samples)) { - sample := samples[i] - fp := s.mapper.mapFP(sample.Metric.FastFingerprint(), sample.Metric) - it := s.preloadChunksForInstant(makeFingerprintSeriesPair(s, fp), sample.Timestamp, sample.Timestamp) - found := it.ValueAtOrBeforeTime(sample.Timestamp) - startTime := it.(*boundedIterator).start - switch { - case found.Timestamp != model.Earliest && sample.Timestamp.Before(startTime): - t.Errorf("Sample #%d %#v: Expected outdated sample to be excluded.", i, sample) - result = false - case found.Timestamp == model.Earliest && !sample.Timestamp.Before(startTime): - t.Errorf("Sample #%d %#v: Expected sample not found.", i, sample) - result = false - case found.Timestamp == model.Earliest && sample.Timestamp.Before(startTime): - // All good. Outdated sample dropped. - case sample.Value != found.Value || sample.Timestamp != found.Timestamp: - t.Errorf( - "Sample #%d %#v: Value (or timestamp) mismatch, want %f (at time %v), got %f (at time %v).", - i, sample, sample.Value, sample.Timestamp, found.Value, found.Timestamp, - ) - result = false - } - it.Close() - } - return result -} - -func verifyStorageSequential(t testing.TB, s *MemorySeriesStorage, samples model.Samples) bool { - s.WaitForIndexing() - var ( - result = true - fp model.Fingerprint - it SeriesIterator - r []model.SamplePair - j int - ) - defer func() { - it.Close() - }() - for i, sample := range samples { - newFP := s.mapper.mapFP(sample.Metric.FastFingerprint(), sample.Metric) - if it == nil || newFP != fp { - fp = newFP - if it != nil { - it.Close() - } - it = s.preloadChunksForRange(makeFingerprintSeriesPair(s, fp), sample.Timestamp, model.Latest) - r = it.RangeValues(metric.Interval{ - OldestInclusive: sample.Timestamp, - NewestInclusive: model.Latest, - }) - j = -1 - } - startTime := it.(*boundedIterator).start - if sample.Timestamp.Before(startTime) { - continue - } - j++ - if j >= len(r) { - t.Errorf( - "Sample #%d %v not found.", - i, sample, - ) - result = false - continue - } - found := r[j] - if sample.Value != found.Value || sample.Timestamp != found.Timestamp { - t.Errorf( - "Sample #%d %v: Value (or timestamp) mismatch, want %f (at time %v), got %f (at time %v).", - i, sample, sample.Value, sample.Timestamp, found.Value, found.Timestamp, - ) - result = false - } - } - return result -} - -func TestAppendOutOfOrder(t *testing.T) { - s, closer := NewTestStorage(t, 2) - defer closer.Close() - - m := model.Metric{ - model.MetricNameLabel: "out_of_order", - } - - tests := []struct { - name string - timestamp model.Time - value model.SampleValue - wantErr error - }{ - { - name: "1st sample", - timestamp: 0, - value: 0, - wantErr: nil, - }, - { - name: "regular append", - timestamp: 2, - value: 1, - wantErr: nil, - }, - { - name: "same timestamp, same value (no-op)", - timestamp: 2, - value: 1, - wantErr: nil, - }, - { - name: "same timestamp, different value", - timestamp: 2, - value: 2, - wantErr: ErrDuplicateSampleForTimestamp, - }, - { - name: "earlier timestamp, same value", - timestamp: 1, - value: 2, - wantErr: ErrOutOfOrderSample, - }, - { - name: "earlier timestamp, different value", - timestamp: 1, - value: 3, - wantErr: ErrOutOfOrderSample, - }, - { - name: "regular append of NaN", - timestamp: 3, - value: model.SampleValue(math.NaN()), - wantErr: nil, - }, - { - name: "no-op append of NaN", - timestamp: 3, - value: model.SampleValue(math.NaN()), - wantErr: nil, - }, - { - name: "append of NaN with earlier timestamp", - timestamp: 2, - value: model.SampleValue(math.NaN()), - wantErr: ErrOutOfOrderSample, - }, - { - name: "append of normal sample after NaN with same timestamp", - timestamp: 3, - value: 3.14, - wantErr: ErrDuplicateSampleForTimestamp, - }, - } - - for _, test := range tests { - gotErr := s.Append(&model.Sample{ - Metric: m, - Timestamp: test.timestamp, - Value: test.value, - }) - if gotErr != test.wantErr { - t.Errorf("%s: got %q, want %q", test.name, gotErr, test.wantErr) - } - } - - fp := s.mapper.mapFP(m.FastFingerprint(), m) - - it := s.preloadChunksForRange(makeFingerprintSeriesPair(s, fp), 0, 2) - defer it.Close() - - want := []model.SamplePair{ - { - Timestamp: 0, - Value: 0, - }, - { - Timestamp: 2, - Value: 1, - }, - { - Timestamp: 3, - Value: model.SampleValue(math.NaN()), - }, - } - got := it.RangeValues(metric.Interval{OldestInclusive: 0, NewestInclusive: 3}) - // Note that we cannot just reflect.DeepEqual(want, got) because it has - // the semantics of NaN != NaN. - for i, gotSamplePair := range got { - wantSamplePair := want[i] - if !wantSamplePair.Equal(&gotSamplePair) { - t.Fatalf("want %v, got %v", wantSamplePair, gotSamplePair) - } - } -} diff --git a/storage/local/storagetool/main.go b/storage/local/storagetool/main.go deleted file mode 100644 index 88d44c44c3..0000000000 --- a/storage/local/storagetool/main.go +++ /dev/null @@ -1,59 +0,0 @@ -// Copyright 2016 The Prometheus Authors -// Licensed under the Apache License, Version 2.0 (the "License"); -// you may not use this file except in compliance with the License. -// You may obtain a copy of the License at -// -// http://www.apache.org/licenses/LICENSE-2.0 -// -// Unless required by applicable law or agreed to in writing, software -// distributed under the License is distributed on an "AS IS" BASIS, -// WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. -// See the License for the specific language governing permissions and -// limitations under the License. - -package main - -import ( - "fmt" - "os" - - "github.com/prometheus/common/version" - "github.com/prometheus/prometheus/storage/local" - "github.com/prometheus/prometheus/util/cli" -) - -// DumpHeadsCmd dumps metadata of a heads.db file. -func DumpHeadsCmd(t cli.Term, args ...string) int { - if len(args) != 1 { - t.Infof("usage: storagetool dump-heads ") - return 2 - } - if err := local.DumpHeads(args[0], t.Out()); err != nil { - t.Errorf(" FAILED: %s", err) - return 1 - } - return 0 -} - -// VersionCmd prints the binaries version information. -func VersionCmd(t cli.Term, _ ...string) int { - fmt.Fprintln(os.Stdout, version.Print("storagetool")) - return 0 -} - -func main() { - app := cli.NewApp("storagetool") - - app.Register("dump-heads", &cli.Command{ - Desc: "dump metadata of a heads.db checkpoint file", - Run: DumpHeadsCmd, - }) - - app.Register("version", &cli.Command{ - Desc: "print the version of this binary", - Run: VersionCmd, - }) - - t := cli.BasicTerm(os.Stdout, os.Stderr) - os.Exit(app.Run(t, os.Args[1:]...)) -} diff --git a/storage/local/test_helpers.go b/storage/local/test_helpers.go deleted file mode 100644 index bdc26c555b..0000000000 --- a/storage/local/test_helpers.go +++ /dev/null @@ -1,72 +0,0 @@ -// Copyright 2014 The Prometheus Authors -// Licensed under the Apache License, Version 2.0 (the "License"); -// you may not use this file except in compliance with the License. -// You may obtain a copy of the License at -// -// http://www.apache.org/licenses/LICENSE-2.0 -// -// Unless required by applicable law or agreed to in writing, software -// distributed under the License is distributed on an "AS IS" BASIS, -// WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. -// See the License for the specific language governing permissions and -// limitations under the License. - -// NOTE ON FILENAME: Do not rename this file helpers_test.go (which might appear -// an obvious choice). We need NewTestStorage in tests outside of the local -// package, too. On the other hand, moving NewTestStorage in its own package -// would cause circular dependencies in the tests in packages local. - -package local - -import ( - "time" - - "github.com/prometheus/common/model" - "github.com/prometheus/prometheus/storage/local/chunk" - "github.com/prometheus/prometheus/util/testutil" -) - -type testStorageCloser struct { - storage Storage - directory testutil.Closer -} - -func (t *testStorageCloser) Close() { - if err := t.storage.Stop(); err != nil { - panic(err) - } - t.directory.Close() -} - -// NewTestStorage creates a storage instance backed by files in a temporary -// directory. The returned storage is already in serving state. Upon closing the -// returned test.Closer, the temporary directory is cleaned up. -func NewTestStorage(t testutil.T, encoding chunk.Encoding) (*MemorySeriesStorage, testutil.Closer) { - chunk.DefaultEncoding = encoding - directory := testutil.NewTemporaryDirectory("test_storage", t) - o := &MemorySeriesStorageOptions{ - MemoryChunks: 1000000, - MaxChunksToPersist: 1000000, - PersistenceRetentionPeriod: 24 * time.Hour * 365 * 100, // Enough to never trigger purging. - PersistenceStoragePath: directory.Path(), - CheckpointInterval: time.Hour, - SyncStrategy: Adaptive, - } - storage := NewMemorySeriesStorage(o) - storage.archiveHighWatermark = model.Latest - if err := storage.Start(); err != nil { - directory.Close() - t.Fatalf("Error creating storage: %s", err) - } - - closer := &testStorageCloser{ - storage: storage, - directory: directory, - } - - return storage, closer -} - -func makeFingerprintSeriesPair(s *MemorySeriesStorage, fp model.Fingerprint) fingerprintSeriesPair { - return fingerprintSeriesPair{fp, s.seriesForRange(fp, model.Earliest, model.Latest)} -}