mirror of
https://github.com/prometheus/prometheus.git
synced 2024-11-11 08:04:04 -08:00
b3e34c6658
This commit introduces to Prometheus a batch database sample curator, which corroborates the high watermarks for sample series against the curation watermark table to see whether a curator of a given type needs to be run. The curator is an abstract executor, which runs various curation strategies across the database. It remarks the progress for each type of curation processor that runs for a given sample series. A curation procesor is responsible for effectuating the underlying batch changes that are request. In this commit, we introduce the CompactionProcessor, which takes several bits of runtime metadata and combine sparse sample entries in the database together to form larger groups. For instance, for a given series it would be possible to have the curator effectuate the following grouping: - Samples Older than Two Weeks: Grouped into Bunches of 10000 - Samples Older than One Week: Grouped into Bunches of 1000 - Samples Older than One Day: Grouped into Bunches of 100 - Samples Older than One Hour: Grouped into Bunches of 10 The benefits hereof of such a compaction are 1. a smaller search space in the database keyspace, 2. better employment of compression for repetious values, and 3. reduced seek times.
389 lines
9.3 KiB
Go
389 lines
9.3 KiB
Go
// Copyright 2013 Prometheus Team
|
|
// Licensed under the Apache License, Version 2.0 (the "License");
|
|
// you may not use this file except in compliance with the License.
|
|
// You may obtain a copy of the License at
|
|
//
|
|
// http://www.apache.org/licenses/LICENSE-2.0
|
|
//
|
|
// Unless required by applicable law or agreed to in writing, software
|
|
// distributed under the License is distributed on an "AS IS" BASIS,
|
|
// WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
|
|
// See the License for the specific language governing permissions and
|
|
// limitations under the License.
|
|
|
|
package leveldb
|
|
|
|
import (
|
|
"flag"
|
|
"fmt"
|
|
"github.com/jmhodges/levigo"
|
|
"github.com/prometheus/prometheus/coding"
|
|
"github.com/prometheus/prometheus/storage"
|
|
"github.com/prometheus/prometheus/storage/raw"
|
|
"time"
|
|
)
|
|
|
|
var (
|
|
leveldbFlushOnMutate = flag.Bool("leveldbFlushOnMutate", false, "Whether LevelDB should flush every operation to disk upon mutation before returning (bool).")
|
|
leveldbUseSnappy = flag.Bool("leveldbUseSnappy", true, "Whether LevelDB attempts to use Snappy for compressing elements (bool).")
|
|
leveldbUseParanoidChecks = flag.Bool("leveldbUseParanoidChecks", true, "Whether LevelDB uses expensive checks (bool).")
|
|
)
|
|
|
|
// LevelDBPersistence is a disk-backed sorted key-value store.
|
|
type LevelDBPersistence struct {
|
|
cache *levigo.Cache
|
|
filterPolicy *levigo.FilterPolicy
|
|
options *levigo.Options
|
|
storage *levigo.DB
|
|
readOptions *levigo.ReadOptions
|
|
writeOptions *levigo.WriteOptions
|
|
}
|
|
|
|
// levigoIterator wraps the LevelDB resources in a convenient manner for uniform
|
|
// resource access and closing through the raw.Iterator protocol.
|
|
type levigoIterator struct {
|
|
// iterator is the receiver of most proxied operation calls.
|
|
iterator *levigo.Iterator
|
|
// readOptions is only set if the iterator is a snapshot of an underlying
|
|
// database. This signals that it needs to be explicitly reaped upon the
|
|
// end of this iterator's life.
|
|
readOptions *levigo.ReadOptions
|
|
// snapshot is only set if the iterator is a snapshot of an underlying
|
|
// database. This signals that it needs to be explicitly reaped upon the
|
|
// end of this this iterator's life.
|
|
snapshot *levigo.Snapshot
|
|
// storage is only set if the iterator is a snapshot of an underlying
|
|
// database. This signals that it needs to be explicitly reaped upon the
|
|
// end of this this iterator's life. The snapshot must be freed in the
|
|
// context of an actual database.
|
|
storage *levigo.DB
|
|
// closed indicates whether the iterator has been closed before.
|
|
closed bool
|
|
// valid indicates whether the iterator may be used. If a LevelDB iterator
|
|
// ever becomes invalid, it must be disposed of and cannot be reused.
|
|
valid bool
|
|
// creationTime provides the time at which the iterator was made.
|
|
creationTime time.Time
|
|
}
|
|
|
|
func (i levigoIterator) String() string {
|
|
var (
|
|
valid = "valid"
|
|
open = "open"
|
|
snapshotted = "snapshotted"
|
|
)
|
|
|
|
if i.closed {
|
|
open = "closed"
|
|
}
|
|
if !i.valid {
|
|
valid = "invalid"
|
|
}
|
|
if i.snapshot == nil {
|
|
snapshotted = "unsnapshotted"
|
|
}
|
|
|
|
return fmt.Sprintf("levigoIterator created at %s that is %s and %s and %s", i.creationTime, open, valid, snapshotted)
|
|
}
|
|
|
|
func (i *levigoIterator) Close() {
|
|
if i.closed {
|
|
return
|
|
}
|
|
|
|
if i.iterator != nil {
|
|
i.iterator.Close()
|
|
}
|
|
if i.readOptions != nil {
|
|
i.readOptions.Close()
|
|
}
|
|
if i.snapshot != nil {
|
|
i.storage.ReleaseSnapshot(i.snapshot)
|
|
}
|
|
|
|
// Explicitly dereference the pointers to prevent cycles, however unlikely.
|
|
i.iterator = nil
|
|
i.readOptions = nil
|
|
i.snapshot = nil
|
|
i.storage = nil
|
|
|
|
i.closed = true
|
|
i.valid = false
|
|
|
|
return
|
|
}
|
|
|
|
func (i *levigoIterator) Seek(key []byte) bool {
|
|
i.iterator.Seek(key)
|
|
|
|
i.valid = i.iterator.Valid()
|
|
|
|
return i.valid
|
|
}
|
|
|
|
func (i *levigoIterator) SeekToFirst() bool {
|
|
i.iterator.SeekToFirst()
|
|
|
|
i.valid = i.iterator.Valid()
|
|
|
|
return i.valid
|
|
}
|
|
|
|
func (i *levigoIterator) SeekToLast() bool {
|
|
i.iterator.SeekToLast()
|
|
|
|
i.valid = i.iterator.Valid()
|
|
|
|
return i.valid
|
|
}
|
|
|
|
func (i *levigoIterator) Next() bool {
|
|
i.iterator.Next()
|
|
|
|
i.valid = i.iterator.Valid()
|
|
|
|
return i.valid
|
|
}
|
|
|
|
func (i *levigoIterator) Previous() bool {
|
|
i.iterator.Prev()
|
|
|
|
i.valid = i.iterator.Valid()
|
|
|
|
return i.valid
|
|
}
|
|
|
|
func (i levigoIterator) Key() (key []byte) {
|
|
return i.iterator.Key()
|
|
}
|
|
|
|
func (i levigoIterator) Value() (value []byte) {
|
|
return i.iterator.Value()
|
|
}
|
|
|
|
func (i levigoIterator) GetError() (err error) {
|
|
return i.iterator.GetError()
|
|
}
|
|
|
|
func NewLevelDBPersistence(storageRoot string, cacheCapacity, bitsPerBloomFilterEncoded int) (p *LevelDBPersistence, err error) {
|
|
options := levigo.NewOptions()
|
|
options.SetCreateIfMissing(true)
|
|
options.SetParanoidChecks(*leveldbUseParanoidChecks)
|
|
compression := levigo.NoCompression
|
|
if *leveldbUseSnappy {
|
|
compression = levigo.SnappyCompression
|
|
}
|
|
options.SetCompression(compression)
|
|
|
|
cache := levigo.NewLRUCache(cacheCapacity)
|
|
options.SetCache(cache)
|
|
|
|
filterPolicy := levigo.NewBloomFilter(bitsPerBloomFilterEncoded)
|
|
options.SetFilterPolicy(filterPolicy)
|
|
|
|
storage, err := levigo.Open(storageRoot, options)
|
|
if err != nil {
|
|
return
|
|
}
|
|
|
|
var (
|
|
readOptions = levigo.NewReadOptions()
|
|
writeOptions = levigo.NewWriteOptions()
|
|
)
|
|
|
|
writeOptions.SetSync(*leveldbFlushOnMutate)
|
|
p = &LevelDBPersistence{
|
|
cache: cache,
|
|
filterPolicy: filterPolicy,
|
|
options: options,
|
|
readOptions: readOptions,
|
|
storage: storage,
|
|
writeOptions: writeOptions,
|
|
}
|
|
|
|
return
|
|
}
|
|
|
|
func (l *LevelDBPersistence) Close() {
|
|
// These are deferred to take advantage of forced closing in case of stack
|
|
// unwinding due to anomalies.
|
|
defer func() {
|
|
if l.storage != nil {
|
|
l.storage.Close()
|
|
}
|
|
}()
|
|
|
|
defer func() {
|
|
if l.filterPolicy != nil {
|
|
l.filterPolicy.Close()
|
|
}
|
|
}()
|
|
|
|
defer func() {
|
|
if l.cache != nil {
|
|
l.cache.Close()
|
|
}
|
|
}()
|
|
|
|
defer func() {
|
|
if l.options != nil {
|
|
l.options.Close()
|
|
}
|
|
}()
|
|
|
|
defer func() {
|
|
if l.readOptions != nil {
|
|
l.readOptions.Close()
|
|
}
|
|
}()
|
|
|
|
defer func() {
|
|
if l.writeOptions != nil {
|
|
l.writeOptions.Close()
|
|
}
|
|
}()
|
|
|
|
return
|
|
}
|
|
|
|
func (l *LevelDBPersistence) Get(value coding.Encoder) (b []byte, err error) {
|
|
key, err := value.Encode()
|
|
if err != nil {
|
|
return
|
|
}
|
|
|
|
return l.storage.Get(l.readOptions, key)
|
|
}
|
|
|
|
func (l *LevelDBPersistence) Has(value coding.Encoder) (h bool, err error) {
|
|
raw, err := l.Get(value)
|
|
if err != nil {
|
|
return
|
|
}
|
|
|
|
h = raw != nil
|
|
|
|
return
|
|
}
|
|
|
|
func (l *LevelDBPersistence) Drop(value coding.Encoder) (err error) {
|
|
key, err := value.Encode()
|
|
if err != nil {
|
|
return
|
|
}
|
|
|
|
err = l.storage.Delete(l.writeOptions, key)
|
|
|
|
return
|
|
}
|
|
|
|
func (l *LevelDBPersistence) Put(key, value coding.Encoder) (err error) {
|
|
keyEncoded, err := key.Encode()
|
|
if err != nil {
|
|
return
|
|
}
|
|
|
|
valueEncoded, err := value.Encode()
|
|
if err != nil {
|
|
return
|
|
}
|
|
|
|
err = l.storage.Put(l.writeOptions, keyEncoded, valueEncoded)
|
|
|
|
return
|
|
}
|
|
|
|
func (l *LevelDBPersistence) Commit(b raw.Batch) (err error) {
|
|
// XXX: This is a wart to clean up later. Ideally, after doing extensive
|
|
// tests, we could create a Batch struct that journals pending
|
|
// operations which the given Persistence implementation could convert
|
|
// to its specific commit requirements.
|
|
batch, ok := b.(*batch)
|
|
if !ok {
|
|
panic("leveldb.batch expected")
|
|
}
|
|
|
|
return l.storage.Write(l.writeOptions, batch.batch)
|
|
}
|
|
|
|
// NewIterator creates a new levigoIterator, which follows the Iterator
|
|
// interface.
|
|
//
|
|
// Important notes:
|
|
//
|
|
// For each of the iterator methods that have a return signature of (ok bool),
|
|
// if ok == false, the iterator may not be used any further and must be closed.
|
|
// Further work with the database requires the creation of a new iterator. This
|
|
// is due to LevelDB and Levigo design. Please refer to Jeff and Sanjay's notes
|
|
// in the LevelDB documentation for this behavior's rationale.
|
|
//
|
|
// The returned iterator must explicitly be closed; otherwise non-managed memory
|
|
// will be leaked.
|
|
//
|
|
// The iterator is optionally snapshotable.
|
|
func (l *LevelDBPersistence) NewIterator(snapshotted bool) Iterator {
|
|
var (
|
|
snapshot *levigo.Snapshot
|
|
readOptions *levigo.ReadOptions
|
|
iterator *levigo.Iterator
|
|
)
|
|
|
|
if snapshotted {
|
|
snapshot = l.storage.NewSnapshot()
|
|
readOptions = levigo.NewReadOptions()
|
|
readOptions.SetSnapshot(snapshot)
|
|
iterator = l.storage.NewIterator(readOptions)
|
|
} else {
|
|
iterator = l.storage.NewIterator(l.readOptions)
|
|
}
|
|
|
|
return &levigoIterator{
|
|
creationTime: time.Now(),
|
|
iterator: iterator,
|
|
readOptions: readOptions,
|
|
snapshot: snapshot,
|
|
storage: l.storage,
|
|
}
|
|
}
|
|
|
|
func (l *LevelDBPersistence) ForEach(decoder storage.RecordDecoder, filter storage.RecordFilter, operator storage.RecordOperator) (scannedEntireCorpus bool, err error) {
|
|
var (
|
|
iterator = l.NewIterator(true)
|
|
valid bool
|
|
)
|
|
defer iterator.Close()
|
|
|
|
for valid = iterator.SeekToFirst(); valid; valid = iterator.Next() {
|
|
err = iterator.GetError()
|
|
if err != nil {
|
|
return
|
|
}
|
|
|
|
decodedKey, decodeErr := decoder.DecodeKey(iterator.Key())
|
|
if decodeErr != nil {
|
|
continue
|
|
}
|
|
decodedValue, decodeErr := decoder.DecodeValue(iterator.Value())
|
|
if decodeErr != nil {
|
|
continue
|
|
}
|
|
|
|
switch filter.Filter(decodedKey, decodedValue) {
|
|
case storage.STOP:
|
|
return
|
|
case storage.SKIP:
|
|
continue
|
|
case storage.ACCEPT:
|
|
opErr := operator.Operate(decodedKey, decodedValue)
|
|
if opErr != nil {
|
|
if opErr.Continuable {
|
|
continue
|
|
}
|
|
break
|
|
}
|
|
}
|
|
}
|
|
scannedEntireCorpus = true
|
|
return
|
|
}
|