mirror of
https://github.com/prometheus/prometheus.git
synced 2025-01-06 03:17:26 -08:00
01f652cb4c
This was initially motivated by wanting to distribute the rule checker tool under `tools/rule_checker`. However, this was not possible without also distributing the LevelDB dynamic libraries because the tool transitively depended on Levigo: rule checker -> query layer -> tiered storage layer -> leveldb This change separates external storage interfaces from the implementation (tiered storage, leveldb storage, memory storage) by putting them into separate packages: - storage/metric: public, implementation-agnostic interfaces - storage/metric/tiered: tiered storage implementation, including memory and LevelDB storage. I initially also considered splitting up the implementation into separate packages for tiered storage, memory storage, and LevelDB storage, but these are currently so intertwined that it would be another major project in itself. The query layers and most other parts of Prometheus now have notion of the storage implementation anymore and just use whatever implementation they get passed in via interfaces. The rule_checker is now a static binary :) Change-Id: I793bbf631a8648ca31790e7e772ecf9c2b92f7a0
700 lines
20 KiB
Go
700 lines
20 KiB
Go
// Copyright 2013 Prometheus Team
|
|
// Licensed under the Apache License, Version 2.0 (the "License");
|
|
// you may not use this file except in compliance with the License.
|
|
// You may obtain a copy of the License at
|
|
//
|
|
// http://www.apache.org/licenses/LICENSE-2.0
|
|
//
|
|
// Unless required by applicable law or agreed to in writing, software
|
|
// distributed under the License is distributed on an "AS IS" BASIS,
|
|
// WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
|
|
// See the License for the specific language governing permissions and
|
|
// limitations under the License.
|
|
|
|
package tiered
|
|
|
|
import (
|
|
"flag"
|
|
"fmt"
|
|
"sort"
|
|
"sync"
|
|
"time"
|
|
|
|
"code.google.com/p/goprotobuf/proto"
|
|
"github.com/golang/glog"
|
|
|
|
clientmodel "github.com/prometheus/client_golang/model"
|
|
|
|
"github.com/prometheus/prometheus/storage"
|
|
"github.com/prometheus/prometheus/storage/metric"
|
|
"github.com/prometheus/prometheus/storage/raw"
|
|
"github.com/prometheus/prometheus/storage/raw/leveldb"
|
|
"github.com/prometheus/prometheus/utility"
|
|
|
|
dto "github.com/prometheus/prometheus/model/generated"
|
|
)
|
|
|
|
const sortConcurrency = 2
|
|
|
|
// LevelDBPersistence is a leveldb-backed persistence layer for metrics.
|
|
type LevelDBPersistence struct {
|
|
CurationRemarks CurationRemarker
|
|
FingerprintToMetrics FingerprintMetricIndex
|
|
LabelNameToLabelValues LabelNameLabelValuesIndex
|
|
LabelPairToFingerprints LabelPairFingerprintIndex
|
|
MetricHighWatermarks HighWatermarker
|
|
MetricMembershipIndex MetricMembershipIndex
|
|
|
|
Indexer MetricIndexer
|
|
|
|
MetricSamples *leveldb.LevelDBPersistence
|
|
|
|
// The remaining indices will be replaced with generalized interface resolvers:
|
|
//
|
|
// type FingerprintResolver interface {
|
|
// GetFingerprintForMetric(clientmodel.Metric) (*clientmodel.Fingerprint, bool, error)
|
|
// GetFingerprintsForLabelMatchers(metric.LabelPair) (clientmodel.Fingerprints, bool, error)
|
|
// }
|
|
|
|
// type MetricResolver interface {
|
|
// GetMetricsForFingerprint(clientmodel.Fingerprints) (FingerprintMetricMapping, bool, error)
|
|
// }
|
|
}
|
|
|
|
var (
|
|
leveldbChunkSize = flag.Int("leveldbChunkSize", 200, "Maximum number of samples stored under one key.")
|
|
|
|
// These flag values are back of the envelope, though they seem
|
|
// sensible. Please re-evaluate based on your own needs.
|
|
curationRemarksCacheSize = flag.Int("curationRemarksCacheSize", 5*1024*1024, "The size for the curation remarks cache (bytes).")
|
|
fingerprintsToLabelPairCacheSize = flag.Int("fingerprintsToLabelPairCacheSizeBytes", 25*1024*1024, "The size for the fingerprint to label pair index (bytes).")
|
|
highWatermarkCacheSize = flag.Int("highWatermarksByFingerprintSizeBytes", 5*1024*1024, "The size for the metric high watermarks (bytes).")
|
|
labelNameToLabelValuesCacheSize = flag.Int("labelNameToLabelValuesCacheSizeBytes", 25*1024*1024, "The size for the label name to label values index (bytes).")
|
|
labelPairToFingerprintsCacheSize = flag.Int("labelPairToFingerprintsCacheSizeBytes", 25*1024*1024, "The size for the label pair to metric fingerprint index (bytes).")
|
|
metricMembershipIndexCacheSize = flag.Int("metricMembershipCacheSizeBytes", 5*1024*1024, "The size for the metric membership index (bytes).")
|
|
samplesByFingerprintCacheSize = flag.Int("samplesByFingerprintCacheSizeBytes", 50*1024*1024, "The size for the samples database (bytes).")
|
|
)
|
|
|
|
type leveldbOpener func()
|
|
|
|
// Close closes all the underlying persistence layers. It implements the
|
|
// Persistence interface.
|
|
func (l *LevelDBPersistence) Close() {
|
|
var persistences = []raw.Database{
|
|
l.CurationRemarks,
|
|
l.FingerprintToMetrics,
|
|
l.LabelNameToLabelValues,
|
|
l.LabelPairToFingerprints,
|
|
l.MetricHighWatermarks,
|
|
l.MetricMembershipIndex,
|
|
l.MetricSamples,
|
|
}
|
|
|
|
closerGroup := sync.WaitGroup{}
|
|
|
|
for _, c := range persistences {
|
|
closerGroup.Add(1)
|
|
go func(c raw.Database) {
|
|
if c != nil {
|
|
if err := c.Close(); err != nil {
|
|
glog.Error("Error closing persistence: ", err)
|
|
}
|
|
}
|
|
closerGroup.Done()
|
|
}(c)
|
|
}
|
|
|
|
closerGroup.Wait()
|
|
}
|
|
|
|
// NewLevelDBPersistence returns a LevelDBPersistence object ready
|
|
// to use.
|
|
func NewLevelDBPersistence(baseDirectory string) (*LevelDBPersistence, error) {
|
|
workers := utility.NewUncertaintyGroup(7)
|
|
|
|
emission := &LevelDBPersistence{}
|
|
|
|
var subsystemOpeners = []struct {
|
|
name string
|
|
opener leveldbOpener
|
|
}{
|
|
{
|
|
"Label Names and Value Pairs by Fingerprint",
|
|
func() {
|
|
var err error
|
|
emission.FingerprintToMetrics, err = NewLevelDBFingerprintMetricIndex(
|
|
leveldb.LevelDBOptions{
|
|
Name: "Metrics by Fingerprint",
|
|
Purpose: "Index",
|
|
Path: baseDirectory + "/label_name_and_value_pairs_by_fingerprint",
|
|
CacheSizeBytes: *fingerprintsToLabelPairCacheSize,
|
|
},
|
|
)
|
|
workers.MayFail(err)
|
|
},
|
|
},
|
|
{
|
|
"Samples by Fingerprint",
|
|
func() {
|
|
var err error
|
|
emission.MetricSamples, err = leveldb.NewLevelDBPersistence(leveldb.LevelDBOptions{
|
|
Name: "Samples",
|
|
Purpose: "Timeseries",
|
|
Path: baseDirectory + "/samples_by_fingerprint",
|
|
CacheSizeBytes: *fingerprintsToLabelPairCacheSize,
|
|
})
|
|
workers.MayFail(err)
|
|
},
|
|
},
|
|
{
|
|
"High Watermarks by Fingerprint",
|
|
func() {
|
|
var err error
|
|
emission.MetricHighWatermarks, err = NewLevelDBHighWatermarker(
|
|
leveldb.LevelDBOptions{
|
|
Name: "High Watermarks",
|
|
Purpose: "The youngest sample in the database per metric.",
|
|
Path: baseDirectory + "/high_watermarks_by_fingerprint",
|
|
CacheSizeBytes: *highWatermarkCacheSize,
|
|
},
|
|
)
|
|
workers.MayFail(err)
|
|
},
|
|
},
|
|
{
|
|
"Fingerprints by Label Name",
|
|
func() {
|
|
var err error
|
|
emission.LabelNameToLabelValues, err = NewLevelDBLabelNameLabelValuesIndex(
|
|
leveldb.LevelDBOptions{
|
|
Name: "Label Values by Label Name",
|
|
Purpose: "Index",
|
|
Path: baseDirectory + "/label_values_by_label_name",
|
|
CacheSizeBytes: *labelNameToLabelValuesCacheSize,
|
|
},
|
|
)
|
|
workers.MayFail(err)
|
|
},
|
|
},
|
|
{
|
|
"Fingerprints by Label Name and Value Pair",
|
|
func() {
|
|
var err error
|
|
emission.LabelPairToFingerprints, err = NewLevelDBLabelSetFingerprintIndex(
|
|
leveldb.LevelDBOptions{
|
|
Name: "Fingerprints by Label Pair",
|
|
Purpose: "Index",
|
|
Path: baseDirectory + "/fingerprints_by_label_name_and_value_pair",
|
|
CacheSizeBytes: *labelPairToFingerprintsCacheSize,
|
|
},
|
|
)
|
|
workers.MayFail(err)
|
|
},
|
|
},
|
|
{
|
|
"Metric Membership Index",
|
|
func() {
|
|
var err error
|
|
emission.MetricMembershipIndex, err = NewLevelDBMetricMembershipIndex(
|
|
leveldb.LevelDBOptions{
|
|
Name: "Metric Membership",
|
|
Purpose: "Index",
|
|
Path: baseDirectory + "/metric_membership_index",
|
|
CacheSizeBytes: *metricMembershipIndexCacheSize,
|
|
},
|
|
)
|
|
workers.MayFail(err)
|
|
},
|
|
},
|
|
{
|
|
"Sample Curation Remarks",
|
|
func() {
|
|
var err error
|
|
emission.CurationRemarks, err = NewLevelDBCurationRemarker(
|
|
leveldb.LevelDBOptions{
|
|
Name: "Sample Curation Remarks",
|
|
Purpose: "Ledger of Progress for Various Curators",
|
|
Path: baseDirectory + "/curation_remarks",
|
|
CacheSizeBytes: *curationRemarksCacheSize,
|
|
},
|
|
)
|
|
workers.MayFail(err)
|
|
},
|
|
},
|
|
}
|
|
|
|
for _, subsystem := range subsystemOpeners {
|
|
opener := subsystem.opener
|
|
go opener()
|
|
}
|
|
|
|
if !workers.Wait() {
|
|
for _, err := range workers.Errors() {
|
|
glog.Error("Could not open storage: ", err)
|
|
}
|
|
|
|
return nil, fmt.Errorf("unable to open metric persistence")
|
|
}
|
|
|
|
emission.Indexer = &TotalIndexer{
|
|
FingerprintToMetric: emission.FingerprintToMetrics,
|
|
LabelNameToLabelValues: emission.LabelNameToLabelValues,
|
|
LabelPairToFingerprint: emission.LabelPairToFingerprints,
|
|
MetricMembership: emission.MetricMembershipIndex,
|
|
}
|
|
|
|
return emission, nil
|
|
}
|
|
|
|
// AppendSample implements the Persistence interface.
|
|
func (l *LevelDBPersistence) AppendSample(sample *clientmodel.Sample) (err error) {
|
|
defer func(begin time.Time) {
|
|
duration := time.Since(begin)
|
|
|
|
recordOutcome(duration, err, map[string]string{operation: appendSample, result: success}, map[string]string{operation: appendSample, result: failure})
|
|
}(time.Now())
|
|
|
|
err = l.AppendSamples(clientmodel.Samples{sample})
|
|
|
|
return
|
|
}
|
|
|
|
// groupByFingerprint collects all of the provided samples, groups them
|
|
// together by their respective metric fingerprint, and finally sorts
|
|
// them chronologically.
|
|
func groupByFingerprint(samples clientmodel.Samples) map[clientmodel.Fingerprint]clientmodel.Samples {
|
|
fingerprintToSamples := map[clientmodel.Fingerprint]clientmodel.Samples{}
|
|
|
|
for _, sample := range samples {
|
|
fingerprint := &clientmodel.Fingerprint{}
|
|
fingerprint.LoadFromMetric(sample.Metric)
|
|
samples := fingerprintToSamples[*fingerprint]
|
|
samples = append(samples, sample)
|
|
fingerprintToSamples[*fingerprint] = samples
|
|
}
|
|
|
|
sortingSemaphore := make(chan bool, sortConcurrency)
|
|
doneSorting := sync.WaitGroup{}
|
|
|
|
for _, samples := range fingerprintToSamples {
|
|
doneSorting.Add(1)
|
|
|
|
sortingSemaphore <- true
|
|
go func(samples clientmodel.Samples) {
|
|
sort.Sort(samples)
|
|
|
|
<-sortingSemaphore
|
|
doneSorting.Done()
|
|
}(samples)
|
|
}
|
|
|
|
doneSorting.Wait()
|
|
|
|
return fingerprintToSamples
|
|
}
|
|
|
|
func (l *LevelDBPersistence) refreshHighWatermarks(groups map[clientmodel.Fingerprint]clientmodel.Samples) (err error) {
|
|
defer func(begin time.Time) {
|
|
duration := time.Since(begin)
|
|
|
|
recordOutcome(duration, err, map[string]string{operation: refreshHighWatermarks, result: success}, map[string]string{operation: refreshHighWatermarks, result: failure})
|
|
}(time.Now())
|
|
|
|
b := FingerprintHighWatermarkMapping{}
|
|
for fp, ss := range groups {
|
|
if len(ss) == 0 {
|
|
continue
|
|
}
|
|
|
|
b[fp] = ss[len(ss)-1].Timestamp
|
|
}
|
|
|
|
return l.MetricHighWatermarks.UpdateBatch(b)
|
|
}
|
|
|
|
// AppendSamples appends the given Samples to the database and indexes them.
|
|
func (l *LevelDBPersistence) AppendSamples(samples clientmodel.Samples) (err error) {
|
|
defer func(begin time.Time) {
|
|
duration := time.Since(begin)
|
|
|
|
recordOutcome(duration, err, map[string]string{operation: appendSamples, result: success}, map[string]string{operation: appendSamples, result: failure})
|
|
}(time.Now())
|
|
|
|
fingerprintToSamples := groupByFingerprint(samples)
|
|
indexErrChan := make(chan error, 1)
|
|
watermarkErrChan := make(chan error, 1)
|
|
|
|
go func(groups map[clientmodel.Fingerprint]clientmodel.Samples) {
|
|
metrics := FingerprintMetricMapping{}
|
|
|
|
for fingerprint, samples := range groups {
|
|
metrics[fingerprint] = samples[0].Metric
|
|
}
|
|
|
|
indexErrChan <- l.Indexer.IndexMetrics(metrics)
|
|
}(fingerprintToSamples)
|
|
|
|
go func(groups map[clientmodel.Fingerprint]clientmodel.Samples) {
|
|
watermarkErrChan <- l.refreshHighWatermarks(groups)
|
|
}(fingerprintToSamples)
|
|
|
|
samplesBatch := leveldb.NewBatch()
|
|
defer samplesBatch.Close()
|
|
|
|
key := &SampleKey{}
|
|
keyDto := &dto.SampleKey{}
|
|
values := make(metric.Values, 0, *leveldbChunkSize)
|
|
|
|
for fingerprint, group := range fingerprintToSamples {
|
|
for {
|
|
values := values[:0]
|
|
lengthOfGroup := len(group)
|
|
|
|
if lengthOfGroup == 0 {
|
|
break
|
|
}
|
|
|
|
take := *leveldbChunkSize
|
|
if lengthOfGroup < take {
|
|
take = lengthOfGroup
|
|
}
|
|
|
|
chunk := group[0:take]
|
|
group = group[take:lengthOfGroup]
|
|
|
|
key.Fingerprint = &fingerprint
|
|
key.FirstTimestamp = chunk[0].Timestamp
|
|
key.LastTimestamp = chunk[take-1].Timestamp
|
|
key.SampleCount = uint32(take)
|
|
|
|
key.Dump(keyDto)
|
|
|
|
for _, sample := range chunk {
|
|
values = append(values, metric.SamplePair{
|
|
Timestamp: sample.Timestamp,
|
|
Value: sample.Value,
|
|
})
|
|
}
|
|
val := marshalValues(values, nil)
|
|
samplesBatch.PutRaw(keyDto, val)
|
|
}
|
|
}
|
|
|
|
err = l.MetricSamples.Commit(samplesBatch)
|
|
if err != nil {
|
|
return
|
|
}
|
|
|
|
err = <-indexErrChan
|
|
if err != nil {
|
|
return
|
|
}
|
|
|
|
err = <-watermarkErrChan
|
|
if err != nil {
|
|
return
|
|
}
|
|
|
|
return
|
|
}
|
|
|
|
func extractSampleKey(i leveldb.Iterator) (*SampleKey, error) {
|
|
k := &dto.SampleKey{}
|
|
if err := i.Key(k); err != nil {
|
|
return nil, err
|
|
}
|
|
|
|
key := &SampleKey{}
|
|
key.Load(k)
|
|
|
|
return key, nil
|
|
}
|
|
|
|
func (l *LevelDBPersistence) hasIndexMetric(m clientmodel.Metric) (value bool, err error) {
|
|
defer func(begin time.Time) {
|
|
duration := time.Since(begin)
|
|
|
|
recordOutcome(duration, err, map[string]string{operation: hasIndexMetric, result: success}, map[string]string{operation: hasIndexMetric, result: failure})
|
|
}(time.Now())
|
|
|
|
return l.MetricMembershipIndex.Has(m)
|
|
}
|
|
|
|
// GetFingerprintsForLabelMatchers returns the Fingerprints for the given
|
|
// LabelMatchers by querying the underlying LabelPairFingerprintIndex and
|
|
// possibly the LabelNameLabelValuesIndex for each matcher. It implements the
|
|
// Persistence interface.
|
|
func (l *LevelDBPersistence) GetFingerprintsForLabelMatchers(labelMatchers metric.LabelMatchers) (fps clientmodel.Fingerprints, err error) {
|
|
defer func(begin time.Time) {
|
|
duration := time.Since(begin)
|
|
|
|
recordOutcome(duration, err, map[string]string{operation: getFingerprintsForLabelMatchers, result: success}, map[string]string{operation: getFingerprintsForLabelMatchers, result: failure})
|
|
}(time.Now())
|
|
|
|
sets := []utility.Set{}
|
|
|
|
for _, matcher := range labelMatchers {
|
|
set := utility.Set{}
|
|
|
|
switch matcher.Type {
|
|
case metric.Equal:
|
|
fps, _, err := l.LabelPairToFingerprints.Lookup(&metric.LabelPair{
|
|
Name: matcher.Name,
|
|
Value: matcher.Value,
|
|
})
|
|
if err != nil {
|
|
return nil, err
|
|
}
|
|
|
|
for _, fp := range fps {
|
|
set.Add(*fp)
|
|
}
|
|
default:
|
|
values, err := l.GetLabelValuesForLabelName(matcher.Name)
|
|
if err != nil {
|
|
return nil, err
|
|
}
|
|
matches := matcher.Filter(values)
|
|
if len(matches) == 0 {
|
|
return nil, nil
|
|
}
|
|
for _, v := range matches {
|
|
fps, _, err := l.LabelPairToFingerprints.Lookup(&metric.LabelPair{
|
|
Name: matcher.Name,
|
|
Value: v,
|
|
})
|
|
if err != nil {
|
|
return nil, err
|
|
}
|
|
for _, fp := range fps {
|
|
set.Add(*fp)
|
|
}
|
|
}
|
|
}
|
|
sets = append(sets, set)
|
|
}
|
|
|
|
numberOfSets := len(sets)
|
|
if numberOfSets == 0 {
|
|
return nil, nil
|
|
}
|
|
|
|
base := sets[0]
|
|
for i := 1; i < numberOfSets; i++ {
|
|
base = base.Intersection(sets[i])
|
|
}
|
|
for _, e := range base.Elements() {
|
|
fingerprint := e.(clientmodel.Fingerprint)
|
|
fps = append(fps, &fingerprint)
|
|
}
|
|
|
|
return fps, nil
|
|
}
|
|
|
|
// GetLabelValuesForLabelName returns the LabelValues for the given LabelName
|
|
// from the underlying LabelNameLabelValuesIndex. It implements the
|
|
// Persistence interface.
|
|
func (l *LevelDBPersistence) GetLabelValuesForLabelName(labelName clientmodel.LabelName) (clientmodel.LabelValues, error) {
|
|
var err error
|
|
defer func(begin time.Time) {
|
|
duration := time.Since(begin)
|
|
|
|
recordOutcome(duration, err, map[string]string{operation: getLabelValuesForLabelName, result: success}, map[string]string{operation: getLabelValuesForLabelName, result: failure})
|
|
}(time.Now())
|
|
|
|
values, _, err := l.LabelNameToLabelValues.Lookup(labelName)
|
|
|
|
return values, err
|
|
}
|
|
|
|
// GetMetricForFingerprint returns the Metric for the given Fingerprint from the
|
|
// underlying FingerprintMetricIndex. It implements the Persistence
|
|
// interface.
|
|
func (l *LevelDBPersistence) GetMetricForFingerprint(f *clientmodel.Fingerprint) (m clientmodel.Metric, err error) {
|
|
defer func(begin time.Time) {
|
|
duration := time.Since(begin)
|
|
|
|
recordOutcome(duration, err, map[string]string{operation: getMetricForFingerprint, result: success}, map[string]string{operation: getMetricForFingerprint, result: failure})
|
|
}(time.Now())
|
|
|
|
// TODO(matt): Update signature to work with ok.
|
|
m, _, err = l.FingerprintToMetrics.Lookup(f)
|
|
|
|
return m, nil
|
|
}
|
|
|
|
// GetAllValuesForLabel gets all label values that are associated with the
|
|
// provided label name.
|
|
func (l *LevelDBPersistence) GetAllValuesForLabel(labelName clientmodel.LabelName) (values clientmodel.LabelValues, err error) {
|
|
filter := &LabelNameFilter{
|
|
labelName: labelName,
|
|
}
|
|
labelValuesOp := &CollectLabelValuesOp{}
|
|
|
|
_, err = l.LabelPairToFingerprints.ForEach(&MetricKeyDecoder{}, filter, labelValuesOp)
|
|
if err != nil {
|
|
return
|
|
}
|
|
|
|
values = labelValuesOp.labelValues
|
|
return
|
|
}
|
|
|
|
// Prune compacts each database's keyspace serially.
|
|
//
|
|
// Beware that it would probably be imprudent to run this on a live user-facing
|
|
// server due to latency implications.
|
|
func (l *LevelDBPersistence) Prune() {
|
|
l.CurationRemarks.Prune()
|
|
l.FingerprintToMetrics.Prune()
|
|
l.LabelNameToLabelValues.Prune()
|
|
l.LabelPairToFingerprints.Prune()
|
|
l.MetricHighWatermarks.Prune()
|
|
l.MetricMembershipIndex.Prune()
|
|
l.MetricSamples.Prune()
|
|
}
|
|
|
|
// Sizes returns the sum of all sizes of the underlying databases.
|
|
func (l *LevelDBPersistence) Sizes() (total uint64, err error) {
|
|
size := uint64(0)
|
|
|
|
if size, err = l.CurationRemarks.Size(); err != nil {
|
|
return 0, err
|
|
}
|
|
total += size
|
|
|
|
if size, err = l.FingerprintToMetrics.Size(); err != nil {
|
|
return 0, err
|
|
}
|
|
total += size
|
|
|
|
if size, err = l.LabelNameToLabelValues.Size(); err != nil {
|
|
return 0, err
|
|
}
|
|
total += size
|
|
|
|
if size, err = l.LabelPairToFingerprints.Size(); err != nil {
|
|
return 0, err
|
|
}
|
|
total += size
|
|
|
|
if size, err = l.MetricHighWatermarks.Size(); err != nil {
|
|
return 0, err
|
|
}
|
|
total += size
|
|
|
|
if size, err = l.MetricMembershipIndex.Size(); err != nil {
|
|
return 0, err
|
|
}
|
|
total += size
|
|
|
|
if size, err = l.MetricSamples.Size(); err != nil {
|
|
return 0, err
|
|
}
|
|
total += size
|
|
|
|
return total, nil
|
|
}
|
|
|
|
// States returns the DatabaseStates of all underlying databases.
|
|
func (l *LevelDBPersistence) States() raw.DatabaseStates {
|
|
return raw.DatabaseStates{
|
|
l.CurationRemarks.State(),
|
|
l.FingerprintToMetrics.State(),
|
|
l.LabelNameToLabelValues.State(),
|
|
l.LabelPairToFingerprints.State(),
|
|
l.MetricHighWatermarks.State(),
|
|
l.MetricMembershipIndex.State(),
|
|
l.MetricSamples.State(),
|
|
}
|
|
}
|
|
|
|
// CollectLabelValuesOp implements storage.RecordOperator. It collects the
|
|
// encountered LabelValues in a slice.
|
|
type CollectLabelValuesOp struct {
|
|
labelValues []clientmodel.LabelValue
|
|
}
|
|
|
|
// Operate implements storage.RecordOperator. 'key' is required to be a
|
|
// LabelPair. Its Value is appended to a slice of collected LabelValues.
|
|
func (op *CollectLabelValuesOp) Operate(key, value interface{}) (err *storage.OperatorError) {
|
|
labelPair := key.(metric.LabelPair)
|
|
op.labelValues = append(op.labelValues, labelPair.Value)
|
|
return
|
|
}
|
|
|
|
// MetricKeyDecoder implements storage.RecordDecoder for LabelPairs.
|
|
type MetricKeyDecoder struct{}
|
|
|
|
// DecodeKey implements storage.RecordDecoder. It requires 'in' to be a
|
|
// LabelPair protobuf. 'out' is a metric.LabelPair.
|
|
func (d *MetricKeyDecoder) DecodeKey(in interface{}) (out interface{}, err error) {
|
|
unmarshaled := dto.LabelPair{}
|
|
err = proto.Unmarshal(in.([]byte), &unmarshaled)
|
|
if err != nil {
|
|
return
|
|
}
|
|
|
|
out = metric.LabelPair{
|
|
Name: clientmodel.LabelName(*unmarshaled.Name),
|
|
Value: clientmodel.LabelValue(*unmarshaled.Value),
|
|
}
|
|
|
|
return
|
|
}
|
|
|
|
// DecodeValue implements storage.RecordDecoder. It is a no-op and always
|
|
// returns (nil, nil).
|
|
func (d *MetricKeyDecoder) DecodeValue(in interface{}) (out interface{}, err error) {
|
|
return
|
|
}
|
|
|
|
// MetricSamplesDecoder implements storage.RecordDecoder for SampleKeys.
|
|
type MetricSamplesDecoder struct{}
|
|
|
|
// DecodeKey implements storage.RecordDecoder. It requires 'in' to be a
|
|
// SampleKey protobuf. 'out' is a metric.SampleKey.
|
|
func (d *MetricSamplesDecoder) DecodeKey(in interface{}) (interface{}, error) {
|
|
key := &dto.SampleKey{}
|
|
err := proto.Unmarshal(in.([]byte), key)
|
|
if err != nil {
|
|
return nil, err
|
|
}
|
|
|
|
sampleKey := &SampleKey{}
|
|
sampleKey.Load(key)
|
|
|
|
return sampleKey, nil
|
|
}
|
|
|
|
// DecodeValue implements storage.RecordDecoder. It requires 'in' to be a
|
|
// SampleValueSeries protobuf. 'out' is of type metric.Values.
|
|
func (d *MetricSamplesDecoder) DecodeValue(in interface{}) (interface{}, error) {
|
|
return unmarshalValues(in.([]byte), nil), nil
|
|
}
|
|
|
|
// AcceptAllFilter implements storage.RecordFilter and accepts all records.
|
|
type AcceptAllFilter struct{}
|
|
|
|
// Filter implements storage.RecordFilter. It always returns ACCEPT.
|
|
func (d *AcceptAllFilter) Filter(_, _ interface{}) storage.FilterResult {
|
|
return storage.Accept
|
|
}
|
|
|
|
// LabelNameFilter implements storage.RecordFilter and filters records matching
|
|
// a LabelName.
|
|
type LabelNameFilter struct {
|
|
labelName clientmodel.LabelName
|
|
}
|
|
|
|
// Filter implements storage.RecordFilter. 'key' is expected to be a
|
|
// LabelPair. The result is ACCEPT if the Name of the LabelPair matches the
|
|
// LabelName of this LabelNameFilter.
|
|
func (f LabelNameFilter) Filter(key, value interface{}) (filterResult storage.FilterResult) {
|
|
labelPair, ok := key.(metric.LabelPair)
|
|
if ok && labelPair.Name == f.labelName {
|
|
return storage.Accept
|
|
}
|
|
return storage.Skip
|
|
}
|