mirror of
https://github.com/prometheus/prometheus.git
synced 2024-11-14 09:34:05 -08:00
2128d9d811
This change will only be submitted when the new client_golang has been moved to the new version. Change-Id: Ifceb59333072a08286a8ac910709a8ba2e3a1581
697 lines
20 KiB
Go
697 lines
20 KiB
Go
// Copyright 2013 Prometheus Team
|
|
// Licensed under the Apache License, Version 2.0 (the "License");
|
|
// you may not use this file except in compliance with the License.
|
|
// You may obtain a copy of the License at
|
|
//
|
|
// http://www.apache.org/licenses/LICENSE-2.0
|
|
//
|
|
// Unless required by applicable law or agreed to in writing, software
|
|
// distributed under the License is distributed on an "AS IS" BASIS,
|
|
// WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
|
|
// See the License for the specific language governing permissions and
|
|
// limitations under the License.
|
|
|
|
package tiered
|
|
|
|
import (
|
|
"flag"
|
|
"fmt"
|
|
"sort"
|
|
"sync"
|
|
"time"
|
|
|
|
"code.google.com/p/goprotobuf/proto"
|
|
"github.com/golang/glog"
|
|
|
|
clientmodel "github.com/prometheus/client_golang/model"
|
|
"github.com/prometheus/client_golang/prometheus"
|
|
|
|
"github.com/prometheus/prometheus/storage"
|
|
"github.com/prometheus/prometheus/storage/metric"
|
|
"github.com/prometheus/prometheus/storage/raw"
|
|
"github.com/prometheus/prometheus/storage/raw/leveldb"
|
|
"github.com/prometheus/prometheus/utility"
|
|
|
|
dto "github.com/prometheus/prometheus/model/generated"
|
|
)
|
|
|
|
const sortConcurrency = 2
|
|
|
|
// LevelDBPersistence is a leveldb-backed persistence layer for metrics.
|
|
type LevelDBPersistence struct {
|
|
CurationRemarks CurationRemarker
|
|
FingerprintToMetrics FingerprintMetricIndex
|
|
LabelNameToLabelValues LabelNameLabelValuesIndex
|
|
LabelPairToFingerprints LabelPairFingerprintIndex
|
|
MetricHighWatermarks HighWatermarker
|
|
MetricMembershipIndex MetricMembershipIndex
|
|
|
|
Indexer MetricIndexer
|
|
|
|
MetricSamples *leveldb.LevelDBPersistence
|
|
|
|
// The remaining indices will be replaced with generalized interface resolvers:
|
|
//
|
|
// type FingerprintResolver interface {
|
|
// GetFingerprintForMetric(clientmodel.Metric) (*clientmodel.Fingerprint, bool, error)
|
|
// GetFingerprintsForLabelMatchers(metric.LabelPair) (clientmodel.Fingerprints, bool, error)
|
|
// }
|
|
|
|
// type MetricResolver interface {
|
|
// GetMetricsForFingerprint(clientmodel.Fingerprints) (FingerprintMetricMapping, bool, error)
|
|
// }
|
|
}
|
|
|
|
var (
|
|
leveldbChunkSize = flag.Int("leveldbChunkSize", 200, "Maximum number of samples stored under one key.")
|
|
|
|
// These flag values are back of the envelope, though they seem
|
|
// sensible. Please re-evaluate based on your own needs.
|
|
curationRemarksCacheSize = flag.Int("curationRemarksCacheSize", 5*1024*1024, "The size for the curation remarks cache (bytes).")
|
|
fingerprintsToLabelPairCacheSize = flag.Int("fingerprintsToLabelPairCacheSizeBytes", 25*1024*1024, "The size for the fingerprint to label pair index (bytes).")
|
|
highWatermarkCacheSize = flag.Int("highWatermarksByFingerprintSizeBytes", 5*1024*1024, "The size for the metric high watermarks (bytes).")
|
|
labelNameToLabelValuesCacheSize = flag.Int("labelNameToLabelValuesCacheSizeBytes", 25*1024*1024, "The size for the label name to label values index (bytes).")
|
|
labelPairToFingerprintsCacheSize = flag.Int("labelPairToFingerprintsCacheSizeBytes", 25*1024*1024, "The size for the label pair to metric fingerprint index (bytes).")
|
|
metricMembershipIndexCacheSize = flag.Int("metricMembershipCacheSizeBytes", 5*1024*1024, "The size for the metric membership index (bytes).")
|
|
samplesByFingerprintCacheSize = flag.Int("samplesByFingerprintCacheSizeBytes", 50*1024*1024, "The size for the samples database (bytes).")
|
|
)
|
|
|
|
type leveldbOpener func()
|
|
|
|
// Close closes all the underlying persistence layers. It implements the
|
|
// Persistence interface.
|
|
func (l *LevelDBPersistence) Close() {
|
|
var persistences = []raw.Database{
|
|
l.CurationRemarks,
|
|
l.FingerprintToMetrics,
|
|
l.LabelNameToLabelValues,
|
|
l.LabelPairToFingerprints,
|
|
l.MetricHighWatermarks,
|
|
l.MetricMembershipIndex,
|
|
l.MetricSamples,
|
|
}
|
|
|
|
closerGroup := sync.WaitGroup{}
|
|
|
|
for _, c := range persistences {
|
|
closerGroup.Add(1)
|
|
go func(c raw.Database) {
|
|
if c != nil {
|
|
if err := c.Close(); err != nil {
|
|
glog.Error("Error closing persistence: ", err)
|
|
}
|
|
}
|
|
closerGroup.Done()
|
|
}(c)
|
|
}
|
|
|
|
closerGroup.Wait()
|
|
}
|
|
|
|
// NewLevelDBPersistence returns a LevelDBPersistence object ready
|
|
// to use.
|
|
func NewLevelDBPersistence(baseDirectory string) (*LevelDBPersistence, error) {
|
|
workers := utility.NewUncertaintyGroup(7)
|
|
|
|
emission := &LevelDBPersistence{}
|
|
|
|
var subsystemOpeners = []struct {
|
|
name string
|
|
opener leveldbOpener
|
|
}{
|
|
{
|
|
"Label Names and Value Pairs by Fingerprint",
|
|
func() {
|
|
var err error
|
|
emission.FingerprintToMetrics, err = NewLevelDBFingerprintMetricIndex(
|
|
leveldb.LevelDBOptions{
|
|
Name: "Metrics by Fingerprint",
|
|
Purpose: "Index",
|
|
Path: baseDirectory + "/label_name_and_value_pairs_by_fingerprint",
|
|
CacheSizeBytes: *fingerprintsToLabelPairCacheSize,
|
|
},
|
|
)
|
|
workers.MayFail(err)
|
|
},
|
|
},
|
|
{
|
|
"Samples by Fingerprint",
|
|
func() {
|
|
var err error
|
|
emission.MetricSamples, err = leveldb.NewLevelDBPersistence(leveldb.LevelDBOptions{
|
|
Name: "Samples",
|
|
Purpose: "Timeseries",
|
|
Path: baseDirectory + "/samples_by_fingerprint",
|
|
CacheSizeBytes: *fingerprintsToLabelPairCacheSize,
|
|
})
|
|
workers.MayFail(err)
|
|
},
|
|
},
|
|
{
|
|
"High Watermarks by Fingerprint",
|
|
func() {
|
|
var err error
|
|
emission.MetricHighWatermarks, err = NewLevelDBHighWatermarker(
|
|
leveldb.LevelDBOptions{
|
|
Name: "High Watermarks",
|
|
Purpose: "The youngest sample in the database per metric.",
|
|
Path: baseDirectory + "/high_watermarks_by_fingerprint",
|
|
CacheSizeBytes: *highWatermarkCacheSize,
|
|
},
|
|
)
|
|
workers.MayFail(err)
|
|
},
|
|
},
|
|
{
|
|
"Fingerprints by Label Name",
|
|
func() {
|
|
var err error
|
|
emission.LabelNameToLabelValues, err = NewLevelDBLabelNameLabelValuesIndex(
|
|
leveldb.LevelDBOptions{
|
|
Name: "Label Values by Label Name",
|
|
Purpose: "Index",
|
|
Path: baseDirectory + "/label_values_by_label_name",
|
|
CacheSizeBytes: *labelNameToLabelValuesCacheSize,
|
|
},
|
|
)
|
|
workers.MayFail(err)
|
|
},
|
|
},
|
|
{
|
|
"Fingerprints by Label Name and Value Pair",
|
|
func() {
|
|
var err error
|
|
emission.LabelPairToFingerprints, err = NewLevelDBLabelSetFingerprintIndex(
|
|
leveldb.LevelDBOptions{
|
|
Name: "Fingerprints by Label Pair",
|
|
Purpose: "Index",
|
|
Path: baseDirectory + "/fingerprints_by_label_name_and_value_pair",
|
|
CacheSizeBytes: *labelPairToFingerprintsCacheSize,
|
|
},
|
|
)
|
|
workers.MayFail(err)
|
|
},
|
|
},
|
|
{
|
|
"Metric Membership Index",
|
|
func() {
|
|
var err error
|
|
emission.MetricMembershipIndex, err = NewLevelDBMetricMembershipIndex(
|
|
leveldb.LevelDBOptions{
|
|
Name: "Metric Membership",
|
|
Purpose: "Index",
|
|
Path: baseDirectory + "/metric_membership_index",
|
|
CacheSizeBytes: *metricMembershipIndexCacheSize,
|
|
},
|
|
)
|
|
workers.MayFail(err)
|
|
},
|
|
},
|
|
{
|
|
"Sample Curation Remarks",
|
|
func() {
|
|
var err error
|
|
emission.CurationRemarks, err = NewLevelDBCurationRemarker(
|
|
leveldb.LevelDBOptions{
|
|
Name: "Sample Curation Remarks",
|
|
Purpose: "Ledger of Progress for Various Curators",
|
|
Path: baseDirectory + "/curation_remarks",
|
|
CacheSizeBytes: *curationRemarksCacheSize,
|
|
},
|
|
)
|
|
workers.MayFail(err)
|
|
},
|
|
},
|
|
}
|
|
|
|
for _, subsystem := range subsystemOpeners {
|
|
opener := subsystem.opener
|
|
go opener()
|
|
}
|
|
|
|
if !workers.Wait() {
|
|
for _, err := range workers.Errors() {
|
|
glog.Error("Could not open storage: ", err)
|
|
}
|
|
|
|
return nil, fmt.Errorf("unable to open metric persistence")
|
|
}
|
|
|
|
emission.Indexer = &TotalIndexer{
|
|
FingerprintToMetric: emission.FingerprintToMetrics,
|
|
LabelNameToLabelValues: emission.LabelNameToLabelValues,
|
|
LabelPairToFingerprint: emission.LabelPairToFingerprints,
|
|
MetricMembership: emission.MetricMembershipIndex,
|
|
}
|
|
|
|
return emission, nil
|
|
}
|
|
|
|
// AppendSample implements the Persistence interface.
|
|
func (l *LevelDBPersistence) AppendSample(sample *clientmodel.Sample) (err error) {
|
|
defer func(begin time.Time) {
|
|
recordOutcome(time.Since(begin), err, appendSample)
|
|
}(time.Now())
|
|
|
|
err = l.AppendSamples(clientmodel.Samples{sample})
|
|
|
|
return
|
|
}
|
|
|
|
// groupByFingerprint collects all of the provided samples, groups them
|
|
// together by their respective metric fingerprint, and finally sorts
|
|
// them chronologically.
|
|
func groupByFingerprint(samples clientmodel.Samples) map[clientmodel.Fingerprint]clientmodel.Samples {
|
|
fingerprintToSamples := map[clientmodel.Fingerprint]clientmodel.Samples{}
|
|
|
|
for _, sample := range samples {
|
|
fingerprint := &clientmodel.Fingerprint{}
|
|
fingerprint.LoadFromMetric(sample.Metric)
|
|
samples := fingerprintToSamples[*fingerprint]
|
|
samples = append(samples, sample)
|
|
fingerprintToSamples[*fingerprint] = samples
|
|
}
|
|
|
|
sortingSemaphore := make(chan bool, sortConcurrency)
|
|
doneSorting := sync.WaitGroup{}
|
|
|
|
for _, samples := range fingerprintToSamples {
|
|
doneSorting.Add(1)
|
|
|
|
sortingSemaphore <- true
|
|
go func(samples clientmodel.Samples) {
|
|
sort.Sort(samples)
|
|
|
|
<-sortingSemaphore
|
|
doneSorting.Done()
|
|
}(samples)
|
|
}
|
|
|
|
doneSorting.Wait()
|
|
|
|
return fingerprintToSamples
|
|
}
|
|
|
|
func (l *LevelDBPersistence) refreshHighWatermarks(groups map[clientmodel.Fingerprint]clientmodel.Samples) (err error) {
|
|
defer func(begin time.Time) {
|
|
recordOutcome(time.Since(begin), err, refreshHighWatermarks)
|
|
}(time.Now())
|
|
|
|
b := FingerprintHighWatermarkMapping{}
|
|
for fp, ss := range groups {
|
|
if len(ss) == 0 {
|
|
continue
|
|
}
|
|
|
|
b[fp] = ss[len(ss)-1].Timestamp
|
|
}
|
|
|
|
return l.MetricHighWatermarks.UpdateBatch(b)
|
|
}
|
|
|
|
// AppendSamples appends the given Samples to the database and indexes them.
|
|
func (l *LevelDBPersistence) AppendSamples(samples clientmodel.Samples) (err error) {
|
|
defer func(begin time.Time) {
|
|
recordOutcome(time.Since(begin), err, appendSamples)
|
|
}(time.Now())
|
|
|
|
fingerprintToSamples := groupByFingerprint(samples)
|
|
indexErrChan := make(chan error, 1)
|
|
watermarkErrChan := make(chan error, 1)
|
|
|
|
go func(groups map[clientmodel.Fingerprint]clientmodel.Samples) {
|
|
metrics := FingerprintMetricMapping{}
|
|
|
|
for fingerprint, samples := range groups {
|
|
metrics[fingerprint] = samples[0].Metric
|
|
}
|
|
|
|
indexErrChan <- l.Indexer.IndexMetrics(metrics)
|
|
}(fingerprintToSamples)
|
|
|
|
go func(groups map[clientmodel.Fingerprint]clientmodel.Samples) {
|
|
watermarkErrChan <- l.refreshHighWatermarks(groups)
|
|
}(fingerprintToSamples)
|
|
|
|
samplesBatch := leveldb.NewBatch()
|
|
defer samplesBatch.Close()
|
|
|
|
key := &SampleKey{}
|
|
keyDto := &dto.SampleKey{}
|
|
values := make(metric.Values, 0, *leveldbChunkSize)
|
|
|
|
for fingerprint, group := range fingerprintToSamples {
|
|
for {
|
|
values := values[:0]
|
|
lengthOfGroup := len(group)
|
|
|
|
if lengthOfGroup == 0 {
|
|
break
|
|
}
|
|
|
|
take := *leveldbChunkSize
|
|
if lengthOfGroup < take {
|
|
take = lengthOfGroup
|
|
}
|
|
|
|
chunk := group[0:take]
|
|
group = group[take:lengthOfGroup]
|
|
|
|
key.Fingerprint = &fingerprint
|
|
key.FirstTimestamp = chunk[0].Timestamp
|
|
key.LastTimestamp = chunk[take-1].Timestamp
|
|
key.SampleCount = uint32(take)
|
|
|
|
key.Dump(keyDto)
|
|
|
|
for _, sample := range chunk {
|
|
values = append(values, metric.SamplePair{
|
|
Timestamp: sample.Timestamp,
|
|
Value: sample.Value,
|
|
})
|
|
}
|
|
val := marshalValues(values, nil)
|
|
samplesBatch.PutRaw(keyDto, val)
|
|
}
|
|
}
|
|
|
|
err = l.MetricSamples.Commit(samplesBatch)
|
|
if err != nil {
|
|
return
|
|
}
|
|
|
|
err = <-indexErrChan
|
|
if err != nil {
|
|
return
|
|
}
|
|
|
|
err = <-watermarkErrChan
|
|
if err != nil {
|
|
return
|
|
}
|
|
|
|
return
|
|
}
|
|
|
|
func extractSampleKey(i leveldb.Iterator) (*SampleKey, error) {
|
|
k := &dto.SampleKey{}
|
|
if err := i.Key(k); err != nil {
|
|
return nil, err
|
|
}
|
|
|
|
key := &SampleKey{}
|
|
key.Load(k)
|
|
|
|
return key, nil
|
|
}
|
|
|
|
func (l *LevelDBPersistence) hasIndexMetric(m clientmodel.Metric) (value bool, err error) {
|
|
defer func(begin time.Time) {
|
|
recordOutcome(time.Since(begin), err, hasIndexMetric)
|
|
}(time.Now())
|
|
|
|
return l.MetricMembershipIndex.Has(m)
|
|
}
|
|
|
|
// GetFingerprintsForLabelMatchers returns the Fingerprints for the given
|
|
// LabelMatchers by querying the underlying LabelPairFingerprintIndex and
|
|
// possibly the LabelNameLabelValuesIndex for each matcher. It implements the
|
|
// Persistence interface.
|
|
func (l *LevelDBPersistence) GetFingerprintsForLabelMatchers(labelMatchers metric.LabelMatchers) (fps clientmodel.Fingerprints, err error) {
|
|
defer func(begin time.Time) {
|
|
recordOutcome(time.Since(begin), err, getFingerprintsForLabelMatchers)
|
|
}(time.Now())
|
|
|
|
sets := []utility.Set{}
|
|
|
|
for _, matcher := range labelMatchers {
|
|
set := utility.Set{}
|
|
|
|
switch matcher.Type {
|
|
case metric.Equal:
|
|
fps, _, err := l.LabelPairToFingerprints.Lookup(&metric.LabelPair{
|
|
Name: matcher.Name,
|
|
Value: matcher.Value,
|
|
})
|
|
if err != nil {
|
|
return nil, err
|
|
}
|
|
|
|
for _, fp := range fps {
|
|
set.Add(*fp)
|
|
}
|
|
default:
|
|
values, err := l.GetLabelValuesForLabelName(matcher.Name)
|
|
if err != nil {
|
|
return nil, err
|
|
}
|
|
matches := matcher.Filter(values)
|
|
if len(matches) == 0 {
|
|
return nil, nil
|
|
}
|
|
for _, v := range matches {
|
|
fps, _, err := l.LabelPairToFingerprints.Lookup(&metric.LabelPair{
|
|
Name: matcher.Name,
|
|
Value: v,
|
|
})
|
|
if err != nil {
|
|
return nil, err
|
|
}
|
|
for _, fp := range fps {
|
|
set.Add(*fp)
|
|
}
|
|
}
|
|
}
|
|
sets = append(sets, set)
|
|
}
|
|
|
|
numberOfSets := len(sets)
|
|
if numberOfSets == 0 {
|
|
return nil, nil
|
|
}
|
|
|
|
base := sets[0]
|
|
for i := 1; i < numberOfSets; i++ {
|
|
base = base.Intersection(sets[i])
|
|
}
|
|
for _, e := range base.Elements() {
|
|
fingerprint := e.(clientmodel.Fingerprint)
|
|
fps = append(fps, &fingerprint)
|
|
}
|
|
|
|
return fps, nil
|
|
}
|
|
|
|
// GetLabelValuesForLabelName returns the LabelValues for the given LabelName
|
|
// from the underlying LabelNameLabelValuesIndex. It implements the
|
|
// Persistence interface.
|
|
func (l *LevelDBPersistence) GetLabelValuesForLabelName(labelName clientmodel.LabelName) (clientmodel.LabelValues, error) {
|
|
var err error
|
|
defer func(begin time.Time) {
|
|
recordOutcome(time.Since(begin), err, getLabelValuesForLabelName)
|
|
}(time.Now())
|
|
|
|
values, _, err := l.LabelNameToLabelValues.Lookup(labelName)
|
|
|
|
return values, err
|
|
}
|
|
|
|
// GetMetricForFingerprint returns the Metric for the given Fingerprint from the
|
|
// underlying FingerprintMetricIndex. It implements the Persistence
|
|
// interface.
|
|
func (l *LevelDBPersistence) GetMetricForFingerprint(f *clientmodel.Fingerprint) (m clientmodel.Metric, err error) {
|
|
defer func(begin time.Time) {
|
|
recordOutcome(time.Since(begin), err, getMetricForFingerprint)
|
|
}(time.Now())
|
|
|
|
// TODO(matt): Update signature to work with ok.
|
|
m, _, err = l.FingerprintToMetrics.Lookup(f)
|
|
|
|
return m, nil
|
|
}
|
|
|
|
// GetAllValuesForLabel gets all label values that are associated with the
|
|
// provided label name.
|
|
func (l *LevelDBPersistence) GetAllValuesForLabel(labelName clientmodel.LabelName) (values clientmodel.LabelValues, err error) {
|
|
filter := &LabelNameFilter{
|
|
labelName: labelName,
|
|
}
|
|
labelValuesOp := &CollectLabelValuesOp{}
|
|
|
|
_, err = l.LabelPairToFingerprints.ForEach(&MetricKeyDecoder{}, filter, labelValuesOp)
|
|
if err != nil {
|
|
return
|
|
}
|
|
|
|
values = labelValuesOp.labelValues
|
|
return
|
|
}
|
|
|
|
// Prune compacts each database's keyspace serially.
|
|
//
|
|
// Beware that it would probably be imprudent to run this on a live user-facing
|
|
// server due to latency implications.
|
|
func (l *LevelDBPersistence) Prune() {
|
|
l.CurationRemarks.Prune()
|
|
l.FingerprintToMetrics.Prune()
|
|
l.LabelNameToLabelValues.Prune()
|
|
l.LabelPairToFingerprints.Prune()
|
|
l.MetricHighWatermarks.Prune()
|
|
l.MetricMembershipIndex.Prune()
|
|
l.MetricSamples.Prune()
|
|
}
|
|
|
|
// Sizes returns the sum of all sizes of the underlying databases.
|
|
func (l *LevelDBPersistence) Sizes() (total uint64, err error) {
|
|
size := uint64(0)
|
|
|
|
if size, err = l.CurationRemarks.Size(); err != nil {
|
|
return 0, err
|
|
}
|
|
total += size
|
|
|
|
if size, err = l.FingerprintToMetrics.Size(); err != nil {
|
|
return 0, err
|
|
}
|
|
total += size
|
|
|
|
if size, err = l.LabelNameToLabelValues.Size(); err != nil {
|
|
return 0, err
|
|
}
|
|
total += size
|
|
|
|
if size, err = l.LabelPairToFingerprints.Size(); err != nil {
|
|
return 0, err
|
|
}
|
|
total += size
|
|
|
|
if size, err = l.MetricHighWatermarks.Size(); err != nil {
|
|
return 0, err
|
|
}
|
|
total += size
|
|
|
|
if size, err = l.MetricMembershipIndex.Size(); err != nil {
|
|
return 0, err
|
|
}
|
|
total += size
|
|
|
|
if size, err = l.MetricSamples.Size(); err != nil {
|
|
return 0, err
|
|
}
|
|
total += size
|
|
|
|
return total, nil
|
|
}
|
|
|
|
// States returns the DatabaseStates of all underlying databases.
|
|
func (l *LevelDBPersistence) States() raw.DatabaseStates {
|
|
return raw.DatabaseStates{
|
|
l.CurationRemarks.State(),
|
|
l.FingerprintToMetrics.State(),
|
|
l.LabelNameToLabelValues.State(),
|
|
l.LabelPairToFingerprints.State(),
|
|
l.MetricHighWatermarks.State(),
|
|
l.MetricMembershipIndex.State(),
|
|
l.MetricSamples.State(),
|
|
}
|
|
}
|
|
|
|
// CollectLabelValuesOp implements storage.RecordOperator. It collects the
|
|
// encountered LabelValues in a slice.
|
|
type CollectLabelValuesOp struct {
|
|
labelValues []clientmodel.LabelValue
|
|
}
|
|
|
|
// Operate implements storage.RecordOperator. 'key' is required to be a
|
|
// LabelPair. Its Value is appended to a slice of collected LabelValues.
|
|
func (op *CollectLabelValuesOp) Operate(key, value interface{}) (err *storage.OperatorError) {
|
|
labelPair := key.(metric.LabelPair)
|
|
op.labelValues = append(op.labelValues, labelPair.Value)
|
|
return
|
|
}
|
|
|
|
// MetricKeyDecoder implements storage.RecordDecoder for LabelPairs.
|
|
type MetricKeyDecoder struct{}
|
|
|
|
// DecodeKey implements storage.RecordDecoder. It requires 'in' to be a
|
|
// LabelPair protobuf. 'out' is a metric.LabelPair.
|
|
func (d *MetricKeyDecoder) DecodeKey(in interface{}) (out interface{}, err error) {
|
|
unmarshaled := dto.LabelPair{}
|
|
err = proto.Unmarshal(in.([]byte), &unmarshaled)
|
|
if err != nil {
|
|
return
|
|
}
|
|
|
|
out = metric.LabelPair{
|
|
Name: clientmodel.LabelName(*unmarshaled.Name),
|
|
Value: clientmodel.LabelValue(*unmarshaled.Value),
|
|
}
|
|
|
|
return
|
|
}
|
|
|
|
// DecodeValue implements storage.RecordDecoder. It is a no-op and always
|
|
// returns (nil, nil).
|
|
func (d *MetricKeyDecoder) DecodeValue(in interface{}) (out interface{}, err error) {
|
|
return
|
|
}
|
|
|
|
// MetricSamplesDecoder implements storage.RecordDecoder for SampleKeys.
|
|
type MetricSamplesDecoder struct{}
|
|
|
|
// DecodeKey implements storage.RecordDecoder. It requires 'in' to be a
|
|
// SampleKey protobuf. 'out' is a metric.SampleKey.
|
|
func (d *MetricSamplesDecoder) DecodeKey(in interface{}) (interface{}, error) {
|
|
key := &dto.SampleKey{}
|
|
err := proto.Unmarshal(in.([]byte), key)
|
|
if err != nil {
|
|
return nil, err
|
|
}
|
|
|
|
sampleKey := &SampleKey{}
|
|
sampleKey.Load(key)
|
|
|
|
return sampleKey, nil
|
|
}
|
|
|
|
// DecodeValue implements storage.RecordDecoder. It requires 'in' to be a
|
|
// SampleValueSeries protobuf. 'out' is of type metric.Values.
|
|
func (d *MetricSamplesDecoder) DecodeValue(in interface{}) (interface{}, error) {
|
|
return unmarshalValues(in.([]byte), nil), nil
|
|
}
|
|
|
|
// AcceptAllFilter implements storage.RecordFilter and accepts all records.
|
|
type AcceptAllFilter struct{}
|
|
|
|
// Filter implements storage.RecordFilter. It always returns ACCEPT.
|
|
func (d *AcceptAllFilter) Filter(_, _ interface{}) storage.FilterResult {
|
|
return storage.Accept
|
|
}
|
|
|
|
// LabelNameFilter implements storage.RecordFilter and filters records matching
|
|
// a LabelName.
|
|
type LabelNameFilter struct {
|
|
labelName clientmodel.LabelName
|
|
}
|
|
|
|
// Filter implements storage.RecordFilter. 'key' is expected to be a
|
|
// LabelPair. The result is ACCEPT if the Name of the LabelPair matches the
|
|
// LabelName of this LabelNameFilter.
|
|
func (f LabelNameFilter) Filter(key, value interface{}) (filterResult storage.FilterResult) {
|
|
labelPair, ok := key.(metric.LabelPair)
|
|
if ok && labelPair.Name == f.labelName {
|
|
return storage.Accept
|
|
}
|
|
return storage.Skip
|
|
}
|
|
|
|
func recordOutcome(duration time.Duration, err error, op string) {
|
|
labels := prometheus.Labels{operation: op}
|
|
if err == nil {
|
|
labels[result] = success
|
|
} else {
|
|
labels[result] = failure
|
|
}
|
|
storageLatency.With(labels).Observe(float64(duration / time.Microsecond))
|
|
}
|