2015-08-24 10:19:21 -07:00
|
|
|
// Copyright 2015 The Prometheus Authors
|
|
|
|
// Licensed under the Apache License, Version 2.0 (the "License");
|
|
|
|
// you may not use this file except in compliance with the License.
|
|
|
|
// You may obtain a copy of the License at
|
|
|
|
//
|
|
|
|
// http://www.apache.org/licenses/LICENSE-2.0
|
|
|
|
//
|
|
|
|
// Unless required by applicable law or agreed to in writing, software
|
|
|
|
// distributed under the License is distributed on an "AS IS" BASIS,
|
|
|
|
// WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
|
|
|
|
// See the License for the specific language governing permissions and
|
|
|
|
// limitations under the License.
|
|
|
|
|
2015-06-22 13:46:55 -07:00
|
|
|
package web
|
|
|
|
|
|
|
|
import (
|
2018-08-17 08:24:35 -07:00
|
|
|
"fmt"
|
2015-06-22 13:46:55 -07:00
|
|
|
"net/http"
|
2016-12-30 10:34:45 -08:00
|
|
|
"sort"
|
2015-06-22 13:46:55 -07:00
|
|
|
|
2021-06-11 09:17:59 -07:00
|
|
|
"github.com/go-kit/log/level"
|
2016-12-30 10:34:45 -08:00
|
|
|
"github.com/gogo/protobuf/proto"
|
2020-04-29 09:16:14 -07:00
|
|
|
"github.com/pkg/errors"
|
2016-12-06 07:09:50 -08:00
|
|
|
"github.com/prometheus/client_golang/prometheus"
|
2016-12-30 10:34:45 -08:00
|
|
|
dto "github.com/prometheus/client_model/go"
|
2015-08-21 04:16:50 -07:00
|
|
|
"github.com/prometheus/common/expfmt"
|
2015-08-20 08:18:46 -07:00
|
|
|
"github.com/prometheus/common/model"
|
2015-09-01 09:47:48 -07:00
|
|
|
|
2023-01-09 03:36:15 -08:00
|
|
|
"github.com/prometheus/prometheus/model/histogram"
|
2021-11-08 06:23:17 -08:00
|
|
|
"github.com/prometheus/prometheus/model/labels"
|
|
|
|
"github.com/prometheus/prometheus/model/timestamp"
|
|
|
|
"github.com/prometheus/prometheus/model/value"
|
2016-07-11 11:27:25 -07:00
|
|
|
"github.com/prometheus/prometheus/promql"
|
2020-02-03 10:06:39 -08:00
|
|
|
"github.com/prometheus/prometheus/promql/parser"
|
2016-12-30 10:34:45 -08:00
|
|
|
"github.com/prometheus/prometheus/storage"
|
2020-10-22 02:00:08 -07:00
|
|
|
"github.com/prometheus/prometheus/tsdb"
|
2021-11-28 23:54:23 -08:00
|
|
|
"github.com/prometheus/prometheus/tsdb/chunkenc"
|
2015-06-22 13:46:55 -07:00
|
|
|
)
|
|
|
|
|
2016-12-06 07:09:50 -08:00
|
|
|
var (
|
|
|
|
federationErrors = prometheus.NewCounter(prometheus.CounterOpts{
|
|
|
|
Name: "prometheus_web_federation_errors_total",
|
|
|
|
Help: "Total number of errors that occurred while sending federation responses.",
|
|
|
|
})
|
2018-11-30 06:27:12 -08:00
|
|
|
federationWarnings = prometheus.NewCounter(prometheus.CounterOpts{
|
|
|
|
Name: "prometheus_web_federation_warnings_total",
|
|
|
|
Help: "Total number of warnings that occurred while sending federation responses.",
|
|
|
|
})
|
2016-12-06 07:09:50 -08:00
|
|
|
)
|
|
|
|
|
2020-04-06 01:05:01 -07:00
|
|
|
func registerFederationMetrics(r prometheus.Registerer) {
|
|
|
|
r.MustRegister(federationWarnings, federationErrors)
|
|
|
|
}
|
|
|
|
|
2015-09-01 09:47:48 -07:00
|
|
|
func (h *Handler) federation(w http.ResponseWriter, req *http.Request) {
|
|
|
|
h.mtx.RLock()
|
|
|
|
defer h.mtx.RUnlock()
|
2015-06-22 13:46:55 -07:00
|
|
|
|
2018-08-17 08:24:35 -07:00
|
|
|
if err := req.ParseForm(); err != nil {
|
|
|
|
http.Error(w, fmt.Sprintf("error parsing form values: %v", err), http.StatusBadRequest)
|
|
|
|
return
|
|
|
|
}
|
2015-06-22 13:46:55 -07:00
|
|
|
|
2016-12-29 00:27:30 -08:00
|
|
|
var matcherSets [][]*labels.Matcher
|
2015-06-22 13:46:55 -07:00
|
|
|
for _, s := range req.Form["match[]"] {
|
2020-02-03 10:06:39 -08:00
|
|
|
matchers, err := parser.ParseMetricSelector(s)
|
2015-06-22 13:46:55 -07:00
|
|
|
if err != nil {
|
|
|
|
http.Error(w, err.Error(), http.StatusBadRequest)
|
|
|
|
return
|
|
|
|
}
|
2016-07-11 11:27:25 -07:00
|
|
|
matcherSets = append(matcherSets, matchers)
|
2015-06-22 13:46:55 -07:00
|
|
|
}
|
|
|
|
|
2015-12-16 04:45:44 -08:00
|
|
|
var (
|
2020-02-09 15:58:23 -08:00
|
|
|
mint = timestamp.FromTime(h.now().Time().Add(-h.lookbackDelta))
|
2016-12-30 10:34:45 -08:00
|
|
|
maxt = timestamp.FromTime(h.now().Time())
|
2016-12-24 15:37:46 -08:00
|
|
|
format = expfmt.Negotiate(req.Header)
|
2016-12-30 10:34:45 -08:00
|
|
|
enc = expfmt.NewEncoder(w, format)
|
2015-12-16 04:45:44 -08:00
|
|
|
)
|
2015-08-21 04:16:50 -07:00
|
|
|
w.Header().Set("Content-Type", string(format))
|
|
|
|
|
2020-04-07 08:42:42 -07:00
|
|
|
q, err := h.localStorage.Querier(req.Context(), mint, maxt)
|
2016-12-30 10:34:45 -08:00
|
|
|
if err != nil {
|
|
|
|
federationErrors.Inc()
|
2020-04-29 09:16:14 -07:00
|
|
|
if errors.Cause(err) == tsdb.ErrNotReady {
|
|
|
|
http.Error(w, err.Error(), http.StatusServiceUnavailable)
|
|
|
|
return
|
|
|
|
}
|
2016-12-30 10:34:45 -08:00
|
|
|
http.Error(w, err.Error(), http.StatusInternalServerError)
|
|
|
|
return
|
|
|
|
}
|
|
|
|
defer q.Close()
|
|
|
|
|
|
|
|
vec := make(promql.Vector, 0, 8000)
|
|
|
|
|
2020-03-12 02:36:09 -07:00
|
|
|
hints := &storage.SelectHints{Start: mint, End: maxt}
|
2018-08-28 03:23:31 -07:00
|
|
|
|
2017-12-10 03:00:23 -08:00
|
|
|
var sets []storage.SeriesSet
|
2016-12-30 10:34:45 -08:00
|
|
|
for _, mset := range matcherSets {
|
2021-06-24 11:32:23 -07:00
|
|
|
s := q.Select(true, hints, mset...)
|
2017-12-10 03:00:23 -08:00
|
|
|
sets = append(sets, s)
|
2017-04-04 02:13:46 -07:00
|
|
|
}
|
2016-12-30 10:34:45 -08:00
|
|
|
|
2020-03-24 13:15:47 -07:00
|
|
|
set := storage.NewMergeSeriesSet(sets, storage.ChainedSeriesMerge)
|
2020-02-09 15:58:23 -08:00
|
|
|
it := storage.NewBuffer(int64(h.lookbackDelta / 1e6))
|
2022-09-20 10:16:45 -07:00
|
|
|
var chkIter chunkenc.Iterator
|
2023-01-09 03:36:15 -08:00
|
|
|
Loop:
|
2017-04-04 02:13:46 -07:00
|
|
|
for set.Next() {
|
|
|
|
s := set.At()
|
|
|
|
|
|
|
|
// TODO(fabxc): allow fast path for most recent sample either
|
|
|
|
// in the storage itself or caching layer in Prometheus.
|
2022-09-20 10:16:45 -07:00
|
|
|
chkIter = s.Iterator(chkIter)
|
|
|
|
it.Reset(chkIter)
|
2017-04-04 02:13:46 -07:00
|
|
|
|
2023-01-09 03:36:15 -08:00
|
|
|
var (
|
|
|
|
t int64
|
|
|
|
v float64
|
|
|
|
h *histogram.FloatHistogram
|
|
|
|
ok bool
|
|
|
|
)
|
2021-11-28 23:54:23 -08:00
|
|
|
valueType := it.Seek(maxt)
|
2023-01-09 03:36:15 -08:00
|
|
|
switch valueType {
|
|
|
|
case chunkenc.ValFloat:
|
2021-11-29 02:53:04 -08:00
|
|
|
t, v = it.At()
|
2023-01-09 03:36:15 -08:00
|
|
|
case chunkenc.ValFloatHistogram, chunkenc.ValHistogram:
|
|
|
|
t, h = it.AtFloatHistogram()
|
|
|
|
default:
|
2021-11-12 10:07:41 -08:00
|
|
|
t, v, _, ok = it.PeekBack(1)
|
2017-04-04 02:13:46 -07:00
|
|
|
if !ok {
|
2023-01-09 03:36:15 -08:00
|
|
|
continue Loop
|
2017-04-04 02:13:46 -07:00
|
|
|
}
|
2016-12-30 10:34:45 -08:00
|
|
|
}
|
2017-05-23 10:03:57 -07:00
|
|
|
// The exposition formats do not support stale markers, so drop them. This
|
|
|
|
// is good enough for staleness handling of federated data, as the
|
|
|
|
// interval-based limits on staleness will do the right thing for supported
|
|
|
|
// use cases (which is to say federating aggregated time series).
|
|
|
|
if value.IsStaleNaN(v) {
|
|
|
|
continue
|
|
|
|
}
|
2017-04-04 02:13:46 -07:00
|
|
|
|
|
|
|
vec = append(vec, promql.Sample{
|
|
|
|
Metric: s.Labels(),
|
2023-01-09 03:36:15 -08:00
|
|
|
Point: promql.Point{T: t, V: v, H: h},
|
2017-04-04 02:13:46 -07:00
|
|
|
})
|
|
|
|
}
|
2020-06-09 09:57:31 -07:00
|
|
|
if ws := set.Warnings(); len(ws) > 0 {
|
|
|
|
level.Debug(h.logger).Log("msg", "Federation select returned warnings", "warnings", ws)
|
|
|
|
federationWarnings.Add(float64(len(ws)))
|
|
|
|
}
|
2017-04-04 02:13:46 -07:00
|
|
|
if set.Err() != nil {
|
|
|
|
federationErrors.Inc()
|
2019-05-03 06:11:28 -07:00
|
|
|
http.Error(w, set.Err().Error(), http.StatusInternalServerError)
|
2017-04-04 02:13:46 -07:00
|
|
|
return
|
2016-12-30 10:34:45 -08:00
|
|
|
}
|
|
|
|
|
|
|
|
sort.Sort(byName(vec))
|
|
|
|
|
2019-03-08 08:29:25 -08:00
|
|
|
externalLabels := h.config.GlobalConfig.ExternalLabels.Map()
|
2017-03-27 08:18:33 -07:00
|
|
|
if _, ok := externalLabels[model.InstanceLabel]; !ok {
|
|
|
|
externalLabels[model.InstanceLabel] = ""
|
|
|
|
}
|
2019-03-08 08:29:25 -08:00
|
|
|
externalLabelNames := make([]string, 0, len(externalLabels))
|
2017-03-27 08:18:33 -07:00
|
|
|
for ln := range externalLabels {
|
|
|
|
externalLabelNames = append(externalLabelNames, ln)
|
|
|
|
}
|
2019-03-08 08:29:25 -08:00
|
|
|
sort.Strings(externalLabelNames)
|
2017-03-27 08:18:33 -07:00
|
|
|
|
2016-12-30 10:34:45 -08:00
|
|
|
var (
|
2023-01-09 03:36:15 -08:00
|
|
|
lastMetricName string
|
|
|
|
lastWasHistogram, lastHistogramWasGauge bool
|
|
|
|
protMetricFam *dto.MetricFamily
|
2016-12-30 10:34:45 -08:00
|
|
|
)
|
|
|
|
for _, s := range vec {
|
2023-01-09 03:36:15 -08:00
|
|
|
isHistogram := s.H != nil
|
|
|
|
if isHistogram &&
|
|
|
|
format != expfmt.FmtProtoDelim && format != expfmt.FmtProtoText && format != expfmt.FmtProtoCompact {
|
|
|
|
// Can't serve the native histogram.
|
|
|
|
// TODO(codesome): Serve them when other protocols get the native histogram support.
|
|
|
|
continue
|
|
|
|
}
|
|
|
|
|
2016-12-30 10:34:45 -08:00
|
|
|
nameSeen := false
|
|
|
|
globalUsed := map[string]struct{}{}
|
2023-01-09 03:36:15 -08:00
|
|
|
protMetric := &dto.Metric{}
|
2016-12-30 10:34:45 -08:00
|
|
|
|
2022-02-27 06:19:11 -08:00
|
|
|
err := s.Metric.Validate(func(l labels.Label) error {
|
2016-12-30 10:34:45 -08:00
|
|
|
if l.Value == "" {
|
|
|
|
// No value means unset. Never consider those labels.
|
|
|
|
// This is also important to protect against nameless metrics.
|
2022-02-27 06:19:11 -08:00
|
|
|
return nil
|
2016-12-30 10:34:45 -08:00
|
|
|
}
|
|
|
|
if l.Name == labels.MetricName {
|
|
|
|
nameSeen = true
|
2023-01-09 03:36:15 -08:00
|
|
|
if l.Value == lastMetricName && // We already have the name in the current MetricFamily, and we ignore nameless metrics.
|
|
|
|
lastWasHistogram == isHistogram && // The sample type matches (float vs histogram).
|
|
|
|
// If it was a histogram, the histogram type (counter vs gauge) also matches.
|
|
|
|
(!isHistogram || lastHistogramWasGauge == (s.H.CounterResetHint == histogram.GaugeType)) {
|
2022-02-27 06:19:11 -08:00
|
|
|
return nil
|
2016-12-30 10:34:45 -08:00
|
|
|
}
|
2023-01-09 03:36:15 -08:00
|
|
|
|
|
|
|
// Since we now check for the sample type and type of histogram above, we will end up
|
|
|
|
// creating multiple metric families for the same metric name. This would technically be
|
|
|
|
// an invalid exposition. But since the consumer of this is Prometheus, and Prometheus can
|
|
|
|
// parse it fine, we allow it and bend the rules to make federation possible in those cases.
|
|
|
|
|
2016-12-30 10:34:45 -08:00
|
|
|
// Need to start a new MetricFamily. Ship off the old one (if any) before
|
|
|
|
// creating the new one.
|
|
|
|
if protMetricFam != nil {
|
|
|
|
if err := enc.Encode(protMetricFam); err != nil {
|
2022-02-27 06:19:11 -08:00
|
|
|
return err
|
2016-12-30 10:34:45 -08:00
|
|
|
}
|
|
|
|
}
|
|
|
|
protMetricFam = &dto.MetricFamily{
|
|
|
|
Type: dto.MetricType_UNTYPED.Enum(),
|
|
|
|
Name: proto.String(l.Value),
|
|
|
|
}
|
2023-01-09 03:36:15 -08:00
|
|
|
if isHistogram {
|
|
|
|
if s.H.CounterResetHint == histogram.GaugeType {
|
|
|
|
protMetricFam.Type = dto.MetricType_GAUGE_HISTOGRAM.Enum()
|
|
|
|
} else {
|
|
|
|
protMetricFam.Type = dto.MetricType_HISTOGRAM.Enum()
|
|
|
|
}
|
|
|
|
}
|
2016-12-30 10:34:45 -08:00
|
|
|
lastMetricName = l.Value
|
2022-02-27 06:19:11 -08:00
|
|
|
return nil
|
2016-12-30 10:34:45 -08:00
|
|
|
}
|
|
|
|
protMetric.Label = append(protMetric.Label, &dto.LabelPair{
|
|
|
|
Name: proto.String(l.Name),
|
|
|
|
Value: proto.String(l.Value),
|
|
|
|
})
|
2019-03-08 08:29:25 -08:00
|
|
|
if _, ok := externalLabels[l.Name]; ok {
|
2016-12-30 10:34:45 -08:00
|
|
|
globalUsed[l.Name] = struct{}{}
|
|
|
|
}
|
2022-02-27 06:19:11 -08:00
|
|
|
return nil
|
|
|
|
})
|
|
|
|
if err != nil {
|
|
|
|
federationErrors.Inc()
|
|
|
|
level.Error(h.logger).Log("msg", "federation failed", "err", err)
|
|
|
|
return
|
2016-12-30 10:34:45 -08:00
|
|
|
}
|
|
|
|
if !nameSeen {
|
2017-08-11 11:45:52 -07:00
|
|
|
level.Warn(h.logger).Log("msg", "Ignoring nameless metric during federation", "metric", s.Metric)
|
2016-12-30 10:34:45 -08:00
|
|
|
continue
|
|
|
|
}
|
|
|
|
// Attach global labels if they do not exist yet.
|
2017-04-05 05:53:34 -07:00
|
|
|
for _, ln := range externalLabelNames {
|
|
|
|
lv := externalLabels[ln]
|
2020-04-07 08:42:42 -07:00
|
|
|
if _, ok := globalUsed[ln]; !ok {
|
2016-12-30 10:34:45 -08:00
|
|
|
protMetric.Label = append(protMetric.Label, &dto.LabelPair{
|
2020-04-07 08:42:42 -07:00
|
|
|
Name: proto.String(ln),
|
|
|
|
Value: proto.String(lv),
|
2016-12-30 10:34:45 -08:00
|
|
|
})
|
|
|
|
}
|
|
|
|
}
|
|
|
|
|
|
|
|
protMetric.TimestampMs = proto.Int64(s.T)
|
2023-01-09 03:36:15 -08:00
|
|
|
if !isHistogram {
|
|
|
|
lastHistogramWasGauge = false
|
|
|
|
protMetric.Untyped = &dto.Untyped{
|
|
|
|
Value: proto.Float64(s.V),
|
|
|
|
}
|
|
|
|
} else {
|
|
|
|
lastHistogramWasGauge = s.H.CounterResetHint == histogram.GaugeType
|
|
|
|
protMetric.Histogram = &dto.Histogram{
|
|
|
|
SampleCountFloat: proto.Float64(s.H.Count),
|
|
|
|
SampleSum: proto.Float64(s.H.Sum),
|
|
|
|
Schema: proto.Int32(s.H.Schema),
|
|
|
|
ZeroThreshold: proto.Float64(s.H.ZeroThreshold),
|
|
|
|
ZeroCountFloat: proto.Float64(s.H.ZeroCount),
|
|
|
|
NegativeCount: s.H.NegativeBuckets,
|
|
|
|
PositiveCount: s.H.PositiveBuckets,
|
|
|
|
}
|
|
|
|
if len(s.H.PositiveSpans) > 0 {
|
|
|
|
protMetric.Histogram.PositiveSpan = make([]*dto.BucketSpan, len(s.H.PositiveSpans))
|
|
|
|
for i, sp := range s.H.PositiveSpans {
|
|
|
|
protMetric.Histogram.PositiveSpan[i] = &dto.BucketSpan{
|
|
|
|
Offset: proto.Int32(sp.Offset),
|
|
|
|
Length: proto.Uint32(sp.Length),
|
|
|
|
}
|
|
|
|
}
|
|
|
|
}
|
|
|
|
if len(s.H.NegativeSpans) > 0 {
|
|
|
|
protMetric.Histogram.NegativeSpan = make([]*dto.BucketSpan, len(s.H.NegativeSpans))
|
|
|
|
for i, sp := range s.H.NegativeSpans {
|
|
|
|
protMetric.Histogram.NegativeSpan[i] = &dto.BucketSpan{
|
|
|
|
Offset: proto.Int32(sp.Offset),
|
|
|
|
Length: proto.Uint32(sp.Length),
|
|
|
|
}
|
|
|
|
}
|
|
|
|
}
|
|
|
|
}
|
|
|
|
lastWasHistogram = isHistogram
|
2016-12-30 10:34:45 -08:00
|
|
|
protMetricFam.Metric = append(protMetricFam.Metric, protMetric)
|
|
|
|
}
|
|
|
|
// Still have to ship off the last MetricFamily, if any.
|
|
|
|
if protMetricFam != nil {
|
|
|
|
if err := enc.Encode(protMetricFam); err != nil {
|
|
|
|
federationErrors.Inc()
|
2017-08-11 11:45:52 -07:00
|
|
|
level.Error(h.logger).Log("msg", "federation failed", "err", err)
|
2016-12-30 10:34:45 -08:00
|
|
|
}
|
|
|
|
}
|
2015-06-22 13:46:55 -07:00
|
|
|
}
|
2016-09-08 08:39:52 -07:00
|
|
|
|
|
|
|
// byName makes a model.Vector sortable by metric name.
|
2016-12-30 10:34:45 -08:00
|
|
|
type byName promql.Vector
|
2016-09-08 08:39:52 -07:00
|
|
|
|
|
|
|
func (vec byName) Len() int { return len(vec) }
|
|
|
|
func (vec byName) Swap(i, j int) { vec[i], vec[j] = vec[j], vec[i] }
|
|
|
|
|
|
|
|
func (vec byName) Less(i, j int) bool {
|
2016-12-30 10:34:45 -08:00
|
|
|
ni := vec[i].Metric.Get(labels.MetricName)
|
|
|
|
nj := vec[j].Metric.Get(labels.MetricName)
|
2016-09-08 08:39:52 -07:00
|
|
|
return ni < nj
|
|
|
|
}
|