mirror of
https://github.com/prometheus/prometheus.git
synced 2025-03-05 20:59:13 -08:00
257 lines
6.6 KiB
Go
257 lines
6.6 KiB
Go
// Copyright 2020 The Prometheus Authors
|
|
// Licensed under the Apache License, Version 2.0 (the "License");
|
|
// you may not use this file except in compliance with the License.
|
|
// You may obtain a copy of the License at
|
|
//
|
|
// http://www.apache.org/licenses/LICENSE-2.0
|
|
//
|
|
// Unless required by applicable law or agreed to in writing, software
|
|
// distributed under the License is distributed on an "AS IS" BASIS,
|
|
// WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
|
|
// See the License for the specific language governing permissions and
|
|
// limitations under the License.
|
|
|
|
package importers
|
|
|
|
import (
|
|
"context"
|
|
"fmt"
|
|
"io/ioutil"
|
|
"math"
|
|
"net/url"
|
|
"os"
|
|
"sort"
|
|
"strconv"
|
|
"time"
|
|
|
|
"github.com/go-kit/kit/log"
|
|
"github.com/pkg/errors"
|
|
"github.com/prometheus/client_golang/api"
|
|
v1 "github.com/prometheus/client_golang/api/prometheus/v1"
|
|
"github.com/prometheus/common/model"
|
|
"github.com/prometheus/prometheus/pkg/labels"
|
|
plabels "github.com/prometheus/prometheus/pkg/labels"
|
|
"github.com/prometheus/prometheus/pkg/rulefmt"
|
|
"github.com/prometheus/prometheus/promql"
|
|
"github.com/prometheus/prometheus/promql/parser"
|
|
"github.com/prometheus/prometheus/rules"
|
|
"github.com/prometheus/prometheus/tsdb/importer/blocks"
|
|
)
|
|
|
|
// RuleImporter is the importer for rules
|
|
type RuleImporter struct {
|
|
logger log.Logger
|
|
|
|
config RuleConfig
|
|
groups map[string]*rules.Group
|
|
|
|
apiClient v1.API
|
|
|
|
writer *blocks.MultiWriter
|
|
}
|
|
|
|
// RuleConfig is the config for the rule importer
|
|
type RuleConfig struct {
|
|
Start string
|
|
End string
|
|
EvalInterval time.Duration
|
|
URL string
|
|
}
|
|
|
|
// NewRuleImporter creates a new rule importer
|
|
func NewRuleImporter(logger log.Logger, config RuleConfig) *RuleImporter {
|
|
return &RuleImporter{
|
|
config: config,
|
|
}
|
|
}
|
|
|
|
// Init initializes the rule importer which creates a new block writer
|
|
// and creates an Prometheus API client
|
|
func (importer *RuleImporter) Init() error {
|
|
// create new block writer
|
|
newBlockDir, err := ioutil.TempDir("", "rule_blocks")
|
|
if err != nil {
|
|
return err
|
|
}
|
|
importer.writer = blocks.NewMultiWriter(importer.logger, newBlockDir, importer.config.EvalInterval.Nanoseconds())
|
|
|
|
// create api client
|
|
config := api.Config{
|
|
Address: importer.config.URL,
|
|
}
|
|
c, err := api.NewClient(config)
|
|
if err != nil {
|
|
return err
|
|
}
|
|
importer.apiClient = v1.NewAPI(c)
|
|
return nil
|
|
}
|
|
|
|
// Close cleans up any open resources
|
|
func (importer *RuleImporter) Close() error {
|
|
return importer.writer.Close()
|
|
}
|
|
|
|
// Parse parses the groups and rules from a list of rules files
|
|
func (importer *RuleImporter) Parse(ctx context.Context, files []string) (errs []error) {
|
|
groups := make(map[string]*rules.Group)
|
|
|
|
for _, file := range files {
|
|
ruleGroups, errs := rulefmt.ParseFile(file)
|
|
if errs != nil {
|
|
return errs
|
|
}
|
|
|
|
for _, ruleGroup := range ruleGroups.Groups {
|
|
itv := importer.config.EvalInterval
|
|
if ruleGroup.Interval != 0 {
|
|
itv = time.Duration(ruleGroup.Interval)
|
|
}
|
|
|
|
rulez := make([]rules.Rule, 0, len(ruleGroup.Rules))
|
|
for _, r := range ruleGroup.Rules {
|
|
expr, err := parser.ParseExpr(r.Expr.Value)
|
|
if err != nil {
|
|
return []error{errors.Wrap(err, file)}
|
|
}
|
|
|
|
rulez = append(rulez, rules.NewRecordingRule(
|
|
r.Record.Value,
|
|
expr,
|
|
labels.FromMap(r.Labels),
|
|
))
|
|
}
|
|
|
|
groups[file+";"+ruleGroup.Name] = rules.NewGroup(rules.GroupOptions{
|
|
Name: ruleGroup.Name,
|
|
File: file,
|
|
Interval: itv,
|
|
Rules: rulez,
|
|
})
|
|
}
|
|
}
|
|
|
|
importer.groups = groups
|
|
return errs
|
|
}
|
|
|
|
// ImportAll evaluates all the groups and rules and creates new time series
|
|
// and stores in new blocks
|
|
func (importer *RuleImporter) ImportAll(ctx context.Context) []error {
|
|
var errs = []error{}
|
|
for _, group := range importer.groups {
|
|
for _, rule := range group.Rules() {
|
|
err := importer.ImportRule(ctx, rule)
|
|
if err != nil {
|
|
errs = append(errs, err)
|
|
}
|
|
}
|
|
}
|
|
err := importer.CreateBlocks()
|
|
if err != nil {
|
|
errs = append(errs, err)
|
|
}
|
|
return errs
|
|
}
|
|
|
|
func (importer *RuleImporter) queryFn(ctx context.Context, q string, t time.Time) (promql.Vector, error) {
|
|
val, warnings, err := importer.apiClient.Query(ctx, q, t)
|
|
if err != nil {
|
|
return promql.Vector{}, err
|
|
}
|
|
if warnings != nil {
|
|
fmt.Fprint(os.Stderr, "warning api.Query:", warnings)
|
|
}
|
|
|
|
switch val.Type() {
|
|
case model.ValVector:
|
|
valVector := val.(model.Vector)
|
|
return modelToPromqlVector(valVector), nil
|
|
case model.ValScalar:
|
|
valScalar := val.(*model.Scalar)
|
|
return promql.Vector{promql.Sample{
|
|
Metric: labels.Labels{},
|
|
Point: promql.Point{T: int64(valScalar.Timestamp), V: float64(valScalar.Value)},
|
|
}}, nil
|
|
default:
|
|
return nil, errors.New("rule result is wrong type")
|
|
}
|
|
}
|
|
|
|
func modelToPromqlVector(modelValue model.Vector) promql.Vector {
|
|
result := make(promql.Vector, 0, len(modelValue))
|
|
|
|
for _, value := range modelValue {
|
|
labels := make(labels.Labels, 0, len(value.Metric))
|
|
|
|
for k, v := range value.Metric {
|
|
labels = append(labels, plabels.Label{
|
|
Name: string(k),
|
|
Value: string(v),
|
|
})
|
|
}
|
|
sort.Sort(labels)
|
|
|
|
result = append(result, promql.Sample{
|
|
Metric: labels,
|
|
Point: promql.Point{T: int64(value.Timestamp), V: float64(value.Value)},
|
|
})
|
|
}
|
|
return result
|
|
}
|
|
|
|
// ImportRule imports the historical data for a single rule
|
|
func (importer *RuleImporter) ImportRule(ctx context.Context, rule rules.Rule) error {
|
|
ts, err := parseTime(importer.config.Start)
|
|
if err != nil {
|
|
return err
|
|
}
|
|
end, err := parseTime(importer.config.End)
|
|
if err != nil {
|
|
return err
|
|
}
|
|
url, err := url.Parse(importer.config.URL)
|
|
if err != nil {
|
|
return err
|
|
}
|
|
|
|
appender := importer.writer.Appender()
|
|
for ts.Before(end) {
|
|
vector, err := rule.Eval(ctx, ts, importer.queryFn, url)
|
|
if err != nil {
|
|
return err
|
|
}
|
|
for _, sample := range vector {
|
|
// we don't AddFast here because we need to maintain the
|
|
// ref for each series bcs rule.Eval could return different labels,
|
|
// so that means you would need to map the ref to metric, but that is what Add does
|
|
// anyways so just use that
|
|
_, err := appender.Add(plabels.Labels{plabels.Label{Name: sample.String()}}, sample.T, sample.V)
|
|
if err != nil {
|
|
return err
|
|
}
|
|
}
|
|
|
|
ts.Add(importer.config.EvalInterval)
|
|
// todo: 2 hr blocks?
|
|
}
|
|
return appender.Commit()
|
|
}
|
|
|
|
func parseTime(s string) (time.Time, error) {
|
|
if t, err := strconv.ParseFloat(s, 64); err == nil {
|
|
s, ns := math.Modf(t)
|
|
return time.Unix(int64(s), int64(ns*float64(time.Second))).UTC(), nil
|
|
}
|
|
if t, err := time.Parse(time.RFC3339Nano, s); err == nil {
|
|
return t, nil
|
|
}
|
|
return time.Time{}, errors.Errorf("cannot parse %q to a valid timestamp", s)
|
|
}
|
|
|
|
// CreateBlocks creates blocks for all the rule data
|
|
func (importer *RuleImporter) CreateBlocks() error {
|
|
_, err := importer.writer.Flush()
|
|
return err
|
|
}
|