mirror of
https://github.com/prometheus/prometheus.git
synced 2024-12-25 13:44:05 -08:00
995d3b831d
This is with `golint -min_confidence=0.5`. I left several lint warnings untouched because they were either incorrect or I felt it was better not to change them at the moment.
273 lines
8.1 KiB
Go
273 lines
8.1 KiB
Go
// Copyright 2013 The Prometheus Authors
|
|
// Licensed under the Apache License, Version 2.0 (the "License");
|
|
// you may not use this file except in compliance with the License.
|
|
// You may obtain a copy of the License at
|
|
//
|
|
// http://www.apache.org/licenses/LICENSE-2.0
|
|
//
|
|
// Unless required by applicable law or agreed to in writing, software
|
|
// distributed under the License is distributed on an "AS IS" BASIS,
|
|
// WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
|
|
// See the License for the specific language governing permissions and
|
|
// limitations under the License.
|
|
|
|
package rules
|
|
|
|
import (
|
|
"fmt"
|
|
"html/template"
|
|
"sync"
|
|
"time"
|
|
|
|
"github.com/prometheus/common/model"
|
|
|
|
"github.com/prometheus/prometheus/promql"
|
|
"github.com/prometheus/prometheus/util/strutil"
|
|
)
|
|
|
|
const (
|
|
// AlertMetricName is the metric name for synthetic alert timeseries.
|
|
alertMetricName model.LabelValue = "ALERTS"
|
|
|
|
// AlertNameLabel is the label name indicating the name of an alert.
|
|
alertNameLabel model.LabelName = "alertname"
|
|
// AlertStateLabel is the label name indicating the state of an alert.
|
|
alertStateLabel model.LabelName = "alertstate"
|
|
)
|
|
|
|
// AlertState denotes the state of an active alert.
|
|
type AlertState int
|
|
|
|
func (s AlertState) String() string {
|
|
switch s {
|
|
case StateInactive:
|
|
return "inactive"
|
|
case StatePending:
|
|
return "pending"
|
|
case StateFiring:
|
|
return "firing"
|
|
default:
|
|
panic("undefined")
|
|
}
|
|
}
|
|
|
|
const (
|
|
// StateInactive is the state of an alert that is either firing nor pending.
|
|
StateInactive AlertState = iota
|
|
// StatePending is the state of an alert that has been active for less than
|
|
// the configured threshold duration.
|
|
StatePending
|
|
// StateFiring is the state of an alert that has been active for longer than
|
|
// the configured threshold duration.
|
|
StateFiring
|
|
)
|
|
|
|
// Alert is used to track active (pending/firing) alerts over time.
|
|
type Alert struct {
|
|
// The name of the alert.
|
|
Name string
|
|
// The vector element labelset triggering this alert.
|
|
Labels model.LabelSet
|
|
// The state of the alert (Pending or Firing).
|
|
State AlertState
|
|
// The time when the alert first transitioned into Pending state.
|
|
ActiveSince model.Time
|
|
// The value of the alert expression for this vector element.
|
|
Value model.SampleValue
|
|
}
|
|
|
|
// sample returns a Sample suitable for recording the alert.
|
|
func (a Alert) sample(timestamp model.Time, value model.SampleValue) *model.Sample {
|
|
recordedMetric := make(model.Metric, len(a.Labels)+3)
|
|
for label, value := range a.Labels {
|
|
recordedMetric[label] = value
|
|
}
|
|
|
|
recordedMetric[model.MetricNameLabel] = alertMetricName
|
|
recordedMetric[alertNameLabel] = model.LabelValue(a.Name)
|
|
recordedMetric[alertStateLabel] = model.LabelValue(a.State.String())
|
|
|
|
return &model.Sample{
|
|
Metric: recordedMetric,
|
|
Value: value,
|
|
Timestamp: timestamp,
|
|
}
|
|
}
|
|
|
|
// An AlertingRule generates alerts from its vector expression.
|
|
type AlertingRule struct {
|
|
// The name of the alert.
|
|
name string
|
|
// The vector expression from which to generate alerts.
|
|
vector promql.Expr
|
|
// The duration for which a labelset needs to persist in the expression
|
|
// output vector before an alert transitions from Pending to Firing state.
|
|
holdDuration time.Duration
|
|
// Extra labels to attach to the resulting alert sample vectors.
|
|
labels model.LabelSet
|
|
// Short alert summary, suitable for email subjects.
|
|
summary string
|
|
// More detailed alert description.
|
|
description string
|
|
// A reference to a runbook for the alert.
|
|
runbook string
|
|
|
|
// Protects the below.
|
|
mutex sync.Mutex
|
|
// A map of alerts which are currently active (Pending or Firing), keyed by
|
|
// the fingerprint of the labelset they correspond to.
|
|
activeAlerts map[model.Fingerprint]*Alert
|
|
}
|
|
|
|
// NewAlertingRule constructs a new AlertingRule.
|
|
func NewAlertingRule(
|
|
name string,
|
|
vector promql.Expr,
|
|
holdDuration time.Duration,
|
|
labels model.LabelSet,
|
|
summary string,
|
|
description string,
|
|
runbook string,
|
|
) *AlertingRule {
|
|
return &AlertingRule{
|
|
name: name,
|
|
vector: vector,
|
|
holdDuration: holdDuration,
|
|
labels: labels,
|
|
summary: summary,
|
|
description: description,
|
|
runbook: runbook,
|
|
|
|
activeAlerts: map[model.Fingerprint]*Alert{},
|
|
}
|
|
}
|
|
|
|
// Name returns the name of the alert.
|
|
func (rule *AlertingRule) Name() string {
|
|
return rule.name
|
|
}
|
|
|
|
// eval evaluates the rule expression and then creates pending alerts and fires
|
|
// or removes previously pending alerts accordingly.
|
|
func (rule *AlertingRule) eval(timestamp model.Time, engine *promql.Engine) (model.Vector, error) {
|
|
query, err := engine.NewInstantQuery(rule.vector.String(), timestamp)
|
|
if err != nil {
|
|
return nil, err
|
|
}
|
|
exprResult, err := query.Exec().Vector()
|
|
if err != nil {
|
|
return nil, err
|
|
}
|
|
|
|
rule.mutex.Lock()
|
|
defer rule.mutex.Unlock()
|
|
|
|
// Create pending alerts for any new vector elements in the alert expression
|
|
// or update the expression value for existing elements.
|
|
resultFPs := map[model.Fingerprint]struct{}{}
|
|
for _, sample := range exprResult {
|
|
fp := sample.Metric.Fingerprint()
|
|
resultFPs[fp] = struct{}{}
|
|
|
|
if alert, ok := rule.activeAlerts[fp]; !ok {
|
|
labels := model.LabelSet(sample.Metric.Clone())
|
|
labels = labels.Merge(rule.labels)
|
|
if _, ok := labels[model.MetricNameLabel]; ok {
|
|
delete(labels, model.MetricNameLabel)
|
|
}
|
|
rule.activeAlerts[fp] = &Alert{
|
|
Name: rule.name,
|
|
Labels: labels,
|
|
State: StatePending,
|
|
ActiveSince: timestamp,
|
|
Value: sample.Value,
|
|
}
|
|
} else {
|
|
alert.Value = sample.Value
|
|
}
|
|
}
|
|
|
|
var vector model.Vector
|
|
|
|
// Check if any pending alerts should be removed or fire now. Write out alert timeseries.
|
|
for fp, activeAlert := range rule.activeAlerts {
|
|
if _, ok := resultFPs[fp]; !ok {
|
|
vector = append(vector, activeAlert.sample(timestamp, 0))
|
|
delete(rule.activeAlerts, fp)
|
|
continue
|
|
}
|
|
|
|
if activeAlert.State == StatePending && timestamp.Sub(activeAlert.ActiveSince) >= rule.holdDuration {
|
|
vector = append(vector, activeAlert.sample(timestamp, 0))
|
|
activeAlert.State = StateFiring
|
|
}
|
|
|
|
vector = append(vector, activeAlert.sample(timestamp, 1))
|
|
}
|
|
|
|
return vector, nil
|
|
}
|
|
|
|
func (rule *AlertingRule) String() string {
|
|
s := fmt.Sprintf("ALERT %s", rule.name)
|
|
s += fmt.Sprintf("\n\tIF %s", rule.vector)
|
|
if rule.holdDuration > 0 {
|
|
s += fmt.Sprintf("\n\tFOR %s", strutil.DurationToString(rule.holdDuration))
|
|
}
|
|
if len(rule.labels) > 0 {
|
|
s += fmt.Sprintf("\n\tWITH %s", rule.labels)
|
|
}
|
|
s += fmt.Sprintf("\n\tSUMMARY %q", rule.summary)
|
|
s += fmt.Sprintf("\n\tDESCRIPTION %q", rule.description)
|
|
s += fmt.Sprintf("\n\tRUNBOOK %q", rule.runbook)
|
|
return s
|
|
}
|
|
|
|
// HTMLSnippet returns an HTML snippet representing this alerting rule. The
|
|
// resulting snippet is expected to be presented in a <pre> element, so that
|
|
// line breaks and other returned whitespace is respected.
|
|
func (rule *AlertingRule) HTMLSnippet(pathPrefix string) template.HTML {
|
|
alertMetric := model.Metric{
|
|
model.MetricNameLabel: alertMetricName,
|
|
alertNameLabel: model.LabelValue(rule.name),
|
|
}
|
|
s := fmt.Sprintf("ALERT <a href=%q>%s</a>", pathPrefix+strutil.GraphLinkForExpression(alertMetric.String()), rule.name)
|
|
s += fmt.Sprintf("\n IF <a href=%q>%s</a>", pathPrefix+strutil.GraphLinkForExpression(rule.vector.String()), rule.vector)
|
|
if rule.holdDuration > 0 {
|
|
s += fmt.Sprintf("\n FOR %s", strutil.DurationToString(rule.holdDuration))
|
|
}
|
|
if len(rule.labels) > 0 {
|
|
s += fmt.Sprintf("\n WITH %s", rule.labels)
|
|
}
|
|
s += fmt.Sprintf("\n SUMMARY %q", rule.summary)
|
|
s += fmt.Sprintf("\n DESCRIPTION %q", rule.description)
|
|
s += fmt.Sprintf("\n RUNBOOK %q", rule.runbook)
|
|
return template.HTML(s)
|
|
}
|
|
|
|
// State returns the "maximum" state: firing > pending > inactive.
|
|
func (rule *AlertingRule) State() AlertState {
|
|
rule.mutex.Lock()
|
|
defer rule.mutex.Unlock()
|
|
|
|
maxState := StateInactive
|
|
for _, activeAlert := range rule.activeAlerts {
|
|
if activeAlert.State > maxState {
|
|
maxState = activeAlert.State
|
|
}
|
|
}
|
|
return maxState
|
|
}
|
|
|
|
// ActiveAlerts returns a slice of active alerts.
|
|
func (rule *AlertingRule) ActiveAlerts() []Alert {
|
|
rule.mutex.Lock()
|
|
defer rule.mutex.Unlock()
|
|
|
|
alerts := make([]Alert, 0, len(rule.activeAlerts))
|
|
for _, alert := range rule.activeAlerts {
|
|
alerts = append(alerts, *alert)
|
|
}
|
|
return alerts
|
|
}
|