2015-01-21 11:07:45 -08:00
|
|
|
// Copyright 2013 The Prometheus Authors
|
2013-02-07 02:49:04 -08:00
|
|
|
// Licensed under the Apache License, Version 2.0 (the "License");
|
|
|
|
// you may not use this file except in compliance with the License.
|
|
|
|
// You may obtain a copy of the License at
|
|
|
|
//
|
|
|
|
// http://www.apache.org/licenses/LICENSE-2.0
|
|
|
|
//
|
|
|
|
// Unless required by applicable law or agreed to in writing, software
|
|
|
|
// distributed under the License is distributed on an "AS IS" BASIS,
|
|
|
|
// WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
|
|
|
|
// See the License for the specific language governing permissions and
|
|
|
|
// limitations under the License.
|
|
|
|
|
2013-01-07 14:24:26 -08:00
|
|
|
package rules
|
|
|
|
|
|
|
|
import (
|
|
|
|
"fmt"
|
2015-02-19 10:56:45 -08:00
|
|
|
"math"
|
2013-04-22 15:26:59 -07:00
|
|
|
"path"
|
2015-02-19 10:56:45 -08:00
|
|
|
"regexp"
|
|
|
|
"strconv"
|
2013-01-07 14:24:26 -08:00
|
|
|
"strings"
|
|
|
|
"testing"
|
2013-03-21 10:06:15 -07:00
|
|
|
"time"
|
2013-06-25 05:02:27 -07:00
|
|
|
|
|
|
|
clientmodel "github.com/prometheus/client_golang/model"
|
|
|
|
|
|
|
|
"github.com/prometheus/prometheus/rules/ast"
|
|
|
|
"github.com/prometheus/prometheus/stats"
|
2014-06-06 02:55:53 -07:00
|
|
|
"github.com/prometheus/prometheus/storage/local"
|
2014-12-12 11:01:32 -08:00
|
|
|
"github.com/prometheus/prometheus/storage/metric"
|
2013-06-25 05:02:27 -07:00
|
|
|
"github.com/prometheus/prometheus/utility/test"
|
2013-01-07 14:24:26 -08:00
|
|
|
)
|
|
|
|
|
2013-04-22 15:26:59 -07:00
|
|
|
var (
|
2013-04-24 02:51:40 -07:00
|
|
|
testEvalTime = testStartTime.Add(testSampleInterval * 10)
|
2013-04-22 15:26:59 -07:00
|
|
|
fixturesPath = "fixtures"
|
2015-02-19 10:56:45 -08:00
|
|
|
|
2015-03-16 07:57:47 -07:00
|
|
|
reSample = regexp.MustCompile(`^(.*)(?: \=\>|:) (\-?\d+\.?\d*(?:e-?\d*)?|[+-]Inf|NaN) \@\[(\d+)\]$`)
|
2015-02-19 10:56:45 -08:00
|
|
|
minNormal = math.Float64frombits(0x0010000000000000) // The smallest positive normal value of type float64.
|
|
|
|
)
|
|
|
|
|
|
|
|
const (
|
|
|
|
epsilon = 0.000001 // Relative error allowed for sample values.
|
2013-04-22 15:26:59 -07:00
|
|
|
)
|
2013-01-07 14:24:26 -08:00
|
|
|
|
Use custom timestamp type for sample timestamps and related code.
So far we've been using Go's native time.Time for anything related to sample
timestamps. Since the range of time.Time is much bigger than what we need, this
has created two problems:
- there could be time.Time values which were out of the range/precision of the
time type that we persist to disk, therefore causing incorrectly ordered keys.
One bug caused by this was:
https://github.com/prometheus/prometheus/issues/367
It would be good to use a timestamp type that's more closely aligned with
what the underlying storage supports.
- sizeof(time.Time) is 192, while Prometheus should be ok with a single 64-bit
Unix timestamp (possibly even a 32-bit one). Since we store samples in large
numbers, this seriously affects memory usage. Furthermore, copying/working
with the data will be faster if it's smaller.
*MEMORY USAGE RESULTS*
Initial memory usage comparisons for a running Prometheus with 1 timeseries and
100,000 samples show roughly a 13% decrease in total (VIRT) memory usage. In my
tests, this advantage for some reason decreased a bit the more samples the
timeseries had (to 5-7% for millions of samples). This I can't fully explain,
but perhaps garbage collection issues were involved.
*WHEN TO USE THE NEW TIMESTAMP TYPE*
The new clientmodel.Timestamp type should be used whenever time
calculations are either directly or indirectly related to sample
timestamps.
For example:
- the timestamp of a sample itself
- all kinds of watermarks
- anything that may become or is compared to a sample timestamp (like the timestamp
passed into Target.Scrape()).
When to still use time.Time:
- for measuring durations/times not related to sample timestamps, like duration
telemetry exporting, timers that indicate how frequently to execute some
action, etc.
*NOTE ON OPERATOR OPTIMIZATION TESTS*
We don't use operator optimization code anymore, but it still lives in
the code as dead code. It still has tests, but I couldn't get all of them to
pass with the new timestamp format. I commented out the failing cases for now,
but we should probably remove the dead code soon. I just didn't want to do that
in the same change as this.
Change-Id: I821787414b0debe85c9fffaeb57abd453727af0f
2013-10-28 06:35:02 -07:00
|
|
|
func annotateWithTime(lines []string, timestamp clientmodel.Timestamp) []string {
|
2013-01-07 14:24:26 -08:00
|
|
|
annotatedLines := []string{}
|
|
|
|
for _, line := range lines {
|
2013-04-24 02:51:40 -07:00
|
|
|
annotatedLines = append(annotatedLines, fmt.Sprintf(line, timestamp))
|
2013-01-07 14:24:26 -08:00
|
|
|
}
|
|
|
|
return annotatedLines
|
|
|
|
}
|
|
|
|
|
|
|
|
func vectorComparisonString(expected []string, actual []string) string {
|
|
|
|
separator := "\n--------------\n"
|
|
|
|
return fmt.Sprintf("Expected:%v%v%v\nActual:%v%v%v ",
|
|
|
|
separator,
|
|
|
|
strings.Join(expected, "\n"),
|
|
|
|
separator,
|
|
|
|
separator,
|
|
|
|
strings.Join(actual, "\n"),
|
|
|
|
separator)
|
|
|
|
}
|
|
|
|
|
2015-02-19 10:56:45 -08:00
|
|
|
// samplesAlmostEqual returns true if the two sample lines only differ by a
|
|
|
|
// small relative error in their sample value.
|
|
|
|
func samplesAlmostEqual(a, b string) bool {
|
|
|
|
if a == b {
|
|
|
|
// Fast path if strings are equal.
|
|
|
|
return true
|
|
|
|
}
|
|
|
|
aMatches := reSample.FindStringSubmatch(a)
|
|
|
|
if aMatches == nil {
|
|
|
|
panic(fmt.Errorf("sample %q did not match regular expression", a))
|
|
|
|
}
|
|
|
|
bMatches := reSample.FindStringSubmatch(b)
|
|
|
|
if bMatches == nil {
|
|
|
|
panic(fmt.Errorf("sample %q did not match regular expression", b))
|
|
|
|
}
|
|
|
|
if aMatches[1] != bMatches[1] {
|
|
|
|
return false // Labels don't match.
|
|
|
|
}
|
2015-03-01 13:49:06 -08:00
|
|
|
if aMatches[3] != bMatches[3] {
|
2015-02-19 10:56:45 -08:00
|
|
|
return false // Timestamps don't match.
|
|
|
|
}
|
|
|
|
// If we are here, we have the diff in the floats.
|
|
|
|
// We have to check if they are almost equal.
|
2015-03-01 13:49:06 -08:00
|
|
|
aVal, err := strconv.ParseFloat(aMatches[2], 64)
|
2015-02-19 10:56:45 -08:00
|
|
|
if err != nil {
|
|
|
|
panic(err)
|
|
|
|
}
|
2015-03-01 13:49:06 -08:00
|
|
|
bVal, err := strconv.ParseFloat(bMatches[2], 64)
|
2015-02-19 10:56:45 -08:00
|
|
|
if err != nil {
|
|
|
|
panic(err)
|
|
|
|
}
|
|
|
|
|
|
|
|
// Cf. http://floating-point-gui.de/errors/comparison/
|
|
|
|
if aVal == bVal {
|
|
|
|
return true
|
|
|
|
}
|
|
|
|
|
|
|
|
diff := math.Abs(aVal - bVal)
|
|
|
|
|
|
|
|
if aVal == 0 || bVal == 0 || diff < minNormal {
|
|
|
|
return diff < epsilon*minNormal
|
|
|
|
}
|
|
|
|
return diff/(math.Abs(aVal)+math.Abs(bVal)) < epsilon
|
|
|
|
}
|
|
|
|
|
2014-10-08 04:49:42 -07:00
|
|
|
func newTestStorage(t testing.TB) (storage local.Storage, closer test.Closer) {
|
2015-03-04 04:40:18 -08:00
|
|
|
storage, closer = local.NewTestStorage(t, 1)
|
2014-04-11 00:27:05 -07:00
|
|
|
storeMatrix(storage, testMatrix)
|
2014-06-06 02:55:53 -07:00
|
|
|
return storage, closer
|
2013-04-24 02:51:40 -07:00
|
|
|
}
|
2013-01-07 14:24:26 -08:00
|
|
|
|
2013-05-08 07:35:16 -07:00
|
|
|
func TestExpressions(t *testing.T) {
|
2013-04-24 02:51:40 -07:00
|
|
|
// Labels in expected output need to be alphabetically sorted.
|
2014-12-12 11:01:32 -08:00
|
|
|
expressionTests := []struct {
|
2015-02-17 17:30:41 -08:00
|
|
|
expr string
|
|
|
|
output []string
|
|
|
|
shouldFail bool
|
|
|
|
checkOrder bool
|
2013-04-24 02:51:40 -07:00
|
|
|
}{
|
|
|
|
{
|
2015-02-17 17:30:41 -08:00
|
|
|
expr: `SUM(http_requests)`,
|
|
|
|
output: []string{`{} => 3600 @[%v]`},
|
2013-12-13 10:20:42 -08:00
|
|
|
}, {
|
|
|
|
expr: `SUM(http_requests{instance="0"}) BY(job)`,
|
|
|
|
output: []string{
|
2014-10-15 08:03:08 -07:00
|
|
|
`{job="api-server"} => 400 @[%v]`,
|
|
|
|
`{job="app-server"} => 1200 @[%v]`,
|
2013-12-13 10:20:42 -08:00
|
|
|
},
|
|
|
|
}, {
|
|
|
|
expr: `SUM(http_requests{instance="0"}) BY(job) KEEPING_EXTRA`,
|
|
|
|
output: []string{
|
2014-10-15 08:03:08 -07:00
|
|
|
`{instance="0", job="api-server"} => 400 @[%v]`,
|
|
|
|
`{instance="0", job="app-server"} => 1200 @[%v]`,
|
2013-12-13 10:20:42 -08:00
|
|
|
},
|
2013-04-24 02:51:40 -07:00
|
|
|
}, {
|
2013-06-11 02:01:36 -07:00
|
|
|
expr: `SUM(http_requests) BY (job)`,
|
2013-04-24 02:51:40 -07:00
|
|
|
output: []string{
|
2014-10-15 08:03:08 -07:00
|
|
|
`{job="api-server"} => 1000 @[%v]`,
|
|
|
|
`{job="app-server"} => 2600 @[%v]`,
|
2013-04-24 02:51:40 -07:00
|
|
|
},
|
2014-01-24 07:05:30 -08:00
|
|
|
}, {
|
|
|
|
// Non-existent labels mentioned in BY-clauses shouldn't propagate to output.
|
|
|
|
expr: `SUM(http_requests) BY (job, nonexistent)`,
|
|
|
|
output: []string{
|
2014-10-15 08:03:08 -07:00
|
|
|
`{job="api-server"} => 1000 @[%v]`,
|
|
|
|
`{job="app-server"} => 2600 @[%v]`,
|
2014-01-24 07:05:30 -08:00
|
|
|
},
|
2013-07-11 09:38:44 -07:00
|
|
|
}, {
|
|
|
|
expr: `
|
|
|
|
// Test comment.
|
|
|
|
SUM(http_requests) BY /* comments shouldn't
|
|
|
|
have any effect */ (job) // another comment`,
|
|
|
|
output: []string{
|
2014-10-15 08:03:08 -07:00
|
|
|
`{job="api-server"} => 1000 @[%v]`,
|
|
|
|
`{job="app-server"} => 2600 @[%v]`,
|
2013-07-11 09:38:44 -07:00
|
|
|
},
|
2013-05-08 07:35:16 -07:00
|
|
|
}, {
|
2013-06-11 02:01:36 -07:00
|
|
|
expr: `COUNT(http_requests) BY (job)`,
|
2013-05-08 07:35:16 -07:00
|
|
|
output: []string{
|
2014-10-15 08:03:08 -07:00
|
|
|
`{job="api-server"} => 4 @[%v]`,
|
|
|
|
`{job="app-server"} => 4 @[%v]`,
|
2013-05-08 07:35:16 -07:00
|
|
|
},
|
2013-04-24 02:51:40 -07:00
|
|
|
}, {
|
2013-06-11 02:01:36 -07:00
|
|
|
expr: `SUM(http_requests) BY (job, group)`,
|
2013-04-24 02:51:40 -07:00
|
|
|
output: []string{
|
2014-10-15 08:03:08 -07:00
|
|
|
`{group="canary", job="api-server"} => 700 @[%v]`,
|
|
|
|
`{group="canary", job="app-server"} => 1500 @[%v]`,
|
|
|
|
`{group="production", job="api-server"} => 300 @[%v]`,
|
|
|
|
`{group="production", job="app-server"} => 1100 @[%v]`,
|
2013-04-24 02:51:40 -07:00
|
|
|
},
|
|
|
|
}, {
|
2013-06-11 02:01:36 -07:00
|
|
|
expr: `AVG(http_requests) BY (job)`,
|
2013-04-24 02:51:40 -07:00
|
|
|
output: []string{
|
2014-10-15 08:03:08 -07:00
|
|
|
`{job="api-server"} => 250 @[%v]`,
|
|
|
|
`{job="app-server"} => 650 @[%v]`,
|
2013-04-24 02:51:40 -07:00
|
|
|
},
|
|
|
|
}, {
|
2013-06-11 02:01:36 -07:00
|
|
|
expr: `MIN(http_requests) BY (job)`,
|
2013-04-24 02:51:40 -07:00
|
|
|
output: []string{
|
2014-10-15 08:03:08 -07:00
|
|
|
`{job="api-server"} => 100 @[%v]`,
|
|
|
|
`{job="app-server"} => 500 @[%v]`,
|
2013-04-24 02:51:40 -07:00
|
|
|
},
|
|
|
|
}, {
|
2013-06-11 02:01:36 -07:00
|
|
|
expr: `MAX(http_requests) BY (job)`,
|
2013-04-24 02:51:40 -07:00
|
|
|
output: []string{
|
2014-10-15 08:03:08 -07:00
|
|
|
`{job="api-server"} => 400 @[%v]`,
|
|
|
|
`{job="app-server"} => 800 @[%v]`,
|
2013-04-24 02:51:40 -07:00
|
|
|
},
|
|
|
|
}, {
|
2013-06-11 02:01:36 -07:00
|
|
|
expr: `SUM(http_requests) BY (job) - COUNT(http_requests) BY (job)`,
|
2013-04-24 02:51:40 -07:00
|
|
|
output: []string{
|
2014-10-15 08:03:08 -07:00
|
|
|
`{job="api-server"} => 996 @[%v]`,
|
|
|
|
`{job="app-server"} => 2596 @[%v]`,
|
2013-04-24 02:51:40 -07:00
|
|
|
},
|
2014-04-08 07:14:35 -07:00
|
|
|
}, {
|
|
|
|
expr: `2 - SUM(http_requests) BY (job)`,
|
|
|
|
output: []string{
|
2014-10-15 08:03:08 -07:00
|
|
|
`{job="api-server"} => -998 @[%v]`,
|
|
|
|
`{job="app-server"} => -2598 @[%v]`,
|
2014-04-08 07:14:35 -07:00
|
|
|
},
|
|
|
|
}, {
|
|
|
|
expr: `1000 / SUM(http_requests) BY (job)`,
|
|
|
|
output: []string{
|
2014-10-15 08:03:08 -07:00
|
|
|
`{job="api-server"} => 1 @[%v]`,
|
|
|
|
`{job="app-server"} => 0.38461538461538464 @[%v]`,
|
2014-04-08 07:14:35 -07:00
|
|
|
},
|
2013-04-24 02:51:40 -07:00
|
|
|
}, {
|
2013-06-11 02:01:36 -07:00
|
|
|
expr: `SUM(http_requests) BY (job) - 2`,
|
2013-04-24 02:51:40 -07:00
|
|
|
output: []string{
|
2014-10-15 08:03:08 -07:00
|
|
|
`{job="api-server"} => 998 @[%v]`,
|
|
|
|
`{job="app-server"} => 2598 @[%v]`,
|
2013-04-24 02:51:40 -07:00
|
|
|
},
|
|
|
|
}, {
|
2013-06-11 02:01:36 -07:00
|
|
|
expr: `SUM(http_requests) BY (job) % 3`,
|
2013-04-24 02:51:40 -07:00
|
|
|
output: []string{
|
2014-10-15 08:03:08 -07:00
|
|
|
`{job="api-server"} => 1 @[%v]`,
|
|
|
|
`{job="app-server"} => 2 @[%v]`,
|
2013-04-24 02:51:40 -07:00
|
|
|
},
|
|
|
|
}, {
|
2013-06-11 02:01:36 -07:00
|
|
|
expr: `SUM(http_requests) BY (job) / 0`,
|
2013-04-24 02:51:40 -07:00
|
|
|
output: []string{
|
2014-10-15 08:03:08 -07:00
|
|
|
`{job="api-server"} => +Inf @[%v]`,
|
|
|
|
`{job="app-server"} => +Inf @[%v]`,
|
2013-04-24 02:51:40 -07:00
|
|
|
},
|
|
|
|
}, {
|
2013-06-11 02:01:36 -07:00
|
|
|
expr: `SUM(http_requests) BY (job) > 1000`,
|
2013-04-24 02:51:40 -07:00
|
|
|
output: []string{
|
2014-10-15 08:03:08 -07:00
|
|
|
`{job="app-server"} => 2600 @[%v]`,
|
2013-04-24 02:51:40 -07:00
|
|
|
},
|
2014-04-08 07:14:35 -07:00
|
|
|
}, {
|
|
|
|
expr: `1000 < SUM(http_requests) BY (job)`,
|
|
|
|
output: []string{
|
2014-10-15 08:03:08 -07:00
|
|
|
`{job="app-server"} => 1000 @[%v]`,
|
2014-04-08 07:14:35 -07:00
|
|
|
},
|
2013-04-24 02:51:40 -07:00
|
|
|
}, {
|
2013-06-11 02:01:36 -07:00
|
|
|
expr: `SUM(http_requests) BY (job) <= 1000`,
|
2013-04-24 02:51:40 -07:00
|
|
|
output: []string{
|
2014-10-15 08:03:08 -07:00
|
|
|
`{job="api-server"} => 1000 @[%v]`,
|
2013-04-24 02:51:40 -07:00
|
|
|
},
|
|
|
|
}, {
|
2013-06-11 02:01:36 -07:00
|
|
|
expr: `SUM(http_requests) BY (job) != 1000`,
|
2013-04-24 02:51:40 -07:00
|
|
|
output: []string{
|
2014-10-15 08:03:08 -07:00
|
|
|
`{job="app-server"} => 2600 @[%v]`,
|
2013-04-24 02:51:40 -07:00
|
|
|
},
|
|
|
|
}, {
|
2013-06-11 02:01:36 -07:00
|
|
|
expr: `SUM(http_requests) BY (job) == 1000`,
|
2013-04-24 02:51:40 -07:00
|
|
|
output: []string{
|
2014-10-15 08:03:08 -07:00
|
|
|
`{job="api-server"} => 1000 @[%v]`,
|
2013-04-24 02:51:40 -07:00
|
|
|
},
|
|
|
|
}, {
|
2013-06-11 02:01:36 -07:00
|
|
|
expr: `SUM(http_requests) BY (job) + SUM(http_requests) BY (job)`,
|
2013-04-24 02:51:40 -07:00
|
|
|
output: []string{
|
2014-10-15 08:03:08 -07:00
|
|
|
`{job="api-server"} => 2000 @[%v]`,
|
|
|
|
`{job="app-server"} => 5200 @[%v]`,
|
2013-04-24 02:51:40 -07:00
|
|
|
},
|
|
|
|
}, {
|
2013-06-11 02:01:36 -07:00
|
|
|
expr: `http_requests{job="api-server", group="canary"}`,
|
2013-04-24 02:51:40 -07:00
|
|
|
output: []string{
|
2013-06-11 02:01:36 -07:00
|
|
|
`http_requests{group="canary", instance="0", job="api-server"} => 300 @[%v]`,
|
|
|
|
`http_requests{group="canary", instance="1", job="api-server"} => 400 @[%v]`,
|
2013-04-24 02:51:40 -07:00
|
|
|
},
|
|
|
|
}, {
|
2015-01-22 03:52:30 -08:00
|
|
|
expr: `http_requests{job="api-server", group="canary"} + rate(http_requests{job="api-server"}[5m]) * 5 * 60`,
|
2013-04-24 02:51:40 -07:00
|
|
|
output: []string{
|
2014-10-15 08:03:08 -07:00
|
|
|
`{group="canary", instance="0", job="api-server"} => 330 @[%v]`,
|
|
|
|
`{group="canary", instance="1", job="api-server"} => 440 @[%v]`,
|
2013-04-24 02:51:40 -07:00
|
|
|
},
|
2015-01-22 03:52:30 -08:00
|
|
|
}, {
|
|
|
|
expr: `rate(http_requests[25m]) * 25 * 60`,
|
|
|
|
output: []string{
|
|
|
|
`{group="canary", instance="0", job="api-server"} => 150 @[%v]`,
|
|
|
|
`{group="canary", instance="0", job="app-server"} => 350 @[%v]`,
|
|
|
|
`{group="canary", instance="1", job="api-server"} => 200 @[%v]`,
|
|
|
|
`{group="canary", instance="1", job="app-server"} => 400 @[%v]`,
|
|
|
|
`{group="production", instance="0", job="api-server"} => 50 @[%v]`,
|
|
|
|
`{group="production", instance="0", job="app-server"} => 249.99999999999997 @[%v]`,
|
|
|
|
`{group="production", instance="1", job="api-server"} => 100 @[%v]`,
|
|
|
|
`{group="production", instance="1", job="app-server"} => 300 @[%v]`,
|
|
|
|
},
|
2013-04-24 02:51:40 -07:00
|
|
|
}, {
|
2013-06-11 02:01:36 -07:00
|
|
|
expr: `delta(http_requests[25m], 1)`,
|
2013-04-24 02:51:40 -07:00
|
|
|
output: []string{
|
2014-10-15 08:03:08 -07:00
|
|
|
`{group="canary", instance="0", job="api-server"} => 150 @[%v]`,
|
|
|
|
`{group="canary", instance="0", job="app-server"} => 350 @[%v]`,
|
|
|
|
`{group="canary", instance="1", job="api-server"} => 200 @[%v]`,
|
|
|
|
`{group="canary", instance="1", job="app-server"} => 400 @[%v]`,
|
|
|
|
`{group="production", instance="0", job="api-server"} => 50 @[%v]`,
|
|
|
|
`{group="production", instance="0", job="app-server"} => 250 @[%v]`,
|
|
|
|
`{group="production", instance="1", job="api-server"} => 100 @[%v]`,
|
|
|
|
`{group="production", instance="1", job="app-server"} => 300 @[%v]`,
|
2013-04-24 02:51:40 -07:00
|
|
|
},
|
|
|
|
}, {
|
2013-06-11 02:01:36 -07:00
|
|
|
expr: `sort(http_requests)`,
|
2013-04-24 02:51:40 -07:00
|
|
|
output: []string{
|
2013-06-11 02:01:36 -07:00
|
|
|
`http_requests{group="production", instance="0", job="api-server"} => 100 @[%v]`,
|
|
|
|
`http_requests{group="production", instance="1", job="api-server"} => 200 @[%v]`,
|
|
|
|
`http_requests{group="canary", instance="0", job="api-server"} => 300 @[%v]`,
|
|
|
|
`http_requests{group="canary", instance="1", job="api-server"} => 400 @[%v]`,
|
|
|
|
`http_requests{group="production", instance="0", job="app-server"} => 500 @[%v]`,
|
|
|
|
`http_requests{group="production", instance="1", job="app-server"} => 600 @[%v]`,
|
|
|
|
`http_requests{group="canary", instance="0", job="app-server"} => 700 @[%v]`,
|
|
|
|
`http_requests{group="canary", instance="1", job="app-server"} => 800 @[%v]`,
|
2013-04-24 02:51:40 -07:00
|
|
|
},
|
2015-02-17 17:30:41 -08:00
|
|
|
checkOrder: true,
|
2013-04-24 02:51:40 -07:00
|
|
|
}, {
|
2013-06-11 02:01:36 -07:00
|
|
|
expr: `sort_desc(http_requests)`,
|
2013-04-24 02:51:40 -07:00
|
|
|
output: []string{
|
2013-06-11 02:01:36 -07:00
|
|
|
`http_requests{group="canary", instance="1", job="app-server"} => 800 @[%v]`,
|
|
|
|
`http_requests{group="canary", instance="0", job="app-server"} => 700 @[%v]`,
|
|
|
|
`http_requests{group="production", instance="1", job="app-server"} => 600 @[%v]`,
|
|
|
|
`http_requests{group="production", instance="0", job="app-server"} => 500 @[%v]`,
|
|
|
|
`http_requests{group="canary", instance="1", job="api-server"} => 400 @[%v]`,
|
|
|
|
`http_requests{group="canary", instance="0", job="api-server"} => 300 @[%v]`,
|
|
|
|
`http_requests{group="production", instance="1", job="api-server"} => 200 @[%v]`,
|
|
|
|
`http_requests{group="production", instance="0", job="api-server"} => 100 @[%v]`,
|
2013-04-24 02:51:40 -07:00
|
|
|
},
|
2015-02-17 17:30:41 -08:00
|
|
|
checkOrder: true,
|
2014-08-05 09:57:47 -07:00
|
|
|
}, {
|
|
|
|
expr: `topk(3, http_requests)`,
|
|
|
|
output: []string{
|
|
|
|
`http_requests{group="canary", instance="1", job="app-server"} => 800 @[%v]`,
|
|
|
|
`http_requests{group="canary", instance="0", job="app-server"} => 700 @[%v]`,
|
|
|
|
`http_requests{group="production", instance="1", job="app-server"} => 600 @[%v]`,
|
|
|
|
},
|
2015-02-17 17:30:41 -08:00
|
|
|
checkOrder: true,
|
2014-08-05 11:14:04 -07:00
|
|
|
}, {
|
|
|
|
expr: `topk(5, http_requests{group="canary",job="app-server"})`,
|
|
|
|
output: []string{
|
|
|
|
`http_requests{group="canary", instance="1", job="app-server"} => 800 @[%v]`,
|
|
|
|
`http_requests{group="canary", instance="0", job="app-server"} => 700 @[%v]`,
|
|
|
|
},
|
2015-02-17 17:30:41 -08:00
|
|
|
checkOrder: true,
|
2014-08-05 09:57:47 -07:00
|
|
|
}, {
|
|
|
|
expr: `bottomk(3, http_requests)`,
|
|
|
|
output: []string{
|
|
|
|
`http_requests{group="production", instance="0", job="api-server"} => 100 @[%v]`,
|
|
|
|
`http_requests{group="production", instance="1", job="api-server"} => 200 @[%v]`,
|
|
|
|
`http_requests{group="canary", instance="0", job="api-server"} => 300 @[%v]`,
|
|
|
|
},
|
2015-02-17 17:30:41 -08:00
|
|
|
checkOrder: true,
|
2014-08-05 11:14:04 -07:00
|
|
|
}, {
|
|
|
|
expr: `bottomk(5, http_requests{group="canary",job="app-server"})`,
|
|
|
|
output: []string{
|
|
|
|
`http_requests{group="canary", instance="0", job="app-server"} => 700 @[%v]`,
|
|
|
|
`http_requests{group="canary", instance="1", job="app-server"} => 800 @[%v]`,
|
|
|
|
},
|
2015-02-17 17:30:41 -08:00
|
|
|
checkOrder: true,
|
2013-04-24 02:51:40 -07:00
|
|
|
}, {
|
|
|
|
// Single-letter label names and values.
|
2013-06-11 02:01:36 -07:00
|
|
|
expr: `x{y="testvalue"}`,
|
2013-04-24 02:51:40 -07:00
|
|
|
output: []string{
|
2013-06-11 02:01:36 -07:00
|
|
|
`x{y="testvalue"} => 100 @[%v]`,
|
2013-04-24 02:51:40 -07:00
|
|
|
},
|
|
|
|
}, {
|
|
|
|
// Lower-cased aggregation operators should work too.
|
2013-06-11 02:01:36 -07:00
|
|
|
expr: `sum(http_requests) by (job) + min(http_requests) by (job) + max(http_requests) by (job) + avg(http_requests) by (job)`,
|
2013-04-24 02:51:40 -07:00
|
|
|
output: []string{
|
2014-10-15 08:03:08 -07:00
|
|
|
`{job="app-server"} => 4550 @[%v]`,
|
|
|
|
`{job="api-server"} => 1750 @[%v]`,
|
2013-04-24 02:51:40 -07:00
|
|
|
},
|
|
|
|
}, {
|
|
|
|
// Deltas should be adjusted for target interval vs. samples under target interval.
|
2015-02-17 17:30:41 -08:00
|
|
|
expr: `delta(http_requests{group="canary", instance="1", job="app-server"}[18m])`,
|
|
|
|
output: []string{`{group="canary", instance="1", job="app-server"} => 288 @[%v]`},
|
2015-01-22 03:52:30 -08:00
|
|
|
}, {
|
|
|
|
// Deltas should perform the same operation when 2nd argument is 0.
|
2015-02-17 17:30:41 -08:00
|
|
|
expr: `delta(http_requests{group="canary", instance="1", job="app-server"}[18m], 0)`,
|
|
|
|
output: []string{`{group="canary", instance="1", job="app-server"} => 288 @[%v]`},
|
2013-04-24 02:51:40 -07:00
|
|
|
}, {
|
2015-01-22 03:52:30 -08:00
|
|
|
// Rates should calculate per-second rates.
|
2015-02-17 17:30:41 -08:00
|
|
|
expr: `rate(http_requests{group="canary", instance="1", job="app-server"}[60m])`,
|
|
|
|
output: []string{`{group="canary", instance="1", job="app-server"} => 0.26666666666666666 @[%v]`},
|
2013-05-11 17:03:16 -07:00
|
|
|
}, {
|
2015-01-22 03:52:30 -08:00
|
|
|
// Deriv should return the same as rate in simple cases.
|
2015-02-17 17:30:41 -08:00
|
|
|
expr: `deriv(http_requests{group="canary", instance="1", job="app-server"}[60m])`,
|
|
|
|
output: []string{`{group="canary", instance="1", job="app-server"} => 0.26666666666666666 @[%v]`},
|
2013-05-11 17:03:16 -07:00
|
|
|
}, {
|
2015-01-22 03:52:30 -08:00
|
|
|
// Counter resets at in the middle of range are handled correctly by rate().
|
2015-02-17 17:30:41 -08:00
|
|
|
expr: `rate(testcounter_reset_middle[60m])`,
|
|
|
|
output: []string{`{} => 0.03 @[%v]`},
|
2013-05-28 08:36:53 -07:00
|
|
|
}, {
|
2015-01-22 03:52:30 -08:00
|
|
|
// Counter resets at end of range are ignored by rate().
|
2015-02-17 17:30:41 -08:00
|
|
|
expr: `rate(testcounter_reset_end[5m])`,
|
|
|
|
output: []string{`{} => 0 @[%v]`},
|
2013-05-28 08:36:53 -07:00
|
|
|
}, {
|
2015-01-22 03:52:30 -08:00
|
|
|
// Deriv should return correct result.
|
2015-02-17 17:30:41 -08:00
|
|
|
expr: `deriv(testcounter_reset_middle[100m])`,
|
|
|
|
output: []string{`{} => 0.010606060606060607 @[%v]`},
|
2014-01-30 04:07:26 -08:00
|
|
|
}, {
|
|
|
|
// count_scalar for a non-empty vector should return scalar element count.
|
2015-02-17 17:30:41 -08:00
|
|
|
expr: `count_scalar(http_requests)`,
|
|
|
|
output: []string{`scalar: 8 @[%v]`},
|
2014-01-30 04:07:26 -08:00
|
|
|
}, {
|
|
|
|
// count_scalar for an empty vector should return scalar 0.
|
2015-02-17 17:30:41 -08:00
|
|
|
expr: `count_scalar(nonexistent)`,
|
|
|
|
output: []string{`scalar: 0 @[%v]`},
|
2013-04-24 02:51:40 -07:00
|
|
|
}, {
|
2014-08-05 11:14:04 -07:00
|
|
|
// Empty expressions shouldn't parse.
|
2013-06-11 02:01:36 -07:00
|
|
|
expr: ``,
|
2013-04-24 02:51:40 -07:00
|
|
|
shouldFail: true,
|
|
|
|
}, {
|
2014-08-05 11:14:04 -07:00
|
|
|
// Interval durations can't be in quotes.
|
2013-06-11 02:01:36 -07:00
|
|
|
expr: `http_requests["1m"]`,
|
2013-04-24 02:51:40 -07:00
|
|
|
shouldFail: true,
|
2014-04-08 07:14:35 -07:00
|
|
|
}, {
|
|
|
|
// Binop arguments need to be scalar or vector.
|
|
|
|
expr: `http_requests - http_requests[1m]`,
|
|
|
|
shouldFail: true,
|
2014-03-28 03:58:47 -07:00
|
|
|
}, {
|
|
|
|
expr: `http_requests{group!="canary"}`,
|
|
|
|
output: []string{
|
|
|
|
`http_requests{group="production", instance="1", job="app-server"} => 600 @[%v]`,
|
|
|
|
`http_requests{group="production", instance="0", job="app-server"} => 500 @[%v]`,
|
|
|
|
`http_requests{group="production", instance="1", job="api-server"} => 200 @[%v]`,
|
|
|
|
`http_requests{group="production", instance="0", job="api-server"} => 100 @[%v]`,
|
|
|
|
},
|
|
|
|
}, {
|
|
|
|
expr: `http_requests{job=~"server",group!="canary"}`,
|
|
|
|
output: []string{
|
|
|
|
`http_requests{group="production", instance="1", job="app-server"} => 600 @[%v]`,
|
|
|
|
`http_requests{group="production", instance="0", job="app-server"} => 500 @[%v]`,
|
|
|
|
`http_requests{group="production", instance="1", job="api-server"} => 200 @[%v]`,
|
|
|
|
`http_requests{group="production", instance="0", job="api-server"} => 100 @[%v]`,
|
|
|
|
},
|
|
|
|
}, {
|
|
|
|
expr: `http_requests{job!~"api",group!="canary"}`,
|
|
|
|
output: []string{
|
|
|
|
`http_requests{group="production", instance="1", job="app-server"} => 600 @[%v]`,
|
|
|
|
`http_requests{group="production", instance="0", job="app-server"} => 500 @[%v]`,
|
|
|
|
},
|
|
|
|
}, {
|
2015-02-17 17:30:41 -08:00
|
|
|
expr: `count_scalar(http_requests{job=~"^server$"})`,
|
|
|
|
output: []string{`scalar: 0 @[%v]`},
|
2014-03-28 03:58:47 -07:00
|
|
|
}, {
|
|
|
|
expr: `http_requests{group="production",job=~"^api"}`,
|
|
|
|
output: []string{
|
2014-07-28 07:12:58 -07:00
|
|
|
`http_requests{group="production", instance="0", job="api-server"} => 100 @[%v]`,
|
2014-03-28 03:58:47 -07:00
|
|
|
`http_requests{group="production", instance="1", job="api-server"} => 200 @[%v]`,
|
2014-07-28 07:12:58 -07:00
|
|
|
},
|
|
|
|
},
|
|
|
|
{
|
|
|
|
expr: `abs(-1 * http_requests{group="production",job="api-server"})`,
|
|
|
|
output: []string{
|
2014-10-15 08:03:08 -07:00
|
|
|
`{group="production", instance="0", job="api-server"} => 100 @[%v]`,
|
|
|
|
`{group="production", instance="1", job="api-server"} => 200 @[%v]`,
|
2014-03-28 03:58:47 -07:00
|
|
|
},
|
2013-04-24 02:51:40 -07:00
|
|
|
},
|
2015-02-04 03:24:00 -08:00
|
|
|
{
|
|
|
|
expr: `floor(0.004 * http_requests{group="production",job="api-server"})`,
|
|
|
|
output: []string{
|
|
|
|
`{group="production", instance="0", job="api-server"} => 0 @[%v]`,
|
|
|
|
`{group="production", instance="1", job="api-server"} => 0 @[%v]`,
|
|
|
|
},
|
|
|
|
},
|
|
|
|
{
|
|
|
|
expr: `ceil(0.004 * http_requests{group="production",job="api-server"})`,
|
|
|
|
output: []string{
|
|
|
|
`{group="production", instance="0", job="api-server"} => 1 @[%v]`,
|
|
|
|
`{group="production", instance="1", job="api-server"} => 1 @[%v]`,
|
|
|
|
},
|
|
|
|
},
|
|
|
|
{
|
|
|
|
expr: `round(0.004 * http_requests{group="production",job="api-server"})`,
|
|
|
|
output: []string{
|
|
|
|
`{group="production", instance="0", job="api-server"} => 0 @[%v]`,
|
|
|
|
`{group="production", instance="1", job="api-server"} => 1 @[%v]`,
|
|
|
|
},
|
|
|
|
},
|
|
|
|
{ // Round should correctly handle negative numbers.
|
|
|
|
expr: `round(-1 * (0.004 * http_requests{group="production",job="api-server"}))`,
|
|
|
|
output: []string{
|
2015-02-05 08:06:35 -08:00
|
|
|
`{group="production", instance="0", job="api-server"} => 0 @[%v]`,
|
2015-02-04 03:24:00 -08:00
|
|
|
`{group="production", instance="1", job="api-server"} => -1 @[%v]`,
|
|
|
|
},
|
|
|
|
},
|
2015-02-05 08:06:35 -08:00
|
|
|
{ // Round should round half up.
|
2015-02-04 03:24:00 -08:00
|
|
|
expr: `round(0.005 * http_requests{group="production",job="api-server"})`,
|
|
|
|
output: []string{
|
2015-02-05 08:06:35 -08:00
|
|
|
`{group="production", instance="0", job="api-server"} => 1 @[%v]`,
|
2015-02-04 03:24:00 -08:00
|
|
|
`{group="production", instance="1", job="api-server"} => 1 @[%v]`,
|
|
|
|
},
|
|
|
|
},
|
|
|
|
{
|
|
|
|
expr: `round(-1 * (0.005 * http_requests{group="production",job="api-server"}))`,
|
|
|
|
output: []string{
|
|
|
|
`{group="production", instance="0", job="api-server"} => 0 @[%v]`,
|
|
|
|
`{group="production", instance="1", job="api-server"} => -1 @[%v]`,
|
|
|
|
},
|
|
|
|
},
|
2015-02-05 08:06:35 -08:00
|
|
|
{
|
2015-02-04 03:24:00 -08:00
|
|
|
expr: `round(1 + 0.005 * http_requests{group="production",job="api-server"})`,
|
|
|
|
output: []string{
|
|
|
|
`{group="production", instance="0", job="api-server"} => 2 @[%v]`,
|
|
|
|
`{group="production", instance="1", job="api-server"} => 2 @[%v]`,
|
|
|
|
},
|
|
|
|
},
|
|
|
|
{
|
|
|
|
expr: `round(-1 * (1 + 0.005 * http_requests{group="production",job="api-server"}))`,
|
|
|
|
output: []string{
|
2015-02-05 08:06:35 -08:00
|
|
|
`{group="production", instance="0", job="api-server"} => -1 @[%v]`,
|
2015-02-04 03:24:00 -08:00
|
|
|
`{group="production", instance="1", job="api-server"} => -2 @[%v]`,
|
|
|
|
},
|
|
|
|
},
|
2015-02-05 08:06:35 -08:00
|
|
|
{ // Round should accept the number to round nearest to.
|
|
|
|
expr: `round(0.0005 * http_requests{group="production",job="api-server"}, 0.1)`,
|
2015-02-04 03:24:00 -08:00
|
|
|
output: []string{
|
2015-02-05 08:06:35 -08:00
|
|
|
`{group="production", instance="0", job="api-server"} => 0.1 @[%v]`,
|
2015-02-04 03:24:00 -08:00
|
|
|
`{group="production", instance="1", job="api-server"} => 0.1 @[%v]`,
|
|
|
|
},
|
|
|
|
},
|
2015-02-05 08:06:35 -08:00
|
|
|
{
|
|
|
|
expr: `round(2.1 + 0.0005 * http_requests{group="production",job="api-server"}, 0.1)`,
|
2015-02-04 03:24:00 -08:00
|
|
|
output: []string{
|
|
|
|
`{group="production", instance="0", job="api-server"} => 2.2 @[%v]`,
|
|
|
|
`{group="production", instance="1", job="api-server"} => 2.2 @[%v]`,
|
|
|
|
},
|
|
|
|
},
|
|
|
|
{
|
2015-02-05 08:06:35 -08:00
|
|
|
expr: `round(5.2 + 0.0005 * http_requests{group="production",job="api-server"}, 0.1)`,
|
2015-02-04 03:24:00 -08:00
|
|
|
output: []string{
|
2015-02-05 08:06:35 -08:00
|
|
|
`{group="production", instance="0", job="api-server"} => 5.3 @[%v]`,
|
2015-02-04 03:24:00 -08:00
|
|
|
`{group="production", instance="1", job="api-server"} => 5.3 @[%v]`,
|
|
|
|
},
|
|
|
|
},
|
|
|
|
{ // Round should work correctly with negative numbers and multiple decimal places.
|
2015-02-05 08:06:35 -08:00
|
|
|
expr: `round(-1 * (5.2 + 0.0005 * http_requests{group="production",job="api-server"}), 0.1)`,
|
2015-02-04 03:24:00 -08:00
|
|
|
output: []string{
|
|
|
|
`{group="production", instance="0", job="api-server"} => -5.2 @[%v]`,
|
|
|
|
`{group="production", instance="1", job="api-server"} => -5.3 @[%v]`,
|
|
|
|
},
|
|
|
|
},
|
2015-02-05 08:06:35 -08:00
|
|
|
{ // Round should work correctly with big toNearests.
|
|
|
|
expr: `round(0.025 * http_requests{group="production",job="api-server"}, 5)`,
|
|
|
|
output: []string{
|
|
|
|
`{group="production", instance="0", job="api-server"} => 5 @[%v]`,
|
|
|
|
`{group="production", instance="1", job="api-server"} => 5 @[%v]`,
|
|
|
|
},
|
|
|
|
},
|
|
|
|
{
|
|
|
|
expr: `round(0.045 * http_requests{group="production",job="api-server"}, 5)`,
|
|
|
|
output: []string{
|
|
|
|
`{group="production", instance="0", job="api-server"} => 5 @[%v]`,
|
|
|
|
`{group="production", instance="1", job="api-server"} => 10 @[%v]`,
|
|
|
|
},
|
|
|
|
},
|
2014-07-28 07:12:58 -07:00
|
|
|
{
|
|
|
|
expr: `avg_over_time(http_requests{group="production",job="api-server"}[1h])`,
|
|
|
|
output: []string{
|
2014-10-15 08:03:08 -07:00
|
|
|
`{group="production", instance="0", job="api-server"} => 50 @[%v]`,
|
|
|
|
`{group="production", instance="1", job="api-server"} => 100 @[%v]`,
|
2014-07-28 07:12:58 -07:00
|
|
|
},
|
|
|
|
},
|
|
|
|
{
|
|
|
|
expr: `count_over_time(http_requests{group="production",job="api-server"}[1h])`,
|
|
|
|
output: []string{
|
2014-10-15 08:03:08 -07:00
|
|
|
`{group="production", instance="0", job="api-server"} => 11 @[%v]`,
|
|
|
|
`{group="production", instance="1", job="api-server"} => 11 @[%v]`,
|
2014-07-28 07:12:58 -07:00
|
|
|
},
|
|
|
|
},
|
|
|
|
{
|
|
|
|
expr: `max_over_time(http_requests{group="production",job="api-server"}[1h])`,
|
|
|
|
output: []string{
|
2014-10-15 08:03:08 -07:00
|
|
|
`{group="production", instance="0", job="api-server"} => 100 @[%v]`,
|
|
|
|
`{group="production", instance="1", job="api-server"} => 200 @[%v]`,
|
2014-07-28 07:12:58 -07:00
|
|
|
},
|
|
|
|
},
|
|
|
|
{
|
|
|
|
expr: `min_over_time(http_requests{group="production",job="api-server"}[1h])`,
|
|
|
|
output: []string{
|
2014-10-15 08:03:08 -07:00
|
|
|
`{group="production", instance="0", job="api-server"} => 0 @[%v]`,
|
|
|
|
`{group="production", instance="1", job="api-server"} => 0 @[%v]`,
|
2014-07-28 07:12:58 -07:00
|
|
|
},
|
|
|
|
},
|
|
|
|
{
|
|
|
|
expr: `sum_over_time(http_requests{group="production",job="api-server"}[1h])`,
|
|
|
|
output: []string{
|
2014-10-15 08:03:08 -07:00
|
|
|
`{group="production", instance="0", job="api-server"} => 550 @[%v]`,
|
|
|
|
`{group="production", instance="1", job="api-server"} => 1100 @[%v]`,
|
2014-07-28 07:12:58 -07:00
|
|
|
},
|
|
|
|
},
|
2014-07-28 09:10:11 -07:00
|
|
|
{
|
2015-02-17 17:30:41 -08:00
|
|
|
expr: `time()`,
|
|
|
|
output: []string{`scalar: 3000 @[%v]`},
|
2014-07-28 09:10:11 -07:00
|
|
|
},
|
2014-08-05 10:56:35 -07:00
|
|
|
{
|
|
|
|
expr: `drop_common_labels(http_requests{group="production",job="api-server"})`,
|
|
|
|
output: []string{
|
|
|
|
`http_requests{instance="0"} => 100 @[%v]`,
|
|
|
|
`http_requests{instance="1"} => 200 @[%v]`,
|
|
|
|
},
|
|
|
|
},
|
2014-10-14 08:21:39 -07:00
|
|
|
{
|
|
|
|
expr: `{` + string(clientmodel.MetricNameLabel) + `=~".*"}`,
|
|
|
|
output: []string{
|
|
|
|
`http_requests{group="canary", instance="0", job="api-server"} => 300 @[%v]`,
|
|
|
|
`http_requests{group="canary", instance="0", job="app-server"} => 700 @[%v]`,
|
|
|
|
`http_requests{group="canary", instance="1", job="api-server"} => 400 @[%v]`,
|
|
|
|
`http_requests{group="canary", instance="1", job="app-server"} => 800 @[%v]`,
|
|
|
|
`http_requests{group="production", instance="0", job="api-server"} => 100 @[%v]`,
|
|
|
|
`http_requests{group="production", instance="0", job="app-server"} => 500 @[%v]`,
|
|
|
|
`http_requests{group="production", instance="1", job="api-server"} => 200 @[%v]`,
|
|
|
|
`http_requests{group="production", instance="1", job="app-server"} => 600 @[%v]`,
|
|
|
|
`testcounter_reset_end => 0 @[%v]`,
|
|
|
|
`testcounter_reset_middle => 50 @[%v]`,
|
|
|
|
`x{y="testvalue"} => 100 @[%v]`,
|
2015-02-21 05:05:50 -08:00
|
|
|
`label_grouping_test{a="a", b="abb"} => 200 @[%v]`,
|
|
|
|
`label_grouping_test{a="aa", b="bb"} => 100 @[%v]`,
|
2015-02-19 10:56:45 -08:00
|
|
|
`testhistogram_bucket{le="0.1", start="positive"} => 50 @[%v]`,
|
|
|
|
`testhistogram_bucket{le=".2", start="positive"} => 70 @[%v]`,
|
|
|
|
`testhistogram_bucket{le="1e0", start="positive"} => 110 @[%v]`,
|
|
|
|
`testhistogram_bucket{le="+Inf", start="positive"} => 120 @[%v]`,
|
|
|
|
`testhistogram_bucket{le="-.2", start="negative"} => 10 @[%v]`,
|
|
|
|
`testhistogram_bucket{le="-0.1", start="negative"} => 20 @[%v]`,
|
|
|
|
`testhistogram_bucket{le="0.3", start="negative"} => 20 @[%v]`,
|
|
|
|
`testhistogram_bucket{le="+Inf", start="negative"} => 30 @[%v]`,
|
|
|
|
`request_duration_seconds_bucket{instance="ins1", job="job1", le="0.1"} => 10 @[%v]`,
|
|
|
|
`request_duration_seconds_bucket{instance="ins1", job="job1", le="0.2"} => 30 @[%v]`,
|
|
|
|
`request_duration_seconds_bucket{instance="ins1", job="job1", le="+Inf"} => 40 @[%v]`,
|
|
|
|
`request_duration_seconds_bucket{instance="ins2", job="job1", le="0.1"} => 20 @[%v]`,
|
|
|
|
`request_duration_seconds_bucket{instance="ins2", job="job1", le="0.2"} => 50 @[%v]`,
|
|
|
|
`request_duration_seconds_bucket{instance="ins2", job="job1", le="+Inf"} => 60 @[%v]`,
|
|
|
|
`request_duration_seconds_bucket{instance="ins1", job="job2", le="0.1"} => 30 @[%v]`,
|
|
|
|
`request_duration_seconds_bucket{instance="ins1", job="job2", le="0.2"} => 40 @[%v]`,
|
|
|
|
`request_duration_seconds_bucket{instance="ins1", job="job2", le="+Inf"} => 60 @[%v]`,
|
|
|
|
`request_duration_seconds_bucket{instance="ins2", job="job2", le="0.1"} => 40 @[%v]`,
|
|
|
|
`request_duration_seconds_bucket{instance="ins2", job="job2", le="0.2"} => 70 @[%v]`,
|
|
|
|
`request_duration_seconds_bucket{instance="ins2", job="job2", le="+Inf"} => 90 @[%v]`,
|
2015-03-03 01:58:28 -08:00
|
|
|
`vector_matching_a{l="x"} => 10 @[%v]`,
|
|
|
|
`vector_matching_a{l="y"} => 20 @[%v]`,
|
|
|
|
`vector_matching_b{l="x"} => 40 @[%v]`,
|
|
|
|
`cpu_count{instance="1", type="smp"} => 200 @[%v]`,
|
|
|
|
`cpu_count{instance="0", type="smp"} => 100 @[%v]`,
|
|
|
|
`cpu_count{instance="0", type="numa"} => 300 @[%v]`,
|
2014-10-14 08:21:39 -07:00
|
|
|
},
|
|
|
|
},
|
|
|
|
{
|
|
|
|
expr: `{job=~"server", job!~"api"}`,
|
|
|
|
output: []string{
|
|
|
|
`http_requests{group="canary", instance="0", job="app-server"} => 700 @[%v]`,
|
|
|
|
`http_requests{group="canary", instance="1", job="app-server"} => 800 @[%v]`,
|
|
|
|
`http_requests{group="production", instance="0", job="app-server"} => 500 @[%v]`,
|
|
|
|
`http_requests{group="production", instance="1", job="app-server"} => 600 @[%v]`,
|
|
|
|
},
|
|
|
|
},
|
2014-10-22 02:03:11 -07:00
|
|
|
{
|
|
|
|
// Test alternative "by"-clause order.
|
|
|
|
expr: `sum by (group) (http_requests{job="api-server"})`,
|
|
|
|
output: []string{
|
2014-11-24 11:12:58 -08:00
|
|
|
`{group="canary"} => 700 @[%v]`,
|
|
|
|
`{group="production"} => 300 @[%v]`,
|
2014-10-22 02:03:11 -07:00
|
|
|
},
|
|
|
|
},
|
|
|
|
{
|
|
|
|
// Test alternative "by"-clause order with "keeping_extra".
|
|
|
|
expr: `sum by (group) keeping_extra (http_requests{job="api-server"})`,
|
|
|
|
output: []string{
|
2014-11-24 11:12:58 -08:00
|
|
|
`{group="canary", job="api-server"} => 700 @[%v]`,
|
|
|
|
`{group="production", job="api-server"} => 300 @[%v]`,
|
2014-10-22 02:03:11 -07:00
|
|
|
},
|
|
|
|
},
|
|
|
|
{
|
|
|
|
// Test both alternative "by"-clause orders in one expression.
|
|
|
|
// Public health warning: stick to one form within an expression (or even
|
|
|
|
// in an organization), or risk serious user confusion.
|
|
|
|
expr: `sum(sum by (group) keeping_extra (http_requests{job="api-server"})) by (job)`,
|
|
|
|
output: []string{
|
2014-11-24 11:12:58 -08:00
|
|
|
`{job="api-server"} => 1000 @[%v]`,
|
2014-10-22 02:03:11 -07:00
|
|
|
},
|
|
|
|
},
|
2015-03-03 01:58:28 -08:00
|
|
|
{
|
|
|
|
expr: `http_requests{group="canary"} and http_requests{instance="0"}`,
|
|
|
|
output: []string{
|
|
|
|
`http_requests{group="canary", instance="0", job="api-server"} => 300 @[%v]`,
|
|
|
|
`http_requests{group="canary", instance="0", job="app-server"} => 700 @[%v]`,
|
|
|
|
},
|
|
|
|
},
|
|
|
|
{
|
|
|
|
expr: `(http_requests{group="canary"} + 1) and http_requests{instance="0"}`,
|
|
|
|
output: []string{
|
|
|
|
`{group="canary", instance="0", job="api-server"} => 301 @[%v]`,
|
|
|
|
`{group="canary", instance="0", job="app-server"} => 701 @[%v]`,
|
|
|
|
},
|
|
|
|
},
|
|
|
|
{
|
|
|
|
expr: `(http_requests{group="canary"} + 1) and on(instance, job) http_requests{instance="0", group="production"}`,
|
|
|
|
output: []string{
|
|
|
|
`{group="canary", instance="0", job="api-server"} => 301 @[%v]`,
|
|
|
|
`{group="canary", instance="0", job="app-server"} => 701 @[%v]`,
|
|
|
|
},
|
|
|
|
},
|
|
|
|
{
|
|
|
|
expr: `(http_requests{group="canary"} + 1) and on(instance) http_requests{instance="0", group="production"}`,
|
|
|
|
output: []string{
|
|
|
|
`{group="canary", instance="0", job="api-server"} => 301 @[%v]`,
|
|
|
|
`{group="canary", instance="0", job="app-server"} => 701 @[%v]`,
|
|
|
|
},
|
|
|
|
},
|
|
|
|
{
|
|
|
|
expr: `http_requests{group="canary"} or http_requests{group="production"}`,
|
|
|
|
output: []string{
|
|
|
|
`http_requests{group="canary", instance="0", job="api-server"} => 300 @[%v]`,
|
|
|
|
`http_requests{group="canary", instance="0", job="app-server"} => 700 @[%v]`,
|
|
|
|
`http_requests{group="canary", instance="1", job="api-server"} => 400 @[%v]`,
|
|
|
|
`http_requests{group="canary", instance="1", job="app-server"} => 800 @[%v]`,
|
|
|
|
`http_requests{group="production", instance="0", job="api-server"} => 100 @[%v]`,
|
|
|
|
`http_requests{group="production", instance="0", job="app-server"} => 500 @[%v]`,
|
|
|
|
`http_requests{group="production", instance="1", job="api-server"} => 200 @[%v]`,
|
|
|
|
`http_requests{group="production", instance="1", job="app-server"} => 600 @[%v]`,
|
|
|
|
},
|
|
|
|
},
|
|
|
|
{
|
|
|
|
// On overlap the rhs samples must be dropped.
|
|
|
|
expr: `(http_requests{group="canary"} + 1) or http_requests{instance="1"}`,
|
|
|
|
output: []string{
|
|
|
|
`{group="canary", instance="0", job="api-server"} => 301 @[%v]`,
|
|
|
|
`{group="canary", instance="0", job="app-server"} => 701 @[%v]`,
|
|
|
|
`{group="canary", instance="1", job="api-server"} => 401 @[%v]`,
|
|
|
|
`{group="canary", instance="1", job="app-server"} => 801 @[%v]`,
|
|
|
|
`http_requests{group="production", instance="1", job="api-server"} => 200 @[%v]`,
|
|
|
|
`http_requests{group="production", instance="1", job="app-server"} => 600 @[%v]`,
|
|
|
|
},
|
|
|
|
},
|
|
|
|
{
|
|
|
|
// Matching only on instance excludes everything that has instance=0/1 but includes
|
|
|
|
// entries without the instance label.
|
|
|
|
expr: `(http_requests{group="canary"} + 1) or on(instance) (http_requests or cpu_count or vector_matching_a)`,
|
|
|
|
output: []string{
|
|
|
|
`{group="canary", instance="0", job="api-server"} => 301 @[%v]`,
|
|
|
|
`{group="canary", instance="0", job="app-server"} => 701 @[%v]`,
|
|
|
|
`{group="canary", instance="1", job="api-server"} => 401 @[%v]`,
|
|
|
|
`{group="canary", instance="1", job="app-server"} => 801 @[%v]`,
|
|
|
|
`vector_matching_a{l="x"} => 10 @[%v]`,
|
|
|
|
`vector_matching_a{l="y"} => 20 @[%v]`,
|
|
|
|
},
|
|
|
|
},
|
|
|
|
{
|
|
|
|
expr: `http_requests{group="canary"} / on(instance,job) http_requests{group="production"}`,
|
|
|
|
output: []string{
|
|
|
|
`{instance="0", job="api-server"} => 3 @[%v]`,
|
|
|
|
`{instance="0", job="app-server"} => 1.4 @[%v]`,
|
|
|
|
`{instance="1", job="api-server"} => 2 @[%v]`,
|
|
|
|
`{instance="1", job="app-server"} => 1.3333333333333333 @[%v]`,
|
|
|
|
},
|
|
|
|
},
|
|
|
|
{
|
|
|
|
// Include labels must guarantee uniquely identifiable time series.
|
|
|
|
expr: `http_requests{group="production"} / on(instance) group_left(group) cpu_count{type="smp"}`,
|
|
|
|
output: []string{}, // Empty result returned on error (see TODOs).
|
|
|
|
},
|
|
|
|
{
|
|
|
|
// Many-to-many matching is not allowed.
|
|
|
|
expr: `http_requests{group="production"} / on(instance) group_left(job,type) cpu_count`,
|
|
|
|
output: []string{}, // Empty result returned on error (see TODOs).
|
|
|
|
},
|
|
|
|
{
|
|
|
|
// Many-to-one matching must be explicit.
|
|
|
|
expr: `http_requests{group="production"} / on(instance) cpu_count{type="smp"}`,
|
|
|
|
output: []string{}, // Empty result returned on error (see TODOs).
|
|
|
|
},
|
|
|
|
{
|
|
|
|
expr: `http_requests{group="production"} / on(instance) group_left(job) cpu_count{type="smp"}`,
|
|
|
|
output: []string{
|
|
|
|
`{instance="1", job="api-server"} => 1 @[%v]`,
|
|
|
|
`{instance="0", job="app-server"} => 5 @[%v]`,
|
|
|
|
`{instance="1", job="app-server"} => 3 @[%v]`,
|
|
|
|
`{instance="0", job="api-server"} => 1 @[%v]`,
|
|
|
|
},
|
|
|
|
},
|
|
|
|
{
|
|
|
|
// Ensure sidedness of grouping preserves operand sides.
|
|
|
|
expr: `cpu_count{type="smp"} / on(instance) group_right(job) http_requests{group="production"}`,
|
|
|
|
output: []string{
|
|
|
|
`{instance="1", job="app-server"} => 0.3333333333333333 @[%v]`,
|
|
|
|
`{instance="0", job="app-server"} => 0.2 @[%v]`,
|
|
|
|
`{instance="1", job="api-server"} => 1 @[%v]`,
|
|
|
|
`{instance="0", job="api-server"} => 1 @[%v]`,
|
|
|
|
},
|
|
|
|
},
|
|
|
|
{
|
|
|
|
// Include labels from both sides.
|
|
|
|
expr: `http_requests{group="production"} / on(instance) group_left(job) cpu_count{type="smp"}`,
|
|
|
|
output: []string{
|
|
|
|
`{instance="1", job="api-server"} => 1 @[%v]`,
|
|
|
|
`{instance="0", job="app-server"} => 5 @[%v]`,
|
|
|
|
`{instance="1", job="app-server"} => 3 @[%v]`,
|
|
|
|
`{instance="0", job="api-server"} => 1 @[%v]`,
|
|
|
|
},
|
|
|
|
},
|
|
|
|
{
|
|
|
|
expr: `http_requests{group="production"} < on(instance,job) http_requests{group="canary"}`,
|
|
|
|
output: []string{
|
|
|
|
`{instance="1", job="app-server"} => 600 @[%v]`,
|
|
|
|
`{instance="0", job="app-server"} => 500 @[%v]`,
|
|
|
|
`{instance="1", job="api-server"} => 200 @[%v]`,
|
|
|
|
`{instance="0", job="api-server"} => 100 @[%v]`,
|
|
|
|
},
|
|
|
|
},
|
|
|
|
{
|
|
|
|
expr: `http_requests{group="production"} > on(instance,job) http_requests{group="canary"}`,
|
|
|
|
output: []string{},
|
|
|
|
},
|
|
|
|
{
|
|
|
|
expr: `http_requests{group="production"} == on(instance,job) http_requests{group="canary"}`,
|
|
|
|
output: []string{},
|
|
|
|
},
|
|
|
|
{
|
|
|
|
expr: `http_requests > on(instance) group_left(group,job) cpu_count{type="smp"}`,
|
|
|
|
output: []string{
|
|
|
|
`{group="canary", instance="0", job="app-server"} => 700 @[%v]`,
|
|
|
|
`{group="canary", instance="1", job="app-server"} => 800 @[%v]`,
|
|
|
|
`{group="canary", instance="0", job="api-server"} => 300 @[%v]`,
|
|
|
|
`{group="canary", instance="1", job="api-server"} => 400 @[%v]`,
|
|
|
|
`{group="production", instance="0", job="app-server"} => 500 @[%v]`,
|
|
|
|
`{group="production", instance="1", job="app-server"} => 600 @[%v]`,
|
|
|
|
},
|
|
|
|
},
|
|
|
|
{
|
|
|
|
expr: `http_requests / on(instance) 3`,
|
|
|
|
shouldFail: true,
|
|
|
|
},
|
|
|
|
{
|
|
|
|
expr: `3 / on(instance) http_requests_total`,
|
|
|
|
shouldFail: true,
|
|
|
|
},
|
|
|
|
{
|
|
|
|
expr: `3 / on(instance) 3`,
|
|
|
|
shouldFail: true,
|
|
|
|
},
|
|
|
|
{
|
|
|
|
// Missing label list for grouping mod.
|
|
|
|
expr: `http_requests{group="production"} / on(instance) group_left cpu_count{type="smp"}`,
|
|
|
|
shouldFail: true,
|
|
|
|
},
|
|
|
|
{
|
|
|
|
// No group mod allowed for logical operations.
|
|
|
|
expr: `http_requests{group="production"} or on(instance) group_left(type) cpu_count{type="smp"}`,
|
|
|
|
shouldFail: true,
|
|
|
|
},
|
|
|
|
{
|
|
|
|
// No group mod allowed for logical operations.
|
|
|
|
expr: `http_requests{group="production"} and on(instance) group_left(type) cpu_count{type="smp"}`,
|
|
|
|
shouldFail: true,
|
|
|
|
},
|
|
|
|
{
|
|
|
|
// No duplicate use of label.
|
|
|
|
expr: `http_requests{group="production"} + on(instance) group_left(job,instance) cpu_count{type="smp"}`,
|
|
|
|
shouldFail: true,
|
|
|
|
},
|
|
|
|
{
|
|
|
|
expr: `{l="x"} + on(__name__) {l="y"}`,
|
|
|
|
output: []string{
|
|
|
|
`vector_matching_a => 30 @[%v]`,
|
|
|
|
},
|
|
|
|
},
|
Add absent() function.
A common problem in Prometheus alerting is to detect when no timeseries
exist for a given metric name and label combination. Unfortunately,
Prometheus alert expressions need to be of vector type, and
"count(nonexistent_metric)" results in an empty vector, yielding no
output vector elements to base an alert on. The newly introduced
absent() function solves this issue:
ALERT FooAbsent IF absent(foo{job="myjob"}) [...]
absent() has the following behavior:
- if the vector passed to it has any elements, it returns an empty
vector.
- if the vector passed to it has no elements, it returns a 1-element
vector with the value 1.
In the second case, absent() tries to be smart about deriving labels of
the 1-element output vector from the input vector:
absent(nonexistent{job="myjob"}) => {job="myjob"}
absent(nonexistent{job="myjob",instance=~".*"}) => {job="myjob"}
absent(sum(nonexistent{job="myjob"})) => {}
That is, if the passed vector is a literal vector selector, it takes all
"=" label matchers as the basis for the output labels, but ignores all
non-equals or regex matchers. Also, if the passed vector results from a
non-selector expression, no labels can be derived.
Change-Id: I948505a1488d50265ab5692a3286bd7c8c70cd78
2014-10-15 10:13:52 -07:00
|
|
|
{
|
|
|
|
expr: `absent(nonexistent)`,
|
|
|
|
output: []string{
|
|
|
|
`{} => 1 @[%v]`,
|
|
|
|
},
|
|
|
|
},
|
|
|
|
{
|
|
|
|
expr: `absent(nonexistent{job="testjob", instance="testinstance", method=~".*"})`,
|
|
|
|
output: []string{
|
|
|
|
`{instance="testinstance", job="testjob"} => 1 @[%v]`,
|
|
|
|
},
|
|
|
|
},
|
|
|
|
{
|
|
|
|
expr: `count_scalar(absent(http_requests))`,
|
|
|
|
output: []string{
|
|
|
|
`scalar: 0 @[%v]`,
|
|
|
|
},
|
|
|
|
},
|
|
|
|
{
|
|
|
|
expr: `count_scalar(absent(sum(http_requests)))`,
|
|
|
|
output: []string{
|
|
|
|
`scalar: 0 @[%v]`,
|
|
|
|
},
|
|
|
|
},
|
|
|
|
{
|
|
|
|
expr: `absent(sum(nonexistent{job="testjob", instance="testinstance"}))`,
|
|
|
|
output: []string{
|
|
|
|
`{} => 1 @[%v]`,
|
|
|
|
},
|
2015-02-17 17:30:41 -08:00
|
|
|
},
|
|
|
|
{
|
|
|
|
expr: `http_requests{group="production",job="api-server"} offset 5m`,
|
|
|
|
output: []string{
|
|
|
|
`http_requests{group="production", instance="0", job="api-server"} => 90 @[%v]`,
|
|
|
|
`http_requests{group="production", instance="1", job="api-server"} => 180 @[%v]`,
|
|
|
|
},
|
|
|
|
},
|
|
|
|
{
|
|
|
|
expr: `rate(http_requests{group="production",job="api-server"}[10m] offset 5m)`,
|
|
|
|
output: []string{
|
|
|
|
`{group="production", instance="0", job="api-server"} => 0.03333333333333333 @[%v]`,
|
|
|
|
`{group="production", instance="1", job="api-server"} => 0.06666666666666667 @[%v]`,
|
|
|
|
},
|
Add absent() function.
A common problem in Prometheus alerting is to detect when no timeseries
exist for a given metric name and label combination. Unfortunately,
Prometheus alert expressions need to be of vector type, and
"count(nonexistent_metric)" results in an empty vector, yielding no
output vector elements to base an alert on. The newly introduced
absent() function solves this issue:
ALERT FooAbsent IF absent(foo{job="myjob"}) [...]
absent() has the following behavior:
- if the vector passed to it has any elements, it returns an empty
vector.
- if the vector passed to it has no elements, it returns a 1-element
vector with the value 1.
In the second case, absent() tries to be smart about deriving labels of
the 1-element output vector from the input vector:
absent(nonexistent{job="myjob"}) => {job="myjob"}
absent(nonexistent{job="myjob",instance=~".*"}) => {job="myjob"}
absent(sum(nonexistent{job="myjob"})) => {}
That is, if the passed vector is a literal vector selector, it takes all
"=" label matchers as the basis for the output labels, but ignores all
non-equals or regex matchers. Also, if the passed vector results from a
non-selector expression, no labels can be derived.
Change-Id: I948505a1488d50265ab5692a3286bd7c8c70cd78
2014-10-15 10:13:52 -07:00
|
|
|
},
|
2015-02-17 17:56:40 -08:00
|
|
|
{
|
|
|
|
expr: `rate(http_requests[10m]) offset 5m`,
|
|
|
|
shouldFail: true,
|
|
|
|
},
|
|
|
|
{
|
|
|
|
expr: `sum(http_requests) offset 5m`,
|
|
|
|
shouldFail: true,
|
|
|
|
},
|
2015-02-21 05:05:50 -08:00
|
|
|
// Regression test for missing separator byte in labelsToGroupingKey.
|
|
|
|
{
|
|
|
|
expr: `sum(label_grouping_test) by (a, b)`,
|
|
|
|
output: []string{
|
|
|
|
`{a="a", b="abb"} => 200 @[%v]`,
|
|
|
|
`{a="aa", b="bb"} => 100 @[%v]`,
|
|
|
|
},
|
|
|
|
},
|
2015-02-19 10:56:45 -08:00
|
|
|
// Quantile too low.
|
|
|
|
{
|
|
|
|
expr: `histogram_quantile(-0.1, testhistogram_bucket)`,
|
|
|
|
output: []string{
|
|
|
|
`{start="positive"} => -Inf @[%v]`,
|
|
|
|
`{start="negative"} => -Inf @[%v]`,
|
|
|
|
},
|
|
|
|
},
|
|
|
|
// Quantile too high.
|
|
|
|
{
|
|
|
|
expr: `histogram_quantile(1.01, testhistogram_bucket)`,
|
|
|
|
output: []string{
|
|
|
|
`{start="positive"} => +Inf @[%v]`,
|
|
|
|
`{start="negative"} => +Inf @[%v]`,
|
|
|
|
},
|
|
|
|
},
|
|
|
|
// Quantile value in lowest bucket, which is positive.
|
|
|
|
{
|
|
|
|
expr: `histogram_quantile(0, testhistogram_bucket{start="positive"})`,
|
|
|
|
output: []string{
|
|
|
|
`{start="positive"} => 0 @[%v]`,
|
|
|
|
},
|
|
|
|
},
|
|
|
|
// Quantile value in lowest bucket, which is negative.
|
|
|
|
{
|
|
|
|
expr: `histogram_quantile(0, testhistogram_bucket{start="negative"})`,
|
|
|
|
output: []string{
|
|
|
|
`{start="negative"} => -0.2 @[%v]`,
|
|
|
|
},
|
|
|
|
},
|
|
|
|
// Quantile value in highest bucket.
|
|
|
|
{
|
|
|
|
expr: `histogram_quantile(1, testhistogram_bucket)`,
|
|
|
|
output: []string{
|
|
|
|
`{start="positive"} => 1 @[%v]`,
|
|
|
|
`{start="negative"} => 0.3 @[%v]`,
|
|
|
|
},
|
|
|
|
},
|
|
|
|
// Finally some useful quantiles.
|
|
|
|
{
|
|
|
|
expr: `histogram_quantile(0.2, testhistogram_bucket)`,
|
|
|
|
output: []string{
|
|
|
|
`{start="positive"} => 0.048 @[%v]`,
|
|
|
|
`{start="negative"} => -0.2 @[%v]`,
|
|
|
|
},
|
|
|
|
},
|
|
|
|
{
|
|
|
|
expr: `histogram_quantile(0.5, testhistogram_bucket)`,
|
|
|
|
output: []string{
|
|
|
|
`{start="positive"} => 0.15 @[%v]`,
|
|
|
|
`{start="negative"} => -0.15 @[%v]`,
|
|
|
|
},
|
|
|
|
},
|
|
|
|
{
|
|
|
|
expr: `histogram_quantile(0.8, testhistogram_bucket)`,
|
|
|
|
output: []string{
|
|
|
|
`{start="positive"} => 0.72 @[%v]`,
|
|
|
|
`{start="negative"} => 0.3 @[%v]`,
|
|
|
|
},
|
|
|
|
},
|
|
|
|
// More realistic with rates.
|
|
|
|
{
|
|
|
|
expr: `histogram_quantile(0.2, rate(testhistogram_bucket[5m]))`,
|
|
|
|
output: []string{
|
|
|
|
`{start="positive"} => 0.048 @[%v]`,
|
|
|
|
`{start="negative"} => -0.2 @[%v]`,
|
|
|
|
},
|
|
|
|
},
|
|
|
|
{
|
|
|
|
expr: `histogram_quantile(0.5, rate(testhistogram_bucket[5m]))`,
|
|
|
|
output: []string{
|
|
|
|
`{start="positive"} => 0.15 @[%v]`,
|
|
|
|
`{start="negative"} => -0.15 @[%v]`,
|
|
|
|
},
|
|
|
|
},
|
|
|
|
{
|
|
|
|
expr: `histogram_quantile(0.8, rate(testhistogram_bucket[5m]))`,
|
|
|
|
output: []string{
|
|
|
|
`{start="positive"} => 0.72 @[%v]`,
|
|
|
|
`{start="negative"} => 0.3 @[%v]`,
|
|
|
|
},
|
|
|
|
},
|
|
|
|
// Aggregated histogram: Everything in one.
|
|
|
|
{
|
|
|
|
expr: `histogram_quantile(0.3, sum(rate(request_duration_seconds_bucket[5m])) by (le))`,
|
|
|
|
output: []string{
|
|
|
|
`{} => 0.075 @[%v]`,
|
|
|
|
},
|
|
|
|
},
|
|
|
|
{
|
|
|
|
expr: `histogram_quantile(0.5, sum(rate(request_duration_seconds_bucket[5m])) by (le))`,
|
|
|
|
output: []string{
|
|
|
|
`{} => 0.1277777777777778 @[%v]`,
|
|
|
|
},
|
|
|
|
},
|
|
|
|
// Aggregated histogram: Everything in one. Now with avg, which does not change anything.
|
|
|
|
{
|
|
|
|
expr: `histogram_quantile(0.3, avg(rate(request_duration_seconds_bucket[5m])) by (le))`,
|
|
|
|
output: []string{
|
|
|
|
`{} => 0.075 @[%v]`,
|
|
|
|
},
|
|
|
|
},
|
|
|
|
{
|
|
|
|
expr: `histogram_quantile(0.5, avg(rate(request_duration_seconds_bucket[5m])) by (le))`,
|
|
|
|
output: []string{
|
|
|
|
`{} => 0.12777777777777778 @[%v]`,
|
|
|
|
},
|
|
|
|
},
|
|
|
|
// Aggregated histogram: By job.
|
|
|
|
{
|
|
|
|
expr: `histogram_quantile(0.3, sum(rate(request_duration_seconds_bucket[5m])) by (le, instance))`,
|
|
|
|
output: []string{
|
|
|
|
`{instance="ins1"} => 0.075 @[%v]`,
|
|
|
|
`{instance="ins2"} => 0.075 @[%v]`,
|
|
|
|
},
|
|
|
|
},
|
|
|
|
{
|
|
|
|
expr: `histogram_quantile(0.5, sum(rate(request_duration_seconds_bucket[5m])) by (le, instance))`,
|
|
|
|
output: []string{
|
|
|
|
`{instance="ins1"} => 0.1333333333 @[%v]`,
|
|
|
|
`{instance="ins2"} => 0.125 @[%v]`,
|
|
|
|
},
|
|
|
|
},
|
|
|
|
// Aggregated histogram: By instance.
|
|
|
|
{
|
|
|
|
expr: `histogram_quantile(0.3, sum(rate(request_duration_seconds_bucket[5m])) by (le, job))`,
|
|
|
|
output: []string{
|
|
|
|
`{job="job1"} => 0.1 @[%v]`,
|
|
|
|
`{job="job2"} => 0.0642857142857143 @[%v]`,
|
|
|
|
},
|
|
|
|
},
|
|
|
|
{
|
|
|
|
expr: `histogram_quantile(0.5, sum(rate(request_duration_seconds_bucket[5m])) by (le, job))`,
|
|
|
|
output: []string{
|
|
|
|
`{job="job1"} => 0.14 @[%v]`,
|
|
|
|
`{job="job2"} => 0.1125 @[%v]`,
|
|
|
|
},
|
|
|
|
},
|
|
|
|
// Aggregated histogram: By job and instance.
|
|
|
|
{
|
|
|
|
expr: `histogram_quantile(0.3, sum(rate(request_duration_seconds_bucket[5m])) by (le, job, instance))`,
|
|
|
|
output: []string{
|
|
|
|
`{instance="ins1", job="job1"} => 0.11 @[%v]`,
|
|
|
|
`{instance="ins2", job="job1"} => 0.09 @[%v]`,
|
|
|
|
`{instance="ins1", job="job2"} => 0.06 @[%v]`,
|
|
|
|
`{instance="ins2", job="job2"} => 0.0675 @[%v]`,
|
|
|
|
},
|
|
|
|
},
|
|
|
|
{
|
|
|
|
expr: `histogram_quantile(0.5, sum(rate(request_duration_seconds_bucket[5m])) by (le, job, instance))`,
|
|
|
|
output: []string{
|
|
|
|
`{instance="ins1", job="job1"} => 0.15 @[%v]`,
|
|
|
|
`{instance="ins2", job="job1"} => 0.1333333333333333 @[%v]`,
|
|
|
|
`{instance="ins1", job="job2"} => 0.1 @[%v]`,
|
|
|
|
`{instance="ins2", job="job2"} => 0.1166666666666667 @[%v]`,
|
|
|
|
},
|
|
|
|
},
|
|
|
|
// The unaggregated histogram for comparison. Same result as the previous one.
|
|
|
|
{
|
|
|
|
expr: `histogram_quantile(0.3, rate(request_duration_seconds_bucket[5m]))`,
|
|
|
|
output: []string{
|
|
|
|
`{instance="ins1", job="job1"} => 0.11 @[%v]`,
|
|
|
|
`{instance="ins2", job="job1"} => 0.09 @[%v]`,
|
|
|
|
`{instance="ins1", job="job2"} => 0.06 @[%v]`,
|
|
|
|
`{instance="ins2", job="job2"} => 0.0675 @[%v]`,
|
|
|
|
},
|
|
|
|
},
|
|
|
|
{
|
|
|
|
expr: `histogram_quantile(0.5, rate(request_duration_seconds_bucket[5m]))`,
|
|
|
|
output: []string{
|
|
|
|
`{instance="ins1", job="job1"} => 0.15 @[%v]`,
|
|
|
|
`{instance="ins2", job="job1"} => 0.13333333333333333 @[%v]`,
|
|
|
|
`{instance="ins1", job="job2"} => 0.1 @[%v]`,
|
|
|
|
`{instance="ins2", job="job2"} => 0.11666666666666667 @[%v]`,
|
|
|
|
},
|
|
|
|
},
|
2015-03-01 10:18:33 -08:00
|
|
|
{
|
|
|
|
expr: `12.34e6`,
|
|
|
|
output: []string{`scalar: 12340000 @[%v]`},
|
|
|
|
},
|
|
|
|
{
|
|
|
|
expr: `12.34e+6`,
|
|
|
|
output: []string{`scalar: 12340000 @[%v]`},
|
|
|
|
},
|
|
|
|
{
|
|
|
|
expr: `12.34e-6`,
|
|
|
|
output: []string{`scalar: 0.00001234 @[%v]`},
|
|
|
|
},
|
2015-03-03 03:16:46 -08:00
|
|
|
{
|
|
|
|
expr: `1+1`,
|
|
|
|
output: []string{`scalar: 2 @[%v]`},
|
|
|
|
},
|
|
|
|
{
|
|
|
|
expr: `1-1`,
|
|
|
|
output: []string{`scalar: 0 @[%v]`},
|
|
|
|
},
|
|
|
|
{
|
|
|
|
expr: `1 - -1`,
|
|
|
|
output: []string{`scalar: 2 @[%v]`},
|
|
|
|
},
|
2015-03-01 10:18:33 -08:00
|
|
|
{
|
|
|
|
expr: `.2`,
|
|
|
|
output: []string{`scalar: 0.2 @[%v]`},
|
|
|
|
},
|
|
|
|
{
|
|
|
|
expr: `+0.2`,
|
|
|
|
output: []string{`scalar: 0.2 @[%v]`},
|
|
|
|
},
|
|
|
|
{
|
|
|
|
expr: `-0.2e-6`,
|
|
|
|
output: []string{`scalar: -0.0000002 @[%v]`},
|
|
|
|
},
|
|
|
|
{
|
|
|
|
expr: `+Inf`,
|
|
|
|
output: []string{`scalar: +Inf @[%v]`},
|
|
|
|
},
|
|
|
|
{
|
|
|
|
expr: `inF`,
|
|
|
|
output: []string{`scalar: +Inf @[%v]`},
|
|
|
|
},
|
|
|
|
{
|
|
|
|
expr: `-inf`,
|
|
|
|
output: []string{`scalar: -Inf @[%v]`},
|
|
|
|
},
|
|
|
|
{
|
|
|
|
expr: `NaN`,
|
|
|
|
output: []string{`scalar: NaN @[%v]`},
|
|
|
|
},
|
|
|
|
{
|
|
|
|
expr: `nan`,
|
|
|
|
output: []string{`scalar: NaN @[%v]`},
|
|
|
|
},
|
|
|
|
{
|
2015-03-01 13:49:06 -08:00
|
|
|
expr: `2.`,
|
|
|
|
output: []string{`scalar: 2 @[%v]`},
|
2015-03-01 10:18:33 -08:00
|
|
|
},
|
|
|
|
{
|
|
|
|
expr: `999999999999999999999999999999999999999999999999999999999999999999999999999999999999999999999999999999999999999999999999999999999999999999999999999999999999999999999999999999999999999999999999999999999999999999999999999999999999999999999999999999999999999999999999999999999999999999999999999999999999999999999`,
|
|
|
|
output: []string{`scalar: +Inf @[%v]`},
|
|
|
|
},
|
2015-03-16 06:23:03 -07:00
|
|
|
{
|
|
|
|
expr: `1 / 0`,
|
|
|
|
output: []string{`scalar: +Inf @[%v]`},
|
|
|
|
},
|
|
|
|
{
|
|
|
|
expr: `-1 / 0`,
|
|
|
|
output: []string{`scalar: -Inf @[%v]`},
|
|
|
|
},
|
|
|
|
{
|
|
|
|
expr: `0 / 0`,
|
|
|
|
output: []string{`scalar: NaN @[%v]`},
|
|
|
|
},
|
|
|
|
{
|
|
|
|
expr: `1 % 0`,
|
|
|
|
output: []string{`scalar: NaN @[%v]`},
|
|
|
|
},
|
|
|
|
{
|
|
|
|
expr: `http_requests{group="canary", instance="0", job="api-server"} / 0`,
|
|
|
|
output: []string{
|
|
|
|
`{group="canary", instance="0", job="api-server"} => +Inf @[%v]`,
|
|
|
|
},
|
|
|
|
},
|
|
|
|
{
|
|
|
|
expr: `-1 * http_requests{group="canary", instance="0", job="api-server"} / 0`,
|
|
|
|
output: []string{
|
|
|
|
`{group="canary", instance="0", job="api-server"} => -Inf @[%v]`,
|
|
|
|
},
|
|
|
|
},
|
|
|
|
{
|
|
|
|
expr: `0 * http_requests{group="canary", instance="0", job="api-server"} / 0`,
|
|
|
|
output: []string{
|
|
|
|
`{group="canary", instance="0", job="api-server"} => NaN @[%v]`,
|
|
|
|
},
|
|
|
|
},
|
|
|
|
{
|
|
|
|
expr: `0 * http_requests{group="canary", instance="0", job="api-server"} % 0`,
|
|
|
|
output: []string{
|
|
|
|
`{group="canary", instance="0", job="api-server"} => NaN @[%v]`,
|
|
|
|
},
|
|
|
|
},
|
2015-03-16 07:57:47 -07:00
|
|
|
{
|
|
|
|
expr: `exp(vector_matching_a)`,
|
|
|
|
output: []string{
|
|
|
|
`{l="x"} => 22026.465794806718 @[%v]`,
|
|
|
|
`{l="y"} => 485165195.4097903 @[%v]`,
|
|
|
|
},
|
|
|
|
},
|
|
|
|
{
|
|
|
|
expr: `exp(vector_matching_a - 10)`,
|
|
|
|
output: []string{
|
|
|
|
`{l="y"} => 22026.465794806718 @[%v]`,
|
|
|
|
`{l="x"} => 1 @[%v]`,
|
|
|
|
},
|
|
|
|
},
|
|
|
|
{
|
|
|
|
expr: `exp(vector_matching_a - 20)`,
|
|
|
|
output: []string{
|
|
|
|
`{l="x"} => 4.5399929762484854e-05 @[%v]`,
|
|
|
|
`{l="y"} => 1 @[%v]`,
|
|
|
|
},
|
|
|
|
},
|
|
|
|
{
|
|
|
|
expr: `ln(vector_matching_a)`,
|
|
|
|
output: []string{
|
|
|
|
`{l="x"} => 2.302585092994046 @[%v]`,
|
|
|
|
`{l="y"} => 2.995732273553991 @[%v]`,
|
|
|
|
},
|
|
|
|
},
|
|
|
|
{
|
|
|
|
expr: `ln(vector_matching_a - 10)`,
|
|
|
|
output: []string{
|
|
|
|
`{l="y"} => 2.302585092994046 @[%v]`,
|
|
|
|
`{l="x"} => -Inf @[%v]`,
|
|
|
|
},
|
|
|
|
},
|
|
|
|
{
|
|
|
|
expr: `ln(vector_matching_a - 20)`,
|
|
|
|
output: []string{
|
|
|
|
`{l="y"} => -Inf @[%v]`,
|
|
|
|
`{l="x"} => NaN @[%v]`,
|
|
|
|
},
|
|
|
|
},
|
|
|
|
{
|
|
|
|
expr: `exp(ln(vector_matching_a))`,
|
|
|
|
output: []string{
|
|
|
|
`{l="y"} => 20 @[%v]`,
|
|
|
|
`{l="x"} => 10 @[%v]`,
|
|
|
|
},
|
|
|
|
},
|
|
|
|
{
|
|
|
|
expr: `log2(vector_matching_a)`,
|
|
|
|
output: []string{
|
|
|
|
`{l="x"} => 3.3219280948873626 @[%v]`,
|
|
|
|
`{l="y"} => 4.321928094887363 @[%v]`,
|
|
|
|
},
|
|
|
|
},
|
|
|
|
{
|
|
|
|
expr: `log2(vector_matching_a - 10)`,
|
|
|
|
output: []string{
|
|
|
|
`{l="y"} => 3.3219280948873626 @[%v]`,
|
|
|
|
`{l="x"} => -Inf @[%v]`,
|
|
|
|
},
|
|
|
|
},
|
|
|
|
{
|
|
|
|
expr: `log2(vector_matching_a - 20)`,
|
|
|
|
output: []string{
|
|
|
|
`{l="x"} => NaN @[%v]`,
|
|
|
|
`{l="y"} => -Inf @[%v]`,
|
|
|
|
},
|
|
|
|
},
|
|
|
|
{
|
|
|
|
expr: `log10(vector_matching_a)`,
|
|
|
|
output: []string{
|
|
|
|
`{l="x"} => 1 @[%v]`,
|
|
|
|
`{l="y"} => 1.301029995663981 @[%v]`,
|
|
|
|
},
|
|
|
|
},
|
|
|
|
{
|
|
|
|
expr: `log10(vector_matching_a - 10)`,
|
|
|
|
output: []string{
|
|
|
|
`{l="y"} => 1 @[%v]`,
|
|
|
|
`{l="x"} => -Inf @[%v]`,
|
|
|
|
},
|
|
|
|
},
|
|
|
|
{
|
|
|
|
expr: `log10(vector_matching_a - 20)`,
|
|
|
|
output: []string{
|
|
|
|
`{l="x"} => NaN @[%v]`,
|
|
|
|
`{l="y"} => -Inf @[%v]`,
|
|
|
|
},
|
|
|
|
},
|
2013-04-24 02:51:40 -07:00
|
|
|
}
|
2013-03-21 10:06:15 -07:00
|
|
|
|
2014-06-06 02:55:53 -07:00
|
|
|
storage, closer := newTestStorage(t)
|
2013-04-24 02:51:40 -07:00
|
|
|
defer closer.Close()
|
2013-01-07 14:24:26 -08:00
|
|
|
|
2013-04-10 01:44:13 -07:00
|
|
|
for i, exprTest := range expressionTests {
|
2013-04-24 02:51:40 -07:00
|
|
|
expectedLines := annotateWithTime(exprTest.output, testEvalTime)
|
2013-01-07 14:24:26 -08:00
|
|
|
|
2013-01-10 16:17:37 -08:00
|
|
|
testExpr, err := LoadExprFromString(exprTest.expr)
|
2013-01-07 14:24:26 -08:00
|
|
|
|
|
|
|
if err != nil {
|
2013-01-10 16:17:37 -08:00
|
|
|
if exprTest.shouldFail {
|
2013-01-07 14:24:26 -08:00
|
|
|
continue
|
|
|
|
}
|
2013-04-24 02:51:40 -07:00
|
|
|
t.Errorf("%d. Error during parsing: %v", i, err)
|
|
|
|
t.Errorf("%d. Expression: %v", i, exprTest.expr)
|
2013-01-07 14:24:26 -08:00
|
|
|
} else {
|
2013-03-21 10:06:15 -07:00
|
|
|
if exprTest.shouldFail {
|
2013-04-24 02:51:40 -07:00
|
|
|
t.Errorf("%d. Test should fail, but didn't", i)
|
2013-03-21 10:06:15 -07:00
|
|
|
}
|
2013-01-07 14:24:26 -08:00
|
|
|
failed := false
|
2015-03-03 01:58:28 -08:00
|
|
|
|
2014-12-24 16:28:35 -08:00
|
|
|
resultStr := ast.EvalToString(testExpr, testEvalTime, ast.Text, storage, stats.NewTimerGroup())
|
2013-01-07 14:24:26 -08:00
|
|
|
resultLines := strings.Split(resultStr, "\n")
|
|
|
|
|
2015-03-03 01:58:28 -08:00
|
|
|
if len(exprTest.output) == 0 && strings.Trim(resultStr, "\n") == "" {
|
|
|
|
// expected and received empty vector, everything is fine
|
|
|
|
continue
|
|
|
|
} else if len(exprTest.output) != len(resultLines) {
|
2013-04-24 02:51:40 -07:00
|
|
|
t.Errorf("%d. Number of samples in expected and actual output don't match", i)
|
2013-01-07 14:24:26 -08:00
|
|
|
failed = true
|
|
|
|
}
|
2013-04-10 01:44:13 -07:00
|
|
|
|
|
|
|
if exprTest.checkOrder {
|
|
|
|
for j, expectedSample := range expectedLines {
|
|
|
|
if resultLines[j] != expectedSample {
|
2013-04-24 02:51:40 -07:00
|
|
|
t.Errorf("%d.%d. Expected sample '%v', got '%v'", i, j, resultLines[j], expectedSample)
|
2013-04-10 01:44:13 -07:00
|
|
|
failed = true
|
2013-01-07 14:24:26 -08:00
|
|
|
}
|
|
|
|
}
|
2013-04-10 01:44:13 -07:00
|
|
|
} else {
|
|
|
|
for j, expectedSample := range expectedLines {
|
|
|
|
found := false
|
|
|
|
for _, actualSample := range resultLines {
|
2015-02-19 10:56:45 -08:00
|
|
|
if samplesAlmostEqual(actualSample, expectedSample) {
|
2013-04-10 01:44:13 -07:00
|
|
|
found = true
|
|
|
|
}
|
|
|
|
}
|
|
|
|
if !found {
|
2013-04-24 02:51:40 -07:00
|
|
|
t.Errorf("%d.%d. Couldn't find expected sample in output: '%v'", i, j, expectedSample)
|
2013-04-10 01:44:13 -07:00
|
|
|
failed = true
|
|
|
|
}
|
2013-01-07 14:24:26 -08:00
|
|
|
}
|
|
|
|
}
|
2013-03-21 10:06:15 -07:00
|
|
|
|
2013-01-07 14:24:26 -08:00
|
|
|
if failed {
|
2013-04-24 02:51:40 -07:00
|
|
|
t.Errorf("%d. Expression: %v\n%v", i, exprTest.expr, vectorComparisonString(expectedLines, resultLines))
|
2013-01-07 14:24:26 -08:00
|
|
|
}
|
|
|
|
}
|
|
|
|
}
|
|
|
|
}
|
2013-04-22 15:26:59 -07:00
|
|
|
|
2014-12-12 11:01:32 -08:00
|
|
|
func TestRangedEvaluationRegressions(t *testing.T) {
|
|
|
|
scenarios := []struct {
|
|
|
|
in ast.Matrix
|
|
|
|
out ast.Matrix
|
|
|
|
expr string
|
|
|
|
}{
|
|
|
|
{
|
|
|
|
// Testing COWMetric behavior in drop_common_labels.
|
|
|
|
in: ast.Matrix{
|
|
|
|
{
|
|
|
|
Metric: clientmodel.COWMetric{
|
|
|
|
Metric: clientmodel.Metric{
|
|
|
|
clientmodel.MetricNameLabel: "testmetric",
|
|
|
|
"testlabel": "1",
|
|
|
|
},
|
|
|
|
},
|
|
|
|
Values: metric.Values{
|
|
|
|
{
|
|
|
|
Timestamp: testStartTime,
|
|
|
|
Value: 1,
|
|
|
|
},
|
|
|
|
{
|
|
|
|
Timestamp: testStartTime.Add(time.Hour),
|
|
|
|
Value: 1,
|
|
|
|
},
|
|
|
|
},
|
|
|
|
},
|
|
|
|
{
|
|
|
|
Metric: clientmodel.COWMetric{
|
|
|
|
Metric: clientmodel.Metric{
|
|
|
|
clientmodel.MetricNameLabel: "testmetric",
|
|
|
|
"testlabel": "2",
|
|
|
|
},
|
|
|
|
},
|
|
|
|
Values: metric.Values{
|
|
|
|
{
|
|
|
|
Timestamp: testStartTime.Add(time.Hour),
|
|
|
|
Value: 2,
|
|
|
|
},
|
|
|
|
},
|
|
|
|
},
|
|
|
|
},
|
|
|
|
out: ast.Matrix{
|
|
|
|
{
|
|
|
|
Metric: clientmodel.COWMetric{
|
|
|
|
Metric: clientmodel.Metric{
|
|
|
|
clientmodel.MetricNameLabel: "testmetric",
|
|
|
|
},
|
|
|
|
},
|
|
|
|
Values: metric.Values{
|
|
|
|
{
|
|
|
|
Timestamp: testStartTime,
|
|
|
|
Value: 1,
|
|
|
|
},
|
|
|
|
},
|
|
|
|
},
|
|
|
|
{
|
|
|
|
Metric: clientmodel.COWMetric{
|
|
|
|
Metric: clientmodel.Metric{
|
|
|
|
clientmodel.MetricNameLabel: "testmetric",
|
|
|
|
"testlabel": "1",
|
|
|
|
},
|
|
|
|
},
|
|
|
|
Values: metric.Values{
|
|
|
|
{
|
|
|
|
Timestamp: testStartTime.Add(time.Hour),
|
|
|
|
Value: 1,
|
|
|
|
},
|
|
|
|
},
|
|
|
|
},
|
|
|
|
{
|
|
|
|
Metric: clientmodel.COWMetric{
|
|
|
|
Metric: clientmodel.Metric{
|
|
|
|
clientmodel.MetricNameLabel: "testmetric",
|
|
|
|
"testlabel": "2",
|
|
|
|
},
|
|
|
|
},
|
|
|
|
Values: metric.Values{
|
|
|
|
{
|
|
|
|
Timestamp: testStartTime.Add(time.Hour),
|
|
|
|
Value: 2,
|
|
|
|
},
|
|
|
|
},
|
|
|
|
},
|
|
|
|
},
|
|
|
|
expr: "drop_common_labels(testmetric)",
|
|
|
|
},
|
|
|
|
{
|
|
|
|
// Testing COWMetric behavior in vector aggregation.
|
|
|
|
in: ast.Matrix{
|
|
|
|
{
|
|
|
|
Metric: clientmodel.COWMetric{
|
|
|
|
Metric: clientmodel.Metric{
|
|
|
|
clientmodel.MetricNameLabel: "testmetric",
|
|
|
|
"testlabel": "1",
|
|
|
|
},
|
|
|
|
},
|
|
|
|
Values: metric.Values{
|
|
|
|
{
|
|
|
|
Timestamp: testStartTime,
|
|
|
|
Value: 1,
|
|
|
|
},
|
|
|
|
{
|
|
|
|
Timestamp: testStartTime.Add(time.Hour),
|
|
|
|
Value: 1,
|
|
|
|
},
|
|
|
|
},
|
|
|
|
},
|
|
|
|
{
|
|
|
|
Metric: clientmodel.COWMetric{
|
|
|
|
Metric: clientmodel.Metric{
|
|
|
|
clientmodel.MetricNameLabel: "testmetric",
|
|
|
|
"testlabel": "2",
|
|
|
|
},
|
|
|
|
},
|
|
|
|
Values: metric.Values{
|
|
|
|
{
|
|
|
|
Timestamp: testStartTime,
|
|
|
|
Value: 2,
|
|
|
|
},
|
|
|
|
},
|
|
|
|
},
|
|
|
|
},
|
|
|
|
out: ast.Matrix{
|
|
|
|
{
|
|
|
|
Metric: clientmodel.COWMetric{
|
|
|
|
Metric: clientmodel.Metric{},
|
|
|
|
},
|
|
|
|
Values: metric.Values{
|
|
|
|
{
|
|
|
|
Timestamp: testStartTime,
|
|
|
|
Value: 3,
|
|
|
|
},
|
|
|
|
},
|
|
|
|
},
|
|
|
|
{
|
|
|
|
Metric: clientmodel.COWMetric{
|
|
|
|
Metric: clientmodel.Metric{
|
|
|
|
"testlabel": "1",
|
|
|
|
},
|
|
|
|
},
|
|
|
|
Values: metric.Values{
|
|
|
|
{
|
|
|
|
Timestamp: testStartTime.Add(time.Hour),
|
|
|
|
Value: 1,
|
|
|
|
},
|
|
|
|
},
|
|
|
|
},
|
|
|
|
},
|
|
|
|
expr: "sum(testmetric) keeping_extra",
|
|
|
|
},
|
2015-02-21 08:45:47 -08:00
|
|
|
{
|
|
|
|
// Testing metric fingerprint grouping behavior.
|
|
|
|
in: ast.Matrix{
|
|
|
|
{
|
|
|
|
Metric: clientmodel.COWMetric{
|
|
|
|
Metric: clientmodel.Metric{
|
|
|
|
clientmodel.MetricNameLabel: "testmetric",
|
|
|
|
"aa": "bb",
|
|
|
|
},
|
|
|
|
},
|
|
|
|
Values: metric.Values{
|
|
|
|
{
|
|
|
|
Timestamp: testStartTime,
|
|
|
|
Value: 1,
|
|
|
|
},
|
|
|
|
},
|
|
|
|
},
|
|
|
|
{
|
|
|
|
Metric: clientmodel.COWMetric{
|
|
|
|
Metric: clientmodel.Metric{
|
|
|
|
clientmodel.MetricNameLabel: "testmetric",
|
|
|
|
"a": "abb",
|
|
|
|
},
|
|
|
|
},
|
|
|
|
Values: metric.Values{
|
|
|
|
{
|
|
|
|
Timestamp: testStartTime,
|
|
|
|
Value: 2,
|
|
|
|
},
|
|
|
|
},
|
|
|
|
},
|
|
|
|
},
|
|
|
|
out: ast.Matrix{
|
|
|
|
{
|
|
|
|
Metric: clientmodel.COWMetric{
|
|
|
|
Metric: clientmodel.Metric{
|
|
|
|
clientmodel.MetricNameLabel: "testmetric",
|
|
|
|
"aa": "bb",
|
|
|
|
},
|
|
|
|
},
|
|
|
|
Values: metric.Values{
|
|
|
|
{
|
|
|
|
Timestamp: testStartTime,
|
|
|
|
Value: 1,
|
|
|
|
},
|
|
|
|
},
|
|
|
|
},
|
|
|
|
{
|
|
|
|
Metric: clientmodel.COWMetric{
|
|
|
|
Metric: clientmodel.Metric{
|
|
|
|
clientmodel.MetricNameLabel: "testmetric",
|
|
|
|
"a": "abb",
|
|
|
|
},
|
|
|
|
},
|
|
|
|
Values: metric.Values{
|
|
|
|
{
|
|
|
|
Timestamp: testStartTime,
|
|
|
|
Value: 2,
|
|
|
|
},
|
|
|
|
},
|
|
|
|
},
|
|
|
|
},
|
|
|
|
expr: "testmetric",
|
|
|
|
},
|
2014-12-12 11:01:32 -08:00
|
|
|
}
|
|
|
|
|
|
|
|
for i, s := range scenarios {
|
2015-03-04 04:40:18 -08:00
|
|
|
storage, closer := local.NewTestStorage(t, 1)
|
2014-12-12 11:01:32 -08:00
|
|
|
storeMatrix(storage, s.in)
|
|
|
|
|
|
|
|
expr, err := LoadExprFromString(s.expr)
|
|
|
|
if err != nil {
|
|
|
|
t.Fatalf("%d. Error parsing expression: %v", i, err)
|
|
|
|
}
|
|
|
|
|
|
|
|
got, err := ast.EvalVectorRange(
|
|
|
|
expr.(ast.VectorNode),
|
|
|
|
testStartTime,
|
|
|
|
testStartTime.Add(time.Hour),
|
|
|
|
time.Hour,
|
|
|
|
storage,
|
|
|
|
stats.NewTimerGroup(),
|
|
|
|
)
|
|
|
|
if err != nil {
|
|
|
|
t.Fatalf("%d. Error evaluating expression: %v", i, err)
|
|
|
|
}
|
|
|
|
|
|
|
|
if got.String() != s.out.String() {
|
|
|
|
t.Fatalf("%d. Expression: %s\n\ngot:\n=====\n%v\n====\n\nwant:\n=====\n%v\n=====\n", i, s.expr, got.String(), s.out.String())
|
|
|
|
}
|
|
|
|
|
|
|
|
closer.Close()
|
|
|
|
}
|
|
|
|
}
|
|
|
|
|
2013-04-22 15:26:59 -07:00
|
|
|
var ruleTests = []struct {
|
|
|
|
inputFile string
|
|
|
|
shouldFail bool
|
|
|
|
errContains string
|
|
|
|
numRecordingRules int
|
|
|
|
numAlertingRules int
|
|
|
|
}{
|
|
|
|
{
|
|
|
|
inputFile: "empty.rules",
|
|
|
|
numRecordingRules: 0,
|
|
|
|
numAlertingRules: 0,
|
|
|
|
}, {
|
|
|
|
inputFile: "mixed.rules",
|
|
|
|
numRecordingRules: 2,
|
|
|
|
numAlertingRules: 2,
|
|
|
|
},
|
|
|
|
{
|
|
|
|
inputFile: "syntax_error.rules",
|
|
|
|
shouldFail: true,
|
2013-07-11 09:38:44 -07:00
|
|
|
errContains: "Error parsing rules at line 5",
|
2013-04-22 15:26:59 -07:00
|
|
|
},
|
|
|
|
{
|
|
|
|
inputFile: "non_vector.rules",
|
|
|
|
shouldFail: true,
|
|
|
|
errContains: "does not evaluate to vector type",
|
|
|
|
},
|
|
|
|
}
|
|
|
|
|
|
|
|
func TestRules(t *testing.T) {
|
|
|
|
for i, ruleTest := range ruleTests {
|
|
|
|
testRules, err := LoadRulesFromFile(path.Join(fixturesPath, ruleTest.inputFile))
|
|
|
|
|
|
|
|
if err != nil {
|
|
|
|
if !ruleTest.shouldFail {
|
|
|
|
t.Fatalf("%d. Error parsing rules file %v: %v", i, ruleTest.inputFile, err)
|
|
|
|
} else {
|
|
|
|
if !strings.Contains(err.Error(), ruleTest.errContains) {
|
|
|
|
t.Fatalf("%d. Expected error containing '%v', got: %v", i, ruleTest.errContains, err)
|
|
|
|
}
|
|
|
|
}
|
|
|
|
} else {
|
|
|
|
numRecordingRules := 0
|
|
|
|
numAlertingRules := 0
|
|
|
|
|
|
|
|
for j, rule := range testRules {
|
|
|
|
switch rule.(type) {
|
|
|
|
case *RecordingRule:
|
|
|
|
numRecordingRules++
|
|
|
|
case *AlertingRule:
|
|
|
|
numAlertingRules++
|
|
|
|
default:
|
|
|
|
t.Fatalf("%d.%d. Unknown rule type!", i, j)
|
|
|
|
}
|
|
|
|
}
|
|
|
|
|
|
|
|
if numRecordingRules != ruleTest.numRecordingRules {
|
|
|
|
t.Fatalf("%d. Expected %d recording rules, got %d", i, ruleTest.numRecordingRules, numRecordingRules)
|
|
|
|
}
|
|
|
|
if numAlertingRules != ruleTest.numAlertingRules {
|
|
|
|
t.Fatalf("%d. Expected %d alerting rules, got %d", i, ruleTest.numAlertingRules, numAlertingRules)
|
|
|
|
}
|
|
|
|
|
|
|
|
// TODO(julius): add more complex checks on the parsed rules here.
|
|
|
|
}
|
|
|
|
}
|
|
|
|
}
|
2013-04-24 02:51:40 -07:00
|
|
|
|
|
|
|
func TestAlertingRule(t *testing.T) {
|
|
|
|
// Labels in expected output need to be alphabetically sorted.
|
|
|
|
var evalOutputs = [][]string{
|
|
|
|
{
|
2013-07-30 08:18:07 -07:00
|
|
|
`ALERTS{alertname="HttpRequestRateLow", alertstate="pending", group="canary", instance="0", job="app-server", severity="critical"} => 1 @[%v]`,
|
|
|
|
`ALERTS{alertname="HttpRequestRateLow", alertstate="pending", group="canary", instance="1", job="app-server", severity="critical"} => 1 @[%v]`,
|
2013-04-24 02:51:40 -07:00
|
|
|
},
|
|
|
|
{
|
2013-07-30 08:18:07 -07:00
|
|
|
`ALERTS{alertname="HttpRequestRateLow", alertstate="pending", group="canary", instance="0", job="app-server", severity="critical"} => 0 @[%v]`,
|
|
|
|
`ALERTS{alertname="HttpRequestRateLow", alertstate="firing", group="canary", instance="0", job="app-server", severity="critical"} => 1 @[%v]`,
|
|
|
|
`ALERTS{alertname="HttpRequestRateLow", alertstate="pending", group="canary", instance="1", job="app-server", severity="critical"} => 0 @[%v]`,
|
|
|
|
`ALERTS{alertname="HttpRequestRateLow", alertstate="firing", group="canary", instance="1", job="app-server", severity="critical"} => 1 @[%v]`,
|
2013-04-24 02:51:40 -07:00
|
|
|
},
|
|
|
|
{
|
2013-07-30 08:18:07 -07:00
|
|
|
`ALERTS{alertname="HttpRequestRateLow", alertstate="firing", group="canary", instance="1", job="app-server", severity="critical"} => 0 @[%v]`,
|
|
|
|
`ALERTS{alertname="HttpRequestRateLow", alertstate="firing", group="canary", instance="0", job="app-server", severity="critical"} => 0 @[%v]`,
|
2013-04-24 02:51:40 -07:00
|
|
|
},
|
|
|
|
{
|
|
|
|
/* empty */
|
|
|
|
},
|
|
|
|
{
|
|
|
|
/* empty */
|
|
|
|
},
|
|
|
|
}
|
|
|
|
|
2014-06-06 02:55:53 -07:00
|
|
|
storage, closer := newTestStorage(t)
|
2013-04-24 02:51:40 -07:00
|
|
|
defer closer.Close()
|
|
|
|
|
2013-06-11 02:01:36 -07:00
|
|
|
alertExpr, err := LoadExprFromString(`http_requests{group="canary", job="app-server"} < 100`)
|
2013-04-24 02:51:40 -07:00
|
|
|
if err != nil {
|
|
|
|
t.Fatalf("Unable to parse alert expression: %s", err)
|
|
|
|
}
|
|
|
|
alertName := "HttpRequestRateLow"
|
2013-06-25 05:02:27 -07:00
|
|
|
alertLabels := clientmodel.LabelSet{
|
2013-07-30 08:18:07 -07:00
|
|
|
"severity": "critical",
|
2013-04-24 02:51:40 -07:00
|
|
|
}
|
2013-07-30 08:18:07 -07:00
|
|
|
rule := NewAlertingRule(alertName, alertExpr.(ast.VectorNode), time.Minute, alertLabels, "summary", "description")
|
2013-04-24 02:51:40 -07:00
|
|
|
|
|
|
|
for i, expected := range evalOutputs {
|
|
|
|
evalTime := testStartTime.Add(testSampleInterval * time.Duration(i))
|
2014-06-06 02:55:53 -07:00
|
|
|
actual, err := rule.Eval(evalTime, storage)
|
2013-04-24 02:51:40 -07:00
|
|
|
if err != nil {
|
|
|
|
t.Fatalf("Error during alerting rule evaluation: %s", err)
|
|
|
|
}
|
|
|
|
actualLines := strings.Split(actual.String(), "\n")
|
|
|
|
expectedLines := annotateWithTime(expected, evalTime)
|
|
|
|
if actualLines[0] == "" {
|
|
|
|
actualLines = []string{}
|
|
|
|
}
|
|
|
|
|
|
|
|
failed := false
|
|
|
|
if len(actualLines) != len(expectedLines) {
|
|
|
|
t.Errorf("%d. Number of samples in expected and actual output don't match (%d vs. %d)", i, len(expectedLines), len(actualLines))
|
|
|
|
failed = true
|
|
|
|
}
|
|
|
|
|
|
|
|
for j, expectedSample := range expectedLines {
|
|
|
|
found := false
|
|
|
|
for _, actualSample := range actualLines {
|
|
|
|
if actualSample == expectedSample {
|
|
|
|
found = true
|
|
|
|
}
|
|
|
|
}
|
|
|
|
if !found {
|
|
|
|
t.Errorf("%d.%d. Couldn't find expected sample in output: '%v'", i, j, expectedSample)
|
|
|
|
failed = true
|
|
|
|
}
|
|
|
|
}
|
|
|
|
|
|
|
|
if failed {
|
|
|
|
t.Fatalf("%d. Expected and actual outputs don't match:\n%v", i, vectorComparisonString(expectedLines, actualLines))
|
|
|
|
}
|
|
|
|
}
|
|
|
|
}
|