2016-04-13 07:08:22 -07:00
|
|
|
// Copyright 2016 The Prometheus Authors
|
|
|
|
// Licensed under the Apache License, Version 2.0 (the "License");
|
|
|
|
// you may not use this file except in compliance with the License.
|
|
|
|
// You may obtain a copy of the License at
|
|
|
|
//
|
|
|
|
// http://www.apache.org/licenses/LICENSE-2.0
|
|
|
|
//
|
|
|
|
// Unless required by applicable law or agreed to in writing, software
|
|
|
|
// distributed under the License is distributed on an "AS IS" BASIS,
|
|
|
|
// WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
|
|
|
|
// See the License for the specific language governing permissions and
|
|
|
|
// limitations under the License.
|
|
|
|
|
2015-03-30 10:13:36 -07:00
|
|
|
package promql
|
|
|
|
|
|
|
|
import (
|
2017-10-04 12:04:15 -07:00
|
|
|
"context"
|
2017-04-18 06:51:10 -07:00
|
|
|
"reflect"
|
2015-03-30 10:13:36 -07:00
|
|
|
"testing"
|
|
|
|
"time"
|
|
|
|
|
2017-08-11 11:45:52 -07:00
|
|
|
"github.com/go-kit/kit/log"
|
2019-03-25 16:01:12 -07:00
|
|
|
"github.com/pkg/errors"
|
|
|
|
|
2017-04-18 06:51:10 -07:00
|
|
|
"github.com/prometheus/prometheus/pkg/labels"
|
2018-02-15 04:08:00 -08:00
|
|
|
"github.com/prometheus/prometheus/storage"
|
2018-09-13 02:57:36 -07:00
|
|
|
"github.com/prometheus/prometheus/util/testutil"
|
2015-03-30 10:13:36 -07:00
|
|
|
)
|
|
|
|
|
2015-09-23 02:06:51 -07:00
|
|
|
func TestQueryConcurrency(t *testing.T) {
|
2018-10-02 04:59:19 -07:00
|
|
|
opts := EngineOpts{
|
|
|
|
Logger: nil,
|
|
|
|
Reg: nil,
|
|
|
|
MaxConcurrent: 10,
|
|
|
|
MaxSamples: 10,
|
|
|
|
Timeout: 100 * time.Second,
|
|
|
|
}
|
2018-01-09 08:44:23 -08:00
|
|
|
|
2018-10-02 04:59:19 -07:00
|
|
|
engine := NewEngine(opts)
|
2016-09-15 15:58:06 -07:00
|
|
|
ctx, cancelCtx := context.WithCancel(context.Background())
|
|
|
|
defer cancelCtx()
|
2015-04-30 15:49:19 -07:00
|
|
|
|
|
|
|
block := make(chan struct{})
|
|
|
|
processing := make(chan struct{})
|
2015-08-10 05:21:24 -07:00
|
|
|
|
|
|
|
f := func(context.Context) error {
|
2015-04-30 15:49:19 -07:00
|
|
|
processing <- struct{}{}
|
|
|
|
<-block
|
|
|
|
return nil
|
2015-08-10 05:21:24 -07:00
|
|
|
}
|
2015-04-30 15:49:19 -07:00
|
|
|
|
2018-10-02 04:59:19 -07:00
|
|
|
for i := 0; i < opts.MaxConcurrent; i++ {
|
2015-08-10 05:21:24 -07:00
|
|
|
q := engine.newTestQuery(f)
|
promql: Allow per-query contexts.
For Weaveworks' Frankenstein, we need to support multitenancy. In
Frankenstein, we initially solved this without modifying the promql
package at all: we constructed a new promql.Engine for every
query and injected a storage implementation into that engine which would
be primed to only collect data for a given user.
This is problematic to upstream, however. Prometheus assumes that there
is only one engine: the query concurrency gate is part of the engine,
and the engine contains one central cancellable context to shut down all
queries. Also, creating a new engine for every query seems like overkill.
Thus, we want to be able to pass per-query contexts into a single engine.
This change gets rid of the promql.Engine's built-in base context and
allows passing in a per-query context instead. Central cancellation of
all queries is still possible by deriving all passed-in contexts from
one central one, but this is now the responsibility of the caller. The
central query context is now created in main() and passed into the
relevant components (web handler / API, rule manager).
In a next step, the per-query context would have to be passed to the
storage implementation, so that the storage can implement multi-tenancy
or other features based on the contextual information.
2016-09-15 04:52:50 -07:00
|
|
|
go q.Exec(ctx)
|
2015-04-30 15:49:19 -07:00
|
|
|
select {
|
|
|
|
case <-processing:
|
|
|
|
// Expected.
|
2015-09-23 02:06:51 -07:00
|
|
|
case <-time.After(20 * time.Millisecond):
|
2015-04-30 15:49:19 -07:00
|
|
|
t.Fatalf("Query within concurrency threshold not being executed")
|
|
|
|
}
|
|
|
|
}
|
|
|
|
|
2015-08-10 05:21:24 -07:00
|
|
|
q := engine.newTestQuery(f)
|
promql: Allow per-query contexts.
For Weaveworks' Frankenstein, we need to support multitenancy. In
Frankenstein, we initially solved this without modifying the promql
package at all: we constructed a new promql.Engine for every
query and injected a storage implementation into that engine which would
be primed to only collect data for a given user.
This is problematic to upstream, however. Prometheus assumes that there
is only one engine: the query concurrency gate is part of the engine,
and the engine contains one central cancellable context to shut down all
queries. Also, creating a new engine for every query seems like overkill.
Thus, we want to be able to pass per-query contexts into a single engine.
This change gets rid of the promql.Engine's built-in base context and
allows passing in a per-query context instead. Central cancellation of
all queries is still possible by deriving all passed-in contexts from
one central one, but this is now the responsibility of the caller. The
central query context is now created in main() and passed into the
relevant components (web handler / API, rule manager).
In a next step, the per-query context would have to be passed to the
storage implementation, so that the storage can implement multi-tenancy
or other features based on the contextual information.
2016-09-15 04:52:50 -07:00
|
|
|
go q.Exec(ctx)
|
2015-04-30 15:49:19 -07:00
|
|
|
|
|
|
|
select {
|
|
|
|
case <-processing:
|
2016-07-11 11:27:25 -07:00
|
|
|
t.Fatalf("Query above concurrency threshold being executed")
|
2015-09-23 02:06:51 -07:00
|
|
|
case <-time.After(20 * time.Millisecond):
|
2015-04-30 15:49:19 -07:00
|
|
|
// Expected.
|
|
|
|
}
|
|
|
|
|
|
|
|
// Terminate a running query.
|
|
|
|
block <- struct{}{}
|
|
|
|
|
|
|
|
select {
|
|
|
|
case <-processing:
|
|
|
|
// Expected.
|
2015-09-23 02:06:51 -07:00
|
|
|
case <-time.After(20 * time.Millisecond):
|
2015-04-30 15:49:19 -07:00
|
|
|
t.Fatalf("Query within concurrency threshold not being executed")
|
|
|
|
}
|
|
|
|
|
|
|
|
// Terminate remaining queries.
|
2018-10-02 04:59:19 -07:00
|
|
|
for i := 0; i < opts.MaxConcurrent; i++ {
|
2015-04-30 15:49:19 -07:00
|
|
|
block <- struct{}{}
|
|
|
|
}
|
|
|
|
}
|
|
|
|
|
2015-03-30 10:13:36 -07:00
|
|
|
func TestQueryTimeout(t *testing.T) {
|
2018-10-02 04:59:19 -07:00
|
|
|
opts := EngineOpts{
|
|
|
|
Logger: nil,
|
|
|
|
Reg: nil,
|
|
|
|
MaxConcurrent: 20,
|
|
|
|
MaxSamples: 10,
|
|
|
|
Timeout: 5 * time.Millisecond,
|
|
|
|
}
|
|
|
|
engine := NewEngine(opts)
|
2016-09-15 15:58:06 -07:00
|
|
|
ctx, cancelCtx := context.WithCancel(context.Background())
|
|
|
|
defer cancelCtx()
|
2015-03-30 10:13:36 -07:00
|
|
|
|
2015-08-10 05:21:24 -07:00
|
|
|
query := engine.newTestQuery(func(ctx context.Context) error {
|
2015-09-14 03:28:27 -07:00
|
|
|
time.Sleep(50 * time.Millisecond)
|
2015-08-10 05:21:24 -07:00
|
|
|
return contextDone(ctx, "test statement execution")
|
2015-03-30 10:13:36 -07:00
|
|
|
})
|
|
|
|
|
promql: Allow per-query contexts.
For Weaveworks' Frankenstein, we need to support multitenancy. In
Frankenstein, we initially solved this without modifying the promql
package at all: we constructed a new promql.Engine for every
query and injected a storage implementation into that engine which would
be primed to only collect data for a given user.
This is problematic to upstream, however. Prometheus assumes that there
is only one engine: the query concurrency gate is part of the engine,
and the engine contains one central cancellable context to shut down all
queries. Also, creating a new engine for every query seems like overkill.
Thus, we want to be able to pass per-query contexts into a single engine.
This change gets rid of the promql.Engine's built-in base context and
allows passing in a per-query context instead. Central cancellation of
all queries is still possible by deriving all passed-in contexts from
one central one, but this is now the responsibility of the caller. The
central query context is now created in main() and passed into the
relevant components (web handler / API, rule manager).
In a next step, the per-query context would have to be passed to the
storage implementation, so that the storage can implement multi-tenancy
or other features based on the contextual information.
2016-09-15 04:52:50 -07:00
|
|
|
res := query.Exec(ctx)
|
2015-03-30 10:13:36 -07:00
|
|
|
if res.Err == nil {
|
|
|
|
t.Fatalf("expected timeout error but got none")
|
|
|
|
}
|
|
|
|
if _, ok := res.Err.(ErrQueryTimeout); res.Err != nil && !ok {
|
|
|
|
t.Fatalf("expected timeout error but got: %s", res.Err)
|
|
|
|
}
|
|
|
|
}
|
|
|
|
|
|
|
|
func TestQueryCancel(t *testing.T) {
|
2018-10-02 04:59:19 -07:00
|
|
|
opts := EngineOpts{
|
|
|
|
Logger: nil,
|
|
|
|
Reg: nil,
|
|
|
|
MaxConcurrent: 10,
|
|
|
|
MaxSamples: 10,
|
|
|
|
Timeout: 10 * time.Second,
|
|
|
|
}
|
|
|
|
engine := NewEngine(opts)
|
2016-09-15 15:58:06 -07:00
|
|
|
ctx, cancelCtx := context.WithCancel(context.Background())
|
|
|
|
defer cancelCtx()
|
2015-03-30 10:13:36 -07:00
|
|
|
|
2015-08-10 05:21:24 -07:00
|
|
|
// Cancel a running query before it completes.
|
|
|
|
block := make(chan struct{})
|
|
|
|
processing := make(chan struct{})
|
2015-03-30 10:13:36 -07:00
|
|
|
|
2015-08-10 05:21:24 -07:00
|
|
|
query1 := engine.newTestQuery(func(ctx context.Context) error {
|
|
|
|
processing <- struct{}{}
|
|
|
|
<-block
|
|
|
|
return contextDone(ctx, "test statement execution")
|
|
|
|
})
|
2015-04-29 02:08:56 -07:00
|
|
|
|
2015-03-30 10:13:36 -07:00
|
|
|
var res *Result
|
|
|
|
|
|
|
|
go func() {
|
promql: Allow per-query contexts.
For Weaveworks' Frankenstein, we need to support multitenancy. In
Frankenstein, we initially solved this without modifying the promql
package at all: we constructed a new promql.Engine for every
query and injected a storage implementation into that engine which would
be primed to only collect data for a given user.
This is problematic to upstream, however. Prometheus assumes that there
is only one engine: the query concurrency gate is part of the engine,
and the engine contains one central cancellable context to shut down all
queries. Also, creating a new engine for every query seems like overkill.
Thus, we want to be able to pass per-query contexts into a single engine.
This change gets rid of the promql.Engine's built-in base context and
allows passing in a per-query context instead. Central cancellation of
all queries is still possible by deriving all passed-in contexts from
one central one, but this is now the responsibility of the caller. The
central query context is now created in main() and passed into the
relevant components (web handler / API, rule manager).
In a next step, the per-query context would have to be passed to the
storage implementation, so that the storage can implement multi-tenancy
or other features based on the contextual information.
2016-09-15 04:52:50 -07:00
|
|
|
res = query1.Exec(ctx)
|
2015-08-10 05:21:24 -07:00
|
|
|
processing <- struct{}{}
|
2015-03-30 10:13:36 -07:00
|
|
|
}()
|
2015-08-10 05:21:24 -07:00
|
|
|
|
|
|
|
<-processing
|
2015-03-30 10:13:36 -07:00
|
|
|
query1.Cancel()
|
2015-08-10 05:21:24 -07:00
|
|
|
block <- struct{}{}
|
|
|
|
<-processing
|
2015-03-30 10:13:36 -07:00
|
|
|
|
|
|
|
if res.Err == nil {
|
|
|
|
t.Fatalf("expected cancellation error for query1 but got none")
|
|
|
|
}
|
2015-08-10 05:21:24 -07:00
|
|
|
if ee := ErrQueryCanceled("test statement execution"); res.Err != ee {
|
2015-08-25 17:04:01 -07:00
|
|
|
t.Fatalf("expected error %q, got %q", ee, res.Err)
|
2015-03-30 10:13:36 -07:00
|
|
|
}
|
|
|
|
|
2015-08-10 05:21:24 -07:00
|
|
|
// Canceling a query before starting it must have no effect.
|
|
|
|
query2 := engine.newTestQuery(func(ctx context.Context) error {
|
|
|
|
return contextDone(ctx, "test statement execution")
|
|
|
|
})
|
|
|
|
|
2015-03-30 10:13:36 -07:00
|
|
|
query2.Cancel()
|
promql: Allow per-query contexts.
For Weaveworks' Frankenstein, we need to support multitenancy. In
Frankenstein, we initially solved this without modifying the promql
package at all: we constructed a new promql.Engine for every
query and injected a storage implementation into that engine which would
be primed to only collect data for a given user.
This is problematic to upstream, however. Prometheus assumes that there
is only one engine: the query concurrency gate is part of the engine,
and the engine contains one central cancellable context to shut down all
queries. Also, creating a new engine for every query seems like overkill.
Thus, we want to be able to pass per-query contexts into a single engine.
This change gets rid of the promql.Engine's built-in base context and
allows passing in a per-query context instead. Central cancellation of
all queries is still possible by deriving all passed-in contexts from
one central one, but this is now the responsibility of the caller. The
central query context is now created in main() and passed into the
relevant components (web handler / API, rule manager).
In a next step, the per-query context would have to be passed to the
storage implementation, so that the storage can implement multi-tenancy
or other features based on the contextual information.
2016-09-15 04:52:50 -07:00
|
|
|
res = query2.Exec(ctx)
|
2015-03-30 10:13:36 -07:00
|
|
|
if res.Err != nil {
|
promql: Allow per-query contexts.
For Weaveworks' Frankenstein, we need to support multitenancy. In
Frankenstein, we initially solved this without modifying the promql
package at all: we constructed a new promql.Engine for every
query and injected a storage implementation into that engine which would
be primed to only collect data for a given user.
This is problematic to upstream, however. Prometheus assumes that there
is only one engine: the query concurrency gate is part of the engine,
and the engine contains one central cancellable context to shut down all
queries. Also, creating a new engine for every query seems like overkill.
Thus, we want to be able to pass per-query contexts into a single engine.
This change gets rid of the promql.Engine's built-in base context and
allows passing in a per-query context instead. Central cancellation of
all queries is still possible by deriving all passed-in contexts from
one central one, but this is now the responsibility of the caller. The
central query context is now created in main() and passed into the
relevant components (web handler / API, rule manager).
In a next step, the per-query context would have to be passed to the
storage implementation, so that the storage can implement multi-tenancy
or other features based on the contextual information.
2016-09-15 04:52:50 -07:00
|
|
|
t.Fatalf("unexpected error on executing query2: %s", res.Err)
|
2015-03-30 10:13:36 -07:00
|
|
|
}
|
|
|
|
}
|
|
|
|
|
2018-02-15 04:08:00 -08:00
|
|
|
// errQuerier implements storage.Querier which always returns error.
|
|
|
|
type errQuerier struct {
|
|
|
|
err error
|
|
|
|
}
|
|
|
|
|
2019-01-02 03:10:13 -08:00
|
|
|
func (q *errQuerier) Select(*storage.SelectParams, ...*labels.Matcher) (storage.SeriesSet, storage.Warnings, error) {
|
|
|
|
return errSeriesSet{err: q.err}, nil, q.err
|
2018-02-15 04:08:00 -08:00
|
|
|
}
|
|
|
|
func (*errQuerier) LabelValues(name string) ([]string, error) { return nil, nil }
|
2018-11-19 02:21:14 -08:00
|
|
|
func (*errQuerier) LabelNames() ([]string, error) { return nil, nil }
|
2018-02-15 04:08:00 -08:00
|
|
|
func (*errQuerier) Close() error { return nil }
|
|
|
|
|
|
|
|
// errSeriesSet implements storage.SeriesSet which always returns error.
|
|
|
|
type errSeriesSet struct {
|
|
|
|
err error
|
|
|
|
}
|
|
|
|
|
|
|
|
func (errSeriesSet) Next() bool { return false }
|
|
|
|
func (errSeriesSet) At() storage.Series { return nil }
|
|
|
|
func (e errSeriesSet) Err() error { return e.err }
|
|
|
|
|
|
|
|
func TestQueryError(t *testing.T) {
|
2018-10-02 04:59:19 -07:00
|
|
|
opts := EngineOpts{
|
|
|
|
Logger: nil,
|
|
|
|
Reg: nil,
|
|
|
|
MaxConcurrent: 10,
|
|
|
|
MaxSamples: 10,
|
|
|
|
Timeout: 10 * time.Second,
|
|
|
|
}
|
|
|
|
engine := NewEngine(opts)
|
2019-03-25 16:01:12 -07:00
|
|
|
errStorage := ErrStorage{errors.New("storage error")}
|
2018-02-15 04:08:00 -08:00
|
|
|
queryable := storage.QueryableFunc(func(ctx context.Context, mint, maxt int64) (storage.Querier, error) {
|
|
|
|
return &errQuerier{err: errStorage}, nil
|
|
|
|
})
|
|
|
|
ctx, cancelCtx := context.WithCancel(context.Background())
|
|
|
|
defer cancelCtx()
|
|
|
|
|
|
|
|
vectorQuery, err := engine.NewInstantQuery(queryable, "foo", time.Unix(1, 0))
|
|
|
|
if err != nil {
|
|
|
|
t.Fatalf("unexpected error creating query: %q", err)
|
|
|
|
}
|
|
|
|
res := vectorQuery.Exec(ctx)
|
|
|
|
if res.Err == nil {
|
|
|
|
t.Fatalf("expected error on failed select but got none")
|
|
|
|
}
|
|
|
|
if res.Err != errStorage {
|
|
|
|
t.Fatalf("expected error %q, got %q", errStorage, res.Err)
|
|
|
|
}
|
|
|
|
|
|
|
|
matrixQuery, err := engine.NewInstantQuery(queryable, "foo[1m]", time.Unix(1, 0))
|
|
|
|
if err != nil {
|
|
|
|
t.Fatalf("unexpected error creating query: %q", err)
|
|
|
|
}
|
|
|
|
res = matrixQuery.Exec(ctx)
|
|
|
|
if res.Err == nil {
|
|
|
|
t.Fatalf("expected error on failed select but got none")
|
|
|
|
}
|
|
|
|
if res.Err != errStorage {
|
|
|
|
t.Fatalf("expected error %q, got %q", errStorage, res.Err)
|
|
|
|
}
|
|
|
|
}
|
|
|
|
|
2019-04-17 05:52:41 -07:00
|
|
|
// paramCheckerQuerier implements storage.Querier which checks the start and end times
|
|
|
|
// in params.
|
|
|
|
type paramCheckerQuerier struct {
|
|
|
|
start int64
|
|
|
|
end int64
|
|
|
|
|
|
|
|
t *testing.T
|
|
|
|
}
|
|
|
|
|
|
|
|
func (q *paramCheckerQuerier) Select(sp *storage.SelectParams, _ ...*labels.Matcher) (storage.SeriesSet, storage.Warnings, error) {
|
|
|
|
testutil.Equals(q.t, q.start, sp.Start)
|
|
|
|
testutil.Equals(q.t, q.end, sp.End)
|
|
|
|
|
|
|
|
return errSeriesSet{err: nil}, nil, nil
|
|
|
|
}
|
|
|
|
func (*paramCheckerQuerier) LabelValues(name string) ([]string, error) { return nil, nil }
|
|
|
|
func (*paramCheckerQuerier) LabelNames() ([]string, error) { return nil, nil }
|
|
|
|
func (*paramCheckerQuerier) Close() error { return nil }
|
|
|
|
|
|
|
|
func TestParamsSetCorrectly(t *testing.T) {
|
|
|
|
opts := EngineOpts{
|
|
|
|
Logger: nil,
|
|
|
|
Reg: nil,
|
|
|
|
MaxConcurrent: 10,
|
|
|
|
MaxSamples: 10,
|
|
|
|
Timeout: 10 * time.Second,
|
|
|
|
}
|
|
|
|
|
|
|
|
// Set the lookback to be smaller and reset at the end.
|
|
|
|
currLookback := LookbackDelta
|
|
|
|
LookbackDelta = 5 * time.Second
|
|
|
|
defer func() {
|
|
|
|
LookbackDelta = currLookback
|
|
|
|
}()
|
|
|
|
|
|
|
|
cases := []struct {
|
|
|
|
query string
|
|
|
|
|
|
|
|
// All times are in seconds.
|
|
|
|
start int64
|
|
|
|
end int64
|
|
|
|
|
|
|
|
paramStart int64
|
|
|
|
paramEnd int64
|
|
|
|
}{{
|
|
|
|
query: "foo",
|
|
|
|
start: 10,
|
|
|
|
|
|
|
|
paramStart: 5,
|
|
|
|
paramEnd: 10,
|
|
|
|
}, {
|
|
|
|
query: "foo[2m]",
|
|
|
|
start: 200,
|
|
|
|
|
|
|
|
paramStart: 80, // 200 - 120
|
|
|
|
paramEnd: 200,
|
|
|
|
}, {
|
|
|
|
query: "foo[2m] offset 2m",
|
|
|
|
start: 300,
|
|
|
|
|
|
|
|
paramStart: 60,
|
|
|
|
paramEnd: 180,
|
|
|
|
}, {
|
|
|
|
query: "foo[2m:1s]",
|
|
|
|
start: 300,
|
|
|
|
|
|
|
|
paramStart: 175, // 300 - 120 - 5
|
|
|
|
paramEnd: 300,
|
|
|
|
}, {
|
|
|
|
query: "count_over_time(foo[2m:1s])",
|
|
|
|
start: 300,
|
|
|
|
|
|
|
|
paramStart: 175, // 300 - 120 - 5
|
|
|
|
paramEnd: 300,
|
|
|
|
}, {
|
|
|
|
query: "count_over_time(foo[2m:1s] offset 10s)",
|
|
|
|
start: 300,
|
|
|
|
|
|
|
|
paramStart: 165, // 300 - 120 - 5 - 10
|
|
|
|
paramEnd: 300,
|
|
|
|
}, {
|
|
|
|
query: "count_over_time((foo offset 10s)[2m:1s] offset 10s)",
|
|
|
|
start: 300,
|
|
|
|
|
|
|
|
paramStart: 155, // 300 - 120 - 5 - 10 - 10
|
|
|
|
paramEnd: 290,
|
|
|
|
}, {
|
|
|
|
// Range queries now.
|
|
|
|
query: "foo",
|
|
|
|
start: 10,
|
|
|
|
end: 20,
|
|
|
|
|
|
|
|
paramStart: 5,
|
|
|
|
paramEnd: 20,
|
|
|
|
}, {
|
|
|
|
query: "rate(foo[2m])",
|
|
|
|
start: 200,
|
|
|
|
end: 500,
|
|
|
|
|
|
|
|
paramStart: 80, // 200 - 120
|
|
|
|
paramEnd: 500,
|
|
|
|
}, {
|
|
|
|
query: "rate(foo[2m] offset 2m)",
|
|
|
|
start: 300,
|
|
|
|
end: 500,
|
|
|
|
|
|
|
|
paramStart: 60,
|
|
|
|
paramEnd: 380,
|
|
|
|
}, {
|
|
|
|
query: "rate(foo[2m:1s])",
|
|
|
|
start: 300,
|
|
|
|
end: 500,
|
|
|
|
|
|
|
|
paramStart: 175, // 300 - 120 - 5
|
|
|
|
paramEnd: 500,
|
|
|
|
}, {
|
|
|
|
query: "count_over_time(foo[2m:1s])",
|
|
|
|
start: 300,
|
|
|
|
end: 500,
|
|
|
|
|
|
|
|
paramStart: 175, // 300 - 120 - 5
|
|
|
|
paramEnd: 500,
|
|
|
|
}, {
|
|
|
|
query: "count_over_time(foo[2m:1s] offset 10s)",
|
|
|
|
start: 300,
|
|
|
|
end: 500,
|
|
|
|
|
|
|
|
paramStart: 165, // 300 - 120 - 5 - 10
|
|
|
|
paramEnd: 500,
|
|
|
|
}, {
|
|
|
|
query: "count_over_time((foo offset 10s)[2m:1s] offset 10s)",
|
|
|
|
start: 300,
|
|
|
|
end: 500,
|
|
|
|
|
|
|
|
paramStart: 155, // 300 - 120 - 5 - 10 - 10
|
|
|
|
paramEnd: 490,
|
|
|
|
}}
|
|
|
|
|
|
|
|
for _, tc := range cases {
|
|
|
|
engine := NewEngine(opts)
|
|
|
|
queryable := storage.QueryableFunc(func(ctx context.Context, mint, maxt int64) (storage.Querier, error) {
|
|
|
|
return ¶mCheckerQuerier{start: tc.paramStart * 1000, end: tc.paramEnd * 1000, t: t}, nil
|
|
|
|
})
|
|
|
|
|
|
|
|
var (
|
|
|
|
query Query
|
|
|
|
err error
|
|
|
|
)
|
|
|
|
if tc.end == 0 {
|
|
|
|
query, err = engine.NewInstantQuery(queryable, tc.query, time.Unix(tc.start, 0))
|
|
|
|
} else {
|
|
|
|
query, err = engine.NewRangeQuery(queryable, tc.query, time.Unix(tc.start, 0), time.Unix(tc.end, 0), time.Second)
|
|
|
|
}
|
|
|
|
testutil.Ok(t, err)
|
|
|
|
|
|
|
|
res := query.Exec(context.Background())
|
|
|
|
testutil.Ok(t, res.Err)
|
|
|
|
}
|
|
|
|
}
|
|
|
|
|
2015-03-30 10:13:36 -07:00
|
|
|
func TestEngineShutdown(t *testing.T) {
|
2018-10-02 04:59:19 -07:00
|
|
|
opts := EngineOpts{
|
|
|
|
Logger: nil,
|
|
|
|
Reg: nil,
|
|
|
|
MaxConcurrent: 10,
|
|
|
|
MaxSamples: 10,
|
|
|
|
Timeout: 10 * time.Second,
|
|
|
|
}
|
|
|
|
engine := NewEngine(opts)
|
2016-09-15 15:58:06 -07:00
|
|
|
ctx, cancelCtx := context.WithCancel(context.Background())
|
2015-03-30 10:13:36 -07:00
|
|
|
|
2015-08-10 05:21:24 -07:00
|
|
|
block := make(chan struct{})
|
|
|
|
processing := make(chan struct{})
|
|
|
|
|
2015-03-30 10:13:36 -07:00
|
|
|
// Shutdown engine on first handler execution. Should handler execution ever become
|
|
|
|
// concurrent this test has to be adjusted accordingly.
|
2015-08-10 05:21:24 -07:00
|
|
|
f := func(ctx context.Context) error {
|
|
|
|
processing <- struct{}{}
|
|
|
|
<-block
|
|
|
|
return contextDone(ctx, "test statement execution")
|
|
|
|
}
|
|
|
|
query1 := engine.newTestQuery(f)
|
2015-03-30 10:13:36 -07:00
|
|
|
|
2015-04-29 02:08:56 -07:00
|
|
|
// Stopping the engine must cancel the base context. While executing queries is
|
|
|
|
// still possible, their context is canceled from the beginning and execution should
|
2015-03-30 10:13:36 -07:00
|
|
|
// terminate immediately.
|
|
|
|
|
2015-08-10 05:21:24 -07:00
|
|
|
var res *Result
|
|
|
|
go func() {
|
promql: Allow per-query contexts.
For Weaveworks' Frankenstein, we need to support multitenancy. In
Frankenstein, we initially solved this without modifying the promql
package at all: we constructed a new promql.Engine for every
query and injected a storage implementation into that engine which would
be primed to only collect data for a given user.
This is problematic to upstream, however. Prometheus assumes that there
is only one engine: the query concurrency gate is part of the engine,
and the engine contains one central cancellable context to shut down all
queries. Also, creating a new engine for every query seems like overkill.
Thus, we want to be able to pass per-query contexts into a single engine.
This change gets rid of the promql.Engine's built-in base context and
allows passing in a per-query context instead. Central cancellation of
all queries is still possible by deriving all passed-in contexts from
one central one, but this is now the responsibility of the caller. The
central query context is now created in main() and passed into the
relevant components (web handler / API, rule manager).
In a next step, the per-query context would have to be passed to the
storage implementation, so that the storage can implement multi-tenancy
or other features based on the contextual information.
2016-09-15 04:52:50 -07:00
|
|
|
res = query1.Exec(ctx)
|
2015-08-10 05:21:24 -07:00
|
|
|
processing <- struct{}{}
|
|
|
|
}()
|
|
|
|
|
|
|
|
<-processing
|
2016-09-15 15:58:06 -07:00
|
|
|
cancelCtx()
|
2015-08-10 05:21:24 -07:00
|
|
|
block <- struct{}{}
|
|
|
|
<-processing
|
|
|
|
|
2015-03-30 10:13:36 -07:00
|
|
|
if res.Err == nil {
|
|
|
|
t.Fatalf("expected error on shutdown during query but got none")
|
|
|
|
}
|
2015-08-10 05:21:24 -07:00
|
|
|
if ee := ErrQueryCanceled("test statement execution"); res.Err != ee {
|
|
|
|
t.Fatalf("expected error %q, got %q", ee, res.Err)
|
2015-03-30 10:13:36 -07:00
|
|
|
}
|
|
|
|
|
2015-08-10 05:21:24 -07:00
|
|
|
query2 := engine.newTestQuery(func(context.Context) error {
|
|
|
|
t.Fatalf("reached query execution unexpectedly")
|
|
|
|
return nil
|
|
|
|
})
|
|
|
|
|
|
|
|
// The second query is started after the engine shut down. It must
|
|
|
|
// be canceled immediately.
|
promql: Allow per-query contexts.
For Weaveworks' Frankenstein, we need to support multitenancy. In
Frankenstein, we initially solved this without modifying the promql
package at all: we constructed a new promql.Engine for every
query and injected a storage implementation into that engine which would
be primed to only collect data for a given user.
This is problematic to upstream, however. Prometheus assumes that there
is only one engine: the query concurrency gate is part of the engine,
and the engine contains one central cancellable context to shut down all
queries. Also, creating a new engine for every query seems like overkill.
Thus, we want to be able to pass per-query contexts into a single engine.
This change gets rid of the promql.Engine's built-in base context and
allows passing in a per-query context instead. Central cancellation of
all queries is still possible by deriving all passed-in contexts from
one central one, but this is now the responsibility of the caller. The
central query context is now created in main() and passed into the
relevant components (web handler / API, rule manager).
In a next step, the per-query context would have to be passed to the
storage implementation, so that the storage can implement multi-tenancy
or other features based on the contextual information.
2016-09-15 04:52:50 -07:00
|
|
|
res2 := query2.Exec(ctx)
|
2015-03-30 10:13:36 -07:00
|
|
|
if res2.Err == nil {
|
promql: Allow per-query contexts.
For Weaveworks' Frankenstein, we need to support multitenancy. In
Frankenstein, we initially solved this without modifying the promql
package at all: we constructed a new promql.Engine for every
query and injected a storage implementation into that engine which would
be primed to only collect data for a given user.
This is problematic to upstream, however. Prometheus assumes that there
is only one engine: the query concurrency gate is part of the engine,
and the engine contains one central cancellable context to shut down all
queries. Also, creating a new engine for every query seems like overkill.
Thus, we want to be able to pass per-query contexts into a single engine.
This change gets rid of the promql.Engine's built-in base context and
allows passing in a per-query context instead. Central cancellation of
all queries is still possible by deriving all passed-in contexts from
one central one, but this is now the responsibility of the caller. The
central query context is now created in main() and passed into the
relevant components (web handler / API, rule manager).
In a next step, the per-query context would have to be passed to the
storage implementation, so that the storage can implement multi-tenancy
or other features based on the contextual information.
2016-09-15 04:52:50 -07:00
|
|
|
t.Fatalf("expected error on querying with canceled context but got none")
|
2015-03-30 10:13:36 -07:00
|
|
|
}
|
2015-08-10 05:21:24 -07:00
|
|
|
if _, ok := res2.Err.(ErrQueryCanceled); !ok {
|
2017-07-18 05:58:00 -07:00
|
|
|
t.Fatalf("expected cancellation error, got %q", res2.Err)
|
2015-03-30 10:13:36 -07:00
|
|
|
}
|
|
|
|
}
|
2015-08-19 06:28:53 -07:00
|
|
|
|
2017-04-18 06:51:10 -07:00
|
|
|
func TestEngineEvalStmtTimestamps(t *testing.T) {
|
|
|
|
test, err := NewTest(t, `
|
|
|
|
load 10s
|
|
|
|
metric 1 2
|
|
|
|
`)
|
|
|
|
if err != nil {
|
|
|
|
t.Fatalf("unexpected error creating test: %q", err)
|
|
|
|
}
|
2018-03-08 08:48:11 -08:00
|
|
|
defer test.Close()
|
|
|
|
|
2017-04-18 06:51:10 -07:00
|
|
|
err = test.Run()
|
|
|
|
if err != nil {
|
|
|
|
t.Fatalf("unexpected error initializing test: %q", err)
|
|
|
|
}
|
|
|
|
|
|
|
|
cases := []struct {
|
2018-09-13 02:57:36 -07:00
|
|
|
Query string
|
|
|
|
Result Value
|
|
|
|
Start time.Time
|
|
|
|
End time.Time
|
|
|
|
Interval time.Duration
|
|
|
|
ShouldError bool
|
2017-04-18 06:51:10 -07:00
|
|
|
}{
|
|
|
|
// Instant queries.
|
|
|
|
{
|
|
|
|
Query: "1",
|
|
|
|
Result: Scalar{V: 1, T: 1000},
|
|
|
|
Start: time.Unix(1, 0),
|
|
|
|
},
|
|
|
|
{
|
|
|
|
Query: "metric",
|
|
|
|
Result: Vector{
|
|
|
|
Sample{Point: Point{V: 1, T: 1000},
|
|
|
|
Metric: labels.FromStrings("__name__", "metric")},
|
|
|
|
},
|
|
|
|
Start: time.Unix(1, 0),
|
|
|
|
},
|
|
|
|
{
|
|
|
|
Query: "metric[20s]",
|
|
|
|
Result: Matrix{Series{
|
|
|
|
Points: []Point{{V: 1, T: 0}, {V: 2, T: 10000}},
|
|
|
|
Metric: labels.FromStrings("__name__", "metric")},
|
|
|
|
},
|
|
|
|
Start: time.Unix(10, 0),
|
|
|
|
},
|
|
|
|
// Range queries.
|
|
|
|
{
|
|
|
|
Query: "1",
|
|
|
|
Result: Matrix{Series{
|
|
|
|
Points: []Point{{V: 1, T: 0}, {V: 1, T: 1000}, {V: 1, T: 2000}},
|
|
|
|
Metric: labels.FromStrings()},
|
|
|
|
},
|
|
|
|
Start: time.Unix(0, 0),
|
|
|
|
End: time.Unix(2, 0),
|
|
|
|
Interval: time.Second,
|
|
|
|
},
|
|
|
|
{
|
|
|
|
Query: "metric",
|
|
|
|
Result: Matrix{Series{
|
2017-06-12 22:22:27 -07:00
|
|
|
Points: []Point{{V: 1, T: 0}, {V: 1, T: 1000}, {V: 1, T: 2000}},
|
2017-04-18 06:51:10 -07:00
|
|
|
Metric: labels.FromStrings("__name__", "metric")},
|
|
|
|
},
|
|
|
|
Start: time.Unix(0, 0),
|
|
|
|
End: time.Unix(2, 0),
|
|
|
|
Interval: time.Second,
|
|
|
|
},
|
2017-06-12 22:22:27 -07:00
|
|
|
{
|
|
|
|
Query: "metric",
|
|
|
|
Result: Matrix{Series{
|
|
|
|
Points: []Point{{V: 1, T: 0}, {V: 1, T: 5000}, {V: 2, T: 10000}},
|
|
|
|
Metric: labels.FromStrings("__name__", "metric")},
|
|
|
|
},
|
|
|
|
Start: time.Unix(0, 0),
|
|
|
|
End: time.Unix(10, 0),
|
|
|
|
Interval: 5 * time.Second,
|
|
|
|
},
|
2018-09-13 02:57:36 -07:00
|
|
|
{
|
|
|
|
Query: `count_values("wrong label!", metric)`,
|
|
|
|
ShouldError: true,
|
|
|
|
},
|
2017-04-18 06:51:10 -07:00
|
|
|
}
|
|
|
|
|
|
|
|
for _, c := range cases {
|
|
|
|
var err error
|
|
|
|
var qry Query
|
|
|
|
if c.Interval == 0 {
|
2018-01-09 08:44:23 -08:00
|
|
|
qry, err = test.QueryEngine().NewInstantQuery(test.Queryable(), c.Query, c.Start)
|
2017-04-18 06:51:10 -07:00
|
|
|
} else {
|
2018-01-09 08:44:23 -08:00
|
|
|
qry, err = test.QueryEngine().NewRangeQuery(test.Queryable(), c.Query, c.Start, c.End, c.Interval)
|
2017-04-18 06:51:10 -07:00
|
|
|
}
|
|
|
|
if err != nil {
|
|
|
|
t.Fatalf("unexpected error creating query: %q", err)
|
|
|
|
}
|
|
|
|
res := qry.Exec(test.Context())
|
2018-09-13 02:57:36 -07:00
|
|
|
if c.ShouldError {
|
|
|
|
testutil.NotOk(t, res.Err, "expected error for the query %q", c.Query)
|
|
|
|
continue
|
|
|
|
}
|
2017-04-18 06:51:10 -07:00
|
|
|
if res.Err != nil {
|
|
|
|
t.Fatalf("unexpected error running query: %q", res.Err)
|
|
|
|
}
|
|
|
|
if !reflect.DeepEqual(res.Value, c.Result) {
|
|
|
|
t.Fatalf("unexpected result for query %q: got %q wanted %q", c.Query, res.Value.String(), c.Result.String())
|
|
|
|
}
|
|
|
|
}
|
|
|
|
|
|
|
|
}
|
|
|
|
|
2018-10-02 04:59:19 -07:00
|
|
|
func TestMaxQuerySamples(t *testing.T) {
|
|
|
|
test, err := NewTest(t, `
|
|
|
|
load 10s
|
|
|
|
metric 1 2
|
|
|
|
`)
|
|
|
|
|
|
|
|
if err != nil {
|
|
|
|
t.Fatalf("unexpected error creating test: %q", err)
|
|
|
|
}
|
|
|
|
defer test.Close()
|
|
|
|
|
|
|
|
err = test.Run()
|
|
|
|
if err != nil {
|
|
|
|
t.Fatalf("unexpected error initializing test: %q", err)
|
|
|
|
}
|
|
|
|
|
|
|
|
cases := []struct {
|
|
|
|
Query string
|
|
|
|
MaxSamples int
|
|
|
|
Result Result
|
|
|
|
Start time.Time
|
|
|
|
End time.Time
|
|
|
|
Interval time.Duration
|
|
|
|
}{
|
|
|
|
// Instant queries.
|
|
|
|
{
|
|
|
|
Query: "1",
|
|
|
|
MaxSamples: 1,
|
|
|
|
Result: Result{
|
|
|
|
nil,
|
2018-11-30 06:27:12 -08:00
|
|
|
Scalar{V: 1, T: 1000},
|
|
|
|
nil},
|
2018-10-02 04:59:19 -07:00
|
|
|
Start: time.Unix(1, 0),
|
|
|
|
},
|
|
|
|
{
|
|
|
|
Query: "1",
|
|
|
|
MaxSamples: 0,
|
|
|
|
Result: Result{
|
|
|
|
ErrTooManySamples(env),
|
|
|
|
nil,
|
2018-11-30 06:27:12 -08:00
|
|
|
nil,
|
2018-10-02 04:59:19 -07:00
|
|
|
},
|
|
|
|
Start: time.Unix(1, 0),
|
|
|
|
},
|
|
|
|
{
|
|
|
|
Query: "metric",
|
|
|
|
MaxSamples: 0,
|
|
|
|
Result: Result{
|
|
|
|
ErrTooManySamples(env),
|
|
|
|
nil,
|
2018-11-30 06:27:12 -08:00
|
|
|
nil,
|
2018-10-02 04:59:19 -07:00
|
|
|
},
|
|
|
|
Start: time.Unix(1, 0),
|
|
|
|
},
|
|
|
|
{
|
|
|
|
Query: "metric",
|
|
|
|
MaxSamples: 1,
|
|
|
|
Result: Result{
|
|
|
|
nil,
|
|
|
|
Vector{
|
|
|
|
Sample{Point: Point{V: 1, T: 1000},
|
|
|
|
Metric: labels.FromStrings("__name__", "metric")},
|
|
|
|
},
|
2018-11-30 06:27:12 -08:00
|
|
|
nil,
|
2018-10-02 04:59:19 -07:00
|
|
|
},
|
|
|
|
Start: time.Unix(1, 0),
|
|
|
|
},
|
|
|
|
{
|
|
|
|
Query: "metric[20s]",
|
|
|
|
MaxSamples: 2,
|
|
|
|
Result: Result{
|
|
|
|
nil,
|
|
|
|
Matrix{Series{
|
|
|
|
Points: []Point{{V: 1, T: 0}, {V: 2, T: 10000}},
|
|
|
|
Metric: labels.FromStrings("__name__", "metric")},
|
|
|
|
},
|
2018-11-30 06:27:12 -08:00
|
|
|
nil,
|
2018-10-02 04:59:19 -07:00
|
|
|
},
|
|
|
|
Start: time.Unix(10, 0),
|
|
|
|
},
|
2018-12-22 05:47:13 -08:00
|
|
|
{
|
|
|
|
Query: "rate(metric[20s])",
|
|
|
|
MaxSamples: 3,
|
|
|
|
Result: Result{
|
|
|
|
nil,
|
|
|
|
Vector{
|
|
|
|
Sample{
|
|
|
|
Point: Point{V: 0.1, T: 10000},
|
|
|
|
Metric: labels.Labels{},
|
|
|
|
},
|
|
|
|
},
|
|
|
|
nil,
|
|
|
|
},
|
|
|
|
Start: time.Unix(10, 0),
|
|
|
|
},
|
|
|
|
{
|
|
|
|
Query: "metric[20s:5s]",
|
|
|
|
MaxSamples: 3,
|
|
|
|
Result: Result{
|
|
|
|
nil,
|
|
|
|
Matrix{Series{
|
|
|
|
Points: []Point{{V: 1, T: 0}, {V: 1, T: 5000}, {V: 2, T: 10000}},
|
|
|
|
Metric: labels.FromStrings("__name__", "metric")},
|
|
|
|
},
|
|
|
|
nil,
|
|
|
|
},
|
|
|
|
Start: time.Unix(10, 0),
|
|
|
|
},
|
2018-10-02 04:59:19 -07:00
|
|
|
{
|
|
|
|
Query: "metric[20s]",
|
|
|
|
MaxSamples: 0,
|
|
|
|
Result: Result{
|
|
|
|
ErrTooManySamples(env),
|
|
|
|
nil,
|
2018-11-30 06:27:12 -08:00
|
|
|
nil,
|
2018-10-02 04:59:19 -07:00
|
|
|
},
|
|
|
|
Start: time.Unix(10, 0),
|
|
|
|
},
|
|
|
|
// Range queries.
|
|
|
|
{
|
|
|
|
Query: "1",
|
|
|
|
MaxSamples: 3,
|
|
|
|
Result: Result{
|
|
|
|
nil,
|
|
|
|
Matrix{Series{
|
|
|
|
Points: []Point{{V: 1, T: 0}, {V: 1, T: 1000}, {V: 1, T: 2000}},
|
|
|
|
Metric: labels.FromStrings()},
|
|
|
|
},
|
2018-11-30 06:27:12 -08:00
|
|
|
nil,
|
2018-10-02 04:59:19 -07:00
|
|
|
},
|
|
|
|
Start: time.Unix(0, 0),
|
|
|
|
End: time.Unix(2, 0),
|
|
|
|
Interval: time.Second,
|
|
|
|
},
|
|
|
|
{
|
|
|
|
Query: "1",
|
|
|
|
MaxSamples: 0,
|
|
|
|
Result: Result{
|
|
|
|
ErrTooManySamples(env),
|
|
|
|
nil,
|
2018-11-30 06:27:12 -08:00
|
|
|
nil,
|
2018-10-02 04:59:19 -07:00
|
|
|
},
|
|
|
|
Start: time.Unix(0, 0),
|
|
|
|
End: time.Unix(2, 0),
|
|
|
|
Interval: time.Second,
|
|
|
|
},
|
|
|
|
{
|
|
|
|
Query: "metric",
|
|
|
|
MaxSamples: 3,
|
|
|
|
Result: Result{
|
|
|
|
nil,
|
|
|
|
Matrix{Series{
|
|
|
|
Points: []Point{{V: 1, T: 0}, {V: 1, T: 1000}, {V: 1, T: 2000}},
|
|
|
|
Metric: labels.FromStrings("__name__", "metric")},
|
|
|
|
},
|
2018-11-30 06:27:12 -08:00
|
|
|
nil,
|
2018-10-02 04:59:19 -07:00
|
|
|
},
|
|
|
|
Start: time.Unix(0, 0),
|
|
|
|
End: time.Unix(2, 0),
|
|
|
|
Interval: time.Second,
|
|
|
|
},
|
|
|
|
{
|
|
|
|
Query: "metric",
|
|
|
|
MaxSamples: 2,
|
|
|
|
Result: Result{
|
|
|
|
ErrTooManySamples(env),
|
|
|
|
nil,
|
2018-11-30 06:27:12 -08:00
|
|
|
nil,
|
2018-10-02 04:59:19 -07:00
|
|
|
},
|
|
|
|
Start: time.Unix(0, 0),
|
|
|
|
End: time.Unix(2, 0),
|
|
|
|
Interval: time.Second,
|
|
|
|
},
|
|
|
|
{
|
|
|
|
Query: "metric",
|
|
|
|
MaxSamples: 3,
|
|
|
|
Result: Result{
|
|
|
|
nil,
|
|
|
|
Matrix{Series{
|
|
|
|
Points: []Point{{V: 1, T: 0}, {V: 1, T: 5000}, {V: 2, T: 10000}},
|
|
|
|
Metric: labels.FromStrings("__name__", "metric")},
|
|
|
|
},
|
2018-11-30 06:27:12 -08:00
|
|
|
nil,
|
2018-10-02 04:59:19 -07:00
|
|
|
},
|
|
|
|
Start: time.Unix(0, 0),
|
|
|
|
End: time.Unix(10, 0),
|
|
|
|
Interval: 5 * time.Second,
|
|
|
|
},
|
|
|
|
{
|
|
|
|
Query: "metric",
|
|
|
|
MaxSamples: 2,
|
|
|
|
Result: Result{
|
|
|
|
ErrTooManySamples(env),
|
|
|
|
nil,
|
2018-11-30 06:27:12 -08:00
|
|
|
nil,
|
2018-10-02 04:59:19 -07:00
|
|
|
},
|
|
|
|
Start: time.Unix(0, 0),
|
|
|
|
End: time.Unix(10, 0),
|
|
|
|
Interval: 5 * time.Second,
|
|
|
|
},
|
|
|
|
}
|
|
|
|
|
|
|
|
engine := test.QueryEngine()
|
|
|
|
for _, c := range cases {
|
|
|
|
var err error
|
|
|
|
var qry Query
|
|
|
|
|
|
|
|
engine.maxSamplesPerQuery = c.MaxSamples
|
|
|
|
|
|
|
|
if c.Interval == 0 {
|
|
|
|
qry, err = engine.NewInstantQuery(test.Queryable(), c.Query, c.Start)
|
|
|
|
} else {
|
|
|
|
qry, err = engine.NewRangeQuery(test.Queryable(), c.Query, c.Start, c.End, c.Interval)
|
|
|
|
}
|
|
|
|
if err != nil {
|
|
|
|
t.Fatalf("unexpected error creating query: %q", err)
|
|
|
|
}
|
|
|
|
res := qry.Exec(test.Context())
|
|
|
|
if res.Err != nil && res.Err != c.Result.Err {
|
|
|
|
t.Fatalf("unexpected error running query: %q, expected to get result: %q", res.Err, c.Result.Value)
|
|
|
|
}
|
|
|
|
if !reflect.DeepEqual(res.Value, c.Result.Value) {
|
|
|
|
t.Fatalf("unexpected result for query %q: got %q wanted %q", c.Query, res.Value.String(), c.Result.String())
|
|
|
|
}
|
|
|
|
}
|
|
|
|
}
|
|
|
|
|
2015-08-19 06:28:53 -07:00
|
|
|
func TestRecoverEvaluatorRuntime(t *testing.T) {
|
2017-08-11 11:45:52 -07:00
|
|
|
ev := &evaluator{logger: log.NewNopLogger()}
|
|
|
|
|
2015-08-19 06:28:53 -07:00
|
|
|
var err error
|
|
|
|
defer ev.recover(&err)
|
|
|
|
|
|
|
|
// Cause a runtime panic.
|
|
|
|
var a []int
|
2019-05-03 06:11:28 -07:00
|
|
|
//nolint:govet
|
2015-08-19 06:28:53 -07:00
|
|
|
a[123] = 1
|
|
|
|
|
|
|
|
if err.Error() != "unexpected error" {
|
|
|
|
t.Fatalf("wrong error message: %q, expected %q", err, "unexpected error")
|
|
|
|
}
|
|
|
|
}
|
|
|
|
|
|
|
|
func TestRecoverEvaluatorError(t *testing.T) {
|
2017-08-11 11:45:52 -07:00
|
|
|
ev := &evaluator{logger: log.NewNopLogger()}
|
2015-08-19 06:28:53 -07:00
|
|
|
var err error
|
|
|
|
|
2019-03-25 16:01:12 -07:00
|
|
|
e := errors.New("custom error")
|
2015-08-19 06:28:53 -07:00
|
|
|
|
2015-08-25 17:04:01 -07:00
|
|
|
defer func() {
|
|
|
|
if err.Error() != e.Error() {
|
|
|
|
t.Fatalf("wrong error message: %q, expected %q", err, e)
|
|
|
|
}
|
|
|
|
}()
|
|
|
|
defer ev.recover(&err)
|
|
|
|
|
|
|
|
panic(e)
|
2015-08-19 06:28:53 -07:00
|
|
|
}
|
2018-12-22 05:47:13 -08:00
|
|
|
|
|
|
|
func TestSubquerySelector(t *testing.T) {
|
|
|
|
tests := []struct {
|
|
|
|
loadString string
|
|
|
|
cases []struct {
|
|
|
|
Query string
|
|
|
|
Result Result
|
|
|
|
Start time.Time
|
|
|
|
}
|
|
|
|
}{
|
|
|
|
{
|
|
|
|
loadString: `load 10s
|
|
|
|
metric 1 2`,
|
|
|
|
cases: []struct {
|
|
|
|
Query string
|
|
|
|
Result Result
|
|
|
|
Start time.Time
|
|
|
|
}{
|
|
|
|
{
|
|
|
|
Query: "metric[20s:10s]",
|
|
|
|
Result: Result{
|
|
|
|
nil,
|
|
|
|
Matrix{Series{
|
|
|
|
Points: []Point{{V: 1, T: 0}, {V: 2, T: 10000}},
|
|
|
|
Metric: labels.FromStrings("__name__", "metric")},
|
|
|
|
},
|
|
|
|
nil,
|
|
|
|
},
|
|
|
|
Start: time.Unix(10, 0),
|
|
|
|
},
|
|
|
|
{
|
|
|
|
Query: "metric[20s:5s]",
|
|
|
|
Result: Result{
|
|
|
|
nil,
|
|
|
|
Matrix{Series{
|
|
|
|
Points: []Point{{V: 1, T: 0}, {V: 1, T: 5000}, {V: 2, T: 10000}},
|
|
|
|
Metric: labels.FromStrings("__name__", "metric")},
|
|
|
|
},
|
|
|
|
nil,
|
|
|
|
},
|
|
|
|
Start: time.Unix(10, 0),
|
|
|
|
},
|
|
|
|
{
|
|
|
|
Query: "metric[20s:5s] offset 2s",
|
|
|
|
Result: Result{
|
|
|
|
nil,
|
|
|
|
Matrix{Series{
|
|
|
|
Points: []Point{{V: 1, T: 0}, {V: 1, T: 5000}, {V: 2, T: 10000}},
|
|
|
|
Metric: labels.FromStrings("__name__", "metric")},
|
|
|
|
},
|
|
|
|
nil,
|
|
|
|
},
|
|
|
|
Start: time.Unix(12, 0),
|
|
|
|
},
|
|
|
|
{
|
|
|
|
Query: "metric[20s:5s] offset 6s",
|
|
|
|
Result: Result{
|
|
|
|
nil,
|
|
|
|
Matrix{Series{
|
|
|
|
Points: []Point{{V: 1, T: 0}, {V: 1, T: 5000}, {V: 2, T: 10000}},
|
|
|
|
Metric: labels.FromStrings("__name__", "metric")},
|
|
|
|
},
|
|
|
|
nil,
|
|
|
|
},
|
|
|
|
Start: time.Unix(20, 0),
|
|
|
|
},
|
|
|
|
{
|
|
|
|
Query: "metric[20s:5s] offset 4s",
|
|
|
|
Result: Result{
|
|
|
|
nil,
|
|
|
|
Matrix{Series{
|
|
|
|
Points: []Point{{V: 2, T: 15000}, {V: 2, T: 20000}, {V: 2, T: 25000}, {V: 2, T: 30000}},
|
|
|
|
Metric: labels.FromStrings("__name__", "metric")},
|
|
|
|
},
|
|
|
|
nil,
|
|
|
|
},
|
|
|
|
Start: time.Unix(35, 0),
|
|
|
|
},
|
|
|
|
{
|
|
|
|
Query: "metric[20s:5s] offset 5s",
|
|
|
|
Result: Result{
|
|
|
|
nil,
|
|
|
|
Matrix{Series{
|
|
|
|
Points: []Point{{V: 2, T: 10000}, {V: 2, T: 15000}, {V: 2, T: 20000}, {V: 2, T: 25000}, {V: 2, T: 30000}},
|
|
|
|
Metric: labels.FromStrings("__name__", "metric")},
|
|
|
|
},
|
|
|
|
nil,
|
|
|
|
},
|
|
|
|
Start: time.Unix(35, 0),
|
|
|
|
},
|
|
|
|
{
|
|
|
|
Query: "metric[20s:5s] offset 6s",
|
|
|
|
Result: Result{
|
|
|
|
nil,
|
|
|
|
Matrix{Series{
|
|
|
|
Points: []Point{{V: 2, T: 10000}, {V: 2, T: 15000}, {V: 2, T: 20000}, {V: 2, T: 25000}},
|
|
|
|
Metric: labels.FromStrings("__name__", "metric")},
|
|
|
|
},
|
|
|
|
nil,
|
|
|
|
},
|
|
|
|
Start: time.Unix(35, 0),
|
|
|
|
},
|
|
|
|
{
|
|
|
|
Query: "metric[20s:5s] offset 7s",
|
|
|
|
Result: Result{
|
|
|
|
nil,
|
|
|
|
Matrix{Series{
|
|
|
|
Points: []Point{{V: 2, T: 10000}, {V: 2, T: 15000}, {V: 2, T: 20000}, {V: 2, T: 25000}},
|
|
|
|
Metric: labels.FromStrings("__name__", "metric")},
|
|
|
|
},
|
|
|
|
nil,
|
|
|
|
},
|
|
|
|
Start: time.Unix(35, 0),
|
|
|
|
},
|
|
|
|
},
|
|
|
|
},
|
|
|
|
{
|
|
|
|
loadString: `load 10s
|
|
|
|
http_requests{job="api-server", instance="0", group="production"} 0+10x1000 100+30x1000
|
|
|
|
http_requests{job="api-server", instance="1", group="production"} 0+20x1000 200+30x1000
|
|
|
|
http_requests{job="api-server", instance="0", group="canary"} 0+30x1000 300+80x1000
|
|
|
|
http_requests{job="api-server", instance="1", group="canary"} 0+40x2000`,
|
|
|
|
cases: []struct {
|
|
|
|
Query string
|
|
|
|
Result Result
|
|
|
|
Start time.Time
|
|
|
|
}{
|
|
|
|
{ // Normal selector.
|
|
|
|
Query: `http_requests{group=~"pro.*",instance="0"}[30s:10s]`,
|
|
|
|
Result: Result{
|
|
|
|
nil,
|
|
|
|
Matrix{Series{
|
|
|
|
Points: []Point{{V: 9990, T: 9990000}, {V: 10000, T: 10000000}, {V: 100, T: 10010000}, {V: 130, T: 10020000}},
|
|
|
|
Metric: labels.FromStrings("__name__", "http_requests", "job", "api-server", "instance", "0", "group", "production")},
|
|
|
|
},
|
|
|
|
nil,
|
|
|
|
},
|
|
|
|
Start: time.Unix(10020, 0),
|
|
|
|
},
|
|
|
|
{ // Default step.
|
|
|
|
Query: `http_requests{group=~"pro.*",instance="0"}[5m:]`,
|
|
|
|
Result: Result{
|
|
|
|
nil,
|
|
|
|
Matrix{Series{
|
|
|
|
Points: []Point{{V: 9840, T: 9840000}, {V: 9900, T: 9900000}, {V: 9960, T: 9960000}, {V: 130, T: 10020000}, {V: 310, T: 10080000}},
|
|
|
|
Metric: labels.FromStrings("__name__", "http_requests", "job", "api-server", "instance", "0", "group", "production")},
|
|
|
|
},
|
|
|
|
nil,
|
|
|
|
},
|
|
|
|
Start: time.Unix(10100, 0),
|
|
|
|
},
|
|
|
|
{ // Checking if high offset (>LookbackDelta) is being taken care of.
|
|
|
|
Query: `http_requests{group=~"pro.*",instance="0"}[5m:] offset 20m`,
|
|
|
|
Result: Result{
|
|
|
|
nil,
|
|
|
|
Matrix{Series{
|
|
|
|
Points: []Point{{V: 8640, T: 8640000}, {V: 8700, T: 8700000}, {V: 8760, T: 8760000}, {V: 8820, T: 8820000}, {V: 8880, T: 8880000}},
|
|
|
|
Metric: labels.FromStrings("__name__", "http_requests", "job", "api-server", "instance", "0", "group", "production")},
|
|
|
|
},
|
|
|
|
nil,
|
|
|
|
},
|
|
|
|
Start: time.Unix(10100, 0),
|
|
|
|
},
|
|
|
|
{
|
|
|
|
Query: `rate(http_requests[1m])[15s:5s]`,
|
|
|
|
Result: Result{
|
|
|
|
nil,
|
|
|
|
Matrix{
|
|
|
|
Series{
|
|
|
|
Points: []Point{{V: 3, T: 7985000}, {V: 3, T: 7990000}, {V: 3, T: 7995000}, {V: 3, T: 8000000}},
|
|
|
|
Metric: labels.FromStrings("job", "api-server", "instance", "0", "group", "canary"),
|
|
|
|
},
|
|
|
|
Series{
|
|
|
|
Points: []Point{{V: 4, T: 7985000}, {V: 4, T: 7990000}, {V: 4, T: 7995000}, {V: 4, T: 8000000}},
|
|
|
|
Metric: labels.FromStrings("job", "api-server", "instance", "1", "group", "canary"),
|
|
|
|
},
|
|
|
|
Series{
|
|
|
|
Points: []Point{{V: 1, T: 7985000}, {V: 1, T: 7990000}, {V: 1, T: 7995000}, {V: 1, T: 8000000}},
|
|
|
|
Metric: labels.FromStrings("job", "api-server", "instance", "0", "group", "production"),
|
|
|
|
},
|
|
|
|
Series{
|
|
|
|
Points: []Point{{V: 2, T: 7985000}, {V: 2, T: 7990000}, {V: 2, T: 7995000}, {V: 2, T: 8000000}},
|
|
|
|
Metric: labels.FromStrings("job", "api-server", "instance", "1", "group", "production"),
|
|
|
|
},
|
|
|
|
},
|
|
|
|
nil,
|
|
|
|
},
|
|
|
|
Start: time.Unix(8000, 0),
|
|
|
|
},
|
|
|
|
{
|
|
|
|
Query: `sum(http_requests{group=~"pro.*"})[30s:10s]`,
|
|
|
|
Result: Result{
|
|
|
|
nil,
|
|
|
|
Matrix{Series{
|
|
|
|
Points: []Point{{V: 270, T: 90000}, {V: 300, T: 100000}, {V: 330, T: 110000}, {V: 360, T: 120000}},
|
|
|
|
Metric: labels.Labels{}},
|
|
|
|
},
|
|
|
|
nil,
|
|
|
|
},
|
|
|
|
Start: time.Unix(120, 0),
|
|
|
|
},
|
|
|
|
{
|
|
|
|
Query: `sum(http_requests)[40s:10s]`,
|
|
|
|
Result: Result{
|
|
|
|
nil,
|
|
|
|
Matrix{Series{
|
|
|
|
Points: []Point{{V: 800, T: 80000}, {V: 900, T: 90000}, {V: 1000, T: 100000}, {V: 1100, T: 110000}, {V: 1200, T: 120000}},
|
|
|
|
Metric: labels.Labels{}},
|
|
|
|
},
|
|
|
|
nil,
|
|
|
|
},
|
|
|
|
Start: time.Unix(120, 0),
|
|
|
|
},
|
|
|
|
{
|
|
|
|
Query: `(sum(http_requests{group=~"p.*"})+sum(http_requests{group=~"c.*"}))[20s:5s]`,
|
|
|
|
Result: Result{
|
|
|
|
nil,
|
|
|
|
Matrix{Series{
|
|
|
|
Points: []Point{{V: 1000, T: 100000}, {V: 1000, T: 105000}, {V: 1100, T: 110000}, {V: 1100, T: 115000}, {V: 1200, T: 120000}},
|
|
|
|
Metric: labels.Labels{}},
|
|
|
|
},
|
|
|
|
nil,
|
|
|
|
},
|
|
|
|
Start: time.Unix(120, 0),
|
|
|
|
},
|
|
|
|
},
|
|
|
|
},
|
|
|
|
}
|
|
|
|
|
|
|
|
SetDefaultEvaluationInterval(1 * time.Minute)
|
|
|
|
for _, tst := range tests {
|
|
|
|
test, err := NewTest(t, tst.loadString)
|
|
|
|
|
|
|
|
if err != nil {
|
|
|
|
t.Fatalf("unexpected error creating test: %q", err)
|
|
|
|
}
|
|
|
|
defer test.Close()
|
|
|
|
|
|
|
|
err = test.Run()
|
|
|
|
if err != nil {
|
|
|
|
t.Fatalf("unexpected error initializing test: %q", err)
|
|
|
|
}
|
|
|
|
|
|
|
|
engine := test.QueryEngine()
|
|
|
|
for _, c := range tst.cases {
|
|
|
|
var err error
|
|
|
|
var qry Query
|
|
|
|
|
|
|
|
qry, err = engine.NewInstantQuery(test.Queryable(), c.Query, c.Start)
|
|
|
|
if err != nil {
|
|
|
|
t.Fatalf("unexpected error creating query: %q", err)
|
|
|
|
}
|
|
|
|
res := qry.Exec(test.Context())
|
|
|
|
if res.Err != nil && res.Err != c.Result.Err {
|
|
|
|
t.Fatalf("unexpected error running query: %q, expected to get result: %q", res.Err, c.Result.Value)
|
|
|
|
}
|
|
|
|
if !reflect.DeepEqual(res.Value, c.Result.Value) {
|
|
|
|
t.Fatalf("unexpected result for query %q: got %q wanted %q", c.Query, res.Value.String(), c.Result.String())
|
|
|
|
}
|
|
|
|
}
|
|
|
|
}
|
|
|
|
}
|