mirror of
https://github.com/prometheus/prometheus.git
synced 2025-03-05 20:59:13 -08:00
Merge pull request #378 from grafana/allow-forcing-usage-of-postings-for-matchers-cache
Allow forcing usage of PostingsForMatchersCache
This commit is contained in:
commit
fa868fd321
|
@ -351,7 +351,7 @@ func OpenBlockWithCache(logger log.Logger, dir string, pool chunkenc.Pool, cache
|
||||||
if err != nil {
|
if err != nil {
|
||||||
return nil, err
|
return nil, err
|
||||||
}
|
}
|
||||||
pfmc := NewPostingsForMatchersCache(defaultPostingsForMatchersCacheTTL, defaultPostingsForMatchersCacheSize)
|
pfmc := NewPostingsForMatchersCache(defaultPostingsForMatchersCacheTTL, defaultPostingsForMatchersCacheSize, false)
|
||||||
ir := indexReaderWithPostingsForMatchers{indexReader, pfmc}
|
ir := indexReaderWithPostingsForMatchers{indexReader, pfmc}
|
||||||
closers = append(closers, ir)
|
closers = append(closers, ir)
|
||||||
|
|
||||||
|
|
15
tsdb/db.go
15
tsdb/db.go
|
@ -86,6 +86,9 @@ func DefaultOptions() *Options {
|
||||||
HeadChunksEndTimeVariance: 0,
|
HeadChunksEndTimeVariance: 0,
|
||||||
HeadChunksWriteQueueSize: chunks.DefaultWriteQueueSize,
|
HeadChunksWriteQueueSize: chunks.DefaultWriteQueueSize,
|
||||||
OutOfOrderCapMax: DefaultOutOfOrderCapMax,
|
OutOfOrderCapMax: DefaultOutOfOrderCapMax,
|
||||||
|
HeadPostingsForMatchersCacheTTL: defaultPostingsForMatchersCacheTTL,
|
||||||
|
HeadPostingsForMatchersCacheSize: defaultPostingsForMatchersCacheSize,
|
||||||
|
HeadPostingsForMatchersCacheForce: false,
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
|
@ -189,6 +192,15 @@ type Options struct {
|
||||||
// OutOfOrderCapMax is maximum capacity for OOO chunks (in samples).
|
// OutOfOrderCapMax is maximum capacity for OOO chunks (in samples).
|
||||||
// If it is <=0, the default value is assumed.
|
// If it is <=0, the default value is assumed.
|
||||||
OutOfOrderCapMax int64
|
OutOfOrderCapMax int64
|
||||||
|
|
||||||
|
// HeadPostingsForMatchersCacheTTL is the TTL of the postings for matchers cache in the Head.
|
||||||
|
// If it's 0, the cache will only deduplicate in-flight requests, deleting the results once the first request has finished.
|
||||||
|
HeadPostingsForMatchersCacheTTL time.Duration
|
||||||
|
// HeadPostingsForMatchersCacheSize is the maximum size of cached postings for matchers elements in the Head.
|
||||||
|
// It's ignored used when HeadPostingsForMatchersCacheTTL is 0.
|
||||||
|
HeadPostingsForMatchersCacheSize int
|
||||||
|
// HeadPostingsForMatchersCacheForce forces the usage of postings for matchers cache for all calls on Head and OOOHead regardless of the `concurrent` param.
|
||||||
|
HeadPostingsForMatchersCacheForce bool
|
||||||
}
|
}
|
||||||
|
|
||||||
type BlocksToDeleteFunc func(blocks []*Block) map[ulid.ULID]struct{}
|
type BlocksToDeleteFunc func(blocks []*Block) map[ulid.ULID]struct{}
|
||||||
|
@ -798,6 +810,9 @@ func open(dir string, l log.Logger, r prometheus.Registerer, opts *Options, rngs
|
||||||
headOpts.EnableNativeHistograms.Store(opts.EnableNativeHistograms)
|
headOpts.EnableNativeHistograms.Store(opts.EnableNativeHistograms)
|
||||||
headOpts.OutOfOrderTimeWindow.Store(opts.OutOfOrderTimeWindow)
|
headOpts.OutOfOrderTimeWindow.Store(opts.OutOfOrderTimeWindow)
|
||||||
headOpts.OutOfOrderCapMax.Store(opts.OutOfOrderCapMax)
|
headOpts.OutOfOrderCapMax.Store(opts.OutOfOrderCapMax)
|
||||||
|
headOpts.PostingsForMatchersCacheTTL = opts.HeadPostingsForMatchersCacheTTL
|
||||||
|
headOpts.PostingsForMatchersCacheSize = opts.HeadPostingsForMatchersCacheSize
|
||||||
|
headOpts.PostingsForMatchersCacheForce = opts.HeadPostingsForMatchersCacheForce
|
||||||
if opts.IsolationDisabled {
|
if opts.IsolationDisabled {
|
||||||
// We only override this flag if isolation is disabled at DB level. We use the default otherwise.
|
// We only override this flag if isolation is disabled at DB level. We use the default otherwise.
|
||||||
headOpts.IsolationDisabled = opts.IsolationDisabled
|
headOpts.IsolationDisabled = opts.IsolationDisabled
|
||||||
|
|
|
@ -170,6 +170,10 @@ type HeadOptions struct {
|
||||||
EnableMemorySnapshotOnShutdown bool
|
EnableMemorySnapshotOnShutdown bool
|
||||||
|
|
||||||
IsolationDisabled bool
|
IsolationDisabled bool
|
||||||
|
|
||||||
|
PostingsForMatchersCacheTTL time.Duration
|
||||||
|
PostingsForMatchersCacheSize int
|
||||||
|
PostingsForMatchersCacheForce bool
|
||||||
}
|
}
|
||||||
|
|
||||||
const (
|
const (
|
||||||
|
@ -188,6 +192,9 @@ func DefaultHeadOptions() *HeadOptions {
|
||||||
StripeSize: DefaultStripeSize,
|
StripeSize: DefaultStripeSize,
|
||||||
SeriesCallback: &noopSeriesLifecycleCallback{},
|
SeriesCallback: &noopSeriesLifecycleCallback{},
|
||||||
IsolationDisabled: defaultIsolationDisabled,
|
IsolationDisabled: defaultIsolationDisabled,
|
||||||
|
PostingsForMatchersCacheTTL: defaultPostingsForMatchersCacheTTL,
|
||||||
|
PostingsForMatchersCacheSize: defaultPostingsForMatchersCacheSize,
|
||||||
|
PostingsForMatchersCacheForce: false,
|
||||||
}
|
}
|
||||||
ho.OutOfOrderCapMax.Store(DefaultOutOfOrderCapMax)
|
ho.OutOfOrderCapMax.Store(DefaultOutOfOrderCapMax)
|
||||||
return ho
|
return ho
|
||||||
|
@ -254,7 +261,7 @@ func NewHead(r prometheus.Registerer, l log.Logger, wal, wbl *wlog.WL, opts *Hea
|
||||||
stats: stats,
|
stats: stats,
|
||||||
reg: r,
|
reg: r,
|
||||||
|
|
||||||
pfmc: NewPostingsForMatchersCache(defaultPostingsForMatchersCacheTTL, defaultPostingsForMatchersCacheSize),
|
pfmc: NewPostingsForMatchersCache(opts.PostingsForMatchersCacheTTL, opts.PostingsForMatchersCacheSize, opts.PostingsForMatchersCacheForce),
|
||||||
}
|
}
|
||||||
if err := h.resetInMemoryState(); err != nil {
|
if err := h.resetInMemoryState(); err != nil {
|
||||||
return nil, err
|
return nil, err
|
||||||
|
|
|
@ -28,13 +28,16 @@ type IndexPostingsReader interface {
|
||||||
}
|
}
|
||||||
|
|
||||||
// NewPostingsForMatchersCache creates a new PostingsForMatchersCache.
|
// NewPostingsForMatchersCache creates a new PostingsForMatchersCache.
|
||||||
func NewPostingsForMatchersCache(ttl time.Duration, cacheSize int) *PostingsForMatchersCache {
|
// If `ttl` is 0, then it only deduplicates in-flight requests.
|
||||||
|
// If `force` is true, then all requests go through cache, regardless of the `concurrent` param provided.
|
||||||
|
func NewPostingsForMatchersCache(ttl time.Duration, cacheSize int, force bool) *PostingsForMatchersCache {
|
||||||
b := &PostingsForMatchersCache{
|
b := &PostingsForMatchersCache{
|
||||||
calls: &sync.Map{},
|
calls: &sync.Map{},
|
||||||
cached: list.New(),
|
cached: list.New(),
|
||||||
|
|
||||||
ttl: ttl,
|
ttl: ttl,
|
||||||
cacheSize: cacheSize,
|
cacheSize: cacheSize,
|
||||||
|
force: force,
|
||||||
|
|
||||||
timeNow: time.Now,
|
timeNow: time.Now,
|
||||||
postingsForMatchers: PostingsForMatchers,
|
postingsForMatchers: PostingsForMatchers,
|
||||||
|
@ -43,7 +46,7 @@ func NewPostingsForMatchersCache(ttl time.Duration, cacheSize int) *PostingsForM
|
||||||
return b
|
return b
|
||||||
}
|
}
|
||||||
|
|
||||||
// PostingsForMatchersCache caches PostingsForMatchers call results when the concurrent hint is passed in.
|
// PostingsForMatchersCache caches PostingsForMatchers call results when the concurrent hint is passed in or force is true.
|
||||||
type PostingsForMatchersCache struct {
|
type PostingsForMatchersCache struct {
|
||||||
calls *sync.Map
|
calls *sync.Map
|
||||||
|
|
||||||
|
@ -52,6 +55,7 @@ type PostingsForMatchersCache struct {
|
||||||
|
|
||||||
ttl time.Duration
|
ttl time.Duration
|
||||||
cacheSize int
|
cacheSize int
|
||||||
|
force bool
|
||||||
|
|
||||||
// timeNow is the time.Now that can be replaced for testing purposes
|
// timeNow is the time.Now that can be replaced for testing purposes
|
||||||
timeNow func() time.Time
|
timeNow func() time.Time
|
||||||
|
@ -60,7 +64,7 @@ type PostingsForMatchersCache struct {
|
||||||
}
|
}
|
||||||
|
|
||||||
func (c *PostingsForMatchersCache) PostingsForMatchers(ix IndexPostingsReader, concurrent bool, ms ...*labels.Matcher) (index.Postings, error) {
|
func (c *PostingsForMatchersCache) PostingsForMatchers(ix IndexPostingsReader, concurrent bool, ms ...*labels.Matcher) (index.Postings, error) {
|
||||||
if !concurrent {
|
if !concurrent && !c.force {
|
||||||
return c.postingsForMatchers(ix, ms...)
|
return c.postingsForMatchers(ix, ms...)
|
||||||
}
|
}
|
||||||
c.expire()
|
c.expire()
|
||||||
|
|
|
@ -16,8 +16,8 @@ import (
|
||||||
func TestPostingsForMatchersCache(t *testing.T) {
|
func TestPostingsForMatchersCache(t *testing.T) {
|
||||||
const testCacheSize = 5
|
const testCacheSize = 5
|
||||||
// newPostingsForMatchersCache tests the NewPostingsForMatcherCache constructor, but overrides the postingsForMatchers func
|
// newPostingsForMatchersCache tests the NewPostingsForMatcherCache constructor, but overrides the postingsForMatchers func
|
||||||
newPostingsForMatchersCache := func(ttl time.Duration, pfm func(ix IndexPostingsReader, ms ...*labels.Matcher) (index.Postings, error), timeMock *timeNowMock) *PostingsForMatchersCache {
|
newPostingsForMatchersCache := func(ttl time.Duration, pfm func(ix IndexPostingsReader, ms ...*labels.Matcher) (index.Postings, error), timeMock *timeNowMock, force bool) *PostingsForMatchersCache {
|
||||||
c := NewPostingsForMatchersCache(ttl, testCacheSize)
|
c := NewPostingsForMatchersCache(ttl, testCacheSize, force)
|
||||||
if c.postingsForMatchers == nil {
|
if c.postingsForMatchers == nil {
|
||||||
t.Fatalf("NewPostingsForMatchersCache() didn't assign postingsForMatchers func")
|
t.Fatalf("NewPostingsForMatchersCache() didn't assign postingsForMatchers func")
|
||||||
}
|
}
|
||||||
|
@ -36,7 +36,7 @@ func TestPostingsForMatchersCache(t *testing.T) {
|
||||||
require.IsType(t, indexForPostingsMock{}, ix, "Incorrect IndexPostingsReader was provided to PostingsForMatchers, expected the mock, was given %v (%T)", ix, ix)
|
require.IsType(t, indexForPostingsMock{}, ix, "Incorrect IndexPostingsReader was provided to PostingsForMatchers, expected the mock, was given %v (%T)", ix, ix)
|
||||||
require.Equal(t, expectedMatchers, ms, "Wrong label matchers provided, expected %v, got %v", expectedMatchers, ms)
|
require.Equal(t, expectedMatchers, ms, "Wrong label matchers provided, expected %v, got %v", expectedMatchers, ms)
|
||||||
return index.ErrPostings(expectedPostingsErr), nil
|
return index.ErrPostings(expectedPostingsErr), nil
|
||||||
}, &timeNowMock{})
|
}, &timeNowMock{}, false)
|
||||||
|
|
||||||
p, err := c.PostingsForMatchers(indexForPostingsMock{}, concurrent, expectedMatchers...)
|
p, err := c.PostingsForMatchers(indexForPostingsMock{}, concurrent, expectedMatchers...)
|
||||||
require.NoError(t, err)
|
require.NoError(t, err)
|
||||||
|
@ -52,7 +52,7 @@ func TestPostingsForMatchersCache(t *testing.T) {
|
||||||
|
|
||||||
c := newPostingsForMatchersCache(defaultPostingsForMatchersCacheTTL, func(ix IndexPostingsReader, ms ...*labels.Matcher) (index.Postings, error) {
|
c := newPostingsForMatchersCache(defaultPostingsForMatchersCacheTTL, func(ix IndexPostingsReader, ms ...*labels.Matcher) (index.Postings, error) {
|
||||||
return nil, expectedErr
|
return nil, expectedErr
|
||||||
}, &timeNowMock{})
|
}, &timeNowMock{}, false)
|
||||||
|
|
||||||
_, err := c.PostingsForMatchers(indexForPostingsMock{}, true, expectedMatchers...)
|
_, err := c.PostingsForMatchers(indexForPostingsMock{}, true, expectedMatchers...)
|
||||||
require.Equal(t, expectedErr, err)
|
require.Equal(t, expectedErr, err)
|
||||||
|
@ -61,6 +61,9 @@ func TestPostingsForMatchersCache(t *testing.T) {
|
||||||
t.Run("happy case multiple concurrent calls: two same one different", func(t *testing.T) {
|
t.Run("happy case multiple concurrent calls: two same one different", func(t *testing.T) {
|
||||||
for _, cacheEnabled := range []bool{true, false} {
|
for _, cacheEnabled := range []bool{true, false} {
|
||||||
t.Run(fmt.Sprintf("cacheEnabled=%t", cacheEnabled), func(t *testing.T) {
|
t.Run(fmt.Sprintf("cacheEnabled=%t", cacheEnabled), func(t *testing.T) {
|
||||||
|
for _, forced := range []bool{true, false} {
|
||||||
|
concurrent := !forced
|
||||||
|
t.Run(fmt.Sprintf("forced=%t", forced), func(t *testing.T) {
|
||||||
calls := [][]*labels.Matcher{
|
calls := [][]*labels.Matcher{
|
||||||
{labels.MustNewMatcher(labels.MatchEqual, "foo", "bar")}, // 1
|
{labels.MustNewMatcher(labels.MatchEqual, "foo", "bar")}, // 1
|
||||||
{labels.MustNewMatcher(labels.MatchEqual, "foo", "bar")}, // 1 same
|
{labels.MustNewMatcher(labels.MatchEqual, "foo", "bar")}, // 1 same
|
||||||
|
@ -102,7 +105,7 @@ func TestPostingsForMatchersCache(t *testing.T) {
|
||||||
}
|
}
|
||||||
<-release
|
<-release
|
||||||
return nil, fmt.Errorf(matchersString(ms))
|
return nil, fmt.Errorf(matchersString(ms))
|
||||||
}, &timeNowMock{})
|
}, &timeNowMock{}, forced)
|
||||||
|
|
||||||
results := make([]string, len(calls))
|
results := make([]string, len(calls))
|
||||||
resultsWg := sync.WaitGroup{}
|
resultsWg := sync.WaitGroup{}
|
||||||
|
@ -111,7 +114,7 @@ func TestPostingsForMatchersCache(t *testing.T) {
|
||||||
// perform all calls
|
// perform all calls
|
||||||
for i := 0; i < len(calls); i++ {
|
for i := 0; i < len(calls); i++ {
|
||||||
go func(i int) {
|
go func(i int) {
|
||||||
_, err := c.PostingsForMatchers(indexForPostingsMock{}, true, calls[i]...)
|
_, err := c.PostingsForMatchers(indexForPostingsMock{}, concurrent, calls[i]...)
|
||||||
results[i] = err.Error()
|
results[i] = err.Error()
|
||||||
resultsWg.Done()
|
resultsWg.Done()
|
||||||
}(i)
|
}(i)
|
||||||
|
@ -135,6 +138,8 @@ func TestPostingsForMatchersCache(t *testing.T) {
|
||||||
})
|
})
|
||||||
}
|
}
|
||||||
})
|
})
|
||||||
|
}
|
||||||
|
})
|
||||||
|
|
||||||
t.Run("with concurrent==false, result is not cached", func(t *testing.T) {
|
t.Run("with concurrent==false, result is not cached", func(t *testing.T) {
|
||||||
expectedMatchers := []*labels.Matcher{labels.MustNewMatcher(labels.MatchEqual, "foo", "bar")}
|
expectedMatchers := []*labels.Matcher{labels.MustNewMatcher(labels.MatchEqual, "foo", "bar")}
|
||||||
|
@ -143,7 +148,7 @@ func TestPostingsForMatchersCache(t *testing.T) {
|
||||||
c := newPostingsForMatchersCache(defaultPostingsForMatchersCacheTTL, func(ix IndexPostingsReader, ms ...*labels.Matcher) (index.Postings, error) {
|
c := newPostingsForMatchersCache(defaultPostingsForMatchersCacheTTL, func(ix IndexPostingsReader, ms ...*labels.Matcher) (index.Postings, error) {
|
||||||
call++
|
call++
|
||||||
return index.ErrPostings(fmt.Errorf("result from call %d", call)), nil
|
return index.ErrPostings(fmt.Errorf("result from call %d", call)), nil
|
||||||
}, &timeNowMock{})
|
}, &timeNowMock{}, false)
|
||||||
|
|
||||||
// first call, fills the cache
|
// first call, fills the cache
|
||||||
p, err := c.PostingsForMatchers(indexForPostingsMock{}, false, expectedMatchers...)
|
p, err := c.PostingsForMatchers(indexForPostingsMock{}, false, expectedMatchers...)
|
||||||
|
@ -163,7 +168,7 @@ func TestPostingsForMatchersCache(t *testing.T) {
|
||||||
c := newPostingsForMatchersCache(0, func(ix IndexPostingsReader, ms ...*labels.Matcher) (index.Postings, error) {
|
c := newPostingsForMatchersCache(0, func(ix IndexPostingsReader, ms ...*labels.Matcher) (index.Postings, error) {
|
||||||
call++
|
call++
|
||||||
return index.ErrPostings(fmt.Errorf("result from call %d", call)), nil
|
return index.ErrPostings(fmt.Errorf("result from call %d", call)), nil
|
||||||
}, &timeNowMock{})
|
}, &timeNowMock{}, false)
|
||||||
|
|
||||||
// first call, fills the cache
|
// first call, fills the cache
|
||||||
p, err := c.PostingsForMatchers(indexForPostingsMock{}, true, expectedMatchers...)
|
p, err := c.PostingsForMatchers(indexForPostingsMock{}, true, expectedMatchers...)
|
||||||
|
@ -186,7 +191,7 @@ func TestPostingsForMatchersCache(t *testing.T) {
|
||||||
c := newPostingsForMatchersCache(defaultPostingsForMatchersCacheTTL, func(ix IndexPostingsReader, ms ...*labels.Matcher) (index.Postings, error) {
|
c := newPostingsForMatchersCache(defaultPostingsForMatchersCacheTTL, func(ix IndexPostingsReader, ms ...*labels.Matcher) (index.Postings, error) {
|
||||||
call++
|
call++
|
||||||
return index.ErrPostings(fmt.Errorf("result from call %d", call)), nil
|
return index.ErrPostings(fmt.Errorf("result from call %d", call)), nil
|
||||||
}, timeNow)
|
}, timeNow, false)
|
||||||
|
|
||||||
// first call, fills the cache
|
// first call, fills the cache
|
||||||
p, err := c.PostingsForMatchers(indexForPostingsMock{}, true, expectedMatchers...)
|
p, err := c.PostingsForMatchers(indexForPostingsMock{}, true, expectedMatchers...)
|
||||||
|
@ -220,7 +225,7 @@ func TestPostingsForMatchersCache(t *testing.T) {
|
||||||
k := matchersKey(ms)
|
k := matchersKey(ms)
|
||||||
callsPerMatchers[k]++
|
callsPerMatchers[k]++
|
||||||
return index.ErrPostings(fmt.Errorf("result from call %d", callsPerMatchers[k])), nil
|
return index.ErrPostings(fmt.Errorf("result from call %d", callsPerMatchers[k])), nil
|
||||||
}, timeNow)
|
}, timeNow, false)
|
||||||
|
|
||||||
// each one of the first testCacheSize calls is cached properly
|
// each one of the first testCacheSize calls is cached properly
|
||||||
for _, matchers := range calls {
|
for _, matchers := range calls {
|
||||||
|
|
Loading…
Reference in a new issue