mirror of
https://github.com/prometheus/prometheus.git
synced 2024-12-27 14:39:40 -08:00
af91fb8e31
This is done by bucketing chunks by fingerprint. If the persisting to disk falls behind, more and more chunks are in the queue. As soon as there are "double hits", we will now persist both chunks in one go, doubling the disk throughput (assuming it is limited by disk seeks). Should even more pile up so that we end wit "triple hits", we will persist those first, and so on. Even if we have millions of time series, this will still help, assuming not all of them are growing with the same speed. Series that get many samples and/or are not very compressable will accumulate chunks faster, and they will soon get double- or triple-writes. To improve the chance of double writes, -storage.local.persistence-queue-capacity could be set to a higher value. However, that will slow down shutdown a lot (as the queue has to be worked through). So we leave it to the user to set it to a really high value. A more fundamental solution would be to checkpoint not only head chunks, but also chunks still in the persist queue. That would be quite complicated for a rather limited use-case (running many time series with high ingestion rate on slow spinning disks).
63 lines
1.9 KiB
Go
63 lines
1.9 KiB
Go
// Copyright 2014 The Prometheus Authors
|
|
// Licensed under the Apache License, Version 2.0 (the "License");
|
|
// you may not use this file except in compliance with the License.
|
|
// You may obtain a copy of the License at
|
|
//
|
|
// http://www.apache.org/licenses/LICENSE-2.0
|
|
//
|
|
// Unless required by applicable law or agreed to in writing, software
|
|
// distributed under the License is distributed on an "AS IS" BASIS,
|
|
// WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
|
|
// See the License for the specific language governing permissions and
|
|
// limitations under the License.
|
|
|
|
// NOTE ON FILENAME: Do not rename this file helpers_test.go (which might appear
|
|
// an obvious choice). We need NewTestStorage in tests outside of the local
|
|
// package, too. On the other hand, moving NewTestStorage in its own package
|
|
// would cause circular dependencies in the tests in packages local.
|
|
|
|
package local
|
|
|
|
import (
|
|
"time"
|
|
|
|
"github.com/prometheus/prometheus/utility/test"
|
|
)
|
|
|
|
type testStorageCloser struct {
|
|
storage Storage
|
|
directory test.Closer
|
|
}
|
|
|
|
func (t *testStorageCloser) Close() {
|
|
t.storage.Stop()
|
|
t.directory.Close()
|
|
}
|
|
|
|
// NewTestStorage creates a storage instance backed by files in a temporary
|
|
// directory. The returned storage is already in serving state. Upon closing the
|
|
// returned test.Closer, the temporary directory is cleaned up.
|
|
func NewTestStorage(t test.T) (Storage, test.Closer) {
|
|
directory := test.NewTemporaryDirectory("test_storage", t)
|
|
o := &MemorySeriesStorageOptions{
|
|
MemoryChunks: 1000000,
|
|
PersistenceRetentionPeriod: 24 * time.Hour * 365 * 100, // Enough to never trigger purging.
|
|
PersistenceStoragePath: directory.Path(),
|
|
CheckpointInterval: time.Hour,
|
|
}
|
|
storage, err := NewMemorySeriesStorage(o)
|
|
if err != nil {
|
|
directory.Close()
|
|
t.Fatalf("Error creating storage: %s", err)
|
|
}
|
|
|
|
storage.Start()
|
|
|
|
closer := &testStorageCloser{
|
|
storage: storage,
|
|
directory: directory,
|
|
}
|
|
|
|
return storage, closer
|
|
}
|