mirror of
https://github.com/prometheus/prometheus.git
synced 2024-11-10 23:54:05 -08:00
Merge pull request #489 from prometheus/beorn7/get-some-todos-done
Beorn7/get some todos done
This commit is contained in:
commit
5f27629985
|
@ -128,24 +128,28 @@ func NewRuleManager(o *RuleManagerOptions) RuleManager {
|
|||
}
|
||||
|
||||
func (m *ruleManager) Run() {
|
||||
defer glog.Info("Rule manager stopped.")
|
||||
|
||||
ticker := time.NewTicker(m.interval)
|
||||
defer ticker.Stop()
|
||||
|
||||
for {
|
||||
// TODO(beorn): This has the same problem as the scraper had
|
||||
// before. If rule evaluation takes longer than the interval,
|
||||
// there is a 50% chance per iteration that - after stopping the
|
||||
// ruleManager - a new evaluation will be started rather than
|
||||
// the ruleManager actually stopped. We need a similar
|
||||
// contraption here as in the scraper.
|
||||
// The outer select clause makes sure that m.done is looked at
|
||||
// first. Otherwise, if m.runIteration takes longer than
|
||||
// m.interval, there is only a 50% chance that m.done will be
|
||||
// looked at before the next m.runIteration call happens.
|
||||
select {
|
||||
case <-ticker.C:
|
||||
start := time.Now()
|
||||
m.runIteration(m.results)
|
||||
iterationDuration.Observe(float64(time.Since(start) / time.Millisecond))
|
||||
case <-m.done:
|
||||
glog.Info("Rule manager stopped.")
|
||||
return
|
||||
default:
|
||||
select {
|
||||
case <-ticker.C:
|
||||
start := time.Now()
|
||||
m.runIteration(m.results)
|
||||
iterationDuration.Observe(float64(time.Since(start) / time.Millisecond))
|
||||
case <-m.done:
|
||||
return
|
||||
}
|
||||
}
|
||||
}
|
||||
}
|
||||
|
|
423
storage/local/crashrecovery.go
Normal file
423
storage/local/crashrecovery.go
Normal file
|
@ -0,0 +1,423 @@
|
|||
// Copyright 2015 The Prometheus Authors
|
||||
// Licensed under the Apache License, Version 2.0 (the "License");
|
||||
// you may not use this file except in compliance with the License.
|
||||
// You may obtain a copy of the License at
|
||||
//
|
||||
// http://www.apache.org/licenses/LICENSE-2.0
|
||||
//
|
||||
// Unless required by applicable law or agreed to in writing, software
|
||||
// distributed under the License is distributed on an "AS IS" BASIS,
|
||||
// WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
|
||||
// See the License for the specific language governing permissions and
|
||||
// limitations under the License.
|
||||
|
||||
package local
|
||||
|
||||
import (
|
||||
"fmt"
|
||||
"io"
|
||||
"math"
|
||||
"os"
|
||||
"path"
|
||||
"strings"
|
||||
|
||||
"github.com/golang/glog"
|
||||
|
||||
clientmodel "github.com/prometheus/client_golang/model"
|
||||
|
||||
"github.com/prometheus/prometheus/storage/local/codable"
|
||||
"github.com/prometheus/prometheus/storage/local/index"
|
||||
)
|
||||
|
||||
// recoverFromCrash is called by loadSeriesMapAndHeads if the persistence
|
||||
// appears to be dirty after the loading (either because the loading resulted in
|
||||
// an error or because the persistence was dirty from the start). Not goroutine
|
||||
// safe. Only call before anything else is running (except index processing
|
||||
// queue as started by newPersistence).
|
||||
func (p *persistence) recoverFromCrash(fingerprintToSeries map[clientmodel.Fingerprint]*memorySeries) error {
|
||||
// TODO(beorn): We need proper tests for the crash recovery.
|
||||
glog.Warning("Starting crash recovery. Prometheus is inoperational until complete.")
|
||||
|
||||
fpsSeen := map[clientmodel.Fingerprint]struct{}{}
|
||||
count := 0
|
||||
seriesDirNameFmt := fmt.Sprintf("%%0%dx", seriesDirNameLen)
|
||||
|
||||
glog.Info("Scanning files.")
|
||||
for i := 0; i < 1<<(seriesDirNameLen*4); i++ {
|
||||
dirname := path.Join(p.basePath, fmt.Sprintf(seriesDirNameFmt, i))
|
||||
dir, err := os.Open(dirname)
|
||||
if os.IsNotExist(err) {
|
||||
continue
|
||||
}
|
||||
if err != nil {
|
||||
return err
|
||||
}
|
||||
defer dir.Close()
|
||||
for fis := []os.FileInfo{}; err != io.EOF; fis, err = dir.Readdir(1024) {
|
||||
if err != nil {
|
||||
return err
|
||||
}
|
||||
for _, fi := range fis {
|
||||
fp, ok := p.sanitizeSeries(dirname, fi, fingerprintToSeries)
|
||||
if ok {
|
||||
fpsSeen[fp] = struct{}{}
|
||||
}
|
||||
count++
|
||||
if count%10000 == 0 {
|
||||
glog.Infof("%d files scanned.", count)
|
||||
}
|
||||
}
|
||||
}
|
||||
}
|
||||
glog.Infof("File scan complete. %d series found.", len(fpsSeen))
|
||||
|
||||
glog.Info("Checking for series without series file.")
|
||||
for fp, s := range fingerprintToSeries {
|
||||
if _, seen := fpsSeen[fp]; !seen {
|
||||
// fp exists in fingerprintToSeries, but has no representation on disk.
|
||||
if s.headChunkPersisted {
|
||||
// Oops, head chunk was persisted, but nothing on disk.
|
||||
// Thus, we lost that series completely. Clean up the remnants.
|
||||
delete(fingerprintToSeries, fp)
|
||||
if err := p.dropArchivedMetric(fp); err != nil {
|
||||
// Dropping the archived metric didn't work, so try
|
||||
// to unindex it, just in case it's in the indexes.
|
||||
p.unindexMetric(fp, s.metric)
|
||||
}
|
||||
glog.Warningf("Lost series detected: fingerprint %v, metric %v.", fp, s.metric)
|
||||
continue
|
||||
}
|
||||
// If we are here, the only chunk we have is the head chunk.
|
||||
// Adjust things accordingly.
|
||||
if len(s.chunkDescs) > 1 || s.chunkDescsOffset != 0 {
|
||||
minLostChunks := len(s.chunkDescs) + s.chunkDescsOffset - 1
|
||||
if minLostChunks <= 0 {
|
||||
glog.Warningf(
|
||||
"Possible loss of chunks for fingerprint %v, metric %v.",
|
||||
fp, s.metric,
|
||||
)
|
||||
} else {
|
||||
glog.Warningf(
|
||||
"Lost at least %d chunks for fingerprint %v, metric %v.",
|
||||
minLostChunks, fp, s.metric,
|
||||
)
|
||||
}
|
||||
s.chunkDescs = s.chunkDescs[len(s.chunkDescs)-1:]
|
||||
s.chunkDescsOffset = 0
|
||||
}
|
||||
fpsSeen[fp] = struct{}{} // Add so that fpsSeen is complete.
|
||||
}
|
||||
}
|
||||
glog.Info("Check for series without series file complete.")
|
||||
|
||||
if err := p.cleanUpArchiveIndexes(fingerprintToSeries, fpsSeen); err != nil {
|
||||
return err
|
||||
}
|
||||
if err := p.rebuildLabelIndexes(fingerprintToSeries); err != nil {
|
||||
return err
|
||||
}
|
||||
|
||||
p.setDirty(false)
|
||||
glog.Warning("Crash recovery complete.")
|
||||
return nil
|
||||
}
|
||||
|
||||
// sanitizeSeries sanitizes a series based on its series file as defined by the
|
||||
// provided directory and FileInfo. The method returns the fingerprint as
|
||||
// derived from the directory and file name, and whether the provided file has
|
||||
// been sanitized. A file that failed to be sanitized is moved into the
|
||||
// "orphaned" sub-directory, if possible.
|
||||
//
|
||||
// The following steps are performed:
|
||||
//
|
||||
// - A file whose name doesn't comply with the naming scheme of a series file is
|
||||
// simply moved into the orphaned directory.
|
||||
//
|
||||
// - If the size of the series file isn't a multiple of the chunk size,
|
||||
// extraneous bytes are truncated. If the truncation fails, the file is
|
||||
// moved into the orphaned directory.
|
||||
//
|
||||
// - A file that is empty (after truncation) is deleted.
|
||||
//
|
||||
// - A series that is not archived (i.e. it is in the fingerprintToSeries map)
|
||||
// is checked for consistency of its various parameters (like head-chunk
|
||||
// persistence state, offset of chunkDescs etc.). In particular, overlap
|
||||
// between an in-memory head chunk with the most recent persisted chunk is
|
||||
// checked. Inconsistencies are rectified.
|
||||
//
|
||||
// - A series that is archived (i.e. it is not in the fingerprintToSeries map)
|
||||
// is checked for its presence in the index of archived series. If it cannot
|
||||
// be found there, it is moved into the orphaned directory.
|
||||
func (p *persistence) sanitizeSeries(dirname string, fi os.FileInfo, fingerprintToSeries map[clientmodel.Fingerprint]*memorySeries) (clientmodel.Fingerprint, bool) {
|
||||
filename := path.Join(dirname, fi.Name())
|
||||
purge := func() {
|
||||
var err error
|
||||
defer func() {
|
||||
if err != nil {
|
||||
glog.Errorf("Failed to move lost series file %s to orphaned directory, deleting it instead. Error was: %s", filename, err)
|
||||
if err = os.Remove(filename); err != nil {
|
||||
glog.Errorf("Even deleting file %s did not work: %s", filename, err)
|
||||
}
|
||||
}
|
||||
}()
|
||||
orphanedDir := path.Join(p.basePath, "orphaned", path.Base(dirname))
|
||||
if err = os.MkdirAll(orphanedDir, 0700); err != nil {
|
||||
return
|
||||
}
|
||||
if err = os.Rename(filename, path.Join(orphanedDir, fi.Name())); err != nil {
|
||||
return
|
||||
}
|
||||
}
|
||||
|
||||
var fp clientmodel.Fingerprint
|
||||
if len(fi.Name()) != fpLen-seriesDirNameLen+len(seriesFileSuffix) ||
|
||||
!strings.HasSuffix(fi.Name(), seriesFileSuffix) {
|
||||
glog.Warningf("Unexpected series file name %s.", filename)
|
||||
purge()
|
||||
return fp, false
|
||||
}
|
||||
if err := fp.LoadFromString(path.Base(dirname) + fi.Name()[:fpLen-seriesDirNameLen]); err != nil {
|
||||
glog.Warningf("Error parsing file name %s: %s", filename, err)
|
||||
purge()
|
||||
return fp, false
|
||||
}
|
||||
|
||||
bytesToTrim := fi.Size() % int64(p.chunkLen+chunkHeaderLen)
|
||||
chunksInFile := int(fi.Size()) / (p.chunkLen + chunkHeaderLen)
|
||||
if bytesToTrim != 0 {
|
||||
glog.Warningf(
|
||||
"Truncating file %s to exactly %d chunks, trimming %d extraneous bytes.",
|
||||
filename, chunksInFile, bytesToTrim,
|
||||
)
|
||||
f, err := os.OpenFile(filename, os.O_WRONLY, 0640)
|
||||
if err != nil {
|
||||
glog.Errorf("Could not open file %s: %s", filename, err)
|
||||
purge()
|
||||
return fp, false
|
||||
}
|
||||
if err := f.Truncate(fi.Size() - bytesToTrim); err != nil {
|
||||
glog.Errorf("Failed to truncate file %s: %s", filename, err)
|
||||
purge()
|
||||
return fp, false
|
||||
}
|
||||
}
|
||||
if chunksInFile == 0 {
|
||||
glog.Warningf("No chunks left in file %s.", filename)
|
||||
purge()
|
||||
return fp, false
|
||||
}
|
||||
|
||||
s, ok := fingerprintToSeries[fp]
|
||||
if ok { // This series is supposed to not be archived.
|
||||
if s == nil {
|
||||
panic("fingerprint mapped to nil pointer")
|
||||
}
|
||||
if bytesToTrim == 0 && s.chunkDescsOffset != -1 &&
|
||||
((s.headChunkPersisted && chunksInFile == s.chunkDescsOffset+len(s.chunkDescs)) ||
|
||||
(!s.headChunkPersisted && chunksInFile == s.chunkDescsOffset+len(s.chunkDescs)-1)) {
|
||||
// Everything is consistent. We are good.
|
||||
return fp, true
|
||||
}
|
||||
// If we are here, something's fishy.
|
||||
if s.headChunkPersisted {
|
||||
// This is the easy case as we don't have a head chunk
|
||||
// in heads.db. Treat this series as a freshly
|
||||
// unarchived one. No chunks or chunkDescs in memory, no
|
||||
// current head chunk.
|
||||
glog.Warningf(
|
||||
"Treating recovered metric %v, fingerprint %v, as freshly unarchived, with %d chunks in series file.",
|
||||
s.metric, fp, chunksInFile,
|
||||
)
|
||||
s.chunkDescs = nil
|
||||
s.chunkDescsOffset = -1
|
||||
return fp, true
|
||||
}
|
||||
// This is the tricky one: We have a head chunk from heads.db,
|
||||
// but the very same head chunk might already be in the series
|
||||
// file. Strategy: Check the first time of both. If it is the
|
||||
// same or newer, assume the latest chunk in the series file
|
||||
// is the most recent head chunk. If not, keep the head chunk
|
||||
// we got from heads.db.
|
||||
// First, assume the head chunk is not yet persisted.
|
||||
s.chunkDescs = s.chunkDescs[len(s.chunkDescs)-1:]
|
||||
s.chunkDescsOffset = -1
|
||||
// Load all the chunk descs (which assumes we have none from the future).
|
||||
cds, err := p.loadChunkDescs(fp, clientmodel.Now())
|
||||
if err != nil {
|
||||
glog.Errorf(
|
||||
"Failed to load chunk descriptors for metric %v, fingerprint %v: %s",
|
||||
s.metric, fp, err,
|
||||
)
|
||||
purge()
|
||||
return fp, false
|
||||
}
|
||||
if cds[len(cds)-1].firstTime().Before(s.head().firstTime()) {
|
||||
s.chunkDescs = append(cds, s.chunkDescs...)
|
||||
glog.Warningf(
|
||||
"Recovered metric %v, fingerprint %v: recovered %d chunks from series file, recovered head chunk from checkpoint.",
|
||||
s.metric, fp, chunksInFile,
|
||||
)
|
||||
} else {
|
||||
glog.Warningf(
|
||||
"Recovered metric %v, fingerprint %v: head chunk found among the %d recovered chunks in series file.",
|
||||
s.metric, fp, chunksInFile,
|
||||
)
|
||||
s.chunkDescs = cds
|
||||
s.headChunkPersisted = true
|
||||
}
|
||||
s.chunkDescsOffset = 0
|
||||
return fp, true
|
||||
}
|
||||
// This series is supposed to be archived.
|
||||
metric, err := p.getArchivedMetric(fp)
|
||||
if err != nil {
|
||||
glog.Errorf(
|
||||
"Fingerprint %v assumed archived but couldn't be looked up in archived index: %s",
|
||||
fp, err,
|
||||
)
|
||||
purge()
|
||||
return fp, false
|
||||
}
|
||||
if metric == nil {
|
||||
glog.Warningf(
|
||||
"Fingerprint %v assumed archived but couldn't be found in archived index.",
|
||||
fp,
|
||||
)
|
||||
purge()
|
||||
return fp, false
|
||||
}
|
||||
// This series looks like a properly archived one.
|
||||
return fp, true
|
||||
}
|
||||
|
||||
func (p *persistence) cleanUpArchiveIndexes(
|
||||
fpToSeries map[clientmodel.Fingerprint]*memorySeries,
|
||||
fpsSeen map[clientmodel.Fingerprint]struct{},
|
||||
) error {
|
||||
glog.Info("Cleaning up archive indexes.")
|
||||
var fp codable.Fingerprint
|
||||
var m codable.Metric
|
||||
count := 0
|
||||
if err := p.archivedFingerprintToMetrics.ForEach(func(kv index.KeyValueAccessor) error {
|
||||
count++
|
||||
if count%10000 == 0 {
|
||||
glog.Infof("%d archived metrics checked.", count)
|
||||
}
|
||||
if err := kv.Key(&fp); err != nil {
|
||||
return err
|
||||
}
|
||||
_, fpSeen := fpsSeen[clientmodel.Fingerprint(fp)]
|
||||
inMemory := false
|
||||
if fpSeen {
|
||||
_, inMemory = fpToSeries[clientmodel.Fingerprint(fp)]
|
||||
}
|
||||
if !fpSeen || inMemory {
|
||||
if inMemory {
|
||||
glog.Warningf("Archive clean-up: Fingerprint %v is not archived. Purging from archive indexes.", clientmodel.Fingerprint(fp))
|
||||
}
|
||||
if !fpSeen {
|
||||
glog.Warningf("Archive clean-up: Fingerprint %v is unknown. Purging from archive indexes.", clientmodel.Fingerprint(fp))
|
||||
}
|
||||
// It's fine if the fp is not in the archive indexes.
|
||||
if _, err := p.archivedFingerprintToMetrics.Delete(fp); err != nil {
|
||||
return err
|
||||
}
|
||||
// Delete from timerange index, too.
|
||||
_, err := p.archivedFingerprintToTimeRange.Delete(fp)
|
||||
return err
|
||||
}
|
||||
// fp is legitimately archived. Make sure it is in timerange index, too.
|
||||
has, err := p.archivedFingerprintToTimeRange.Has(fp)
|
||||
if err != nil {
|
||||
return err
|
||||
}
|
||||
if has {
|
||||
return nil // All good.
|
||||
}
|
||||
glog.Warningf("Archive clean-up: Fingerprint %v is not in time-range index. Unarchiving it for recovery.")
|
||||
// Again, it's fine if fp is not in the archive index.
|
||||
if _, err := p.archivedFingerprintToMetrics.Delete(fp); err != nil {
|
||||
return err
|
||||
}
|
||||
if err := kv.Value(&m); err != nil {
|
||||
return err
|
||||
}
|
||||
series := newMemorySeries(clientmodel.Metric(m), false, math.MinInt64)
|
||||
cds, err := p.loadChunkDescs(clientmodel.Fingerprint(fp), clientmodel.Now())
|
||||
if err != nil {
|
||||
return err
|
||||
}
|
||||
series.chunkDescs = cds
|
||||
series.chunkDescsOffset = 0
|
||||
fpToSeries[clientmodel.Fingerprint(fp)] = series
|
||||
return nil
|
||||
}); err != nil {
|
||||
return err
|
||||
}
|
||||
count = 0
|
||||
if err := p.archivedFingerprintToTimeRange.ForEach(func(kv index.KeyValueAccessor) error {
|
||||
count++
|
||||
if count%10000 == 0 {
|
||||
glog.Infof("%d archived time ranges checked.", count)
|
||||
}
|
||||
if err := kv.Key(&fp); err != nil {
|
||||
return err
|
||||
}
|
||||
has, err := p.archivedFingerprintToMetrics.Has(fp)
|
||||
if err != nil {
|
||||
return err
|
||||
}
|
||||
if has {
|
||||
return nil // All good.
|
||||
}
|
||||
glog.Warningf("Archive clean-up: Purging unknown fingerprint %v in time-range index.", fp)
|
||||
deleted, err := p.archivedFingerprintToTimeRange.Delete(fp)
|
||||
if err != nil {
|
||||
return err
|
||||
}
|
||||
if !deleted {
|
||||
glog.Errorf("Fingerprint %s to be deleted from archivedFingerprintToTimeRange not found. This should never happen.", fp)
|
||||
}
|
||||
return nil
|
||||
}); err != nil {
|
||||
return err
|
||||
}
|
||||
glog.Info("Clean-up of archive indexes complete.")
|
||||
return nil
|
||||
}
|
||||
|
||||
func (p *persistence) rebuildLabelIndexes(
|
||||
fpToSeries map[clientmodel.Fingerprint]*memorySeries,
|
||||
) error {
|
||||
count := 0
|
||||
glog.Info("Rebuilding label indexes.")
|
||||
glog.Info("Indexing metrics in memory.")
|
||||
for fp, s := range fpToSeries {
|
||||
p.indexMetric(fp, s.metric)
|
||||
count++
|
||||
if count%10000 == 0 {
|
||||
glog.Infof("%d metrics queued for indexing.", count)
|
||||
}
|
||||
}
|
||||
glog.Info("Indexing archived metrics.")
|
||||
var fp codable.Fingerprint
|
||||
var m codable.Metric
|
||||
if err := p.archivedFingerprintToMetrics.ForEach(func(kv index.KeyValueAccessor) error {
|
||||
if err := kv.Key(&fp); err != nil {
|
||||
return err
|
||||
}
|
||||
if err := kv.Value(&m); err != nil {
|
||||
return err
|
||||
}
|
||||
p.indexMetric(clientmodel.Fingerprint(fp), clientmodel.Metric(m))
|
||||
count++
|
||||
if count%10000 == 0 {
|
||||
glog.Infof("%d metrics queued for indexing.", count)
|
||||
}
|
||||
return nil
|
||||
}); err != nil {
|
||||
return err
|
||||
}
|
||||
glog.Info("All requests for rebuilding the label indexes queued. (Actual processing may lag behind.)")
|
||||
return nil
|
||||
}
|
|
@ -35,7 +35,6 @@ const (
|
|||
)
|
||||
|
||||
var (
|
||||
// TODO: Tweak default values.
|
||||
fingerprintToMetricCacheSize = flag.Int("storage.local.index-cache-size.fingerprint-to-metric", 10*1024*1024, "The size in bytes for the fingerprint to metric index cache.")
|
||||
fingerprintTimeRangeCacheSize = flag.Int("storage.local.index-cache-size.fingerprint-to-timerange", 5*1024*1024, "The size in bytes for the metric time range index cache.")
|
||||
labelNameToLabelValuesCacheSize = flag.Int("storage.local.index-cache-size.label-name-to-label-values", 10*1024*1024, "The size in bytes for the label name to label values index cache.")
|
||||
|
|
|
@ -29,8 +29,8 @@ type KeyValueStore interface {
|
|||
// could be found for key. If value is nil, Get behaves like Has.
|
||||
Get(key encoding.BinaryMarshaler, value encoding.BinaryUnmarshaler) (bool, error)
|
||||
Has(key encoding.BinaryMarshaler) (bool, error)
|
||||
// Delete returns an error if key does not exist.
|
||||
Delete(key encoding.BinaryMarshaler) error
|
||||
// Delete returns (false, nil) if key does not exist.
|
||||
Delete(key encoding.BinaryMarshaler) (bool, error)
|
||||
|
||||
NewBatch() Batch
|
||||
Commit(b Batch) error
|
||||
|
|
|
@ -104,12 +104,19 @@ func (l *LevelDB) Has(key encoding.BinaryMarshaler) (has bool, err error) {
|
|||
}
|
||||
|
||||
// Delete implements KeyValueStore.
|
||||
func (l *LevelDB) Delete(key encoding.BinaryMarshaler) error {
|
||||
func (l *LevelDB) Delete(key encoding.BinaryMarshaler) (bool, error) {
|
||||
k, err := key.MarshalBinary()
|
||||
if err != nil {
|
||||
return err
|
||||
return false, err
|
||||
}
|
||||
return l.storage.Delete(k, l.writeOpts)
|
||||
err = l.storage.Delete(k, l.writeOpts)
|
||||
if err == leveldb.ErrNotFound {
|
||||
return false, nil
|
||||
}
|
||||
if err != nil {
|
||||
return false, err
|
||||
}
|
||||
return true, nil
|
||||
}
|
||||
|
||||
// Put implements KeyValueStore.
|
||||
|
|
|
@ -18,11 +18,9 @@ import (
|
|||
"encoding/binary"
|
||||
"fmt"
|
||||
"io"
|
||||
"math"
|
||||
"os"
|
||||
"path"
|
||||
"path/filepath"
|
||||
"strings"
|
||||
"sync"
|
||||
"sync/atomic"
|
||||
"time"
|
||||
|
@ -261,382 +259,6 @@ func (p *persistence) setDirty(dirty bool) {
|
|||
}
|
||||
}
|
||||
|
||||
// recoverFromCrash is called by loadSeriesMapAndHeads if the persistence
|
||||
// appears to be dirty after the loading (either because the loading resulted in
|
||||
// an error or because the persistence was dirty from the start). Not goroutine
|
||||
// safe. Only call before anything else is running (except index processing
|
||||
// queue as started by newPersistence).
|
||||
func (p *persistence) recoverFromCrash(fingerprintToSeries map[clientmodel.Fingerprint]*memorySeries) error {
|
||||
// TODO(beorn): We need proper tests for the crash recovery.
|
||||
glog.Warning("Starting crash recovery. Prometheus is inoperational until complete.")
|
||||
|
||||
fpsSeen := map[clientmodel.Fingerprint]struct{}{}
|
||||
count := 0
|
||||
seriesDirNameFmt := fmt.Sprintf("%%0%dx", seriesDirNameLen)
|
||||
|
||||
glog.Info("Scanning files.")
|
||||
for i := 0; i < 1<<(seriesDirNameLen*4); i++ {
|
||||
dirname := path.Join(p.basePath, fmt.Sprintf(seriesDirNameFmt, i))
|
||||
dir, err := os.Open(dirname)
|
||||
if os.IsNotExist(err) {
|
||||
continue
|
||||
}
|
||||
if err != nil {
|
||||
return err
|
||||
}
|
||||
defer dir.Close()
|
||||
for fis := []os.FileInfo{}; err != io.EOF; fis, err = dir.Readdir(1024) {
|
||||
if err != nil {
|
||||
return err
|
||||
}
|
||||
for _, fi := range fis {
|
||||
fp, ok := p.sanitizeSeries(dirname, fi, fingerprintToSeries)
|
||||
if ok {
|
||||
fpsSeen[fp] = struct{}{}
|
||||
}
|
||||
count++
|
||||
if count%10000 == 0 {
|
||||
glog.Infof("%d files scanned.", count)
|
||||
}
|
||||
}
|
||||
}
|
||||
}
|
||||
glog.Infof("File scan complete. %d series found.", len(fpsSeen))
|
||||
|
||||
glog.Info("Checking for series without series file.")
|
||||
for fp, s := range fingerprintToSeries {
|
||||
if _, seen := fpsSeen[fp]; !seen {
|
||||
// fp exists in fingerprintToSeries, but has no representation on disk.
|
||||
if s.headChunkPersisted {
|
||||
// Oops, head chunk was persisted, but nothing on disk.
|
||||
// Thus, we lost that series completely. Clean up the remnants.
|
||||
delete(fingerprintToSeries, fp)
|
||||
if err := p.dropArchivedMetric(fp); err != nil {
|
||||
// Dropping the archived metric didn't work, so try
|
||||
// to unindex it, just in case it's in the indexes.
|
||||
p.unindexMetric(fp, s.metric)
|
||||
}
|
||||
glog.Warningf("Lost series detected: fingerprint %v, metric %v.", fp, s.metric)
|
||||
continue
|
||||
}
|
||||
// If we are here, the only chunk we have is the head chunk.
|
||||
// Adjust things accordingly.
|
||||
if len(s.chunkDescs) > 1 || s.chunkDescsOffset != 0 {
|
||||
minLostChunks := len(s.chunkDescs) + s.chunkDescsOffset - 1
|
||||
if minLostChunks <= 0 {
|
||||
glog.Warningf(
|
||||
"Possible loss of chunks for fingerprint %v, metric %v.",
|
||||
fp, s.metric,
|
||||
)
|
||||
} else {
|
||||
glog.Warningf(
|
||||
"Lost at least %d chunks for fingerprint %v, metric %v.",
|
||||
minLostChunks, fp, s.metric,
|
||||
)
|
||||
}
|
||||
s.chunkDescs = s.chunkDescs[len(s.chunkDescs)-1:]
|
||||
s.chunkDescsOffset = 0
|
||||
}
|
||||
fpsSeen[fp] = struct{}{} // Add so that fpsSeen is complete.
|
||||
}
|
||||
}
|
||||
glog.Info("Check for series without series file complete.")
|
||||
|
||||
if err := p.cleanUpArchiveIndexes(fingerprintToSeries, fpsSeen); err != nil {
|
||||
return err
|
||||
}
|
||||
if err := p.rebuildLabelIndexes(fingerprintToSeries); err != nil {
|
||||
return err
|
||||
}
|
||||
|
||||
p.setDirty(false)
|
||||
glog.Warning("Crash recovery complete.")
|
||||
return nil
|
||||
}
|
||||
|
||||
// sanitizeSeries sanitizes a series based on its series file as defined by the
|
||||
// provided directory and FileInfo. The method returns the fingerprint as
|
||||
// derived from the directory and file name, and whether the provided file has
|
||||
// been sanitized. A file that failed to be sanitized is deleted, if possible.
|
||||
//
|
||||
// The following steps are performed:
|
||||
//
|
||||
// - A file whose name doesn't comply with the naming scheme of a series file is
|
||||
// simply deleted.
|
||||
//
|
||||
// - If the size of the series file isn't a multiple of the chunk size,
|
||||
// extraneous bytes are truncated. If the truncation fails, the file is
|
||||
// deleted instead.
|
||||
//
|
||||
// - A file that is empty (after truncation) is deleted.
|
||||
//
|
||||
// - A series that is not archived (i.e. it is in the fingerprintToSeries map)
|
||||
// is checked for consistency of its various parameters (like head-chunk
|
||||
// persistence state, offset of chunkDescs etc.). In particular, overlap
|
||||
// between an in-memory head chunk with the most recent persisted chunk is
|
||||
// checked. Inconsistencies are rectified.
|
||||
//
|
||||
// - A series this in archived (i.e. it is not in the fingerprintToSeries map)
|
||||
// is checked for its presence in the index of archived series. If it cannot
|
||||
// be found there, it is deleted.
|
||||
func (p *persistence) sanitizeSeries(dirname string, fi os.FileInfo, fingerprintToSeries map[clientmodel.Fingerprint]*memorySeries) (clientmodel.Fingerprint, bool) {
|
||||
filename := path.Join(dirname, fi.Name())
|
||||
purge := func() {
|
||||
glog.Warningf("Deleting lost series file %s.", filename) // TODO: Move to lost+found directory?
|
||||
os.Remove(filename)
|
||||
}
|
||||
|
||||
var fp clientmodel.Fingerprint
|
||||
if len(fi.Name()) != fpLen-seriesDirNameLen+len(seriesFileSuffix) ||
|
||||
!strings.HasSuffix(fi.Name(), seriesFileSuffix) {
|
||||
glog.Warningf("Unexpected series file name %s.", filename)
|
||||
purge()
|
||||
return fp, false
|
||||
}
|
||||
if err := fp.LoadFromString(path.Base(dirname) + fi.Name()[:fpLen-seriesDirNameLen]); err != nil {
|
||||
glog.Warningf("Error parsing file name %s: %s", filename, err)
|
||||
purge()
|
||||
return fp, false
|
||||
}
|
||||
|
||||
bytesToTrim := fi.Size() % int64(p.chunkLen+chunkHeaderLen)
|
||||
chunksInFile := int(fi.Size()) / (p.chunkLen + chunkHeaderLen)
|
||||
if bytesToTrim != 0 {
|
||||
glog.Warningf(
|
||||
"Truncating file %s to exactly %d chunks, trimming %d extraneous bytes.",
|
||||
filename, chunksInFile, bytesToTrim,
|
||||
)
|
||||
f, err := os.OpenFile(filename, os.O_WRONLY, 0640)
|
||||
if err != nil {
|
||||
glog.Errorf("Could not open file %s: %s", filename, err)
|
||||
purge()
|
||||
return fp, false
|
||||
}
|
||||
if err := f.Truncate(fi.Size() - bytesToTrim); err != nil {
|
||||
glog.Errorf("Failed to truncate file %s: %s", filename, err)
|
||||
purge()
|
||||
return fp, false
|
||||
}
|
||||
}
|
||||
if chunksInFile == 0 {
|
||||
glog.Warningf("No chunks left in file %s.", filename)
|
||||
purge()
|
||||
return fp, false
|
||||
}
|
||||
|
||||
s, ok := fingerprintToSeries[fp]
|
||||
if ok { // This series is supposed to not be archived.
|
||||
if s == nil {
|
||||
panic("fingerprint mapped to nil pointer")
|
||||
}
|
||||
if bytesToTrim == 0 && s.chunkDescsOffset != -1 &&
|
||||
((s.headChunkPersisted && chunksInFile == s.chunkDescsOffset+len(s.chunkDescs)) ||
|
||||
(!s.headChunkPersisted && chunksInFile == s.chunkDescsOffset+len(s.chunkDescs)-1)) {
|
||||
// Everything is consistent. We are good.
|
||||
return fp, true
|
||||
}
|
||||
// If we are here, something's fishy.
|
||||
if s.headChunkPersisted {
|
||||
// This is the easy case as we don't have a head chunk
|
||||
// in heads.db. Treat this series as a freshly
|
||||
// unarchived one. No chunks or chunkDescs in memory, no
|
||||
// current head chunk.
|
||||
glog.Warningf(
|
||||
"Treating recovered metric %v, fingerprint %v, as freshly unarchived, with %d chunks in series file.",
|
||||
s.metric, fp, chunksInFile,
|
||||
)
|
||||
s.chunkDescs = nil
|
||||
s.chunkDescsOffset = -1
|
||||
return fp, true
|
||||
}
|
||||
// This is the tricky one: We have a head chunk from heads.db,
|
||||
// but the very same head chunk might already be in the series
|
||||
// file. Strategy: Check the first time of both. If it is the
|
||||
// same or newer, assume the latest chunk in the series file
|
||||
// is the most recent head chunk. If not, keep the head chunk
|
||||
// we got from heads.db.
|
||||
// First, assume the head chunk is not yet persisted.
|
||||
s.chunkDescs = s.chunkDescs[len(s.chunkDescs)-1:]
|
||||
s.chunkDescsOffset = -1
|
||||
// Load all the chunk descs (which assumes we have none from the future).
|
||||
cds, err := p.loadChunkDescs(fp, clientmodel.Now())
|
||||
if err != nil {
|
||||
glog.Errorf(
|
||||
"Failed to load chunk descriptors for metric %v, fingerprint %v: %s",
|
||||
s.metric, fp, err,
|
||||
)
|
||||
purge()
|
||||
return fp, false
|
||||
}
|
||||
if cds[len(cds)-1].firstTime().Before(s.head().firstTime()) {
|
||||
s.chunkDescs = append(cds, s.chunkDescs...)
|
||||
glog.Warningf(
|
||||
"Recovered metric %v, fingerprint %v: recovered %d chunks from series file, recovered head chunk from checkpoint.",
|
||||
s.metric, fp, chunksInFile,
|
||||
)
|
||||
} else {
|
||||
glog.Warningf(
|
||||
"Recovered metric %v, fingerprint %v: head chunk found among the %d recovered chunks in series file.",
|
||||
s.metric, fp, chunksInFile,
|
||||
)
|
||||
s.chunkDescs = cds
|
||||
s.headChunkPersisted = true
|
||||
}
|
||||
s.chunkDescsOffset = 0
|
||||
return fp, true
|
||||
}
|
||||
// This series is supposed to be archived.
|
||||
metric, err := p.getArchivedMetric(fp)
|
||||
if err != nil {
|
||||
glog.Errorf(
|
||||
"Fingerprint %v assumed archived but couldn't be looked up in archived index: %s",
|
||||
fp, err,
|
||||
)
|
||||
purge()
|
||||
return fp, false
|
||||
}
|
||||
if metric == nil {
|
||||
glog.Warningf(
|
||||
"Fingerprint %v assumed archived but couldn't be found in archived index.",
|
||||
fp,
|
||||
)
|
||||
purge()
|
||||
return fp, false
|
||||
}
|
||||
// This series looks like a properly archived one.
|
||||
return fp, true
|
||||
}
|
||||
|
||||
func (p *persistence) cleanUpArchiveIndexes(
|
||||
fpToSeries map[clientmodel.Fingerprint]*memorySeries,
|
||||
fpsSeen map[clientmodel.Fingerprint]struct{},
|
||||
) error {
|
||||
glog.Info("Cleaning up archive indexes.")
|
||||
var fp codable.Fingerprint
|
||||
var m codable.Metric
|
||||
count := 0
|
||||
if err := p.archivedFingerprintToMetrics.ForEach(func(kv index.KeyValueAccessor) error {
|
||||
count++
|
||||
if count%10000 == 0 {
|
||||
glog.Infof("%d archived metrics checked.", count)
|
||||
}
|
||||
if err := kv.Key(&fp); err != nil {
|
||||
return err
|
||||
}
|
||||
_, fpSeen := fpsSeen[clientmodel.Fingerprint(fp)]
|
||||
inMemory := false
|
||||
if fpSeen {
|
||||
_, inMemory = fpToSeries[clientmodel.Fingerprint(fp)]
|
||||
}
|
||||
if !fpSeen || inMemory {
|
||||
if inMemory {
|
||||
glog.Warningf("Archive clean-up: Fingerprint %v is not archived. Purging from archive indexes.", clientmodel.Fingerprint(fp))
|
||||
}
|
||||
if !fpSeen {
|
||||
glog.Warningf("Archive clean-up: Fingerprint %v is unknown. Purging from archive indexes.", clientmodel.Fingerprint(fp))
|
||||
}
|
||||
if err := p.archivedFingerprintToMetrics.Delete(fp); err != nil {
|
||||
return err
|
||||
}
|
||||
// Delete from timerange index, too.
|
||||
p.archivedFingerprintToTimeRange.Delete(fp)
|
||||
// TODO: Ignoring errors here as fp might not be in
|
||||
// timerange index (which is good) but which would
|
||||
// return an error. Delete signature could be changed
|
||||
// like the Get signature to detect a real error.
|
||||
return nil
|
||||
}
|
||||
// fp is legitimately archived. Make sure it is in timerange index, too.
|
||||
has, err := p.archivedFingerprintToTimeRange.Has(fp)
|
||||
if err != nil {
|
||||
return err
|
||||
}
|
||||
if has {
|
||||
return nil // All good.
|
||||
}
|
||||
glog.Warningf("Archive clean-up: Fingerprint %v is not in time-range index. Unarchiving it for recovery.")
|
||||
if err := p.archivedFingerprintToMetrics.Delete(fp); err != nil {
|
||||
return err
|
||||
}
|
||||
if err := kv.Value(&m); err != nil {
|
||||
return err
|
||||
}
|
||||
series := newMemorySeries(clientmodel.Metric(m), false, math.MinInt64)
|
||||
cds, err := p.loadChunkDescs(clientmodel.Fingerprint(fp), clientmodel.Now())
|
||||
if err != nil {
|
||||
return err
|
||||
}
|
||||
series.chunkDescs = cds
|
||||
series.chunkDescsOffset = 0
|
||||
fpToSeries[clientmodel.Fingerprint(fp)] = series
|
||||
return nil
|
||||
}); err != nil {
|
||||
return err
|
||||
}
|
||||
count = 0
|
||||
if err := p.archivedFingerprintToTimeRange.ForEach(func(kv index.KeyValueAccessor) error {
|
||||
count++
|
||||
if count%10000 == 0 {
|
||||
glog.Infof("%d archived time ranges checked.", count)
|
||||
}
|
||||
if err := kv.Key(&fp); err != nil {
|
||||
return err
|
||||
}
|
||||
has, err := p.archivedFingerprintToMetrics.Has(fp)
|
||||
if err != nil {
|
||||
return err
|
||||
}
|
||||
if has {
|
||||
return nil // All good.
|
||||
}
|
||||
glog.Warningf("Archive clean-up: Purging unknown fingerprint %v in time-range index.", fp)
|
||||
if err := p.archivedFingerprintToTimeRange.Delete(fp); err != nil {
|
||||
return err
|
||||
}
|
||||
return nil
|
||||
}); err != nil {
|
||||
return err
|
||||
}
|
||||
glog.Info("Clean-up of archive indexes complete.")
|
||||
return nil
|
||||
}
|
||||
|
||||
func (p *persistence) rebuildLabelIndexes(
|
||||
fpToSeries map[clientmodel.Fingerprint]*memorySeries,
|
||||
) error {
|
||||
count := 0
|
||||
glog.Info("Rebuilding label indexes.")
|
||||
glog.Info("Indexing metrics in memory.")
|
||||
for fp, s := range fpToSeries {
|
||||
p.indexMetric(fp, s.metric)
|
||||
count++
|
||||
if count%10000 == 0 {
|
||||
glog.Infof("%d metrics queued for indexing.", count)
|
||||
}
|
||||
}
|
||||
glog.Info("Indexing archived metrics.")
|
||||
var fp codable.Fingerprint
|
||||
var m codable.Metric
|
||||
if err := p.archivedFingerprintToMetrics.ForEach(func(kv index.KeyValueAccessor) error {
|
||||
if err := kv.Key(&fp); err != nil {
|
||||
return err
|
||||
}
|
||||
if err := kv.Value(&m); err != nil {
|
||||
return err
|
||||
}
|
||||
p.indexMetric(clientmodel.Fingerprint(fp), clientmodel.Metric(m))
|
||||
count++
|
||||
if count%10000 == 0 {
|
||||
glog.Infof("%d metrics queued for indexing.", count)
|
||||
}
|
||||
return nil
|
||||
}); err != nil {
|
||||
return err
|
||||
}
|
||||
glog.Info("All requests for rebuilding the label indexes queued. (Actual processing may lag behind.)")
|
||||
return nil
|
||||
}
|
||||
|
||||
// getFingerprintsForLabelPair returns the fingerprints for the given label
|
||||
// pair. This method is goroutine-safe but take into account that metrics queued
|
||||
// for indexing with IndexMetric might not have made it into the index
|
||||
|
@ -1303,12 +925,20 @@ func (p *persistence) dropArchivedMetric(fp clientmodel.Fingerprint) (err error)
|
|||
if err != nil || metric == nil {
|
||||
return err
|
||||
}
|
||||
if err := p.archivedFingerprintToMetrics.Delete(codable.Fingerprint(fp)); err != nil {
|
||||
deleted, err := p.archivedFingerprintToMetrics.Delete(codable.Fingerprint(fp))
|
||||
if err != nil {
|
||||
return err
|
||||
}
|
||||
if err := p.archivedFingerprintToTimeRange.Delete(codable.Fingerprint(fp)); err != nil {
|
||||
if !deleted {
|
||||
glog.Errorf("Tried to delete non-archived fingerprint %s from archivedFingerprintToMetrics index. This should never happen.", fp)
|
||||
}
|
||||
deleted, err = p.archivedFingerprintToTimeRange.Delete(codable.Fingerprint(fp))
|
||||
if err != nil {
|
||||
return err
|
||||
}
|
||||
if !deleted {
|
||||
glog.Errorf("Tried to delete non-archived fingerprint %s from archivedFingerprintToTimeRange index. This should never happen.", fp)
|
||||
}
|
||||
p.unindexMetric(fp, metric)
|
||||
return nil
|
||||
}
|
||||
|
@ -1332,12 +962,20 @@ func (p *persistence) unarchiveMetric(fp clientmodel.Fingerprint) (
|
|||
if err != nil || !has {
|
||||
return false, firstTime, err
|
||||
}
|
||||
if err := p.archivedFingerprintToMetrics.Delete(codable.Fingerprint(fp)); err != nil {
|
||||
deleted, err := p.archivedFingerprintToMetrics.Delete(codable.Fingerprint(fp))
|
||||
if err != nil {
|
||||
return false, firstTime, err
|
||||
}
|
||||
if err := p.archivedFingerprintToTimeRange.Delete(codable.Fingerprint(fp)); err != nil {
|
||||
if !deleted {
|
||||
glog.Errorf("Tried to delete non-archived fingerprint %s from archivedFingerprintToMetrics index. This should never happen.", fp)
|
||||
}
|
||||
deleted, err = p.archivedFingerprintToTimeRange.Delete(codable.Fingerprint(fp))
|
||||
if err != nil {
|
||||
return false, firstTime, err
|
||||
}
|
||||
if !deleted {
|
||||
glog.Errorf("Tried to delete non-archived fingerprint %s from archivedFingerprintToTimeRange index. This should never happen.", fp)
|
||||
}
|
||||
return true, firstTime, nil
|
||||
}
|
||||
|
||||
|
@ -1544,16 +1182,3 @@ loop:
|
|||
}
|
||||
close(p.indexingStopped)
|
||||
}
|
||||
|
||||
// exists returns true when the given file or directory exists.
|
||||
func exists(path string) (bool, error) {
|
||||
_, err := os.Stat(path)
|
||||
if err == nil {
|
||||
return true, nil
|
||||
}
|
||||
if os.IsNotExist(err) {
|
||||
return false, nil
|
||||
}
|
||||
|
||||
return false, err
|
||||
}
|
||||
|
|
Loading…
Reference in a new issue