mirror of
https://github.com/prometheus/prometheus.git
synced 2025-01-13 14:57:40 -08:00
7e85711df0
This reintroduces a LevelDB-based metrics index. Change-Id: I4111540301c52255a07b2f570761707a32f72c05
420 lines
12 KiB
Go
420 lines
12 KiB
Go
package storage_ng
|
|
|
|
import (
|
|
"bufio"
|
|
"encoding/binary"
|
|
"flag"
|
|
"fmt"
|
|
"io"
|
|
"os"
|
|
"path"
|
|
|
|
"github.com/golang/glog"
|
|
|
|
//"github.com/prometheus/client_golang/prometheus"
|
|
|
|
clientmodel "github.com/prometheus/client_golang/model"
|
|
|
|
"github.com/prometheus/prometheus/storage/local/index"
|
|
)
|
|
|
|
const (
|
|
seriesFileName = "series.db"
|
|
seriesTempFileName = "series.db.tmp"
|
|
headsFileName = "heads.db"
|
|
indexDirName = "index"
|
|
|
|
fingerprintToMetricDir = "fingerprint_to_metric"
|
|
labelNameToLabelValuesDir = "labelname_to_labelvalues"
|
|
labelPairToFingerprintsDir = "labelpair_to_fingerprints"
|
|
fingerprintMembershipDir = "fingerprint_membership"
|
|
|
|
chunkHeaderLen = 17
|
|
chunkHeaderTypeOffset = 0
|
|
chunkHeaderFirstTimeOffset = 1
|
|
chunkHeaderLastTimeOffset = 9
|
|
|
|
headsHeaderLen = 9
|
|
headsHeaderFingerprintOffset = 0
|
|
headsHeaderTypeOffset = 8
|
|
)
|
|
|
|
var (
|
|
fingerprintToMetricCacheSize = flag.Int("storage.fingerprintToMetricCacheSizeBytes", 25*1024*1024, "The size in bytes for the fingerprint to metric index cache.")
|
|
labelNameToLabelValuesCacheSize = flag.Int("storage.labelNameToLabelValuesCacheSizeBytes", 25*1024*1024, "The size in bytes for the label name to label values index.")
|
|
labelPairToFingerprintsCacheSize = flag.Int("storage.labelPairToFingerprintsCacheSizeBytes", 25*1024*1024, "The size in bytes for the label pair to fingerprints index.")
|
|
fingerprintMembershipCacheSize = flag.Int("storage.fingerprintMembershipCacheSizeBytes", 5*1024*1024, "The size in bytes for the metric membership index.")
|
|
)
|
|
|
|
type diskPersistence struct {
|
|
index.MetricIndexer
|
|
indexDBs []index.KeyValueStore
|
|
|
|
basePath string
|
|
chunkLen int
|
|
buf []byte // Staging space for persisting indexes.
|
|
}
|
|
|
|
func NewDiskPersistence(basePath string, chunkLen int) (Persistence, error) {
|
|
err := os.MkdirAll(basePath, 0700)
|
|
if err != nil {
|
|
return nil, err
|
|
}
|
|
|
|
fingerprintToMetricDB, err := index.NewLevelDB(index.LevelDBOptions{
|
|
Path: path.Join(basePath, fingerprintToMetricDir),
|
|
CacheSizeBytes: *fingerprintToMetricCacheSize,
|
|
})
|
|
if err != nil {
|
|
return nil, err
|
|
}
|
|
labelNameToLabelValuesDB, err := index.NewLevelDB(index.LevelDBOptions{
|
|
Path: path.Join(basePath, labelNameToLabelValuesDir),
|
|
CacheSizeBytes: *labelNameToLabelValuesCacheSize,
|
|
})
|
|
if err != nil {
|
|
return nil, err
|
|
}
|
|
labelPairToFingerprintsDB, err := index.NewLevelDB(index.LevelDBOptions{
|
|
Path: path.Join(basePath, labelPairToFingerprintsDir),
|
|
CacheSizeBytes: *labelPairToFingerprintsCacheSize,
|
|
})
|
|
if err != nil {
|
|
return nil, err
|
|
}
|
|
fingerprintMembershipDB, err := index.NewLevelDB(index.LevelDBOptions{
|
|
Path: path.Join(basePath, fingerprintMembershipDir),
|
|
CacheSizeBytes: *fingerprintMembershipCacheSize,
|
|
})
|
|
if err != nil {
|
|
return nil, err
|
|
}
|
|
|
|
return &diskPersistence{
|
|
basePath: basePath,
|
|
chunkLen: chunkLen,
|
|
buf: make([]byte, binary.MaxVarintLen64), // Also sufficient for uint64.
|
|
MetricIndexer: &index.TotalIndexer{
|
|
FingerprintToMetric: index.NewFingerprintMetricIndex(fingerprintToMetricDB),
|
|
LabelNameToLabelValues: index.NewLabelNameLabelValuesIndex(labelNameToLabelValuesDB),
|
|
LabelPairToFingerprints: index.NewLabelPairFingerprintIndex(labelPairToFingerprintsDB),
|
|
FingerprintMembership: index.NewFingerprintMembershipIndex(fingerprintMembershipDB),
|
|
},
|
|
indexDBs: []index.KeyValueStore{
|
|
fingerprintToMetricDB,
|
|
labelNameToLabelValuesDB,
|
|
labelPairToFingerprintsDB,
|
|
fingerprintMembershipDB,
|
|
},
|
|
}, nil
|
|
}
|
|
|
|
func (p *diskPersistence) dirForFingerprint(fp clientmodel.Fingerprint) string {
|
|
fpStr := fp.String()
|
|
return fmt.Sprintf("%s/%c%c/%s", p.basePath, fpStr[0], fpStr[1], fpStr[2:])
|
|
}
|
|
|
|
// exists returns true when the given file or directory exists.
|
|
func exists(path string) (bool, error) {
|
|
_, err := os.Stat(path)
|
|
if err == nil {
|
|
return true, nil
|
|
}
|
|
if os.IsNotExist(err) {
|
|
return false, nil
|
|
}
|
|
|
|
return false, err
|
|
}
|
|
|
|
func (p *diskPersistence) openChunkFileForWriting(fp clientmodel.Fingerprint) (*os.File, error) {
|
|
dirname := p.dirForFingerprint(fp)
|
|
ex, err := exists(dirname)
|
|
if err != nil {
|
|
return nil, err
|
|
}
|
|
if !ex {
|
|
if err := os.MkdirAll(dirname, 0700); err != nil {
|
|
return nil, err
|
|
}
|
|
}
|
|
return os.OpenFile(path.Join(dirname, seriesFileName), os.O_WRONLY|os.O_APPEND|os.O_CREATE, 0640)
|
|
}
|
|
|
|
func (p *diskPersistence) openChunkFileForReading(fp clientmodel.Fingerprint) (*os.File, error) {
|
|
dirname := p.dirForFingerprint(fp)
|
|
return os.Open(path.Join(dirname, seriesFileName))
|
|
}
|
|
|
|
func writeChunkHeader(w io.Writer, c chunk) error {
|
|
header := make([]byte, chunkHeaderLen)
|
|
header[chunkHeaderTypeOffset] = chunkType(c)
|
|
binary.LittleEndian.PutUint64(header[chunkHeaderFirstTimeOffset:], uint64(c.firstTime()))
|
|
binary.LittleEndian.PutUint64(header[chunkHeaderLastTimeOffset:], uint64(c.lastTime()))
|
|
_, err := w.Write(header)
|
|
return err
|
|
}
|
|
|
|
func (p *diskPersistence) PersistChunk(fp clientmodel.Fingerprint, c chunk) error {
|
|
// 1. Open chunk file.
|
|
f, err := p.openChunkFileForWriting(fp)
|
|
if err != nil {
|
|
return err
|
|
}
|
|
defer f.Close()
|
|
|
|
b := bufio.NewWriterSize(f, chunkHeaderLen+p.chunkLen)
|
|
defer b.Flush()
|
|
|
|
// 2. Write the header (chunk type and first/last times).
|
|
err = writeChunkHeader(b, c)
|
|
if err != nil {
|
|
return err
|
|
}
|
|
|
|
// 3. Write chunk into file.
|
|
return c.marshal(b)
|
|
}
|
|
|
|
func (p *diskPersistence) offsetForChunkIndex(i int) int64 {
|
|
return int64(i * (chunkHeaderLen + p.chunkLen))
|
|
}
|
|
|
|
func (p *diskPersistence) LoadChunks(fp clientmodel.Fingerprint, indexes []int) (chunks, error) {
|
|
// TODO: we need to verify at some point that file length is a multiple of
|
|
// the chunk size. When is the best time to do this, and where to remember
|
|
// it? Right now, we only do it when loading chunkDescs.
|
|
f, err := p.openChunkFileForReading(fp)
|
|
if err != nil {
|
|
return nil, err
|
|
}
|
|
defer f.Close()
|
|
|
|
chunks := make(chunks, 0, len(indexes))
|
|
defer func() {
|
|
if err == nil {
|
|
return
|
|
}
|
|
}()
|
|
|
|
typeBuf := make([]byte, 1)
|
|
for _, idx := range indexes {
|
|
_, err := f.Seek(p.offsetForChunkIndex(idx), os.SEEK_SET)
|
|
if err != nil {
|
|
return nil, err
|
|
}
|
|
// TODO: check seek offset too?
|
|
|
|
n, err := f.Read(typeBuf)
|
|
if err != nil {
|
|
return nil, err
|
|
}
|
|
if n != 1 {
|
|
// Shouldn't happen?
|
|
panic("read returned != 1 bytes")
|
|
}
|
|
|
|
_, err = f.Seek(chunkHeaderLen-1, os.SEEK_CUR)
|
|
if err != nil {
|
|
return nil, err
|
|
}
|
|
chunk := chunkForType(typeBuf[0])
|
|
chunk.unmarshal(f)
|
|
chunks = append(chunks, chunk)
|
|
}
|
|
return chunks, nil
|
|
}
|
|
|
|
func (p *diskPersistence) LoadChunkDescs(fp clientmodel.Fingerprint, beforeTime clientmodel.Timestamp) (chunkDescs, error) {
|
|
f, err := p.openChunkFileForReading(fp)
|
|
if os.IsNotExist(err) {
|
|
return nil, nil
|
|
}
|
|
if err != nil {
|
|
return nil, err
|
|
}
|
|
defer f.Close()
|
|
|
|
fi, err := f.Stat()
|
|
if err != nil {
|
|
return nil, err
|
|
}
|
|
totalChunkLen := chunkHeaderLen + p.chunkLen
|
|
if fi.Size()%int64(totalChunkLen) != 0 {
|
|
// TODO: record number of encountered corrupt series files in a metric?
|
|
|
|
// Truncate the file size to the nearest multiple of chunkLen.
|
|
truncateTo := fi.Size() - fi.Size()%int64(totalChunkLen)
|
|
glog.Infof("Bad series file size for %s: %d bytes (no multiple of %d). Truncating to %d bytes.", fp, fi.Size(), totalChunkLen, truncateTo)
|
|
// TODO: this doesn't work, as this is a read-only file handle.
|
|
if err := f.Truncate(truncateTo); err != nil {
|
|
return nil, err
|
|
}
|
|
}
|
|
|
|
numChunks := int(fi.Size()) / totalChunkLen
|
|
cds := make(chunkDescs, 0, numChunks)
|
|
for i := 0; i < numChunks; i++ {
|
|
_, err := f.Seek(p.offsetForChunkIndex(i)+chunkHeaderFirstTimeOffset, os.SEEK_SET)
|
|
if err != nil {
|
|
return nil, err
|
|
}
|
|
|
|
chunkTimesBuf := make([]byte, 16)
|
|
_, err = io.ReadAtLeast(f, chunkTimesBuf, 16)
|
|
if err != nil {
|
|
return nil, err
|
|
}
|
|
cd := &chunkDesc{
|
|
firstTimeField: clientmodel.Timestamp(binary.LittleEndian.Uint64(chunkTimesBuf)),
|
|
lastTimeField: clientmodel.Timestamp(binary.LittleEndian.Uint64(chunkTimesBuf[8:])),
|
|
}
|
|
if !cd.firstTime().Before(beforeTime) {
|
|
// From here on, we have chunkDescs in memory already.
|
|
break
|
|
}
|
|
cds = append(cds, cd)
|
|
}
|
|
return cds, nil
|
|
}
|
|
|
|
func (p *diskPersistence) headsPath() string {
|
|
return path.Join(p.basePath, headsFileName)
|
|
}
|
|
|
|
func (p *diskPersistence) PersistHeads(fpToSeries map[clientmodel.Fingerprint]*memorySeries) error {
|
|
f, err := os.OpenFile(p.headsPath(), os.O_WRONLY|os.O_TRUNC|os.O_CREATE, 0640)
|
|
if err != nil {
|
|
return err
|
|
}
|
|
header := make([]byte, 9)
|
|
for fp, series := range fpToSeries {
|
|
head := series.head().chunk
|
|
|
|
binary.LittleEndian.PutUint64(header[headsHeaderFingerprintOffset:], uint64(fp))
|
|
header[headsHeaderTypeOffset] = chunkType(head)
|
|
_, err := f.Write(header)
|
|
if err != nil {
|
|
return err
|
|
}
|
|
err = head.marshal(f)
|
|
if err != nil {
|
|
return err
|
|
}
|
|
}
|
|
return nil
|
|
}
|
|
|
|
func (p *diskPersistence) DropChunks(fp clientmodel.Fingerprint, beforeTime clientmodel.Timestamp) error {
|
|
f, err := p.openChunkFileForReading(fp)
|
|
if os.IsNotExist(err) {
|
|
return nil
|
|
}
|
|
if err != nil {
|
|
return err
|
|
}
|
|
defer f.Close()
|
|
|
|
// Find the first chunk that should be kept.
|
|
for i := 0; ; i++ {
|
|
_, err := f.Seek(p.offsetForChunkIndex(i)+chunkHeaderLastTimeOffset, os.SEEK_SET)
|
|
if err != nil {
|
|
return err
|
|
}
|
|
lastTimeBuf := make([]byte, 8)
|
|
_, err = io.ReadAtLeast(f, lastTimeBuf, 8)
|
|
if err == io.EOF {
|
|
// We ran into the end of the file without finding any chunks that should
|
|
// be kept. Remove the whole file.
|
|
if err := os.Remove(f.Name()); err != nil {
|
|
return err
|
|
}
|
|
return nil
|
|
}
|
|
if err != nil {
|
|
return err
|
|
}
|
|
lastTime := clientmodel.Timestamp(binary.LittleEndian.Uint64(lastTimeBuf))
|
|
if !lastTime.Before(beforeTime) {
|
|
break
|
|
}
|
|
}
|
|
|
|
// We've found the first chunk that should be kept. Seek backwards to the
|
|
// beginning of its header and start copying everything from there into a new
|
|
// file.
|
|
_, err = f.Seek(-(chunkHeaderLastTimeOffset + 8), os.SEEK_CUR)
|
|
if err != nil {
|
|
return err
|
|
}
|
|
|
|
dirname := p.dirForFingerprint(fp)
|
|
temp, err := os.OpenFile(path.Join(dirname, seriesTempFileName), os.O_WRONLY|os.O_CREATE, 0640)
|
|
if err != nil {
|
|
return err
|
|
}
|
|
defer temp.Close()
|
|
|
|
if _, err := io.Copy(temp, f); err != nil {
|
|
return err
|
|
}
|
|
|
|
os.Rename(path.Join(dirname, seriesTempFileName), path.Join(dirname, seriesFileName))
|
|
return nil
|
|
}
|
|
|
|
func (p *diskPersistence) LoadHeads(fpToSeries map[clientmodel.Fingerprint]*memorySeries) error {
|
|
f, err := os.Open(p.headsPath())
|
|
if os.IsNotExist(err) {
|
|
// TODO: this should only happen if there never was a shutdown before. In
|
|
// that case, all heads should be in order already, since the series got
|
|
// created during this process' runtime.
|
|
// Still, make this more robust.
|
|
return nil
|
|
}
|
|
|
|
header := make([]byte, headsHeaderLen)
|
|
for {
|
|
_, err := io.ReadAtLeast(f, header, headsHeaderLen)
|
|
if err == io.ErrUnexpectedEOF {
|
|
// TODO: this should only be ok if n is 0.
|
|
break
|
|
}
|
|
if err != nil {
|
|
return nil
|
|
}
|
|
// TODO: this relies on the implementation (uint64) of Fingerprint.
|
|
fp := clientmodel.Fingerprint(binary.LittleEndian.Uint64(header[headsHeaderFingerprintOffset:]))
|
|
chunk := chunkForType(header[headsHeaderTypeOffset])
|
|
chunk.unmarshal(f)
|
|
fpToSeries[fp].chunkDescs = append(fpToSeries[fp].chunkDescs, &chunkDesc{
|
|
chunk: chunk,
|
|
refCount: 1,
|
|
})
|
|
}
|
|
return nil
|
|
}
|
|
|
|
func (d *diskPersistence) Close() {
|
|
for _, db := range d.indexDBs {
|
|
if err := db.Close(); err != nil {
|
|
glog.Error("Error closing index DB: ", err)
|
|
}
|
|
}
|
|
}
|
|
|
|
// Get all of the label values that are associated with a given label name.
|
|
func (d *diskPersistence) GetFingerprintsForLabelPair(l clientmodel.LabelName, v clientmodel.LabelValue) (clientmodel.Fingerprints, error) {
|
|
return nil, nil
|
|
}
|
|
|
|
// Get all label values that are associated with a given label name.
|
|
func (d *diskPersistence) GetLabelValuesForLabelName(clientmodel.LabelName) (clientmodel.LabelValues, error) {
|
|
return nil, nil
|
|
}
|
|
|
|
// Get the metric associated with the provided fingerprint.
|
|
func (d *diskPersistence) GetMetricForFingerprint(clientmodel.Fingerprint) (clientmodel.Metric, error) {
|
|
return nil, nil
|
|
}
|