prometheus/block.go

327 lines
7.1 KiB
Go
Raw Normal View History

2017-04-10 11:59:45 -07:00
// Copyright 2017 The Prometheus Authors
// Licensed under the Apache License, Version 2.0 (the "License");
// you may not use this file except in compliance with the License.
// You may obtain a copy of the License at
//
// http://www.apache.org/licenses/LICENSE-2.0
//
// Unless required by applicable law or agreed to in writing, software
// distributed under the License is distributed on an "AS IS" BASIS,
// WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
// See the License for the specific language governing permissions and
// limitations under the License.
package tsdb
2016-12-14 23:31:26 -08:00
import (
"encoding/json"
"io/ioutil"
2016-12-14 23:31:26 -08:00
"os"
"path/filepath"
2016-12-22 06:54:39 -08:00
2017-02-27 01:46:15 -08:00
"github.com/oklog/ulid"
2016-12-22 06:54:39 -08:00
"github.com/pkg/errors"
"github.com/prometheus/tsdb/labels"
)
// DiskBlock handles reads against a Block of time series data.
type DiskBlock interface {
// Directory where block data is stored.
Dir() string
// Stats returns statistics about the block.
Meta() BlockMeta
// Index returns an IndexReader over the block's data.
Index() IndexReader
2017-04-28 06:41:42 -07:00
// Chunks returns a ChunkReader over the block's data.
Chunks() ChunkReader
// Tombstones returns a TombstoneReader over the block's deleted data.
Tombstones() TombstoneReader
// Delete deletes data from the block.
Delete(mint, maxt int64, ms ...labels.Matcher) error
// Close releases all underlying resources of the block.
2017-01-17 21:18:32 -08:00
Close() error
2016-12-13 06:26:58 -08:00
}
// Block is an interface to a DiskBlock that can also be queried.
type Block interface {
DiskBlock
Queryable
}
// headBlock is a regular block that can still be appended to.
type headBlock interface {
Block
Appendable
Snapshottable
}
// Snapshottable defines an entity that can be backedup online.
type Snapshottable interface {
Snapshot(dir string) error
}
// Appendable defines an entity to which data can be appended.
type Appendable interface {
// Appender returns a new Appender against an underlying store.
Appender() Appender
// Busy returns whether there are any currently active appenders.
Busy() bool
}
// Queryable defines an entity which provides a Querier.
type Queryable interface {
Querier(mint, maxt int64) Querier
}
// BlockMeta provides meta information about a block.
type BlockMeta struct {
2017-02-27 01:46:15 -08:00
// Unique identifier for the block and its contents. Changes on compaction.
ULID ulid.ULID `json:"ulid"`
2017-01-19 05:01:38 -08:00
// MinTime and MaxTime specify the time range all samples
// in the block are in.
MinTime int64 `json:"minTime"`
MaxTime int64 `json:"maxTime"`
2017-01-07 09:02:17 -08:00
// Stats about the contents of the block.
Stats struct {
NumSamples uint64 `json:"numSamples,omitempty"`
NumSeries uint64 `json:"numSeries,omitempty"`
NumChunks uint64 `json:"numChunks,omitempty"`
NumTombstones uint64 `json:"numTombstones,omitempty"`
} `json:"stats,omitempty"`
2017-01-19 10:45:52 -08:00
// Information on compactions the block was created from.
2017-01-19 10:45:52 -08:00
Compaction struct {
Generation int `json:"generation"`
} `json:"compaction"`
}
const (
flagNone = 0
flagStd = 1
)
type blockMeta struct {
2017-01-19 05:01:38 -08:00
Version int `json:"version"`
*BlockMeta
}
const metaFilename = "meta.json"
2017-01-19 05:01:38 -08:00
func readMetaFile(dir string) (*BlockMeta, error) {
b, err := ioutil.ReadFile(filepath.Join(dir, metaFilename))
if err != nil {
return nil, err
}
var m blockMeta
if err := json.Unmarshal(b, &m); err != nil {
return nil, err
}
if m.Version != 1 {
return nil, errors.Errorf("unexpected meta file version %d", m.Version)
}
return m.BlockMeta, nil
}
func writeMetaFile(dir string, meta *BlockMeta) error {
2017-03-01 08:19:57 -08:00
// Make any changes to the file appear atomic.
path := filepath.Join(dir, metaFilename)
tmp := path + ".tmp"
f, err := os.Create(tmp)
2017-01-19 05:01:38 -08:00
if err != nil {
return err
}
enc := json.NewEncoder(f)
enc.SetIndent("", "\t")
var merr MultiError
if merr.Add(enc.Encode(&blockMeta{Version: 1, BlockMeta: meta})); merr.Err() != nil {
merr.Add(f.Close())
return merr
2017-01-19 05:01:38 -08:00
}
if err := f.Close(); err != nil {
return err
}
2017-03-01 08:19:57 -08:00
return renameFile(tmp, path)
2017-01-19 05:01:38 -08:00
}
type persistedBlock struct {
dir string
meta BlockMeta
chunkr *chunkReader
indexr *indexReader
tombstones tombstoneReader
}
func newPersistedBlock(dir string) (*persistedBlock, error) {
2017-01-19 10:45:52 -08:00
meta, err := readMetaFile(dir)
if err != nil {
return nil, err
}
2016-12-14 23:31:26 -08:00
2017-02-27 01:46:15 -08:00
cr, err := newChunkReader(chunkDir(dir))
2016-12-14 23:31:26 -08:00
if err != nil {
return nil, err
2016-12-14 23:31:26 -08:00
}
ir, err := newIndexReader(dir)
2016-12-14 23:31:26 -08:00
if err != nil {
return nil, err
2016-12-14 23:31:26 -08:00
}
tr, err := readTombstones(dir)
if err != nil {
return nil, err
}
2016-12-14 23:31:26 -08:00
pb := &persistedBlock{
dir: dir,
meta: *meta,
chunkr: cr,
indexr: ir,
tombstones: tr,
2016-12-14 23:31:26 -08:00
}
return pb, nil
}
func (pb *persistedBlock) Close() error {
2017-02-27 01:46:15 -08:00
var merr MultiError
2016-12-14 23:31:26 -08:00
2017-02-27 01:46:15 -08:00
merr.Add(pb.chunkr.Close())
merr.Add(pb.indexr.Close())
return merr.Err()
2016-12-14 23:31:26 -08:00
}
func (pb *persistedBlock) String() string {
return pb.meta.ULID.String()
}
func (pb *persistedBlock) Querier(mint, maxt int64) Querier {
return &blockQuerier{
mint: mint,
maxt: maxt,
index: pb.Index(),
chunks: pb.Chunks(),
tombstones: pb.Tombstones(),
}
}
func (pb *persistedBlock) Dir() string { return pb.dir }
func (pb *persistedBlock) Index() IndexReader { return pb.indexr }
func (pb *persistedBlock) Chunks() ChunkReader { return pb.chunkr }
func (pb *persistedBlock) Tombstones() TombstoneReader {
return pb.tombstones
}
func (pb *persistedBlock) Meta() BlockMeta { return pb.meta }
func (pb *persistedBlock) Delete(mint, maxt int64, ms ...labels.Matcher) error {
pr := newPostingsReader(pb.indexr)
p, absent := pr.Select(ms...)
ir := pb.indexr
// Choose only valid postings which have chunks in the time-range.
stones := map[uint32]intervals{}
Outer:
for p.Next() {
lset, chunks, err := ir.Series(p.At())
if err != nil {
return err
}
for _, abs := range absent {
if lset.Get(abs) != "" {
continue Outer
}
}
for _, chk := range chunks {
if intervalOverlap(mint, maxt, chk.MinTime, chk.MaxTime) {
// Delete only until the current vlaues and not beyond.
tmin, tmax := clampInterval(mint, maxt, chunks[0].MinTime, chunks[len(chunks)-1].MaxTime)
stones[p.At()] = intervals{{tmin, tmax}}
continue Outer
}
}
}
if p.Err() != nil {
return p.Err()
}
// Merge the current and new tombstones.
for k, v := range stones {
pb.tombstones.add(k, v[0])
}
if err := writeTombstoneFile(pb.dir, pb.tombstones); err != nil {
return err
}
pb.meta.Stats.NumTombstones = uint64(len(pb.tombstones))
return writeMetaFile(pb.dir, &pb.meta)
}
2017-02-27 01:46:15 -08:00
func chunkDir(dir string) string { return filepath.Join(dir, "chunks") }
func walDir(dir string) string { return filepath.Join(dir, "wal") }
2016-12-14 23:31:26 -08:00
func clampInterval(a, b, mint, maxt int64) (int64, int64) {
if a < mint {
a = mint
}
if b > maxt {
b = maxt
}
return a, b
}
2016-12-14 23:31:26 -08:00
type mmapFile struct {
f *os.File
2016-12-14 23:31:26 -08:00
b []byte
}
func openMmapFile(path string) (*mmapFile, error) {
f, err := os.Open(path)
2016-12-14 23:31:26 -08:00
if err != nil {
return nil, errors.Wrap(err, "try lock file")
2016-12-14 23:31:26 -08:00
}
info, err := f.Stat()
if err != nil {
return nil, errors.Wrap(err, "stat")
2016-12-14 23:31:26 -08:00
}
b, err := mmap(f, int(info.Size()))
2016-12-14 23:31:26 -08:00
if err != nil {
return nil, errors.Wrap(err, "mmap")
2016-12-14 23:31:26 -08:00
}
return &mmapFile{f: f, b: b}, nil
}
func (f *mmapFile) Close() error {
err0 := munmap(f.b)
err1 := f.f.Close()
if err0 != nil {
return err0
}
return err1
}