mirror of
https://github.com/prometheus/prometheus.git
synced 2025-03-05 20:59:13 -08:00
Export ErrNotFound
This commit is contained in:
parent
9cf49f68e9
commit
dd0b69fe1b
2
db.go
2
db.go
|
@ -369,7 +369,7 @@ func (a *dbAppender) Add(ref uint64, t int64, v float64) error {
|
||||||
gen := uint8((ref << 16) >> 56)
|
gen := uint8((ref << 16) >> 56)
|
||||||
|
|
||||||
if gen != a.gen {
|
if gen != a.gen {
|
||||||
return errNotFound
|
return ErrNotFound
|
||||||
}
|
}
|
||||||
return a.head.Add(ref, t, v)
|
return a.head.Add(ref, t, v)
|
||||||
}
|
}
|
||||||
|
|
9
head.go
9
head.go
|
@ -2,6 +2,7 @@ package tsdb
|
||||||
|
|
||||||
import (
|
import (
|
||||||
"errors"
|
"errors"
|
||||||
|
"fmt"
|
||||||
"math"
|
"math"
|
||||||
"math/rand"
|
"math/rand"
|
||||||
"sort"
|
"sort"
|
||||||
|
@ -182,7 +183,7 @@ func (a *headAppender) Add(ref uint64, t int64, v float64) error {
|
||||||
// this transaction.
|
// this transaction.
|
||||||
if ref&(1<<32) > 0 {
|
if ref&(1<<32) > 0 {
|
||||||
if _, ok := a.newSeries[ref]; !ok {
|
if _, ok := a.newSeries[ref]; !ok {
|
||||||
return errNotFound
|
return ErrNotFound
|
||||||
}
|
}
|
||||||
// TODO(fabxc): we also have to validate here that the
|
// TODO(fabxc): we also have to validate here that the
|
||||||
// sample sequence is valid.
|
// sample sequence is valid.
|
||||||
|
@ -198,7 +199,7 @@ func (a *headAppender) Add(ref uint64, t int64, v float64) error {
|
||||||
|
|
||||||
ms := a.series[int(ref)]
|
ms := a.series[int(ref)]
|
||||||
if ms == nil {
|
if ms == nil {
|
||||||
return errNotFound
|
return ErrNotFound
|
||||||
}
|
}
|
||||||
c := ms.head()
|
c := ms.head()
|
||||||
|
|
||||||
|
@ -371,7 +372,7 @@ func (h *headIndexReader) Series(ref uint32) (labels.Labels, []ChunkMeta, error)
|
||||||
defer h.mtx.RUnlock()
|
defer h.mtx.RUnlock()
|
||||||
|
|
||||||
if int(ref) >= len(h.series) {
|
if int(ref) >= len(h.series) {
|
||||||
return nil, nil, errNotFound
|
return nil, nil, ErrNotFound
|
||||||
}
|
}
|
||||||
s := h.series[ref]
|
s := h.series[ref]
|
||||||
metas := make([]ChunkMeta, 0, len(s.chunks))
|
metas := make([]ChunkMeta, 0, len(s.chunks))
|
||||||
|
@ -426,6 +427,7 @@ func (h *headBlock) create(hash uint64, lset labels.Labels) *memSeries {
|
||||||
lset: lset,
|
lset: lset,
|
||||||
ref: uint32(len(h.series)),
|
ref: uint32(len(h.series)),
|
||||||
}
|
}
|
||||||
|
fmt.Println("create", lset)
|
||||||
|
|
||||||
// Allocate empty space until we can insert at the given index.
|
// Allocate empty space until we can insert at the given index.
|
||||||
h.series = append(h.series, s)
|
h.series = append(h.series, s)
|
||||||
|
@ -449,6 +451,7 @@ func (h *headBlock) create(hash uint64, lset labels.Labels) *memSeries {
|
||||||
}
|
}
|
||||||
|
|
||||||
var (
|
var (
|
||||||
|
ErrNotFound = fmt.Errorf("not found")
|
||||||
// ErrOutOfOrderSample is returned if an appended sample has a
|
// ErrOutOfOrderSample is returned if an appended sample has a
|
||||||
// timestamp larger than the most recent sample.
|
// timestamp larger than the most recent sample.
|
||||||
ErrOutOfOrderSample = errors.New("out of order sample")
|
ErrOutOfOrderSample = errors.New("out of order sample")
|
||||||
|
|
|
@ -91,7 +91,6 @@ type indexReader struct {
|
||||||
var (
|
var (
|
||||||
errInvalidSize = fmt.Errorf("invalid size")
|
errInvalidSize = fmt.Errorf("invalid size")
|
||||||
errInvalidFlag = fmt.Errorf("invalid flag")
|
errInvalidFlag = fmt.Errorf("invalid flag")
|
||||||
errNotFound = fmt.Errorf("not found")
|
|
||||||
)
|
)
|
||||||
|
|
||||||
func newIndexReader(s SeriesReader, b []byte) (*indexReader, error) {
|
func newIndexReader(s SeriesReader, b []byte) (*indexReader, error) {
|
||||||
|
@ -333,7 +332,7 @@ func (r *indexReader) Postings(name, value string) (Postings, error) {
|
||||||
|
|
||||||
off, ok := r.postings[key]
|
off, ok := r.postings[key]
|
||||||
if !ok {
|
if !ok {
|
||||||
return nil, errNotFound
|
return nil, ErrNotFound
|
||||||
}
|
}
|
||||||
|
|
||||||
flag, b, err := r.section(off)
|
flag, b, err := r.section(off)
|
||||||
|
|
Loading…
Reference in a new issue