Convert metric.Values to slice of values.

The initial impetus for this was that it made unmarshalling sample
values much faster.

Other relevant benchmark changes in ns/op:

Benchmark                                 old        new   speedup
==================================================================
BenchmarkMarshal                       179170     127996     1.4x
BenchmarkUnmarshal                     404984     132186     3.1x

BenchmarkMemoryGetValueAtTime           57801      50050     1.2x
BenchmarkMemoryGetBoundaryValues        64496      53194     1.2x
BenchmarkMemoryGetRangeValues           66585      54065     1.2x

BenchmarkStreamAdd                       45.0       75.3     0.6x
BenchmarkAppendSample1                   1157       1587     0.7x
BenchmarkAppendSample10                  4090       4284     0.95x
BenchmarkAppendSample100                45660      44066     1.0x
BenchmarkAppendSample1000              579084     582380     1.0x
BenchmarkMemoryAppendRepeatingValues 22796594   22005502     1.0x

Overall, this gives us good speedups in the areas where they matter
most: decoding values from disk and accessing the memory storage (which
is also used for views).

Some of the smaller append examples take minimally longer, but the cost
seems to get amortized over larger appends, so I'm not worried about
these. Also, we're currently not bottlenecked on the write path and have
plenty of other optimizations available in that area if it becomes
necessary.

Memory allocations during appends don't change measurably at all.

Change-Id: I7dc7394edea09506976765551f35b138518db9e8
This commit is contained in:
Julius Volz 2014-03-10 17:55:17 +01:00
parent a7d0973fe3
commit 86fc13a52e
10 changed files with 27 additions and 27 deletions

View file

@ -384,7 +384,7 @@ func EvalVectorRange(node VectorNode, start clientmodel.Timestamp, end clientmod
for t := start; t.Before(end); t = t.Add(interval) {
vector := node.Eval(t, viewAdapter)
for _, sample := range vector {
samplePair := &metric.SamplePair{
samplePair := metric.SamplePair{
Value: sample.Value,
Timestamp: sample.Timestamp,
}

View file

@ -82,7 +82,7 @@ func (v *viewAdapter) chooseClosestSample(samples metric.Values, timestamp clien
continue
}
sample := candidate
closestBefore = sample
closestBefore = &sample
}
// Samples after target time.
@ -96,7 +96,7 @@ func (v *viewAdapter) chooseClosestSample(samples metric.Values, timestamp clien
continue
}
sample := candidate
closestAfter = sample
closestAfter = &sample
}
}

View file

@ -28,7 +28,7 @@ var testStartTime = clientmodel.Timestamp(0)
func getTestValueStream(startVal clientmodel.SampleValue, endVal clientmodel.SampleValue, stepVal clientmodel.SampleValue, startTime clientmodel.Timestamp) (resultValues metric.Values) {
currentTime := startTime
for currentVal := startVal; currentVal <= endVal; currentVal += stepVal {
sample := &metric.SamplePair{
sample := metric.SamplePair{
Value: currentVal,
Timestamp: currentTime,
}

View file

@ -350,7 +350,7 @@ func (l *LevelDBMetricPersistence) AppendSamples(samples clientmodel.Samples) (e
key.Dump(keyDto)
for _, sample := range chunk {
values = append(values, &SamplePair{
values = append(values, SamplePair{
Timestamp: sample.Timestamp,
Value: sample.Value,
})

View file

@ -27,7 +27,7 @@ import (
const initialSeriesArenaSize = 4 * 60
type stream interface {
add(...*SamplePair)
add(Values)
clone() Values
expunge(age clientmodel.Timestamp) Values
@ -53,7 +53,7 @@ func (s *arrayStream) metric() clientmodel.Metric {
return s.m
}
func (s *arrayStream) add(v ...*SamplePair) {
func (s *arrayStream) add(v Values) {
s.Lock()
defer s.Unlock()
@ -202,9 +202,11 @@ func (s *memorySeriesStorage) AppendSample(sample *clientmodel.Sample) error {
fingerprint := &clientmodel.Fingerprint{}
fingerprint.LoadFromMetric(sample.Metric)
series := s.getOrCreateSeries(sample.Metric, fingerprint)
series.add(&SamplePair{
series.add(Values{
SamplePair{
Value: sample.Value,
Timestamp: sample.Timestamp,
},
})
if s.wmCache != nil {
@ -325,7 +327,7 @@ func (s *memorySeriesStorage) appendSamplesWithoutIndexing(fingerprint *clientmo
s.fingerprintToSeries[*fingerprint] = series
}
series.add(samples...)
series.add(samples)
}
func (s *memorySeriesStorage) GetFingerprintsForLabelSet(l clientmodel.LabelSet) (clientmodel.Fingerprints, error) {

View file

@ -27,7 +27,7 @@ func BenchmarkStreamAdd(b *testing.B) {
s := newArrayStream(clientmodel.Metric{})
samples := make(Values, b.N)
for i := 0; i < b.N; i++ {
samples = append(samples, &SamplePair{
samples = append(samples, SamplePair{
Timestamp: clientmodel.TimestampFromTime(time.Date(i, 0, 0, 0, 0, 0, 0, time.UTC)),
Value: clientmodel.SampleValue(i),
})
@ -38,7 +38,7 @@ func BenchmarkStreamAdd(b *testing.B) {
var pre runtime.MemStats
runtime.ReadMemStats(&pre)
s.add(samples...)
s.add(samples)
var post runtime.MemStats
runtime.ReadMemStats(&post)

View file

@ -211,7 +211,7 @@ func TestGetValuesAtTimeOp(t *testing.T) {
t.Fatalf("%d. expected length %d, got %d", i, len(scenario.out), len(actual))
}
for j, out := range scenario.out {
if !out.Equal(actual[j]) {
if !out.Equal(&actual[j]) {
t.Fatalf("%d. expected output %v, got %v", i, scenario.out, actual)
}
}
@ -464,7 +464,7 @@ func TestGetValuesAtIntervalOp(t *testing.T) {
}
for j, out := range scenario.out {
if !out.Equal(actual[j]) {
if !out.Equal(&actual[j]) {
t.Fatalf("%d. expected output %v, got %v", i, scenario.out, actual)
}
}
@ -639,7 +639,7 @@ func TestGetValuesAlongRangeOp(t *testing.T) {
t.Fatalf("%d. expected length %d, got %d: %v", i, len(scenario.out), len(actual), actual)
}
for j, out := range scenario.out {
if !out.Equal(actual[j]) {
if !out.Equal(&actual[j]) {
t.Fatalf("%d. expected output %v, got %v", i, scenario.out, actual)
}
}
@ -796,7 +796,7 @@ func TestGetValueRangeAtIntervalOp(t *testing.T) {
t.Fatalf("%d. expected length %d, got %d: %v", i, len(scenario.out), len(actual), actual)
}
for j, out := range scenario.out {
if !out.Equal(actual[j]) {
if !out.Equal(&actual[j]) {
t.Fatalf("%d. expected output %v, got %v", i, scenario.out, actual)
}
}

View file

@ -57,9 +57,9 @@ func (s *SamplePair) String() string {
return fmt.Sprintf("SamplePair at %s of %s", s.Timestamp, s.Value)
}
// Values is a sortable slice of SamplePair pointers (as in: it implements
// Values is a sortable slice of SamplePairs (as in: it implements
// sort.Interface). Sorting happens by Timestamp.
type Values []*SamplePair
type Values []SamplePair
// Len implements sort.Interface.
func (v Values) Len() int {
@ -84,7 +84,7 @@ func (v Values) Equal(o Values) bool {
}
for i, expected := range v {
if !expected.Equal(o[i]) {
if !expected.Equal(&o[i]) {
return false
}
}
@ -182,10 +182,8 @@ func unmarshalValues(buf []byte) Values {
}
for i := 0; i < n; i++ {
offset := formatVersionSize + i*sampleSize
v[i] = &SamplePair{
Timestamp: clientmodel.TimestampFromUnix(int64(binary.LittleEndian.Uint64(buf[offset:]))),
Value: clientmodel.SampleValue(math.Float64frombits(binary.LittleEndian.Uint64(buf[offset+8:]))),
}
v[i].Timestamp = clientmodel.TimestampFromUnix(int64(binary.LittleEndian.Uint64(buf[offset:])))
v[i].Value = clientmodel.SampleValue(math.Float64frombits(binary.LittleEndian.Uint64(buf[offset+8:])))
}
return v
}

View file

@ -17,7 +17,7 @@ func TestValuesMarshalAndUnmarshal(t *testing.T) {
for i, expected := range values {
actual := unmarshalled[i]
if !actual.Equal(expected) {
if !actual.Equal(&expected) {
t.Fatalf("%d. got: %v, expected: %v", i, actual, expected)
}
}
@ -26,7 +26,7 @@ func TestValuesMarshalAndUnmarshal(t *testing.T) {
func randomValues(numSamples int) Values {
v := make(Values, 0, numSamples)
for i := 0; i < numSamples; i++ {
v = append(v, &SamplePair{
v = append(v, SamplePair{
Timestamp: clientmodel.Timestamp(rand.Int63()),
Value: clientmodel.SampleValue(rand.NormFloat64()),
})

View file

@ -719,7 +719,7 @@ func testTruncateBefore(t test.Tester) {
}
for j, actualValue := range actual {
if !actualValue.Equal(scenario.out[j]) {
if !actualValue.Equal(&scenario.out[j]) {
t.Fatalf("%d.%d. expected %s, got %s", i, j, scenario.out[j], actualValue)
}
}