mirror of
https://github.com/prometheus/prometheus.git
synced 2024-11-14 01:24:04 -08:00
147 lines
3.9 KiB
Go
147 lines
3.9 KiB
Go
|
// Copyright 2014 The Go Authors. All rights reserved.
|
||
|
// Use of this source code is governed by a BSD-style
|
||
|
// license that can be found in the LICENSE file.
|
||
|
|
||
|
package http2
|
||
|
|
||
|
import (
|
||
|
"errors"
|
||
|
"fmt"
|
||
|
"sync"
|
||
|
)
|
||
|
|
||
|
// Buffer chunks are allocated from a pool to reduce pressure on GC.
|
||
|
// The maximum wasted space per dataBuffer is 2x the largest size class,
|
||
|
// which happens when the dataBuffer has multiple chunks and there is
|
||
|
// one unread byte in both the first and last chunks. We use a few size
|
||
|
// classes to minimize overheads for servers that typically receive very
|
||
|
// small request bodies.
|
||
|
//
|
||
|
// TODO: Benchmark to determine if the pools are necessary. The GC may have
|
||
|
// improved enough that we can instead allocate chunks like this:
|
||
|
// make([]byte, max(16<<10, expectedBytesRemaining))
|
||
|
var (
|
||
|
dataChunkSizeClasses = []int{
|
||
|
1 << 10,
|
||
|
2 << 10,
|
||
|
4 << 10,
|
||
|
8 << 10,
|
||
|
16 << 10,
|
||
|
}
|
||
|
dataChunkPools = [...]sync.Pool{
|
||
|
{New: func() interface{} { return make([]byte, 1<<10) }},
|
||
|
{New: func() interface{} { return make([]byte, 2<<10) }},
|
||
|
{New: func() interface{} { return make([]byte, 4<<10) }},
|
||
|
{New: func() interface{} { return make([]byte, 8<<10) }},
|
||
|
{New: func() interface{} { return make([]byte, 16<<10) }},
|
||
|
}
|
||
|
)
|
||
|
|
||
|
func getDataBufferChunk(size int64) []byte {
|
||
|
i := 0
|
||
|
for ; i < len(dataChunkSizeClasses)-1; i++ {
|
||
|
if size <= int64(dataChunkSizeClasses[i]) {
|
||
|
break
|
||
|
}
|
||
|
}
|
||
|
return dataChunkPools[i].Get().([]byte)
|
||
|
}
|
||
|
|
||
|
func putDataBufferChunk(p []byte) {
|
||
|
for i, n := range dataChunkSizeClasses {
|
||
|
if len(p) == n {
|
||
|
dataChunkPools[i].Put(p)
|
||
|
return
|
||
|
}
|
||
|
}
|
||
|
panic(fmt.Sprintf("unexpected buffer len=%v", len(p)))
|
||
|
}
|
||
|
|
||
|
// dataBuffer is an io.ReadWriter backed by a list of data chunks.
|
||
|
// Each dataBuffer is used to read DATA frames on a single stream.
|
||
|
// The buffer is divided into chunks so the server can limit the
|
||
|
// total memory used by a single connection without limiting the
|
||
|
// request body size on any single stream.
|
||
|
type dataBuffer struct {
|
||
|
chunks [][]byte
|
||
|
r int // next byte to read is chunks[0][r]
|
||
|
w int // next byte to write is chunks[len(chunks)-1][w]
|
||
|
size int // total buffered bytes
|
||
|
expected int64 // we expect at least this many bytes in future Write calls (ignored if <= 0)
|
||
|
}
|
||
|
|
||
|
var errReadEmpty = errors.New("read from empty dataBuffer")
|
||
|
|
||
|
// Read copies bytes from the buffer into p.
|
||
|
// It is an error to read when no data is available.
|
||
|
func (b *dataBuffer) Read(p []byte) (int, error) {
|
||
|
if b.size == 0 {
|
||
|
return 0, errReadEmpty
|
||
|
}
|
||
|
var ntotal int
|
||
|
for len(p) > 0 && b.size > 0 {
|
||
|
readFrom := b.bytesFromFirstChunk()
|
||
|
n := copy(p, readFrom)
|
||
|
p = p[n:]
|
||
|
ntotal += n
|
||
|
b.r += n
|
||
|
b.size -= n
|
||
|
// If the first chunk has been consumed, advance to the next chunk.
|
||
|
if b.r == len(b.chunks[0]) {
|
||
|
putDataBufferChunk(b.chunks[0])
|
||
|
end := len(b.chunks) - 1
|
||
|
copy(b.chunks[:end], b.chunks[1:])
|
||
|
b.chunks[end] = nil
|
||
|
b.chunks = b.chunks[:end]
|
||
|
b.r = 0
|
||
|
}
|
||
|
}
|
||
|
return ntotal, nil
|
||
|
}
|
||
|
|
||
|
func (b *dataBuffer) bytesFromFirstChunk() []byte {
|
||
|
if len(b.chunks) == 1 {
|
||
|
return b.chunks[0][b.r:b.w]
|
||
|
}
|
||
|
return b.chunks[0][b.r:]
|
||
|
}
|
||
|
|
||
|
// Len returns the number of bytes of the unread portion of the buffer.
|
||
|
func (b *dataBuffer) Len() int {
|
||
|
return b.size
|
||
|
}
|
||
|
|
||
|
// Write appends p to the buffer.
|
||
|
func (b *dataBuffer) Write(p []byte) (int, error) {
|
||
|
ntotal := len(p)
|
||
|
for len(p) > 0 {
|
||
|
// If the last chunk is empty, allocate a new chunk. Try to allocate
|
||
|
// enough to fully copy p plus any additional bytes we expect to
|
||
|
// receive. However, this may allocate less than len(p).
|
||
|
want := int64(len(p))
|
||
|
if b.expected > want {
|
||
|
want = b.expected
|
||
|
}
|
||
|
chunk := b.lastChunkOrAlloc(want)
|
||
|
n := copy(chunk[b.w:], p)
|
||
|
p = p[n:]
|
||
|
b.w += n
|
||
|
b.size += n
|
||
|
b.expected -= int64(n)
|
||
|
}
|
||
|
return ntotal, nil
|
||
|
}
|
||
|
|
||
|
func (b *dataBuffer) lastChunkOrAlloc(want int64) []byte {
|
||
|
if len(b.chunks) != 0 {
|
||
|
last := b.chunks[len(b.chunks)-1]
|
||
|
if b.w < len(last) {
|
||
|
return last
|
||
|
}
|
||
|
}
|
||
|
chunk := getDataBufferChunk(want)
|
||
|
b.chunks = append(b.chunks, chunk)
|
||
|
b.w = 0
|
||
|
return chunk
|
||
|
}
|