mirror of
https://github.com/prometheus/prometheus.git
synced 2025-03-05 20:59:13 -08:00
Merge remote-tracking branch 'prometheus/main' into arve/sync-prometheus
This commit is contained in:
commit
b5f0920506
|
@ -52,7 +52,8 @@ Release cadence of first pre-releases being cut is 6 weeks.
|
||||||
| v2.45 LTS | 2023-05-31 | Jesus Vazquez (Github: @jesusvazquez) |
|
| v2.45 LTS | 2023-05-31 | Jesus Vazquez (Github: @jesusvazquez) |
|
||||||
| v2.46 | 2023-07-12 | Julien Pivotto (GitHub: @roidelapluie) |
|
| v2.46 | 2023-07-12 | Julien Pivotto (GitHub: @roidelapluie) |
|
||||||
| v2.47 | 2023-08-23 | Bryan Boreham (GitHub: @bboreham) |
|
| v2.47 | 2023-08-23 | Bryan Boreham (GitHub: @bboreham) |
|
||||||
| v2.48 | 2023-10-04 | **searching for volunteer** |
|
| v2.48 | 2023-10-04 | Levi Harrison (GitHub: @LeviHarrison) |
|
||||||
|
| v2.49 | 2023-11-15 | **searching for volunteer** |
|
||||||
|
|
||||||
If you are interested in volunteering please create a pull request against the [prometheus/prometheus](https://github.com/prometheus/prometheus) repository and propose yourself for the release series of your choice.
|
If you are interested in volunteering please create a pull request against the [prometheus/prometheus](https://github.com/prometheus/prometheus) repository and propose yourself for the release series of your choice.
|
||||||
|
|
||||||
|
|
|
@ -174,6 +174,7 @@ func (m *Gauge) GetValue() float64 {
|
||||||
type Counter struct {
|
type Counter struct {
|
||||||
Value float64 `protobuf:"fixed64,1,opt,name=value,proto3" json:"value,omitempty"`
|
Value float64 `protobuf:"fixed64,1,opt,name=value,proto3" json:"value,omitempty"`
|
||||||
Exemplar *Exemplar `protobuf:"bytes,2,opt,name=exemplar,proto3" json:"exemplar,omitempty"`
|
Exemplar *Exemplar `protobuf:"bytes,2,opt,name=exemplar,proto3" json:"exemplar,omitempty"`
|
||||||
|
CreatedTimestamp *types.Timestamp `protobuf:"bytes,3,opt,name=created_timestamp,json=createdTimestamp,proto3" json:"created_timestamp,omitempty"`
|
||||||
XXX_NoUnkeyedLiteral struct{} `json:"-"`
|
XXX_NoUnkeyedLiteral struct{} `json:"-"`
|
||||||
XXX_unrecognized []byte `json:"-"`
|
XXX_unrecognized []byte `json:"-"`
|
||||||
XXX_sizecache int32 `json:"-"`
|
XXX_sizecache int32 `json:"-"`
|
||||||
|
@ -226,6 +227,13 @@ func (m *Counter) GetExemplar() *Exemplar {
|
||||||
return nil
|
return nil
|
||||||
}
|
}
|
||||||
|
|
||||||
|
func (m *Counter) GetCreatedTimestamp() *types.Timestamp {
|
||||||
|
if m != nil {
|
||||||
|
return m.CreatedTimestamp
|
||||||
|
}
|
||||||
|
return nil
|
||||||
|
}
|
||||||
|
|
||||||
type Quantile struct {
|
type Quantile struct {
|
||||||
Quantile float64 `protobuf:"fixed64,1,opt,name=quantile,proto3" json:"quantile,omitempty"`
|
Quantile float64 `protobuf:"fixed64,1,opt,name=quantile,proto3" json:"quantile,omitempty"`
|
||||||
Value float64 `protobuf:"fixed64,2,opt,name=value,proto3" json:"value,omitempty"`
|
Value float64 `protobuf:"fixed64,2,opt,name=value,proto3" json:"value,omitempty"`
|
||||||
|
@ -285,6 +293,7 @@ type Summary struct {
|
||||||
SampleCount uint64 `protobuf:"varint,1,opt,name=sample_count,json=sampleCount,proto3" json:"sample_count,omitempty"`
|
SampleCount uint64 `protobuf:"varint,1,opt,name=sample_count,json=sampleCount,proto3" json:"sample_count,omitempty"`
|
||||||
SampleSum float64 `protobuf:"fixed64,2,opt,name=sample_sum,json=sampleSum,proto3" json:"sample_sum,omitempty"`
|
SampleSum float64 `protobuf:"fixed64,2,opt,name=sample_sum,json=sampleSum,proto3" json:"sample_sum,omitempty"`
|
||||||
Quantile []Quantile `protobuf:"bytes,3,rep,name=quantile,proto3" json:"quantile"`
|
Quantile []Quantile `protobuf:"bytes,3,rep,name=quantile,proto3" json:"quantile"`
|
||||||
|
CreatedTimestamp *types.Timestamp `protobuf:"bytes,4,opt,name=created_timestamp,json=createdTimestamp,proto3" json:"created_timestamp,omitempty"`
|
||||||
XXX_NoUnkeyedLiteral struct{} `json:"-"`
|
XXX_NoUnkeyedLiteral struct{} `json:"-"`
|
||||||
XXX_unrecognized []byte `json:"-"`
|
XXX_unrecognized []byte `json:"-"`
|
||||||
XXX_sizecache int32 `json:"-"`
|
XXX_sizecache int32 `json:"-"`
|
||||||
|
@ -344,6 +353,13 @@ func (m *Summary) GetQuantile() []Quantile {
|
||||||
return nil
|
return nil
|
||||||
}
|
}
|
||||||
|
|
||||||
|
func (m *Summary) GetCreatedTimestamp() *types.Timestamp {
|
||||||
|
if m != nil {
|
||||||
|
return m.CreatedTimestamp
|
||||||
|
}
|
||||||
|
return nil
|
||||||
|
}
|
||||||
|
|
||||||
type Untyped struct {
|
type Untyped struct {
|
||||||
Value float64 `protobuf:"fixed64,1,opt,name=value,proto3" json:"value,omitempty"`
|
Value float64 `protobuf:"fixed64,1,opt,name=value,proto3" json:"value,omitempty"`
|
||||||
XXX_NoUnkeyedLiteral struct{} `json:"-"`
|
XXX_NoUnkeyedLiteral struct{} `json:"-"`
|
||||||
|
@ -397,6 +413,7 @@ type Histogram struct {
|
||||||
SampleSum float64 `protobuf:"fixed64,2,opt,name=sample_sum,json=sampleSum,proto3" json:"sample_sum,omitempty"`
|
SampleSum float64 `protobuf:"fixed64,2,opt,name=sample_sum,json=sampleSum,proto3" json:"sample_sum,omitempty"`
|
||||||
// Buckets for the conventional histogram.
|
// Buckets for the conventional histogram.
|
||||||
Bucket []Bucket `protobuf:"bytes,3,rep,name=bucket,proto3" json:"bucket"`
|
Bucket []Bucket `protobuf:"bytes,3,rep,name=bucket,proto3" json:"bucket"`
|
||||||
|
CreatedTimestamp *types.Timestamp `protobuf:"bytes,15,opt,name=created_timestamp,json=createdTimestamp,proto3" json:"created_timestamp,omitempty"`
|
||||||
// schema defines the bucket schema. Currently, valid numbers are -4 <= n <= 8.
|
// schema defines the bucket schema. Currently, valid numbers are -4 <= n <= 8.
|
||||||
// They are all for base-2 bucket schemas, where 1 is a bucket boundary in each case, and
|
// They are all for base-2 bucket schemas, where 1 is a bucket boundary in each case, and
|
||||||
// then each power of two is divided into 2^n logarithmic buckets.
|
// then each power of two is divided into 2^n logarithmic buckets.
|
||||||
|
@ -489,6 +506,13 @@ func (m *Histogram) GetBucket() []Bucket {
|
||||||
return nil
|
return nil
|
||||||
}
|
}
|
||||||
|
|
||||||
|
func (m *Histogram) GetCreatedTimestamp() *types.Timestamp {
|
||||||
|
if m != nil {
|
||||||
|
return m.CreatedTimestamp
|
||||||
|
}
|
||||||
|
return nil
|
||||||
|
}
|
||||||
|
|
||||||
func (m *Histogram) GetSchema() int32 {
|
func (m *Histogram) GetSchema() int32 {
|
||||||
if m != nil {
|
if m != nil {
|
||||||
return m.Schema
|
return m.Schema
|
||||||
|
@ -941,65 +965,68 @@ func init() {
|
||||||
}
|
}
|
||||||
|
|
||||||
var fileDescriptor_d1e5ddb18987a258 = []byte{
|
var fileDescriptor_d1e5ddb18987a258 = []byte{
|
||||||
// 923 bytes of a gzipped FileDescriptorProto
|
// 963 bytes of a gzipped FileDescriptorProto
|
||||||
0x1f, 0x8b, 0x08, 0x00, 0x00, 0x00, 0x00, 0x00, 0x02, 0xff, 0xa4, 0x56, 0xdd, 0x8e, 0xdb, 0x44,
|
0x1f, 0x8b, 0x08, 0x00, 0x00, 0x00, 0x00, 0x00, 0x02, 0xff, 0xa4, 0x56, 0xdd, 0x6e, 0x1b, 0x45,
|
||||||
0x18, 0xad, 0x1b, 0xe7, 0xc7, 0x5f, 0x36, 0xdb, 0x74, 0x88, 0x2a, 0x6b, 0x61, 0x37, 0xc1, 0x12,
|
0x14, 0xee, 0x76, 0xfd, 0x93, 0x3d, 0x8e, 0x93, 0xcd, 0x60, 0x55, 0xab, 0x40, 0x62, 0xb3, 0x12,
|
||||||
0xd2, 0x82, 0x50, 0x22, 0xa0, 0x08, 0x54, 0x40, 0x62, 0xb7, 0xdd, 0x6e, 0x51, 0x49, 0x5b, 0x26,
|
0x52, 0x40, 0xc8, 0x16, 0x50, 0x04, 0x2a, 0x45, 0x22, 0x69, 0xd3, 0x14, 0x15, 0xb7, 0x65, 0x6c,
|
||||||
0xc9, 0x45, 0xe1, 0xc2, 0x9a, 0x64, 0x67, 0x1d, 0x0b, 0xdb, 0x63, 0xec, 0x71, 0xc5, 0x72, 0xcf,
|
0x5f, 0x94, 0x9b, 0xd5, 0xd8, 0x9e, 0xac, 0x57, 0xec, 0xee, 0x2c, 0xfb, 0x53, 0x11, 0xee, 0x79,
|
||||||
0x25, 0xd7, 0xbc, 0x02, 0x4f, 0x82, 0x7a, 0xc9, 0x13, 0x20, 0xb4, 0xef, 0xc0, 0x3d, 0x9a, 0x3f,
|
0x06, 0x5e, 0x01, 0xf1, 0x1c, 0x08, 0xf5, 0x92, 0x07, 0x40, 0x08, 0xe5, 0x49, 0xd0, 0xfc, 0xed,
|
||||||
0x3b, 0x5b, 0x39, 0x85, 0x15, 0x77, 0x33, 0xc7, 0xe7, 0xfb, 0xe6, 0x9c, 0x99, 0xc9, 0x99, 0x80,
|
0x3a, 0xd5, 0xba, 0x90, 0xf6, 0x6e, 0xe6, 0xf3, 0x77, 0xce, 0x7c, 0xe7, 0x9b, 0xf1, 0x39, 0x0b,
|
||||||
0x17, 0xb2, 0x49, 0x9a, 0xb1, 0x98, 0xf2, 0x35, 0x2d, 0xf2, 0xc9, 0x2a, 0x0a, 0x69, 0xc2, 0x27,
|
0x6e, 0xc0, 0x46, 0x49, 0xca, 0x22, 0x9a, 0xaf, 0x68, 0x91, 0x8d, 0x16, 0x61, 0x40, 0xe3, 0x7c,
|
||||||
0x31, 0xe5, 0x59, 0xb8, 0xca, 0xc7, 0x69, 0xc6, 0x38, 0x43, 0x83, 0x90, 0x8d, 0x2b, 0xce, 0x58,
|
0x14, 0xd1, 0x3c, 0x0d, 0x16, 0xd9, 0x30, 0x49, 0x59, 0xce, 0x50, 0x2f, 0x60, 0xc3, 0x8a, 0x33,
|
||||||
0x71, 0xf6, 0x06, 0x01, 0x0b, 0x98, 0x24, 0x4c, 0xc4, 0x48, 0x71, 0xf7, 0x86, 0x01, 0x63, 0x41,
|
0x94, 0x9c, 0xfd, 0x9e, 0xcf, 0x7c, 0x26, 0x08, 0x23, 0xbe, 0x92, 0xdc, 0xfd, 0xbe, 0xcf, 0x98,
|
||||||
0x44, 0x27, 0x72, 0xb6, 0x2c, 0xce, 0x27, 0x3c, 0x8c, 0x69, 0xce, 0x49, 0x9c, 0x2a, 0x82, 0xf7,
|
0x1f, 0xd2, 0x91, 0xd8, 0xcd, 0x8b, 0xf3, 0x51, 0x1e, 0x44, 0x34, 0xcb, 0x49, 0x94, 0x48, 0x82,
|
||||||
0x31, 0x38, 0x5f, 0x93, 0x25, 0x8d, 0x9e, 0x91, 0x30, 0x43, 0x08, 0xec, 0x84, 0xc4, 0xd4, 0xb5,
|
0xfb, 0x29, 0x58, 0xdf, 0x90, 0x39, 0x0d, 0x9f, 0x92, 0x20, 0x45, 0x08, 0x1a, 0x31, 0x89, 0xa8,
|
||||||
0x46, 0xd6, 0xa1, 0x83, 0xe5, 0x18, 0x0d, 0xa0, 0xf9, 0x82, 0x44, 0x05, 0x75, 0x6f, 0x4a, 0x50,
|
0x63, 0x0c, 0x8c, 0x23, 0x0b, 0x8b, 0x35, 0xea, 0x41, 0xf3, 0x39, 0x09, 0x0b, 0xea, 0xdc, 0x14,
|
||||||
0x4d, 0xbc, 0x7d, 0x68, 0x9e, 0x92, 0x22, 0xd8, 0xf8, 0x2c, 0x6a, 0x2c, 0xf3, 0xf9, 0x3b, 0x68,
|
0xa0, 0xdc, 0xb8, 0x07, 0xd0, 0x3c, 0x23, 0x85, 0xbf, 0xf6, 0x33, 0x8f, 0x31, 0xf4, 0xcf, 0xbf,
|
||||||
0xdf, 0x67, 0x45, 0xc2, 0x69, 0x56, 0x4f, 0x40, 0xf7, 0xa0, 0x43, 0x7f, 0xa4, 0x71, 0x1a, 0x91,
|
0x19, 0xd0, 0xbe, 0xc7, 0x8a, 0x38, 0xa7, 0x69, 0x3d, 0x03, 0xdd, 0x81, 0x2d, 0xfa, 0x23, 0x8d,
|
||||||
0x4c, 0x36, 0xee, 0x7e, 0x78, 0x30, 0xae, 0xb3, 0x35, 0x3e, 0xd1, 0x2c, 0x5c, 0xf2, 0xbd, 0xcf,
|
0x92, 0x90, 0xa4, 0x22, 0x73, 0xe7, 0xe3, 0xc3, 0x61, 0x5d, 0x5d, 0xc3, 0x53, 0xc5, 0xc2, 0x25,
|
||||||
0xa1, 0xf3, 0x4d, 0x41, 0x12, 0x1e, 0x46, 0x14, 0xed, 0x41, 0xe7, 0x07, 0x3d, 0xd6, 0x0b, 0x94,
|
0x1f, 0x8d, 0x61, 0x6f, 0x91, 0x52, 0x92, 0xd3, 0xa5, 0x57, 0x96, 0xe3, 0x98, 0x22, 0xc9, 0xfe,
|
||||||
0xf3, 0xab, 0xca, 0x4b, 0x69, 0xbf, 0x58, 0xd0, 0x9e, 0x15, 0x71, 0x4c, 0xb2, 0x0b, 0xf4, 0x36,
|
0x50, 0x16, 0x3c, 0xd4, 0x05, 0x0f, 0xa7, 0x9a, 0x71, 0xd2, 0x78, 0xf1, 0x77, 0xdf, 0xc0, 0xb6,
|
||||||
0xec, 0xe4, 0x24, 0x4e, 0x23, 0xea, 0xaf, 0x84, 0x5a, 0xd9, 0xc1, 0xc6, 0x5d, 0x85, 0x49, 0x03,
|
0x0a, 0x2d, 0x71, 0xf7, 0x2e, 0x6c, 0x7d, 0x5b, 0x90, 0x38, 0x0f, 0x42, 0x8a, 0xf6, 0x61, 0xeb,
|
||||||
0x68, 0x1f, 0x40, 0x53, 0xf2, 0x22, 0xd6, 0x9d, 0x1c, 0x85, 0xcc, 0x8a, 0x18, 0x7d, 0xb9, 0xb1,
|
0x07, 0xb5, 0x56, 0x7a, 0xcb, 0xfd, 0x55, 0x27, 0xca, 0x52, 0xff, 0x32, 0xa0, 0x3d, 0x29, 0xa2,
|
||||||
0x7e, 0x63, 0xd4, 0xd8, 0xee, 0xc3, 0x28, 0x3e, 0xb6, 0x5f, 0xfe, 0x39, 0xbc, 0x51, 0xa9, 0xf4,
|
0x88, 0xa4, 0x17, 0xe8, 0x5d, 0xd8, 0xce, 0x48, 0x94, 0x84, 0xd4, 0x5b, 0xf0, 0xe2, 0x45, 0x86,
|
||||||
0x86, 0xd0, 0x5e, 0x24, 0xfc, 0x22, 0xa5, 0x67, 0x5b, 0xf6, 0xf2, 0x6f, 0x1b, 0x9c, 0x47, 0x61,
|
0x06, 0xee, 0x48, 0x4c, 0xf8, 0x81, 0x0e, 0x00, 0x14, 0x25, 0x2b, 0x22, 0x95, 0xc9, 0x92, 0xc8,
|
||||||
0xce, 0x59, 0x90, 0x91, 0xf8, 0xbf, 0x48, 0x7e, 0x1f, 0xd0, 0x26, 0xc5, 0x3f, 0x8f, 0x18, 0xe1,
|
0xa4, 0x88, 0xd0, 0x57, 0x6b, 0xe7, 0x9b, 0x03, 0x73, 0xb3, 0x2d, 0x5a, 0xb1, 0xa8, 0xea, 0xc6,
|
||||||
0xae, 0x2d, 0x7b, 0xf6, 0x37, 0x88, 0x0f, 0x05, 0xfe, 0x6f, 0x06, 0xef, 0x41, 0x6b, 0x59, 0xac,
|
0x9a, 0xca, 0x5a, 0x73, 0x1a, 0xaf, 0x6d, 0x4e, 0x1f, 0xda, 0xb3, 0x38, 0xbf, 0x48, 0xe8, 0x72,
|
||||||
0xbe, 0xa7, 0x5c, 0xdb, 0x7b, 0xab, 0xde, 0xde, 0xb1, 0xe4, 0x68, 0x73, 0xba, 0x02, 0xdd, 0x81,
|
0xc3, 0x55, 0xff, 0xde, 0x04, 0xeb, 0x61, 0x90, 0xe5, 0xcc, 0x4f, 0x49, 0xf4, 0x7f, 0x1c, 0xf8,
|
||||||
0x56, 0xbe, 0x5a, 0xd3, 0x98, 0xb8, 0xcd, 0x91, 0x75, 0x78, 0x1b, 0xeb, 0x19, 0x7a, 0x07, 0x76,
|
0x10, 0xd0, 0x3a, 0xc5, 0x3b, 0x0f, 0x19, 0xc9, 0x85, 0x42, 0x03, 0xdb, 0x6b, 0xc4, 0x07, 0x1c,
|
||||||
0x7f, 0xa2, 0x19, 0xf3, 0xf9, 0x3a, 0xa3, 0xf9, 0x9a, 0x45, 0x67, 0x6e, 0x4b, 0x2e, 0xdb, 0x13,
|
0xff, 0x2f, 0xbf, 0xee, 0x40, 0x6b, 0x5e, 0x2c, 0xbe, 0xa7, 0xb9, 0x72, 0xeb, 0x9d, 0x7a, 0xb7,
|
||||||
0xe8, 0xdc, 0x80, 0x42, 0x99, 0xa4, 0x29, 0xa3, 0x6d, 0x69, 0xd4, 0x11, 0x88, 0xb2, 0x79, 0x08,
|
0x4e, 0x04, 0x47, 0x79, 0xa5, 0x22, 0xea, 0x9d, 0xda, 0x7d, 0x5d, 0xa7, 0xd0, 0x2d, 0x68, 0x65,
|
||||||
0xfd, 0xea, 0xb3, 0x36, 0xd9, 0x91, 0x7d, 0x76, 0x4b, 0x92, 0xb2, 0xf8, 0x18, 0x7a, 0x09, 0x0d,
|
0x8b, 0x15, 0x8d, 0x88, 0xd3, 0x1c, 0x18, 0x47, 0x7b, 0x58, 0xed, 0xd0, 0x7b, 0xb0, 0xf3, 0x13,
|
||||||
0x08, 0x0f, 0x5f, 0x50, 0x3f, 0x4f, 0x49, 0xe2, 0x3a, 0xd2, 0xca, 0xe8, 0x75, 0x56, 0x66, 0x29,
|
0x4d, 0x99, 0x97, 0xaf, 0x52, 0x9a, 0xad, 0x58, 0xb8, 0x74, 0x5a, 0xa2, 0x8a, 0x2e, 0x47, 0xa7,
|
||||||
0x49, 0xb4, 0x9d, 0x1d, 0x53, 0x2c, 0x30, 0x21, 0xbe, 0x6c, 0x76, 0x46, 0x23, 0x4e, 0x5c, 0x18,
|
0x1a, 0xe4, 0x85, 0x0a, 0x9a, 0xf4, 0xad, 0x2d, 0x7c, 0xb3, 0x38, 0x22, 0x5d, 0x3b, 0x02, 0xbb,
|
||||||
0x35, 0x0e, 0x11, 0x2e, 0x97, 0x78, 0x20, 0xc0, 0x2b, 0x34, 0x65, 0xa0, 0x3b, 0x6a, 0x08, 0x8f,
|
0xfa, 0x59, 0x79, 0xb6, 0x25, 0xf2, 0xec, 0x94, 0x24, 0xe9, 0xd8, 0x23, 0xe8, 0xc6, 0xd4, 0x27,
|
||||||
0x06, 0x55, 0x26, 0x1e, 0x43, 0x2f, 0x65, 0x79, 0x58, 0x49, 0xdb, 0xb9, 0x9e, 0x34, 0x53, 0x6c,
|
0x79, 0xf0, 0x9c, 0x7a, 0x59, 0x42, 0x62, 0xc7, 0x12, 0xce, 0x0c, 0x5e, 0xe5, 0xcc, 0x24, 0x21,
|
||||||
0xa4, 0x95, 0xcd, 0x94, 0xb4, 0x9e, 0x92, 0x66, 0xd0, 0x52, 0x5a, 0x49, 0x53, 0xd2, 0x76, 0x95,
|
0xb1, 0x72, 0x67, 0x5b, 0x07, 0x73, 0x8c, 0x8b, 0x2f, 0x93, 0x2d, 0x69, 0x98, 0x13, 0x07, 0x06,
|
||||||
0x34, 0x83, 0x4a, 0x69, 0xde, 0xef, 0x16, 0xb4, 0xd4, 0x82, 0xe8, 0x5d, 0xe8, 0xaf, 0x8a, 0xb8,
|
0xe6, 0x11, 0xc2, 0xe5, 0x11, 0xf7, 0x39, 0x78, 0x85, 0x26, 0x0b, 0xe8, 0x0c, 0x4c, 0x5e, 0xa3,
|
||||||
0x88, 0x36, 0xed, 0xa8, 0x8b, 0x77, 0xab, 0xc2, 0x95, 0xa1, 0xbb, 0x70, 0xe7, 0x55, 0xea, 0x95,
|
0x46, 0x65, 0x11, 0x8f, 0xa0, 0x9b, 0xb0, 0x2c, 0xa8, 0xa4, 0x6d, 0x5f, 0x4f, 0x9a, 0x0e, 0xd6,
|
||||||
0x0b, 0x38, 0x78, 0xa5, 0x40, 0x9d, 0xd0, 0x10, 0xba, 0x45, 0x9a, 0xd2, 0xcc, 0x5f, 0xb2, 0x22,
|
0xd2, 0xca, 0x64, 0x52, 0x5a, 0x57, 0x4a, 0xd3, 0x68, 0x29, 0xad, 0xa4, 0x49, 0x69, 0x3b, 0x52,
|
||||||
0x39, 0xd3, 0xb7, 0x10, 0x24, 0x74, 0x2c, 0x90, 0x2b, 0x79, 0xd1, 0xb8, 0x76, 0x5e, 0x40, 0xb5,
|
0x9a, 0x46, 0x85, 0x34, 0xf7, 0x0f, 0x03, 0x5a, 0xf2, 0x40, 0xf4, 0x3e, 0xd8, 0x8b, 0x22, 0x2a,
|
||||||
0x71, 0xe2, 0x52, 0xb2, 0xf3, 0xf3, 0x9c, 0x2a, 0x07, 0xb7, 0xb1, 0x9e, 0x09, 0x3c, 0xa2, 0x49,
|
0xc2, 0xf5, 0x72, 0xe4, 0x3b, 0xde, 0xad, 0x70, 0x59, 0xd0, 0x6d, 0xb8, 0xf5, 0x32, 0xf5, 0xca,
|
||||||
0xc0, 0xd7, 0x72, 0xf5, 0x1e, 0xd6, 0x33, 0xef, 0x57, 0x0b, 0x3a, 0xa6, 0x29, 0xfa, 0x0c, 0x9a,
|
0x7b, 0xee, 0xbd, 0x14, 0x20, 0x6f, 0xa8, 0x0f, 0x9d, 0x22, 0x49, 0x68, 0xea, 0xcd, 0x59, 0x11,
|
||||||
0x91, 0x48, 0x4b, 0xd7, 0x92, 0xc7, 0x34, 0xac, 0xd7, 0x50, 0x06, 0xaa, 0x3e, 0x25, 0x55, 0x53,
|
0x2f, 0xd5, 0xa3, 0x06, 0x01, 0x9d, 0x70, 0xe4, 0x4a, 0x73, 0x34, 0xaf, 0xd7, 0x1c, 0xdd, 0xbb,
|
||||||
0x9f, 0x47, 0xe8, 0x53, 0x70, 0xca, 0x4c, 0xd6, 0xd6, 0xf6, 0xc6, 0x2a, 0xb5, 0xc7, 0x26, 0xb5,
|
0x00, 0x95, 0x71, 0xfc, 0x51, 0xb2, 0xf3, 0xf3, 0x8c, 0xca, 0x0a, 0xf6, 0xb0, 0xda, 0x71, 0x3c,
|
||||||
0xc7, 0x73, 0xc3, 0xc0, 0x15, 0xd9, 0xfb, 0xb9, 0x01, 0xad, 0xa9, 0x7c, 0x19, 0xfe, 0x9f, 0xae,
|
0xa4, 0xb1, 0x9f, 0xaf, 0xc4, 0xe9, 0x5d, 0xac, 0x76, 0xee, 0x2f, 0x06, 0x6c, 0xe9, 0xa4, 0xe8,
|
||||||
0x0f, 0xa0, 0x19, 0x88, 0x2c, 0xd7, 0x41, 0xfc, 0x66, 0x7d, 0xb1, 0x8c, 0x7b, 0xac, 0x98, 0xe8,
|
0x0b, 0x68, 0x86, 0x7c, 0x36, 0x38, 0x86, 0xb8, 0xa6, 0x7e, 0xbd, 0x86, 0x72, 0x7c, 0xa8, 0x5b,
|
||||||
0x13, 0x68, 0xaf, 0x54, 0xbe, 0x6b, 0xc9, 0xfb, 0xf5, 0x45, 0xfa, 0x11, 0xc0, 0x86, 0x2d, 0x0a,
|
0x92, 0x31, 0xf5, 0xdd, 0x12, 0x7d, 0x0e, 0xd6, 0x35, 0x5a, 0x36, 0xae, 0xc8, 0xee, 0xcf, 0x26,
|
||||||
0x73, 0x15, 0xbe, 0xf2, 0x3e, 0x6c, 0x2d, 0xd4, 0x09, 0x8d, 0x0d, 0x5b, 0x14, 0x16, 0x2a, 0x26,
|
0xb4, 0xc6, 0x62, 0x0e, 0xbe, 0x99, 0xae, 0x8f, 0xa0, 0xe9, 0xf3, 0xc9, 0xa5, 0xa6, 0xce, 0xdb,
|
||||||
0x65, 0x98, 0x6c, 0x2d, 0xd4, 0x59, 0x8a, 0x0d, 0x1b, 0x7d, 0x01, 0xce, 0xda, 0xa4, 0xa7, 0x0c,
|
0xf5, 0xc1, 0x62, 0xb8, 0x61, 0xc9, 0x44, 0x9f, 0x41, 0x7b, 0x21, 0x87, 0x99, 0x92, 0x7c, 0x50,
|
||||||
0x91, 0xad, 0xdb, 0x53, 0x86, 0x2c, 0xae, 0x2a, 0x44, 0xde, 0x96, 0x3b, 0xee, 0xc7, 0xb9, 0x4c,
|
0x1f, 0xa4, 0x26, 0x1e, 0xd6, 0x6c, 0x1e, 0x98, 0xc9, 0xd1, 0xa0, 0x3a, 0xf0, 0x86, 0x40, 0x35,
|
||||||
0xaa, 0x06, 0xee, 0x96, 0xd8, 0x34, 0xf7, 0x7e, 0xb3, 0x60, 0x47, 0x9d, 0xc3, 0x43, 0x12, 0x87,
|
0x3f, 0xb0, 0x66, 0xf3, 0xc0, 0x42, 0x76, 0x5d, 0xd1, 0x4c, 0x36, 0x06, 0xaa, 0xd6, 0x8c, 0x35,
|
||||||
0xd1, 0x45, 0xed, 0x33, 0x8a, 0xc0, 0x5e, 0xd3, 0x28, 0xd5, 0xaf, 0xa8, 0x1c, 0xa3, 0xbb, 0x60,
|
0x1b, 0x7d, 0x09, 0xd6, 0x4a, 0x37, 0x63, 0xd1, 0x44, 0x36, 0xda, 0x53, 0xf6, 0x6c, 0x5c, 0x45,
|
||||||
0x0b, 0x8d, 0x72, 0x0b, 0x77, 0xb7, 0xfd, 0xe6, 0x55, 0xe7, 0xf9, 0x45, 0x4a, 0xb1, 0x64, 0x8b,
|
0xf0, 0xf6, 0x5d, 0x3a, 0xee, 0x45, 0x99, 0xe8, 0x54, 0x26, 0xee, 0x94, 0xd8, 0x38, 0x73, 0x7f,
|
||||||
0x44, 0x56, 0xff, 0x07, 0x5c, 0xfb, 0x75, 0x89, 0xac, 0xea, 0x4c, 0x22, 0xab, 0x8a, 0xf7, 0x96,
|
0x35, 0x60, 0x5b, 0xde, 0xc3, 0x03, 0x12, 0x05, 0xe1, 0x45, 0xed, 0x47, 0x03, 0x82, 0xc6, 0x8a,
|
||||||
0x00, 0x55, 0x3f, 0xd4, 0x85, 0xf6, 0xfd, 0xa7, 0x8b, 0x27, 0xf3, 0x13, 0xdc, 0xbf, 0x81, 0x1c,
|
0x86, 0x89, 0xfa, 0x66, 0x10, 0x6b, 0x74, 0x1b, 0x1a, 0x5c, 0xa3, 0xb0, 0x70, 0x67, 0xd3, 0x7f,
|
||||||
0x68, 0x9e, 0x1e, 0x2d, 0x4e, 0x4f, 0xfa, 0x96, 0xc0, 0x67, 0x8b, 0xe9, 0xf4, 0x08, 0x3f, 0xef,
|
0x5e, 0x66, 0x9e, 0x5e, 0x24, 0x14, 0x0b, 0x36, 0x6f, 0xf0, 0xf2, 0xeb, 0xc7, 0x69, 0xbc, 0xaa,
|
||||||
0xdf, 0x14, 0x93, 0xc5, 0x93, 0xf9, 0xf3, 0x67, 0x27, 0x0f, 0xfa, 0x0d, 0xd4, 0x03, 0xe7, 0xd1,
|
0xc1, 0xcb, 0x38, 0xdd, 0xe0, 0x65, 0xc4, 0x07, 0x73, 0x80, 0x2a, 0x1f, 0xea, 0x40, 0xfb, 0xde,
|
||||||
0x57, 0xb3, 0xf9, 0xd3, 0x53, 0x7c, 0x34, 0xed, 0xdb, 0xe8, 0x0d, 0xb8, 0x25, 0x6b, 0xfc, 0x0a,
|
0x93, 0xd9, 0xe3, 0xe9, 0x29, 0xb6, 0x6f, 0x20, 0x0b, 0x9a, 0x67, 0xc7, 0xb3, 0xb3, 0x53, 0xdb,
|
||||||
0x6c, 0x1e, 0x7b, 0x2f, 0x2f, 0x0f, 0xac, 0x3f, 0x2e, 0x0f, 0xac, 0xbf, 0x2e, 0x0f, 0xac, 0x6f,
|
0xe0, 0xf8, 0x64, 0x36, 0x1e, 0x1f, 0xe3, 0x67, 0xf6, 0x4d, 0xbe, 0x99, 0x3d, 0x9e, 0x3e, 0x7b,
|
||||||
0x07, 0x21, 0xf3, 0x2b, 0x71, 0xbe, 0x12, 0xb7, 0x6c, 0xc9, 0x9b, 0xfd, 0xd1, 0x3f, 0x01, 0x00,
|
0x7a, 0x7a, 0xdf, 0x36, 0x51, 0x17, 0xac, 0x87, 0x5f, 0x4f, 0xa6, 0x4f, 0xce, 0xf0, 0xf1, 0xd8,
|
||||||
0x00, 0xff, 0xff, 0x52, 0x2d, 0xb5, 0x31, 0xef, 0x08, 0x00, 0x00,
|
0x6e, 0xa0, 0xb7, 0x60, 0x57, 0xc4, 0x78, 0x15, 0xd8, 0x3c, 0x71, 0x5f, 0x5c, 0x1e, 0x1a, 0x7f,
|
||||||
|
0x5e, 0x1e, 0x1a, 0xff, 0x5c, 0x1e, 0x1a, 0xdf, 0xf5, 0x02, 0xe6, 0x55, 0xe2, 0x3c, 0x29, 0x6e,
|
||||||
|
0xde, 0x12, 0x2f, 0xfb, 0x93, 0x7f, 0x03, 0x00, 0x00, 0xff, 0xff, 0x68, 0x3f, 0xd9, 0x07, 0xdd,
|
||||||
|
0x09, 0x00, 0x00,
|
||||||
}
|
}
|
||||||
|
|
||||||
func (m *LabelPair) Marshal() (dAtA []byte, err error) {
|
func (m *LabelPair) Marshal() (dAtA []byte, err error) {
|
||||||
|
@ -1100,6 +1127,18 @@ func (m *Counter) MarshalToSizedBuffer(dAtA []byte) (int, error) {
|
||||||
i -= len(m.XXX_unrecognized)
|
i -= len(m.XXX_unrecognized)
|
||||||
copy(dAtA[i:], m.XXX_unrecognized)
|
copy(dAtA[i:], m.XXX_unrecognized)
|
||||||
}
|
}
|
||||||
|
if m.CreatedTimestamp != nil {
|
||||||
|
{
|
||||||
|
size, err := m.CreatedTimestamp.MarshalToSizedBuffer(dAtA[:i])
|
||||||
|
if err != nil {
|
||||||
|
return 0, err
|
||||||
|
}
|
||||||
|
i -= size
|
||||||
|
i = encodeVarintMetrics(dAtA, i, uint64(size))
|
||||||
|
}
|
||||||
|
i--
|
||||||
|
dAtA[i] = 0x1a
|
||||||
|
}
|
||||||
if m.Exemplar != nil {
|
if m.Exemplar != nil {
|
||||||
{
|
{
|
||||||
size, err := m.Exemplar.MarshalToSizedBuffer(dAtA[:i])
|
size, err := m.Exemplar.MarshalToSizedBuffer(dAtA[:i])
|
||||||
|
@ -1184,6 +1223,18 @@ func (m *Summary) MarshalToSizedBuffer(dAtA []byte) (int, error) {
|
||||||
i -= len(m.XXX_unrecognized)
|
i -= len(m.XXX_unrecognized)
|
||||||
copy(dAtA[i:], m.XXX_unrecognized)
|
copy(dAtA[i:], m.XXX_unrecognized)
|
||||||
}
|
}
|
||||||
|
if m.CreatedTimestamp != nil {
|
||||||
|
{
|
||||||
|
size, err := m.CreatedTimestamp.MarshalToSizedBuffer(dAtA[:i])
|
||||||
|
if err != nil {
|
||||||
|
return 0, err
|
||||||
|
}
|
||||||
|
i -= size
|
||||||
|
i = encodeVarintMetrics(dAtA, i, uint64(size))
|
||||||
|
}
|
||||||
|
i--
|
||||||
|
dAtA[i] = 0x22
|
||||||
|
}
|
||||||
if len(m.Quantile) > 0 {
|
if len(m.Quantile) > 0 {
|
||||||
for iNdEx := len(m.Quantile) - 1; iNdEx >= 0; iNdEx-- {
|
for iNdEx := len(m.Quantile) - 1; iNdEx >= 0; iNdEx-- {
|
||||||
{
|
{
|
||||||
|
@ -1269,32 +1320,44 @@ func (m *Histogram) MarshalToSizedBuffer(dAtA []byte) (int, error) {
|
||||||
i -= len(m.XXX_unrecognized)
|
i -= len(m.XXX_unrecognized)
|
||||||
copy(dAtA[i:], m.XXX_unrecognized)
|
copy(dAtA[i:], m.XXX_unrecognized)
|
||||||
}
|
}
|
||||||
|
if m.CreatedTimestamp != nil {
|
||||||
|
{
|
||||||
|
size, err := m.CreatedTimestamp.MarshalToSizedBuffer(dAtA[:i])
|
||||||
|
if err != nil {
|
||||||
|
return 0, err
|
||||||
|
}
|
||||||
|
i -= size
|
||||||
|
i = encodeVarintMetrics(dAtA, i, uint64(size))
|
||||||
|
}
|
||||||
|
i--
|
||||||
|
dAtA[i] = 0x7a
|
||||||
|
}
|
||||||
if len(m.PositiveCount) > 0 {
|
if len(m.PositiveCount) > 0 {
|
||||||
for iNdEx := len(m.PositiveCount) - 1; iNdEx >= 0; iNdEx-- {
|
for iNdEx := len(m.PositiveCount) - 1; iNdEx >= 0; iNdEx-- {
|
||||||
f2 := math.Float64bits(float64(m.PositiveCount[iNdEx]))
|
f5 := math.Float64bits(float64(m.PositiveCount[iNdEx]))
|
||||||
i -= 8
|
i -= 8
|
||||||
encoding_binary.LittleEndian.PutUint64(dAtA[i:], uint64(f2))
|
encoding_binary.LittleEndian.PutUint64(dAtA[i:], uint64(f5))
|
||||||
}
|
}
|
||||||
i = encodeVarintMetrics(dAtA, i, uint64(len(m.PositiveCount)*8))
|
i = encodeVarintMetrics(dAtA, i, uint64(len(m.PositiveCount)*8))
|
||||||
i--
|
i--
|
||||||
dAtA[i] = 0x72
|
dAtA[i] = 0x72
|
||||||
}
|
}
|
||||||
if len(m.PositiveDelta) > 0 {
|
if len(m.PositiveDelta) > 0 {
|
||||||
var j3 int
|
var j6 int
|
||||||
dAtA5 := make([]byte, len(m.PositiveDelta)*10)
|
dAtA8 := make([]byte, len(m.PositiveDelta)*10)
|
||||||
for _, num := range m.PositiveDelta {
|
for _, num := range m.PositiveDelta {
|
||||||
x4 := (uint64(num) << 1) ^ uint64((num >> 63))
|
x7 := (uint64(num) << 1) ^ uint64((num >> 63))
|
||||||
for x4 >= 1<<7 {
|
for x7 >= 1<<7 {
|
||||||
dAtA5[j3] = uint8(uint64(x4)&0x7f | 0x80)
|
dAtA8[j6] = uint8(uint64(x7)&0x7f | 0x80)
|
||||||
j3++
|
j6++
|
||||||
x4 >>= 7
|
x7 >>= 7
|
||||||
}
|
}
|
||||||
dAtA5[j3] = uint8(x4)
|
dAtA8[j6] = uint8(x7)
|
||||||
j3++
|
j6++
|
||||||
}
|
}
|
||||||
i -= j3
|
i -= j6
|
||||||
copy(dAtA[i:], dAtA5[:j3])
|
copy(dAtA[i:], dAtA8[:j6])
|
||||||
i = encodeVarintMetrics(dAtA, i, uint64(j3))
|
i = encodeVarintMetrics(dAtA, i, uint64(j6))
|
||||||
i--
|
i--
|
||||||
dAtA[i] = 0x6a
|
dAtA[i] = 0x6a
|
||||||
}
|
}
|
||||||
|
@ -1314,30 +1377,30 @@ func (m *Histogram) MarshalToSizedBuffer(dAtA []byte) (int, error) {
|
||||||
}
|
}
|
||||||
if len(m.NegativeCount) > 0 {
|
if len(m.NegativeCount) > 0 {
|
||||||
for iNdEx := len(m.NegativeCount) - 1; iNdEx >= 0; iNdEx-- {
|
for iNdEx := len(m.NegativeCount) - 1; iNdEx >= 0; iNdEx-- {
|
||||||
f6 := math.Float64bits(float64(m.NegativeCount[iNdEx]))
|
f9 := math.Float64bits(float64(m.NegativeCount[iNdEx]))
|
||||||
i -= 8
|
i -= 8
|
||||||
encoding_binary.LittleEndian.PutUint64(dAtA[i:], uint64(f6))
|
encoding_binary.LittleEndian.PutUint64(dAtA[i:], uint64(f9))
|
||||||
}
|
}
|
||||||
i = encodeVarintMetrics(dAtA, i, uint64(len(m.NegativeCount)*8))
|
i = encodeVarintMetrics(dAtA, i, uint64(len(m.NegativeCount)*8))
|
||||||
i--
|
i--
|
||||||
dAtA[i] = 0x5a
|
dAtA[i] = 0x5a
|
||||||
}
|
}
|
||||||
if len(m.NegativeDelta) > 0 {
|
if len(m.NegativeDelta) > 0 {
|
||||||
var j7 int
|
var j10 int
|
||||||
dAtA9 := make([]byte, len(m.NegativeDelta)*10)
|
dAtA12 := make([]byte, len(m.NegativeDelta)*10)
|
||||||
for _, num := range m.NegativeDelta {
|
for _, num := range m.NegativeDelta {
|
||||||
x8 := (uint64(num) << 1) ^ uint64((num >> 63))
|
x11 := (uint64(num) << 1) ^ uint64((num >> 63))
|
||||||
for x8 >= 1<<7 {
|
for x11 >= 1<<7 {
|
||||||
dAtA9[j7] = uint8(uint64(x8)&0x7f | 0x80)
|
dAtA12[j10] = uint8(uint64(x11)&0x7f | 0x80)
|
||||||
j7++
|
j10++
|
||||||
x8 >>= 7
|
x11 >>= 7
|
||||||
}
|
}
|
||||||
dAtA9[j7] = uint8(x8)
|
dAtA12[j10] = uint8(x11)
|
||||||
j7++
|
j10++
|
||||||
}
|
}
|
||||||
i -= j7
|
i -= j10
|
||||||
copy(dAtA[i:], dAtA9[:j7])
|
copy(dAtA[i:], dAtA12[:j10])
|
||||||
i = encodeVarintMetrics(dAtA, i, uint64(j7))
|
i = encodeVarintMetrics(dAtA, i, uint64(j10))
|
||||||
i--
|
i--
|
||||||
dAtA[i] = 0x52
|
dAtA[i] = 0x52
|
||||||
}
|
}
|
||||||
|
@ -1788,6 +1851,10 @@ func (m *Counter) Size() (n int) {
|
||||||
l = m.Exemplar.Size()
|
l = m.Exemplar.Size()
|
||||||
n += 1 + l + sovMetrics(uint64(l))
|
n += 1 + l + sovMetrics(uint64(l))
|
||||||
}
|
}
|
||||||
|
if m.CreatedTimestamp != nil {
|
||||||
|
l = m.CreatedTimestamp.Size()
|
||||||
|
n += 1 + l + sovMetrics(uint64(l))
|
||||||
|
}
|
||||||
if m.XXX_unrecognized != nil {
|
if m.XXX_unrecognized != nil {
|
||||||
n += len(m.XXX_unrecognized)
|
n += len(m.XXX_unrecognized)
|
||||||
}
|
}
|
||||||
|
@ -1830,6 +1897,10 @@ func (m *Summary) Size() (n int) {
|
||||||
n += 1 + l + sovMetrics(uint64(l))
|
n += 1 + l + sovMetrics(uint64(l))
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
if m.CreatedTimestamp != nil {
|
||||||
|
l = m.CreatedTimestamp.Size()
|
||||||
|
n += 1 + l + sovMetrics(uint64(l))
|
||||||
|
}
|
||||||
if m.XXX_unrecognized != nil {
|
if m.XXX_unrecognized != nil {
|
||||||
n += len(m.XXX_unrecognized)
|
n += len(m.XXX_unrecognized)
|
||||||
}
|
}
|
||||||
|
@ -1916,6 +1987,10 @@ func (m *Histogram) Size() (n int) {
|
||||||
if len(m.PositiveCount) > 0 {
|
if len(m.PositiveCount) > 0 {
|
||||||
n += 1 + sovMetrics(uint64(len(m.PositiveCount)*8)) + len(m.PositiveCount)*8
|
n += 1 + sovMetrics(uint64(len(m.PositiveCount)*8)) + len(m.PositiveCount)*8
|
||||||
}
|
}
|
||||||
|
if m.CreatedTimestamp != nil {
|
||||||
|
l = m.CreatedTimestamp.Size()
|
||||||
|
n += 1 + l + sovMetrics(uint64(l))
|
||||||
|
}
|
||||||
if m.XXX_unrecognized != nil {
|
if m.XXX_unrecognized != nil {
|
||||||
n += len(m.XXX_unrecognized)
|
n += len(m.XXX_unrecognized)
|
||||||
}
|
}
|
||||||
|
@ -2319,6 +2394,42 @@ func (m *Counter) Unmarshal(dAtA []byte) error {
|
||||||
return err
|
return err
|
||||||
}
|
}
|
||||||
iNdEx = postIndex
|
iNdEx = postIndex
|
||||||
|
case 3:
|
||||||
|
if wireType != 2 {
|
||||||
|
return fmt.Errorf("proto: wrong wireType = %d for field CreatedTimestamp", wireType)
|
||||||
|
}
|
||||||
|
var msglen int
|
||||||
|
for shift := uint(0); ; shift += 7 {
|
||||||
|
if shift >= 64 {
|
||||||
|
return ErrIntOverflowMetrics
|
||||||
|
}
|
||||||
|
if iNdEx >= l {
|
||||||
|
return io.ErrUnexpectedEOF
|
||||||
|
}
|
||||||
|
b := dAtA[iNdEx]
|
||||||
|
iNdEx++
|
||||||
|
msglen |= int(b&0x7F) << shift
|
||||||
|
if b < 0x80 {
|
||||||
|
break
|
||||||
|
}
|
||||||
|
}
|
||||||
|
if msglen < 0 {
|
||||||
|
return ErrInvalidLengthMetrics
|
||||||
|
}
|
||||||
|
postIndex := iNdEx + msglen
|
||||||
|
if postIndex < 0 {
|
||||||
|
return ErrInvalidLengthMetrics
|
||||||
|
}
|
||||||
|
if postIndex > l {
|
||||||
|
return io.ErrUnexpectedEOF
|
||||||
|
}
|
||||||
|
if m.CreatedTimestamp == nil {
|
||||||
|
m.CreatedTimestamp = &types.Timestamp{}
|
||||||
|
}
|
||||||
|
if err := m.CreatedTimestamp.Unmarshal(dAtA[iNdEx:postIndex]); err != nil {
|
||||||
|
return err
|
||||||
|
}
|
||||||
|
iNdEx = postIndex
|
||||||
default:
|
default:
|
||||||
iNdEx = preIndex
|
iNdEx = preIndex
|
||||||
skippy, err := skipMetrics(dAtA[iNdEx:])
|
skippy, err := skipMetrics(dAtA[iNdEx:])
|
||||||
|
@ -2507,6 +2618,42 @@ func (m *Summary) Unmarshal(dAtA []byte) error {
|
||||||
return err
|
return err
|
||||||
}
|
}
|
||||||
iNdEx = postIndex
|
iNdEx = postIndex
|
||||||
|
case 4:
|
||||||
|
if wireType != 2 {
|
||||||
|
return fmt.Errorf("proto: wrong wireType = %d for field CreatedTimestamp", wireType)
|
||||||
|
}
|
||||||
|
var msglen int
|
||||||
|
for shift := uint(0); ; shift += 7 {
|
||||||
|
if shift >= 64 {
|
||||||
|
return ErrIntOverflowMetrics
|
||||||
|
}
|
||||||
|
if iNdEx >= l {
|
||||||
|
return io.ErrUnexpectedEOF
|
||||||
|
}
|
||||||
|
b := dAtA[iNdEx]
|
||||||
|
iNdEx++
|
||||||
|
msglen |= int(b&0x7F) << shift
|
||||||
|
if b < 0x80 {
|
||||||
|
break
|
||||||
|
}
|
||||||
|
}
|
||||||
|
if msglen < 0 {
|
||||||
|
return ErrInvalidLengthMetrics
|
||||||
|
}
|
||||||
|
postIndex := iNdEx + msglen
|
||||||
|
if postIndex < 0 {
|
||||||
|
return ErrInvalidLengthMetrics
|
||||||
|
}
|
||||||
|
if postIndex > l {
|
||||||
|
return io.ErrUnexpectedEOF
|
||||||
|
}
|
||||||
|
if m.CreatedTimestamp == nil {
|
||||||
|
m.CreatedTimestamp = &types.Timestamp{}
|
||||||
|
}
|
||||||
|
if err := m.CreatedTimestamp.Unmarshal(dAtA[iNdEx:postIndex]); err != nil {
|
||||||
|
return err
|
||||||
|
}
|
||||||
|
iNdEx = postIndex
|
||||||
default:
|
default:
|
||||||
iNdEx = preIndex
|
iNdEx = preIndex
|
||||||
skippy, err := skipMetrics(dAtA[iNdEx:])
|
skippy, err := skipMetrics(dAtA[iNdEx:])
|
||||||
|
@ -3089,6 +3236,42 @@ func (m *Histogram) Unmarshal(dAtA []byte) error {
|
||||||
} else {
|
} else {
|
||||||
return fmt.Errorf("proto: wrong wireType = %d for field PositiveCount", wireType)
|
return fmt.Errorf("proto: wrong wireType = %d for field PositiveCount", wireType)
|
||||||
}
|
}
|
||||||
|
case 15:
|
||||||
|
if wireType != 2 {
|
||||||
|
return fmt.Errorf("proto: wrong wireType = %d for field CreatedTimestamp", wireType)
|
||||||
|
}
|
||||||
|
var msglen int
|
||||||
|
for shift := uint(0); ; shift += 7 {
|
||||||
|
if shift >= 64 {
|
||||||
|
return ErrIntOverflowMetrics
|
||||||
|
}
|
||||||
|
if iNdEx >= l {
|
||||||
|
return io.ErrUnexpectedEOF
|
||||||
|
}
|
||||||
|
b := dAtA[iNdEx]
|
||||||
|
iNdEx++
|
||||||
|
msglen |= int(b&0x7F) << shift
|
||||||
|
if b < 0x80 {
|
||||||
|
break
|
||||||
|
}
|
||||||
|
}
|
||||||
|
if msglen < 0 {
|
||||||
|
return ErrInvalidLengthMetrics
|
||||||
|
}
|
||||||
|
postIndex := iNdEx + msglen
|
||||||
|
if postIndex < 0 {
|
||||||
|
return ErrInvalidLengthMetrics
|
||||||
|
}
|
||||||
|
if postIndex > l {
|
||||||
|
return io.ErrUnexpectedEOF
|
||||||
|
}
|
||||||
|
if m.CreatedTimestamp == nil {
|
||||||
|
m.CreatedTimestamp = &types.Timestamp{}
|
||||||
|
}
|
||||||
|
if err := m.CreatedTimestamp.Unmarshal(dAtA[iNdEx:postIndex]); err != nil {
|
||||||
|
return err
|
||||||
|
}
|
||||||
|
iNdEx = postIndex
|
||||||
default:
|
default:
|
||||||
iNdEx = preIndex
|
iNdEx = preIndex
|
||||||
skippy, err := skipMetrics(dAtA[iNdEx:])
|
skippy, err := skipMetrics(dAtA[iNdEx:])
|
||||||
|
|
|
@ -51,6 +51,8 @@ message Gauge {
|
||||||
message Counter {
|
message Counter {
|
||||||
double value = 1;
|
double value = 1;
|
||||||
Exemplar exemplar = 2;
|
Exemplar exemplar = 2;
|
||||||
|
|
||||||
|
google.protobuf.Timestamp created_timestamp = 3 [(gogoproto.nullable) = true];
|
||||||
}
|
}
|
||||||
|
|
||||||
message Quantile {
|
message Quantile {
|
||||||
|
@ -62,6 +64,8 @@ message Summary {
|
||||||
uint64 sample_count = 1;
|
uint64 sample_count = 1;
|
||||||
double sample_sum = 2;
|
double sample_sum = 2;
|
||||||
repeated Quantile quantile = 3 [(gogoproto.nullable) = false];
|
repeated Quantile quantile = 3 [(gogoproto.nullable) = false];
|
||||||
|
|
||||||
|
google.protobuf.Timestamp created_timestamp = 4 [(gogoproto.nullable) = true];
|
||||||
}
|
}
|
||||||
|
|
||||||
message Untyped {
|
message Untyped {
|
||||||
|
@ -75,6 +79,8 @@ message Histogram {
|
||||||
// Buckets for the conventional histogram.
|
// Buckets for the conventional histogram.
|
||||||
repeated Bucket bucket = 3 [(gogoproto.nullable) = false]; // Ordered in increasing order of upper_bound, +Inf bucket is optional.
|
repeated Bucket bucket = 3 [(gogoproto.nullable) = false]; // Ordered in increasing order of upper_bound, +Inf bucket is optional.
|
||||||
|
|
||||||
|
google.protobuf.Timestamp created_timestamp = 15 [(gogoproto.nullable) = true];
|
||||||
|
|
||||||
// Everything below here is for native histograms (also known as sparse histograms).
|
// Everything below here is for native histograms (also known as sparse histograms).
|
||||||
// Native histograms are an experimental feature without stability guarantees.
|
// Native histograms are an experimental feature without stability guarantees.
|
||||||
|
|
||||||
|
|
|
@ -1168,10 +1168,14 @@ func funcHistogramQuantile(vals []parser.Value, args parser.Expressions, enh *Ev
|
||||||
|
|
||||||
for _, mb := range enh.signatureToMetricWithBuckets {
|
for _, mb := range enh.signatureToMetricWithBuckets {
|
||||||
if len(mb.buckets) > 0 {
|
if len(mb.buckets) > 0 {
|
||||||
|
res, forcedMonotonicity := bucketQuantile(q, mb.buckets)
|
||||||
enh.Out = append(enh.Out, Sample{
|
enh.Out = append(enh.Out, Sample{
|
||||||
Metric: mb.metric,
|
Metric: mb.metric,
|
||||||
F: bucketQuantile(q, mb.buckets),
|
F: res,
|
||||||
})
|
})
|
||||||
|
if forcedMonotonicity {
|
||||||
|
annos.Add(annotations.NewHistogramQuantileForcedMonotonicityInfo(mb.metric.Get(labels.MetricName), args[1].PositionRange()))
|
||||||
|
}
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
|
|
|
@ -71,15 +71,17 @@ type metricWithBuckets struct {
|
||||||
// If q<0, -Inf is returned.
|
// If q<0, -Inf is returned.
|
||||||
//
|
//
|
||||||
// If q>1, +Inf is returned.
|
// If q>1, +Inf is returned.
|
||||||
func bucketQuantile(q float64, buckets buckets) float64 {
|
//
|
||||||
|
// We also return a bool to indicate if monotonicity needed to be forced.
|
||||||
|
func bucketQuantile(q float64, buckets buckets) (float64, bool) {
|
||||||
if math.IsNaN(q) {
|
if math.IsNaN(q) {
|
||||||
return math.NaN()
|
return math.NaN(), false
|
||||||
}
|
}
|
||||||
if q < 0 {
|
if q < 0 {
|
||||||
return math.Inf(-1)
|
return math.Inf(-1), false
|
||||||
}
|
}
|
||||||
if q > 1 {
|
if q > 1 {
|
||||||
return math.Inf(+1)
|
return math.Inf(+1), false
|
||||||
}
|
}
|
||||||
slices.SortFunc(buckets, func(a, b bucket) int {
|
slices.SortFunc(buckets, func(a, b bucket) int {
|
||||||
// We don't expect the bucket boundary to be a NaN.
|
// We don't expect the bucket boundary to be a NaN.
|
||||||
|
@ -92,27 +94,27 @@ func bucketQuantile(q float64, buckets buckets) float64 {
|
||||||
return 0
|
return 0
|
||||||
})
|
})
|
||||||
if !math.IsInf(buckets[len(buckets)-1].upperBound, +1) {
|
if !math.IsInf(buckets[len(buckets)-1].upperBound, +1) {
|
||||||
return math.NaN()
|
return math.NaN(), false
|
||||||
}
|
}
|
||||||
|
|
||||||
buckets = coalesceBuckets(buckets)
|
buckets = coalesceBuckets(buckets)
|
||||||
ensureMonotonic(buckets)
|
forcedMonotonic := ensureMonotonic(buckets)
|
||||||
|
|
||||||
if len(buckets) < 2 {
|
if len(buckets) < 2 {
|
||||||
return math.NaN()
|
return math.NaN(), false
|
||||||
}
|
}
|
||||||
observations := buckets[len(buckets)-1].count
|
observations := buckets[len(buckets)-1].count
|
||||||
if observations == 0 {
|
if observations == 0 {
|
||||||
return math.NaN()
|
return math.NaN(), false
|
||||||
}
|
}
|
||||||
rank := q * observations
|
rank := q * observations
|
||||||
b := sort.Search(len(buckets)-1, func(i int) bool { return buckets[i].count >= rank })
|
b := sort.Search(len(buckets)-1, func(i int) bool { return buckets[i].count >= rank })
|
||||||
|
|
||||||
if b == len(buckets)-1 {
|
if b == len(buckets)-1 {
|
||||||
return buckets[len(buckets)-2].upperBound
|
return buckets[len(buckets)-2].upperBound, forcedMonotonic
|
||||||
}
|
}
|
||||||
if b == 0 && buckets[0].upperBound <= 0 {
|
if b == 0 && buckets[0].upperBound <= 0 {
|
||||||
return buckets[0].upperBound
|
return buckets[0].upperBound, forcedMonotonic
|
||||||
}
|
}
|
||||||
var (
|
var (
|
||||||
bucketStart float64
|
bucketStart float64
|
||||||
|
@ -124,7 +126,7 @@ func bucketQuantile(q float64, buckets buckets) float64 {
|
||||||
count -= buckets[b-1].count
|
count -= buckets[b-1].count
|
||||||
rank -= buckets[b-1].count
|
rank -= buckets[b-1].count
|
||||||
}
|
}
|
||||||
return bucketStart + (bucketEnd-bucketStart)*(rank/count)
|
return bucketStart + (bucketEnd-bucketStart)*(rank/count), forcedMonotonic
|
||||||
}
|
}
|
||||||
|
|
||||||
// histogramQuantile calculates the quantile 'q' based on the given histogram.
|
// histogramQuantile calculates the quantile 'q' based on the given histogram.
|
||||||
|
@ -342,37 +344,24 @@ func coalesceBuckets(buckets buckets) buckets {
|
||||||
// The assumption that bucket counts increase monotonically with increasing
|
// The assumption that bucket counts increase monotonically with increasing
|
||||||
// upperBound may be violated during:
|
// upperBound may be violated during:
|
||||||
//
|
//
|
||||||
// * Recording rule evaluation of histogram_quantile, especially when rate()
|
// - Circumstances where data is already inconsistent at the target's side.
|
||||||
// has been applied to the underlying bucket timeseries.
|
// - Ingestion via the remote write receiver that Prometheus implements.
|
||||||
// * Evaluation of histogram_quantile computed over federated bucket
|
// - Optimisation of query execution where precision is sacrificed for other
|
||||||
// timeseries, especially when rate() has been applied.
|
// benefits, not by Prometheus but by systems built on top of it.
|
||||||
//
|
|
||||||
// This is because scraped data is not made available to rule evaluation or
|
|
||||||
// federation atomically, so some buckets are computed with data from the
|
|
||||||
// most recent scrapes, but the other buckets are missing data from the most
|
|
||||||
// recent scrape.
|
|
||||||
//
|
//
|
||||||
// Monotonicity is usually guaranteed because if a bucket with upper bound
|
// Monotonicity is usually guaranteed because if a bucket with upper bound
|
||||||
// u1 has count c1, then any bucket with a higher upper bound u > u1 must
|
// u1 has count c1, then any bucket with a higher upper bound u > u1 must
|
||||||
// have counted all c1 observations and perhaps more, so that c >= c1.
|
// have counted all c1 observations and perhaps more, so that c >= c1.
|
||||||
//
|
//
|
||||||
// Randomly interspersed partial sampling breaks that guarantee, and rate()
|
|
||||||
// exacerbates it. Specifically, suppose bucket le=1000 has a count of 10 from
|
|
||||||
// 4 samples but the bucket with le=2000 has a count of 7 from 3 samples. The
|
|
||||||
// monotonicity is broken. It is exacerbated by rate() because under normal
|
|
||||||
// operation, cumulative counting of buckets will cause the bucket counts to
|
|
||||||
// diverge such that small differences from missing samples are not a problem.
|
|
||||||
// rate() removes this divergence.)
|
|
||||||
//
|
|
||||||
// bucketQuantile depends on that monotonicity to do a binary search for the
|
// bucketQuantile depends on that monotonicity to do a binary search for the
|
||||||
// bucket with the φ-quantile count, so breaking the monotonicity
|
// bucket with the φ-quantile count, so breaking the monotonicity
|
||||||
// guarantee causes bucketQuantile() to return undefined (nonsense) results.
|
// guarantee causes bucketQuantile() to return undefined (nonsense) results.
|
||||||
//
|
//
|
||||||
// As a somewhat hacky solution until ingestion is atomic per scrape, we
|
// As a somewhat hacky solution, we calculate the "envelope" of the histogram
|
||||||
// calculate the "envelope" of the histogram buckets, essentially removing
|
// buckets, essentially removing any decreases in the count between successive
|
||||||
// any decreases in the count between successive buckets.
|
// buckets. We return a bool to indicate if this monotonicity was forced or not.
|
||||||
|
func ensureMonotonic(buckets buckets) bool {
|
||||||
func ensureMonotonic(buckets buckets) {
|
forced := false
|
||||||
max := buckets[0].count
|
max := buckets[0].count
|
||||||
for i := 1; i < len(buckets); i++ {
|
for i := 1; i < len(buckets); i++ {
|
||||||
switch {
|
switch {
|
||||||
|
@ -380,8 +369,10 @@ func ensureMonotonic(buckets buckets) {
|
||||||
max = buckets[i].count
|
max = buckets[i].count
|
||||||
case buckets[i].count < max:
|
case buckets[i].count < max:
|
||||||
buckets[i].count = max
|
buckets[i].count = max
|
||||||
|
forced = true
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
return forced
|
||||||
}
|
}
|
||||||
|
|
||||||
// quantile calculates the given quantile of a vector of samples.
|
// quantile calculates the given quantile of a vector of samples.
|
||||||
|
|
|
@ -785,7 +785,8 @@ func appender(app storage.Appender, sampleLimit, bucketLimit int) storage.Append
|
||||||
|
|
||||||
// A scraper retrieves samples and accepts a status report at the end.
|
// A scraper retrieves samples and accepts a status report at the end.
|
||||||
type scraper interface {
|
type scraper interface {
|
||||||
scrape(ctx context.Context, w io.Writer) (string, error)
|
scrape(ctx context.Context) (*http.Response, error)
|
||||||
|
readResponse(ctx context.Context, resp *http.Response, w io.Writer) (string, error)
|
||||||
Report(start time.Time, dur time.Duration, err error)
|
Report(start time.Time, dur time.Duration, err error)
|
||||||
offset(interval time.Duration, offsetSeed uint64) time.Duration
|
offset(interval time.Duration, offsetSeed uint64) time.Duration
|
||||||
}
|
}
|
||||||
|
@ -814,11 +815,11 @@ const (
|
||||||
|
|
||||||
var UserAgent = fmt.Sprintf("Prometheus/%s", version.Version)
|
var UserAgent = fmt.Sprintf("Prometheus/%s", version.Version)
|
||||||
|
|
||||||
func (s *targetScraper) scrape(ctx context.Context, w io.Writer) (string, error) {
|
func (s *targetScraper) scrape(ctx context.Context) (*http.Response, error) {
|
||||||
if s.req == nil {
|
if s.req == nil {
|
||||||
req, err := http.NewRequest("GET", s.URL().String(), nil)
|
req, err := http.NewRequest("GET", s.URL().String(), nil)
|
||||||
if err != nil {
|
if err != nil {
|
||||||
return "", err
|
return nil, err
|
||||||
}
|
}
|
||||||
req.Header.Add("Accept", s.acceptHeader)
|
req.Header.Add("Accept", s.acceptHeader)
|
||||||
req.Header.Add("Accept-Encoding", "gzip")
|
req.Header.Add("Accept-Encoding", "gzip")
|
||||||
|
@ -828,10 +829,10 @@ func (s *targetScraper) scrape(ctx context.Context, w io.Writer) (string, error)
|
||||||
s.req = req
|
s.req = req
|
||||||
}
|
}
|
||||||
|
|
||||||
resp, err := s.client.Do(s.req.WithContext(ctx))
|
return s.client.Do(s.req.WithContext(ctx))
|
||||||
if err != nil {
|
}
|
||||||
return "", err
|
|
||||||
}
|
func (s *targetScraper) readResponse(ctx context.Context, resp *http.Response, w io.Writer) (string, error) {
|
||||||
defer func() {
|
defer func() {
|
||||||
io.Copy(io.Discard, resp.Body)
|
io.Copy(io.Discard, resp.Body)
|
||||||
resp.Body.Close()
|
resp.Body.Close()
|
||||||
|
@ -858,13 +859,14 @@ func (s *targetScraper) scrape(ctx context.Context, w io.Writer) (string, error)
|
||||||
|
|
||||||
if s.gzipr == nil {
|
if s.gzipr == nil {
|
||||||
s.buf = bufio.NewReader(resp.Body)
|
s.buf = bufio.NewReader(resp.Body)
|
||||||
|
var err error
|
||||||
s.gzipr, err = gzip.NewReader(s.buf)
|
s.gzipr, err = gzip.NewReader(s.buf)
|
||||||
if err != nil {
|
if err != nil {
|
||||||
return "", err
|
return "", err
|
||||||
}
|
}
|
||||||
} else {
|
} else {
|
||||||
s.buf.Reset(resp.Body)
|
s.buf.Reset(resp.Body)
|
||||||
if err = s.gzipr.Reset(s.buf); err != nil {
|
if err := s.gzipr.Reset(s.buf); err != nil {
|
||||||
return "", err
|
return "", err
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
@ -1326,11 +1328,7 @@ func (sl *scrapeLoop) scrapeAndReport(last, appendTime time.Time, errc chan<- er
|
||||||
)
|
)
|
||||||
}
|
}
|
||||||
|
|
||||||
b := sl.buffers.Get(sl.lastScrapeSize).([]byte)
|
var total, added, seriesAdded, bytesRead int
|
||||||
defer sl.buffers.Put(b)
|
|
||||||
buf := bytes.NewBuffer(b)
|
|
||||||
|
|
||||||
var total, added, seriesAdded, bytes int
|
|
||||||
var err, appErr, scrapeErr error
|
var err, appErr, scrapeErr error
|
||||||
|
|
||||||
app := sl.appender(sl.appenderCtx)
|
app := sl.appender(sl.appenderCtx)
|
||||||
|
@ -1346,7 +1344,7 @@ func (sl *scrapeLoop) scrapeAndReport(last, appendTime time.Time, errc chan<- er
|
||||||
}()
|
}()
|
||||||
|
|
||||||
defer func() {
|
defer func() {
|
||||||
if err = sl.report(app, appendTime, time.Since(start), total, added, seriesAdded, bytes, scrapeErr); err != nil {
|
if err = sl.report(app, appendTime, time.Since(start), total, added, seriesAdded, bytesRead, scrapeErr); err != nil {
|
||||||
level.Warn(sl.l).Log("msg", "Appending scrape report failed", "err", err)
|
level.Warn(sl.l).Log("msg", "Appending scrape report failed", "err", err)
|
||||||
}
|
}
|
||||||
}()
|
}()
|
||||||
|
@ -1367,8 +1365,17 @@ func (sl *scrapeLoop) scrapeAndReport(last, appendTime time.Time, errc chan<- er
|
||||||
}
|
}
|
||||||
|
|
||||||
var contentType string
|
var contentType string
|
||||||
|
var resp *http.Response
|
||||||
|
var b []byte
|
||||||
|
var buf *bytes.Buffer
|
||||||
scrapeCtx, cancel := context.WithTimeout(sl.parentCtx, sl.timeout)
|
scrapeCtx, cancel := context.WithTimeout(sl.parentCtx, sl.timeout)
|
||||||
contentType, scrapeErr = sl.scraper.scrape(scrapeCtx, buf)
|
resp, scrapeErr = sl.scraper.scrape(scrapeCtx)
|
||||||
|
if scrapeErr == nil {
|
||||||
|
b = sl.buffers.Get(sl.lastScrapeSize).([]byte)
|
||||||
|
defer sl.buffers.Put(b)
|
||||||
|
buf = bytes.NewBuffer(b)
|
||||||
|
contentType, scrapeErr = sl.scraper.readResponse(scrapeCtx, resp, buf)
|
||||||
|
}
|
||||||
cancel()
|
cancel()
|
||||||
|
|
||||||
if scrapeErr == nil {
|
if scrapeErr == nil {
|
||||||
|
@ -1379,14 +1386,14 @@ func (sl *scrapeLoop) scrapeAndReport(last, appendTime time.Time, errc chan<- er
|
||||||
if len(b) > 0 {
|
if len(b) > 0 {
|
||||||
sl.lastScrapeSize = len(b)
|
sl.lastScrapeSize = len(b)
|
||||||
}
|
}
|
||||||
bytes = len(b)
|
bytesRead = len(b)
|
||||||
} else {
|
} else {
|
||||||
level.Debug(sl.l).Log("msg", "Scrape failed", "err", scrapeErr)
|
level.Debug(sl.l).Log("msg", "Scrape failed", "err", scrapeErr)
|
||||||
if errc != nil {
|
if errc != nil {
|
||||||
errc <- scrapeErr
|
errc <- scrapeErr
|
||||||
}
|
}
|
||||||
if errors.Is(scrapeErr, errBodySizeLimit) {
|
if errors.Is(scrapeErr, errBodySizeLimit) {
|
||||||
bytes = -1
|
bytesRead = -1
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
|
|
|
@ -2619,7 +2619,9 @@ func TestTargetScraperScrapeOK(t *testing.T) {
|
||||||
}
|
}
|
||||||
var buf bytes.Buffer
|
var buf bytes.Buffer
|
||||||
|
|
||||||
contentType, err := ts.scrape(context.Background(), &buf)
|
resp, err := ts.scrape(context.Background())
|
||||||
|
require.NoError(t, err)
|
||||||
|
contentType, err := ts.readResponse(context.Background(), resp, &buf)
|
||||||
require.NoError(t, err)
|
require.NoError(t, err)
|
||||||
require.Equal(t, "text/plain; version=0.0.4", contentType)
|
require.Equal(t, "text/plain; version=0.0.4", contentType)
|
||||||
require.Equal(t, "metric_a 1\nmetric_b 2\n", buf.String())
|
require.Equal(t, "metric_a 1\nmetric_b 2\n", buf.String())
|
||||||
|
@ -2665,7 +2667,7 @@ func TestTargetScrapeScrapeCancel(t *testing.T) {
|
||||||
}()
|
}()
|
||||||
|
|
||||||
go func() {
|
go func() {
|
||||||
_, err := ts.scrape(ctx, io.Discard)
|
_, err := ts.scrape(ctx)
|
||||||
switch {
|
switch {
|
||||||
case err == nil:
|
case err == nil:
|
||||||
errc <- errors.New("Expected error but got nil")
|
errc <- errors.New("Expected error but got nil")
|
||||||
|
@ -2711,7 +2713,9 @@ func TestTargetScrapeScrapeNotFound(t *testing.T) {
|
||||||
acceptHeader: scrapeAcceptHeader,
|
acceptHeader: scrapeAcceptHeader,
|
||||||
}
|
}
|
||||||
|
|
||||||
_, err = ts.scrape(context.Background(), io.Discard)
|
resp, err := ts.scrape(context.Background())
|
||||||
|
require.NoError(t, err)
|
||||||
|
_, err = ts.readResponse(context.Background(), resp, io.Discard)
|
||||||
require.Contains(t, err.Error(), "404", "Expected \"404 NotFound\" error but got: %s", err)
|
require.Contains(t, err.Error(), "404", "Expected \"404 NotFound\" error but got: %s", err)
|
||||||
}
|
}
|
||||||
|
|
||||||
|
@ -2755,26 +2759,34 @@ func TestTargetScraperBodySizeLimit(t *testing.T) {
|
||||||
var buf bytes.Buffer
|
var buf bytes.Buffer
|
||||||
|
|
||||||
// Target response uncompressed body, scrape with body size limit.
|
// Target response uncompressed body, scrape with body size limit.
|
||||||
_, err = ts.scrape(context.Background(), &buf)
|
resp, err := ts.scrape(context.Background())
|
||||||
|
require.NoError(t, err)
|
||||||
|
_, err = ts.readResponse(context.Background(), resp, &buf)
|
||||||
require.ErrorIs(t, err, errBodySizeLimit)
|
require.ErrorIs(t, err, errBodySizeLimit)
|
||||||
require.Equal(t, bodySizeLimit, buf.Len())
|
require.Equal(t, bodySizeLimit, buf.Len())
|
||||||
// Target response gzip compressed body, scrape with body size limit.
|
// Target response gzip compressed body, scrape with body size limit.
|
||||||
gzipResponse = true
|
gzipResponse = true
|
||||||
buf.Reset()
|
buf.Reset()
|
||||||
_, err = ts.scrape(context.Background(), &buf)
|
resp, err = ts.scrape(context.Background())
|
||||||
|
require.NoError(t, err)
|
||||||
|
_, err = ts.readResponse(context.Background(), resp, &buf)
|
||||||
require.ErrorIs(t, err, errBodySizeLimit)
|
require.ErrorIs(t, err, errBodySizeLimit)
|
||||||
require.Equal(t, bodySizeLimit, buf.Len())
|
require.Equal(t, bodySizeLimit, buf.Len())
|
||||||
// Target response uncompressed body, scrape without body size limit.
|
// Target response uncompressed body, scrape without body size limit.
|
||||||
gzipResponse = false
|
gzipResponse = false
|
||||||
buf.Reset()
|
buf.Reset()
|
||||||
ts.bodySizeLimit = 0
|
ts.bodySizeLimit = 0
|
||||||
_, err = ts.scrape(context.Background(), &buf)
|
resp, err = ts.scrape(context.Background())
|
||||||
|
require.NoError(t, err)
|
||||||
|
_, err = ts.readResponse(context.Background(), resp, &buf)
|
||||||
require.NoError(t, err)
|
require.NoError(t, err)
|
||||||
require.Equal(t, len(responseBody), buf.Len())
|
require.Equal(t, len(responseBody), buf.Len())
|
||||||
// Target response gzip compressed body, scrape without body size limit.
|
// Target response gzip compressed body, scrape without body size limit.
|
||||||
gzipResponse = true
|
gzipResponse = true
|
||||||
buf.Reset()
|
buf.Reset()
|
||||||
_, err = ts.scrape(context.Background(), &buf)
|
resp, err = ts.scrape(context.Background())
|
||||||
|
require.NoError(t, err)
|
||||||
|
_, err = ts.readResponse(context.Background(), resp, &buf)
|
||||||
require.NoError(t, err)
|
require.NoError(t, err)
|
||||||
require.Equal(t, len(responseBody), buf.Len())
|
require.Equal(t, len(responseBody), buf.Len())
|
||||||
}
|
}
|
||||||
|
@ -2802,7 +2814,11 @@ func (ts *testScraper) Report(start time.Time, duration time.Duration, err error
|
||||||
ts.lastError = err
|
ts.lastError = err
|
||||||
}
|
}
|
||||||
|
|
||||||
func (ts *testScraper) scrape(ctx context.Context, w io.Writer) (string, error) {
|
func (ts *testScraper) scrape(ctx context.Context) (*http.Response, error) {
|
||||||
|
return nil, ts.scrapeErr
|
||||||
|
}
|
||||||
|
|
||||||
|
func (ts *testScraper) readResponse(ctx context.Context, resp *http.Response, w io.Writer) (string, error) {
|
||||||
if ts.scrapeFunc != nil {
|
if ts.scrapeFunc != nil {
|
||||||
return "", ts.scrapeFunc(ctx, w)
|
return "", ts.scrapeFunc(ctx, w)
|
||||||
}
|
}
|
||||||
|
|
|
@ -413,6 +413,7 @@ type Postings interface {
|
||||||
Seek(v storage.SeriesRef) bool
|
Seek(v storage.SeriesRef) bool
|
||||||
|
|
||||||
// At returns the value at the current iterator position.
|
// At returns the value at the current iterator position.
|
||||||
|
// At should only be called after a successful call to Next or Seek.
|
||||||
At() storage.SeriesRef
|
At() storage.SeriesRef
|
||||||
|
|
||||||
// Err returns the last error of the iterator.
|
// Err returns the last error of the iterator.
|
||||||
|
@ -747,7 +748,7 @@ func (rp *removedPostings) Err() error {
|
||||||
// ListPostings implements the Postings interface over a plain list.
|
// ListPostings implements the Postings interface over a plain list.
|
||||||
type ListPostings struct {
|
type ListPostings struct {
|
||||||
list []storage.SeriesRef
|
list []storage.SeriesRef
|
||||||
pos int
|
cur storage.SeriesRef
|
||||||
}
|
}
|
||||||
|
|
||||||
func NewListPostings(list []storage.SeriesRef) Postings {
|
func NewListPostings(list []storage.SeriesRef) Postings {
|
||||||
|
@ -759,34 +760,39 @@ func newListPostings(list ...storage.SeriesRef) *ListPostings {
|
||||||
}
|
}
|
||||||
|
|
||||||
func (it *ListPostings) At() storage.SeriesRef {
|
func (it *ListPostings) At() storage.SeriesRef {
|
||||||
return it.list[it.pos-1]
|
return it.cur
|
||||||
}
|
}
|
||||||
|
|
||||||
func (it *ListPostings) Next() bool {
|
func (it *ListPostings) Next() bool {
|
||||||
if it.pos < len(it.list) {
|
if len(it.list) > 0 {
|
||||||
it.pos++
|
it.cur = it.list[0]
|
||||||
|
it.list = it.list[1:]
|
||||||
return true
|
return true
|
||||||
}
|
}
|
||||||
|
it.cur = 0
|
||||||
return false
|
return false
|
||||||
}
|
}
|
||||||
|
|
||||||
func (it *ListPostings) Seek(x storage.SeriesRef) bool {
|
func (it *ListPostings) Seek(x storage.SeriesRef) bool {
|
||||||
if it.pos == 0 {
|
|
||||||
it.pos++
|
|
||||||
}
|
|
||||||
if it.pos > len(it.list) {
|
|
||||||
return false
|
|
||||||
}
|
|
||||||
// If the current value satisfies, then return.
|
// If the current value satisfies, then return.
|
||||||
if it.list[it.pos-1] >= x {
|
if it.cur >= x {
|
||||||
return true
|
return true
|
||||||
}
|
}
|
||||||
|
if len(it.list) == 0 {
|
||||||
|
return false
|
||||||
|
}
|
||||||
|
|
||||||
// Do binary search between current position and end.
|
// Do binary search between current position and end.
|
||||||
it.pos = sort.Search(len(it.list[it.pos-1:]), func(i int) bool {
|
i := sort.Search(len(it.list), func(i int) bool {
|
||||||
return it.list[i+it.pos-1] >= x
|
return it.list[i] >= x
|
||||||
}) + it.pos
|
})
|
||||||
return it.pos-1 < len(it.list)
|
if i < len(it.list) {
|
||||||
|
it.cur = it.list[i]
|
||||||
|
it.list = it.list[i+1:]
|
||||||
|
return true
|
||||||
|
}
|
||||||
|
it.list = nil
|
||||||
|
return false
|
||||||
}
|
}
|
||||||
|
|
||||||
func (it *ListPostings) Err() error {
|
func (it *ListPostings) Err() error {
|
||||||
|
|
|
@ -1353,6 +1353,8 @@ func TestListPostings(t *testing.T) {
|
||||||
})
|
})
|
||||||
}
|
}
|
||||||
|
|
||||||
|
// BenchmarkListPostings benchmarks ListPostings by iterating Next/At sequentially.
|
||||||
|
// See also BenchmarkIntersect as it performs more `At` calls than `Next` calls when intersecting.
|
||||||
func BenchmarkListPostings(b *testing.B) {
|
func BenchmarkListPostings(b *testing.B) {
|
||||||
const maxCount = 1e6
|
const maxCount = 1e6
|
||||||
input := make([]storage.SeriesRef, maxCount)
|
input := make([]storage.SeriesRef, maxCount)
|
||||||
|
|
|
@ -106,6 +106,7 @@ var (
|
||||||
MixedClassicNativeHistogramsWarning = fmt.Errorf("%w: vector contains a mix of classic and native histograms for metric name", PromQLWarning)
|
MixedClassicNativeHistogramsWarning = fmt.Errorf("%w: vector contains a mix of classic and native histograms for metric name", PromQLWarning)
|
||||||
|
|
||||||
PossibleNonCounterInfo = fmt.Errorf("%w: metric might not be a counter, name does not end in _total/_sum/_count:", PromQLInfo)
|
PossibleNonCounterInfo = fmt.Errorf("%w: metric might not be a counter, name does not end in _total/_sum/_count:", PromQLInfo)
|
||||||
|
HistogramQuantileForcedMonotonicityInfo = fmt.Errorf("%w: input to histogram_quantile needed to be fixed for monotonicity (and may give inaccurate results) for metric name", PromQLInfo)
|
||||||
)
|
)
|
||||||
|
|
||||||
type annoErr struct {
|
type annoErr struct {
|
||||||
|
@ -163,3 +164,12 @@ func NewPossibleNonCounterInfo(metricName string, pos posrange.PositionRange) an
|
||||||
Err: fmt.Errorf("%w %q", PossibleNonCounterInfo, metricName),
|
Err: fmt.Errorf("%w %q", PossibleNonCounterInfo, metricName),
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
|
// NewHistogramQuantileForcedMonotonicityInfo is used when the input (classic histograms) to
|
||||||
|
// histogram_quantile needs to be forced to be monotonic.
|
||||||
|
func NewHistogramQuantileForcedMonotonicityInfo(metricName string, pos posrange.PositionRange) annoErr {
|
||||||
|
return annoErr{
|
||||||
|
PositionRange: pos,
|
||||||
|
Err: fmt.Errorf("%w %q", HistogramQuantileForcedMonotonicityInfo, metricName),
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
|
@ -13,6 +13,7 @@ interface RulesContentProps {
|
||||||
interface RuleGroup {
|
interface RuleGroup {
|
||||||
name: string;
|
name: string;
|
||||||
file: string;
|
file: string;
|
||||||
|
interval: string;
|
||||||
rules: Rule[];
|
rules: Rule[];
|
||||||
evaluationTime: string;
|
evaluationTime: string;
|
||||||
lastEvaluation: string;
|
lastEvaluation: string;
|
||||||
|
@ -58,13 +59,16 @@ export const RulesContent: FC<RulesContentProps> = ({ response }) => {
|
||||||
<Table bordered key={i}>
|
<Table bordered key={i}>
|
||||||
<thead>
|
<thead>
|
||||||
<tr>
|
<tr>
|
||||||
<td colSpan={3}>
|
<td>
|
||||||
<a href={'#' + g.name}>
|
<a href={'#' + g.name}>
|
||||||
<h4 id={g.name} className="text-break">
|
<h4 id={g.name} className="text-break">
|
||||||
{g.name}
|
{g.name}
|
||||||
</h4>
|
</h4>
|
||||||
</a>
|
</a>
|
||||||
</td>
|
</td>
|
||||||
|
<td colSpan={2}>
|
||||||
|
<h4>Interval: {humanizeDuration(parseFloat(g.interval) * 1000)}</h4>
|
||||||
|
</td>
|
||||||
<td>
|
<td>
|
||||||
<h4>{formatRelative(g.lastEvaluation, now())}</h4>
|
<h4>{formatRelative(g.lastEvaluation, now())}</h4>
|
||||||
</td>
|
</td>
|
||||||
|
|
Loading…
Reference in a new issue