Async marker handler

This commit is contained in:
Paulin Todev 2023-06-28 13:57:21 +01:00
parent b9ae36c3c3
commit 1ea38884bf
3 changed files with 81 additions and 74 deletions

View file

@ -47,6 +47,7 @@ func NewMarkerFileHandler(logger log.Logger, walDir, markerId string) (MarkerFil
lastMarkedSegmentFilePath: filepath.Join(markerDir, "segment"),
}
//TODO: Should this be in a separate Start() function?
go mfh.markSegmentAsync()
return mfh, nil

View file

@ -1,24 +1,29 @@
package remote
import (
"sync"
"github.com/prometheus/prometheus/tsdb/wlog"
)
type MarkerHandler interface {
wlog.Marker
UpdatePendingData(dataCount, dataSegment int)
ProcessConsumedData(data map[int]int)
UpdateReceivedData(segmentId, dataCount int) // Data queued for sending
UpdateSentData(segmentId, dataCount int) // Data which was sent or given up on sending
Stop()
}
type markerHandler struct {
markerFileHandler MarkerFileHandler
pendingDataMut sync.Mutex
latestDataSegment int
lastMarkedSegment int
pendingDataSegments map[int]int // Map of segment index to pending count
dataIOUpdate chan data
quit chan struct{}
}
// TODO: Rename this struct
type data struct {
segmentId int
dataCount int
}
var (
@ -27,10 +32,11 @@ var (
func NewMarkerHandler(mfh MarkerFileHandler) MarkerHandler {
mh := &markerHandler{
latestDataSegment: -1,
lastMarkedSegment: -1,
pendingDataSegments: make(map[int]int),
lastMarkedSegment: -1, // Segment ID last marked on disk.
markerFileHandler: mfh,
//TODO: What is a good size for the channel?
dataIOUpdate: make(chan data, 100),
quit: make(chan struct{}),
}
// Load the last marked segment from disk (if it exists).
@ -38,6 +44,9 @@ func NewMarkerHandler(mfh MarkerFileHandler) MarkerHandler {
mh.lastMarkedSegment = lastSegment
}
//TODO: Should this be in a separate Start() function?
go mh.updatePendingData()
return mh
}
@ -45,45 +54,57 @@ func (mh *markerHandler) LastMarkedSegment() int {
return mh.markerFileHandler.LastMarkedSegment()
}
// updatePendingData updates a counter for how much data is yet to be sent from each segment.
// "dataCount" will be added to the segment with ID "dataSegment".
func (mh *markerHandler) UpdatePendingData(dataCount, dataSegment int) {
mh.pendingDataMut.Lock()
defer mh.pendingDataMut.Unlock()
mh.pendingDataSegments[dataSegment] += dataCount
// We want to save segments whose data has been fully processed by the
// QueueManager. To avoid too much overhead when appending data, we'll only
// examine pending data per segment whenever a new segment is detected.
//
// This could cause our saved segment to lag behind multiple segments
// depending on the rate that new segments are created and how long
// data in other segments takes to be processed.
if dataSegment <= mh.latestDataSegment {
return
func (mh *markerHandler) UpdateReceivedData(segmentId, dataCount int) {
mh.dataIOUpdate <- data{
segmentId: segmentId,
dataCount: dataCount,
}
}
// We've received data for a new segment. We'll use this as a signal to see
// if older segments have been fully processed.
//
// We want to mark the highest segment which has no more pending data and is
// newer than our last mark.
mh.latestDataSegment = dataSegment
func (mh *markerHandler) UpdateSentData(segmentId, dataCount int) {
mh.dataIOUpdate <- data{
segmentId: segmentId,
dataCount: -1 * dataCount,
}
}
var markableSegment int
for segment, count := range mh.pendingDataSegments {
if segment >= dataSegment || count > 0 {
func (mh *markerHandler) Stop() {
// Firstly stop the Marker Handler, because it might want to use the Marker File Handler.
mh.quit <- struct{}{}
// Finally, stop the File Handler.
mh.markerFileHandler.Stop()
}
// updatePendingData updates a counter for how much data is yet to be sent from each segment.
// "dataCount" will be added to the segment with ID "dataSegment".
func (mh *markerHandler) updatePendingData() {
batchSegmentCount := make(map[int]int)
for {
select {
case <-mh.quit:
return
case dataUpdate := <-mh.dataIOUpdate:
batchSegmentCount[dataUpdate.segmentId] += dataUpdate.dataCount
}
markableSegment := -1
for segment, count := range batchSegmentCount {
//TODO: If count is less than 0, then log an error and remove the entry from the map?
if count != 0 {
continue
}
//TODO: Is it save to assume that just because a segment is 0 inside the map,
// all samples from it have been processed?
if segment > markableSegment {
markableSegment = segment
}
// Clean up the pending map: the current segment has been completely
// consumed and doesn't need to be considered for marking again.
delete(mh.pendingDataSegments, segment)
delete(batchSegmentCount, segment)
}
if markableSegment > mh.lastMarkedSegment {
@ -91,22 +112,4 @@ func (mh *markerHandler) UpdatePendingData(dataCount, dataSegment int) {
mh.lastMarkedSegment = markableSegment
}
}
// processConsumedData updates a counter for how many samples have been sent for each segment.
// "data" is a map of segment index to consumed data count (e.g. number of samples).
func (mh *markerHandler) ProcessConsumedData(data map[int]int) {
mh.pendingDataMut.Lock()
defer mh.pendingDataMut.Unlock()
for segment, count := range data {
if _, tracked := mh.pendingDataSegments[segment]; !tracked {
//TODO: How is it possible to not track it? This should log an error?
continue
}
mh.pendingDataSegments[segment] -= count
}
}
func (mh *markerHandler) Stop() {
mh.markerFileHandler.Stop()
}

View file

@ -616,7 +616,7 @@ outer:
sType: tSample,
segment: segment,
}) {
t.markerHandler.UpdatePendingData(len(samples), segment)
t.markerHandler.UpdateReceivedData(segment, len(samples))
continue outer
}
@ -670,7 +670,7 @@ outer:
sType: tExemplar,
segment: segment,
}) {
t.markerHandler.UpdatePendingData(len(exemplars), segment)
t.markerHandler.UpdateReceivedData(segment, len(exemplars))
continue outer
}
@ -719,7 +719,7 @@ outer:
sType: tHistogram,
segment: segment,
}) {
t.markerHandler.UpdatePendingData(len(histograms), segment)
t.markerHandler.UpdateReceivedData(segment, len(histograms))
continue outer
}
@ -768,7 +768,7 @@ outer:
sType: tFloatHistogram,
segment: segment,
}) {
t.markerHandler.UpdatePendingData(len(floatHistograms), segment)
t.markerHandler.UpdateReceivedData(segment, len(floatHistograms))
continue outer
}
@ -1500,7 +1500,10 @@ func (s *shards) sendSamples(ctx context.Context, samples []prompb.TimeSeries, b
// Inform our queue manager about the data that got processed and clear out
// our map to prepare for the next batch.
s.qm.markerHandler.ProcessConsumedData(batchSegmentCount)
// Even if the sending failed, we still have to move on with the WAL marker
for segment, count := range batchSegmentCount {
s.qm.markerHandler.UpdateSentData(segment, count)
}
for segment := range batchSegmentCount {
delete(batchSegmentCount, segment)
}