163 lines
3.5 KiB
Go
163 lines
3.5 KiB
Go
package drainer
|
|
|
|
import (
|
|
"context"
|
|
"sync"
|
|
"time"
|
|
)
|
|
|
|
// DrainDeadlineNotifier allows batch notification of nodes that have reached
|
|
// their drain deadline.
|
|
type DrainDeadlineNotifier interface {
|
|
// NextBatch returns the next batch of nodes that have reached their
|
|
// deadline.
|
|
NextBatch() <-chan []string
|
|
|
|
// Remove removes the given node from being tracked for a deadline.
|
|
Remove(nodeID string)
|
|
|
|
// Watch marks the given node for being watched for its deadline.
|
|
Watch(nodeID string, deadline time.Time)
|
|
}
|
|
|
|
// deadlineHeap implements the DrainDeadlineNotifier and is backed by a min-heap
|
|
// to efficiently determine the next deadlining node. It also supports
|
|
// coalescing several deadlines into a single emission.
|
|
type deadlineHeap struct {
|
|
ctx context.Context
|
|
coalesceWindow time.Duration
|
|
batch chan []string
|
|
nodes map[string]time.Time
|
|
trigger chan struct{}
|
|
mu sync.Mutex
|
|
}
|
|
|
|
// NewDeadlineHeap returns a new deadline heap that coalesces for the given
|
|
// duration and will stop watching when the passed context is cancelled.
|
|
func NewDeadlineHeap(ctx context.Context, coalesceWindow time.Duration) *deadlineHeap {
|
|
d := &deadlineHeap{
|
|
ctx: ctx,
|
|
coalesceWindow: coalesceWindow,
|
|
batch: make(chan []string),
|
|
nodes: make(map[string]time.Time, 64),
|
|
trigger: make(chan struct{}, 1),
|
|
}
|
|
|
|
go d.watch()
|
|
return d
|
|
}
|
|
|
|
func (d *deadlineHeap) watch() {
|
|
timer := time.NewTimer(0)
|
|
timer.Stop()
|
|
select {
|
|
case <-timer.C:
|
|
default:
|
|
}
|
|
|
|
var nextDeadline time.Time
|
|
defer timer.Stop()
|
|
|
|
for {
|
|
select {
|
|
case <-d.ctx.Done():
|
|
return
|
|
case <-timer.C:
|
|
if nextDeadline.IsZero() {
|
|
continue
|
|
}
|
|
|
|
var batch []string
|
|
|
|
d.mu.Lock()
|
|
for nodeID, nodeDeadline := range d.nodes {
|
|
if !nodeDeadline.After(nextDeadline) {
|
|
batch = append(batch, nodeID)
|
|
delete(d.nodes, nodeID)
|
|
}
|
|
}
|
|
d.mu.Unlock()
|
|
|
|
if len(batch) > 0 {
|
|
// Send the batch
|
|
select {
|
|
case d.batch <- batch:
|
|
case <-d.ctx.Done():
|
|
return
|
|
}
|
|
}
|
|
|
|
case <-d.trigger:
|
|
}
|
|
|
|
// Calculate the next deadline
|
|
deadline, ok := d.calculateNextDeadline()
|
|
if !ok {
|
|
continue
|
|
}
|
|
|
|
if !deadline.Equal(nextDeadline) {
|
|
timer.Reset(deadline.Sub(time.Now()))
|
|
nextDeadline = deadline
|
|
}
|
|
}
|
|
}
|
|
|
|
// calculateNextDeadline returns the next deadline in which to scan for
|
|
// deadlined nodes. It applies the coalesce window.
|
|
func (d *deadlineHeap) calculateNextDeadline() (time.Time, bool) {
|
|
d.mu.Lock()
|
|
defer d.mu.Unlock()
|
|
|
|
if len(d.nodes) == 0 {
|
|
return time.Time{}, false
|
|
}
|
|
|
|
// Calculate the new timer value
|
|
var deadline time.Time
|
|
for _, v := range d.nodes {
|
|
if deadline.IsZero() || v.Before(deadline) {
|
|
deadline = v
|
|
}
|
|
}
|
|
|
|
var maxWithinWindow time.Time
|
|
coalescedDeadline := deadline.Add(d.coalesceWindow)
|
|
for _, nodeDeadline := range d.nodes {
|
|
if nodeDeadline.Before(coalescedDeadline) {
|
|
if maxWithinWindow.IsZero() || nodeDeadline.After(maxWithinWindow) {
|
|
maxWithinWindow = nodeDeadline
|
|
}
|
|
}
|
|
}
|
|
|
|
return maxWithinWindow, true
|
|
}
|
|
|
|
// NextBatch returns the next batch of nodes to be drained.
|
|
func (d *deadlineHeap) NextBatch() <-chan []string {
|
|
return d.batch
|
|
}
|
|
|
|
func (d *deadlineHeap) Remove(nodeID string) {
|
|
d.mu.Lock()
|
|
defer d.mu.Unlock()
|
|
delete(d.nodes, nodeID)
|
|
|
|
select {
|
|
case d.trigger <- struct{}{}:
|
|
default:
|
|
}
|
|
}
|
|
|
|
func (d *deadlineHeap) Watch(nodeID string, deadline time.Time) {
|
|
d.mu.Lock()
|
|
defer d.mu.Unlock()
|
|
d.nodes[nodeID] = deadline
|
|
|
|
select {
|
|
case d.trigger <- struct{}{}:
|
|
default:
|
|
}
|
|
}
|