786 lines
22 KiB
Go
786 lines
22 KiB
Go
package nomad
|
|
|
|
import (
|
|
"sync"
|
|
"time"
|
|
|
|
"github.com/armon/go-metrics"
|
|
"github.com/hashicorp/go-hclog"
|
|
"github.com/hashicorp/nomad/helper"
|
|
"github.com/hashicorp/nomad/nomad/structs"
|
|
)
|
|
|
|
const (
|
|
// unblockBuffer is the buffer size for the unblock channel. The buffer
|
|
// should be large to ensure that the FSM doesn't block when calling Unblock
|
|
// as this would apply back-pressure on Raft.
|
|
unblockBuffer = 8096
|
|
|
|
// pruneInterval is the interval at which we prune objects from the
|
|
// BlockedEvals tracker
|
|
pruneInterval = 5 * time.Minute
|
|
|
|
// pruneThreshold is the threshold after which objects will be pruned.
|
|
pruneThreshold = 15 * time.Minute
|
|
)
|
|
|
|
// BlockedEvals is used to track evaluations that shouldn't be queued until a
|
|
// certain class of nodes becomes available. An evaluation is put into the
|
|
// blocked state when it is run through the scheduler and produced failed
|
|
// allocations. It is unblocked when the capacity of a node that could run the
|
|
// failed allocation becomes available.
|
|
type BlockedEvals struct {
|
|
// logger is the logger to use by the blocked eval tracker.
|
|
logger hclog.Logger
|
|
|
|
evalBroker *EvalBroker
|
|
enabled bool
|
|
stats *BlockedStats
|
|
l sync.RWMutex
|
|
|
|
// captured is the set of evaluations that are captured by computed node
|
|
// classes.
|
|
captured map[string]wrappedEval
|
|
|
|
// escaped is the set of evaluations that have escaped computed node
|
|
// classes.
|
|
escaped map[string]wrappedEval
|
|
|
|
// system is the set of system evaluations that failed to start on nodes because of
|
|
// resource constraints.
|
|
system *systemEvals
|
|
|
|
// unblockCh is used to buffer unblocking of evaluations.
|
|
capacityChangeCh chan *capacityUpdate
|
|
|
|
// jobs is the map of blocked job and is used to ensure that only one
|
|
// blocked eval exists for each job. The value is the blocked evaluation ID.
|
|
jobs map[structs.NamespacedID]string
|
|
|
|
// unblockIndexes maps computed node classes or quota name to the index in
|
|
// which they were unblocked. This is used to check if an evaluation could
|
|
// have been unblocked between the time they were in the scheduler and the
|
|
// time they are being blocked.
|
|
unblockIndexes map[string]uint64
|
|
|
|
// duplicates is the set of evaluations for jobs that had pre-existing
|
|
// blocked evaluations. These should be marked as cancelled since only one
|
|
// blocked eval is needed per job.
|
|
duplicates []*structs.Evaluation
|
|
|
|
// duplicateCh is used to signal that a duplicate eval was added to the
|
|
// duplicate set. It can be used to unblock waiting callers looking for
|
|
// duplicates.
|
|
duplicateCh chan struct{}
|
|
|
|
// timetable is used to correlate indexes with their insertion time. This
|
|
// allows us to prune based on time.
|
|
timetable *TimeTable
|
|
|
|
// stopCh is used to stop any created goroutines.
|
|
stopCh chan struct{}
|
|
}
|
|
|
|
// capacityUpdate stores unblock data.
|
|
type capacityUpdate struct {
|
|
computedClass string
|
|
quotaChange string
|
|
index uint64
|
|
}
|
|
|
|
// wrappedEval captures both the evaluation and the optional token
|
|
type wrappedEval struct {
|
|
eval *structs.Evaluation
|
|
token string
|
|
}
|
|
|
|
// NewBlockedEvals creates a new blocked eval tracker that will enqueue
|
|
// unblocked evals into the passed broker.
|
|
func NewBlockedEvals(evalBroker *EvalBroker, logger hclog.Logger) *BlockedEvals {
|
|
return &BlockedEvals{
|
|
logger: logger.Named("blocked_evals"),
|
|
evalBroker: evalBroker,
|
|
captured: make(map[string]wrappedEval),
|
|
escaped: make(map[string]wrappedEval),
|
|
system: newSystemEvals(),
|
|
jobs: make(map[structs.NamespacedID]string),
|
|
unblockIndexes: make(map[string]uint64),
|
|
capacityChangeCh: make(chan *capacityUpdate, unblockBuffer),
|
|
duplicateCh: make(chan struct{}, 1),
|
|
stopCh: make(chan struct{}),
|
|
stats: NewBlockedStats(),
|
|
}
|
|
}
|
|
|
|
// Enabled is used to check if the broker is enabled.
|
|
func (b *BlockedEvals) Enabled() bool {
|
|
b.l.RLock()
|
|
defer b.l.RUnlock()
|
|
return b.enabled
|
|
}
|
|
|
|
// SetEnabled is used to control if the blocked eval tracker is enabled. The
|
|
// tracker should only be enabled on the active leader.
|
|
func (b *BlockedEvals) SetEnabled(enabled bool) {
|
|
b.l.Lock()
|
|
if b.enabled == enabled {
|
|
// No-op
|
|
b.l.Unlock()
|
|
return
|
|
} else if enabled {
|
|
go b.watchCapacity(b.stopCh, b.capacityChangeCh)
|
|
go b.prune(b.stopCh)
|
|
} else {
|
|
close(b.stopCh)
|
|
}
|
|
b.enabled = enabled
|
|
b.l.Unlock()
|
|
if !enabled {
|
|
b.Flush()
|
|
}
|
|
}
|
|
|
|
func (b *BlockedEvals) SetTimetable(timetable *TimeTable) {
|
|
b.l.Lock()
|
|
b.timetable = timetable
|
|
b.l.Unlock()
|
|
}
|
|
|
|
// Block tracks the passed evaluation and enqueues it into the eval broker when
|
|
// a suitable node calls unblock.
|
|
func (b *BlockedEvals) Block(eval *structs.Evaluation) {
|
|
b.processBlock(eval, "")
|
|
}
|
|
|
|
// Reblock tracks the passed evaluation and enqueues it into the eval broker when
|
|
// a suitable node calls unblock. Reblock should be used over Block when the
|
|
// blocking is occurring by an outstanding evaluation. The token is the
|
|
// evaluation's token.
|
|
func (b *BlockedEvals) Reblock(eval *structs.Evaluation, token string) {
|
|
b.processBlock(eval, token)
|
|
}
|
|
|
|
// processBlock is the implementation of blocking an evaluation. It supports
|
|
// taking an optional evaluation token to use when reblocking an evaluation that
|
|
// may be outstanding.
|
|
func (b *BlockedEvals) processBlock(eval *structs.Evaluation, token string) {
|
|
b.l.Lock()
|
|
defer b.l.Unlock()
|
|
|
|
// Do nothing if not enabled
|
|
if !b.enabled {
|
|
return
|
|
}
|
|
|
|
// Handle the new evaluation being for a job we are already tracking.
|
|
if b.processBlockJobDuplicate(eval) {
|
|
// If process block job duplicate returns true, the new evaluation has
|
|
// been marked as a duplicate and we have nothing to do, so return
|
|
// early.
|
|
return
|
|
}
|
|
|
|
// Check if the eval missed an unblock while it was in the scheduler at an
|
|
// older index. The scheduler could have been invoked with a snapshot of
|
|
// state that was prior to additional capacity being added or allocations
|
|
// becoming terminal.
|
|
if b.missedUnblock(eval) {
|
|
// Just re-enqueue the eval immediately. We pass the token so that the
|
|
// eval_broker can properly handle the case in which the evaluation is
|
|
// still outstanding.
|
|
b.evalBroker.EnqueueAll(map[*structs.Evaluation]string{eval: token})
|
|
return
|
|
}
|
|
|
|
// Mark the job as tracked.
|
|
b.jobs[structs.NewNamespacedID(eval.JobID, eval.Namespace)] = eval.ID
|
|
b.stats.Block(eval)
|
|
|
|
// Track that the evaluation is being added due to reaching the quota limit
|
|
if eval.QuotaLimitReached != "" {
|
|
b.stats.TotalQuotaLimit++
|
|
}
|
|
|
|
// Wrap the evaluation, capturing its token.
|
|
wrapped := wrappedEval{
|
|
eval: eval,
|
|
token: token,
|
|
}
|
|
|
|
// If the eval has escaped, meaning computed node classes could not capture
|
|
// the constraints of the job, we store the eval separately as we have to
|
|
// unblock it whenever node capacity changes. This is because we don't know
|
|
// what node class is feasible for the jobs constraints.
|
|
if eval.EscapedComputedClass {
|
|
b.escaped[eval.ID] = wrapped
|
|
b.stats.TotalEscaped++
|
|
return
|
|
}
|
|
|
|
// System evals are indexed by node and re-processed on utilization changes in
|
|
// existing nodes
|
|
if eval.Type == structs.JobTypeSystem {
|
|
b.system.Add(eval, token)
|
|
}
|
|
|
|
// Add the eval to the set of blocked evals whose jobs constraints are
|
|
// captured by computed node class.
|
|
b.captured[eval.ID] = wrapped
|
|
}
|
|
|
|
// processBlockJobDuplicate handles the case where the new eval is for a job
|
|
// that we are already tracking. If the eval is a duplicate, we add the older
|
|
// evaluation by Raft index to the list of duplicates such that it can be
|
|
// cancelled. We only ever want one blocked evaluation per job, otherwise we
|
|
// would create unnecessary work for the scheduler as multiple evals for the
|
|
// same job would be run, all producing the same outcome. It is critical to
|
|
// prefer the newer evaluation, since it will contain the most up to date set of
|
|
// class eligibility. The return value is set to true, if the passed evaluation
|
|
// is cancelled. This should be called with the lock held.
|
|
func (b *BlockedEvals) processBlockJobDuplicate(eval *structs.Evaluation) (newCancelled bool) {
|
|
existingID, hasExisting := b.jobs[structs.NewNamespacedID(eval.JobID, eval.Namespace)]
|
|
if !hasExisting {
|
|
return
|
|
}
|
|
|
|
var dup *structs.Evaluation
|
|
existingW, ok := b.captured[existingID]
|
|
if ok {
|
|
if latestEvalIndex(existingW.eval) <= latestEvalIndex(eval) {
|
|
delete(b.captured, existingID)
|
|
dup = existingW.eval
|
|
b.stats.Unblock(dup)
|
|
} else {
|
|
dup = eval
|
|
newCancelled = true
|
|
}
|
|
} else {
|
|
existingW, ok = b.escaped[existingID]
|
|
if !ok {
|
|
// This is a programming error
|
|
b.logger.Error("existing blocked evaluation is neither tracked as captured or escaped", "existing_id", existingID)
|
|
delete(b.jobs, structs.NewNamespacedID(eval.JobID, eval.Namespace))
|
|
return
|
|
}
|
|
|
|
if latestEvalIndex(existingW.eval) <= latestEvalIndex(eval) {
|
|
delete(b.escaped, existingID)
|
|
b.stats.TotalEscaped--
|
|
dup = existingW.eval
|
|
} else {
|
|
dup = eval
|
|
newCancelled = true
|
|
}
|
|
}
|
|
|
|
b.duplicates = append(b.duplicates, dup)
|
|
|
|
// Unblock any waiter.
|
|
select {
|
|
case b.duplicateCh <- struct{}{}:
|
|
default:
|
|
}
|
|
|
|
return
|
|
}
|
|
|
|
// latestEvalIndex returns the max of the evaluations create and snapshot index
|
|
func latestEvalIndex(eval *structs.Evaluation) uint64 {
|
|
if eval == nil {
|
|
return 0
|
|
}
|
|
|
|
return helper.Uint64Max(eval.CreateIndex, eval.SnapshotIndex)
|
|
}
|
|
|
|
// missedUnblock returns whether an evaluation missed an unblock while it was in
|
|
// the scheduler. Since the scheduler can operate at an index in the past, the
|
|
// evaluation may have been processed missing data that would allow it to
|
|
// complete. This method returns if that is the case and should be called with
|
|
// the lock held.
|
|
func (b *BlockedEvals) missedUnblock(eval *structs.Evaluation) bool {
|
|
var max uint64 = 0
|
|
for id, index := range b.unblockIndexes {
|
|
// Calculate the max unblock index
|
|
if max < index {
|
|
max = index
|
|
}
|
|
|
|
// The evaluation is blocked because it has hit a quota limit not class
|
|
// eligibility
|
|
if eval.QuotaLimitReached != "" {
|
|
if eval.QuotaLimitReached != id {
|
|
// Not a match
|
|
continue
|
|
} else if eval.SnapshotIndex < index {
|
|
// The evaluation was processed before the quota specification was
|
|
// updated, so unblock the evaluation.
|
|
return true
|
|
}
|
|
|
|
// The evaluation was processed having seen all changes to the quota
|
|
return false
|
|
}
|
|
|
|
elig, ok := eval.ClassEligibility[id]
|
|
if !ok && eval.SnapshotIndex < index {
|
|
// The evaluation was processed and did not encounter this class
|
|
// because it was added after it was processed. Thus for correctness
|
|
// we need to unblock it.
|
|
return true
|
|
}
|
|
|
|
// The evaluation could use the computed node class and the eval was
|
|
// processed before the last unblock.
|
|
if elig && eval.SnapshotIndex < index {
|
|
return true
|
|
}
|
|
}
|
|
|
|
// If the evaluation has escaped, and the map contains an index older than
|
|
// the evaluations, it should be unblocked.
|
|
if eval.EscapedComputedClass && eval.SnapshotIndex < max {
|
|
return true
|
|
}
|
|
|
|
// The evaluation is ahead of all recent unblocks.
|
|
return false
|
|
}
|
|
|
|
// Untrack causes any blocked evaluation for the passed job to be no longer
|
|
// tracked. Untrack is called when there is a successful evaluation for the job
|
|
// and a blocked evaluation is no longer needed.
|
|
func (b *BlockedEvals) Untrack(jobID, namespace string) {
|
|
b.l.Lock()
|
|
defer b.l.Unlock()
|
|
|
|
// Do nothing if not enabled
|
|
if !b.enabled {
|
|
return
|
|
}
|
|
|
|
nsID := structs.NewNamespacedID(jobID, namespace)
|
|
|
|
if evals, ok := b.system.JobEvals(nsID); ok {
|
|
for _, e := range evals {
|
|
b.system.Remove(e)
|
|
b.stats.Unblock(e)
|
|
}
|
|
return
|
|
}
|
|
|
|
// Get the evaluation ID to cancel
|
|
evalID, ok := b.jobs[nsID]
|
|
if !ok {
|
|
// No blocked evaluation so exit
|
|
return
|
|
}
|
|
|
|
// Attempt to delete the evaluation
|
|
if w, ok := b.captured[evalID]; ok {
|
|
delete(b.jobs, nsID)
|
|
delete(b.captured, evalID)
|
|
b.stats.Unblock(w.eval)
|
|
if w.eval.QuotaLimitReached != "" {
|
|
b.stats.TotalQuotaLimit--
|
|
}
|
|
}
|
|
|
|
if w, ok := b.escaped[evalID]; ok {
|
|
delete(b.jobs, nsID)
|
|
delete(b.escaped, evalID)
|
|
b.stats.TotalEscaped--
|
|
b.stats.Unblock(w.eval)
|
|
if w.eval.QuotaLimitReached != "" {
|
|
b.stats.TotalQuotaLimit--
|
|
}
|
|
}
|
|
}
|
|
|
|
// Unblock causes any evaluation that could potentially make progress on a
|
|
// capacity change on the passed computed node class to be enqueued into the
|
|
// eval broker.
|
|
func (b *BlockedEvals) Unblock(computedClass string, index uint64) {
|
|
b.l.Lock()
|
|
|
|
// Do nothing if not enabled
|
|
if !b.enabled {
|
|
b.l.Unlock()
|
|
return
|
|
}
|
|
|
|
// Store the index in which the unblock happened. We use this on subsequent
|
|
// block calls in case the evaluation was in the scheduler when a trigger
|
|
// occurred.
|
|
b.unblockIndexes[computedClass] = index
|
|
b.l.Unlock()
|
|
|
|
b.capacityChangeCh <- &capacityUpdate{
|
|
computedClass: computedClass,
|
|
index: index,
|
|
}
|
|
}
|
|
|
|
// UnblockQuota causes any evaluation that could potentially make progress on a
|
|
// capacity change on the passed quota to be enqueued into the eval broker.
|
|
func (b *BlockedEvals) UnblockQuota(quota string, index uint64) {
|
|
// Nothing to do
|
|
if quota == "" {
|
|
return
|
|
}
|
|
|
|
b.l.Lock()
|
|
|
|
// Do nothing if not enabled
|
|
if !b.enabled {
|
|
b.l.Unlock()
|
|
return
|
|
}
|
|
|
|
// Store the index in which the unblock happened. We use this on subsequent
|
|
// block calls in case the evaluation was in the scheduler when a trigger
|
|
// occurred.
|
|
b.unblockIndexes[quota] = index
|
|
b.l.Unlock()
|
|
|
|
b.capacityChangeCh <- &capacityUpdate{
|
|
quotaChange: quota,
|
|
index: index,
|
|
}
|
|
}
|
|
|
|
// UnblockClassAndQuota causes any evaluation that could potentially make
|
|
// progress on a capacity change on the passed computed node class or quota to
|
|
// be enqueued into the eval broker.
|
|
func (b *BlockedEvals) UnblockClassAndQuota(class, quota string, index uint64) {
|
|
b.l.Lock()
|
|
|
|
// Do nothing if not enabled
|
|
if !b.enabled {
|
|
b.l.Unlock()
|
|
return
|
|
}
|
|
|
|
// Store the index in which the unblock happened. We use this on subsequent
|
|
// block calls in case the evaluation was in the scheduler when a trigger
|
|
// occurred.
|
|
if quota != "" {
|
|
b.unblockIndexes[quota] = index
|
|
}
|
|
b.unblockIndexes[class] = index
|
|
|
|
// Capture chan inside the lock to prevent a race with it getting reset
|
|
// in Flush.
|
|
ch := b.capacityChangeCh
|
|
b.l.Unlock()
|
|
|
|
ch <- &capacityUpdate{
|
|
computedClass: class,
|
|
quotaChange: quota,
|
|
index: index,
|
|
}
|
|
}
|
|
|
|
// UnblockNode finds any blocked evalution that's node specific (system jobs) and enqueues
|
|
// it on the eval broker
|
|
func (b *BlockedEvals) UnblockNode(nodeID string, index uint64) {
|
|
b.l.Lock()
|
|
defer b.l.Unlock()
|
|
|
|
evals, ok := b.system.NodeEvals(nodeID)
|
|
|
|
// Do nothing if not enabled
|
|
if !b.enabled || !ok || len(evals) == 0 {
|
|
return
|
|
}
|
|
|
|
for e := range evals {
|
|
b.system.Remove(e)
|
|
b.stats.Unblock(e)
|
|
}
|
|
|
|
b.evalBroker.EnqueueAll(evals)
|
|
}
|
|
|
|
// watchCapacity is a long lived function that watches for capacity changes in
|
|
// nodes and unblocks the correct set of evals.
|
|
func (b *BlockedEvals) watchCapacity(stopCh <-chan struct{}, changeCh <-chan *capacityUpdate) {
|
|
for {
|
|
select {
|
|
case <-stopCh:
|
|
return
|
|
case update := <-changeCh:
|
|
b.unblock(update.computedClass, update.quotaChange, update.index)
|
|
}
|
|
}
|
|
}
|
|
|
|
func (b *BlockedEvals) unblock(computedClass, quota string, index uint64) {
|
|
b.l.Lock()
|
|
defer b.l.Unlock()
|
|
|
|
// Protect against the case of a flush.
|
|
if !b.enabled {
|
|
return
|
|
}
|
|
|
|
// Every eval that has escaped computed node class has to be unblocked
|
|
// because any node could potentially be feasible.
|
|
numEscaped := len(b.escaped)
|
|
numQuotaLimit := 0
|
|
unblocked := make(map[*structs.Evaluation]string, helper.MaxInt(numEscaped, 4))
|
|
|
|
if numEscaped != 0 && computedClass != "" {
|
|
for id, wrapped := range b.escaped {
|
|
unblocked[wrapped.eval] = wrapped.token
|
|
delete(b.escaped, id)
|
|
delete(b.jobs, structs.NewNamespacedID(wrapped.eval.JobID, wrapped.eval.Namespace))
|
|
|
|
if wrapped.eval.QuotaLimitReached != "" {
|
|
numQuotaLimit++
|
|
}
|
|
}
|
|
}
|
|
|
|
// We unblock any eval that is explicitly eligible for the computed class
|
|
// and also any eval that is not eligible or uneligible. This signifies that
|
|
// when the evaluation was originally run through the scheduler, that it
|
|
// never saw a node with the given computed class and thus needs to be
|
|
// unblocked for correctness.
|
|
for id, wrapped := range b.captured {
|
|
if quota != "" && wrapped.eval.QuotaLimitReached != quota {
|
|
// We are unblocking based on quota and this eval doesn't match
|
|
continue
|
|
} else if elig, ok := wrapped.eval.ClassEligibility[computedClass]; ok && !elig {
|
|
// Can skip because the eval has explicitly marked the node class
|
|
// as ineligible.
|
|
continue
|
|
}
|
|
|
|
// Unblock the evaluation because it is either for the matching quota,
|
|
// is eligible based on the computed node class, or never seen the
|
|
// computed node class.
|
|
unblocked[wrapped.eval] = wrapped.token
|
|
delete(b.jobs, structs.NewNamespacedID(wrapped.eval.JobID, wrapped.eval.Namespace))
|
|
delete(b.captured, id)
|
|
if wrapped.eval.QuotaLimitReached != "" {
|
|
numQuotaLimit++
|
|
}
|
|
}
|
|
|
|
if len(unblocked) != 0 {
|
|
// Update the counters
|
|
b.stats.TotalEscaped = 0
|
|
b.stats.TotalQuotaLimit -= numQuotaLimit
|
|
for eval := range unblocked {
|
|
b.stats.Unblock(eval)
|
|
}
|
|
|
|
// Enqueue all the unblocked evals into the broker.
|
|
b.evalBroker.EnqueueAll(unblocked)
|
|
}
|
|
}
|
|
|
|
// UnblockFailed unblocks all blocked evaluation that were due to scheduler
|
|
// failure.
|
|
func (b *BlockedEvals) UnblockFailed() {
|
|
b.l.Lock()
|
|
defer b.l.Unlock()
|
|
|
|
// Do nothing if not enabled
|
|
if !b.enabled {
|
|
return
|
|
}
|
|
|
|
quotaLimit := 0
|
|
unblocked := make(map[*structs.Evaluation]string, 4)
|
|
for id, wrapped := range b.captured {
|
|
if wrapped.eval.TriggeredBy == structs.EvalTriggerMaxPlans {
|
|
unblocked[wrapped.eval] = wrapped.token
|
|
delete(b.captured, id)
|
|
delete(b.jobs, structs.NewNamespacedID(wrapped.eval.JobID, wrapped.eval.Namespace))
|
|
if wrapped.eval.QuotaLimitReached != "" {
|
|
quotaLimit++
|
|
}
|
|
}
|
|
}
|
|
|
|
for id, wrapped := range b.escaped {
|
|
if wrapped.eval.TriggeredBy == structs.EvalTriggerMaxPlans {
|
|
unblocked[wrapped.eval] = wrapped.token
|
|
delete(b.escaped, id)
|
|
delete(b.jobs, structs.NewNamespacedID(wrapped.eval.JobID, wrapped.eval.Namespace))
|
|
b.stats.TotalEscaped -= 1
|
|
if wrapped.eval.QuotaLimitReached != "" {
|
|
quotaLimit++
|
|
}
|
|
}
|
|
}
|
|
|
|
if len(unblocked) > 0 {
|
|
b.stats.TotalQuotaLimit -= quotaLimit
|
|
for eval := range unblocked {
|
|
b.stats.Unblock(eval)
|
|
}
|
|
|
|
b.evalBroker.EnqueueAll(unblocked)
|
|
}
|
|
}
|
|
|
|
// GetDuplicates returns all the duplicate evaluations and blocks until the
|
|
// passed timeout.
|
|
func (b *BlockedEvals) GetDuplicates(timeout time.Duration) []*structs.Evaluation {
|
|
var timeoutTimer *time.Timer
|
|
var timeoutCh <-chan time.Time
|
|
SCAN:
|
|
b.l.Lock()
|
|
if len(b.duplicates) != 0 {
|
|
dups := b.duplicates
|
|
b.duplicates = nil
|
|
b.l.Unlock()
|
|
return dups
|
|
}
|
|
|
|
// Capture chans inside the lock to prevent a race with them getting
|
|
// reset in Flush
|
|
dupCh := b.duplicateCh
|
|
stopCh := b.stopCh
|
|
b.l.Unlock()
|
|
|
|
// Create the timer
|
|
if timeoutTimer == nil && timeout != 0 {
|
|
timeoutTimer = time.NewTimer(timeout)
|
|
timeoutCh = timeoutTimer.C
|
|
defer timeoutTimer.Stop()
|
|
}
|
|
|
|
select {
|
|
case <-stopCh:
|
|
return nil
|
|
case <-timeoutCh:
|
|
return nil
|
|
case <-dupCh:
|
|
goto SCAN
|
|
}
|
|
}
|
|
|
|
// Flush is used to clear the state of blocked evaluations.
|
|
func (b *BlockedEvals) Flush() {
|
|
b.l.Lock()
|
|
defer b.l.Unlock()
|
|
|
|
// Reset the blocked eval tracker.
|
|
b.stats.TotalEscaped = 0
|
|
b.stats.TotalBlocked = 0
|
|
b.stats.TotalQuotaLimit = 0
|
|
b.stats.BlockedResources = NewBlockedResourcesStats()
|
|
b.captured = make(map[string]wrappedEval)
|
|
b.escaped = make(map[string]wrappedEval)
|
|
b.jobs = make(map[structs.NamespacedID]string)
|
|
b.unblockIndexes = make(map[string]uint64)
|
|
b.timetable = nil
|
|
b.duplicates = nil
|
|
b.capacityChangeCh = make(chan *capacityUpdate, unblockBuffer)
|
|
b.stopCh = make(chan struct{})
|
|
b.duplicateCh = make(chan struct{}, 1)
|
|
b.system = newSystemEvals()
|
|
}
|
|
|
|
// Stats is used to query the state of the blocked eval tracker.
|
|
func (b *BlockedEvals) Stats() *BlockedStats {
|
|
// Allocate a new stats struct
|
|
stats := NewBlockedStats()
|
|
|
|
b.l.RLock()
|
|
defer b.l.RUnlock()
|
|
|
|
// Copy all the stats
|
|
stats.TotalEscaped = b.stats.TotalEscaped
|
|
stats.TotalBlocked = b.stats.TotalBlocked
|
|
stats.TotalQuotaLimit = b.stats.TotalQuotaLimit
|
|
stats.BlockedResources = b.stats.BlockedResources.Copy()
|
|
|
|
return stats
|
|
}
|
|
|
|
// EmitStats is used to export metrics about the blocked eval tracker while enabled
|
|
func (b *BlockedEvals) EmitStats(period time.Duration, stopCh <-chan struct{}) {
|
|
timer, stop := helper.NewSafeTimer(period)
|
|
defer stop()
|
|
|
|
for {
|
|
timer.Reset(period)
|
|
|
|
select {
|
|
case <-timer.C:
|
|
stats := b.Stats()
|
|
metrics.SetGauge([]string{"nomad", "blocked_evals", "total_quota_limit"}, float32(stats.TotalQuotaLimit))
|
|
metrics.SetGauge([]string{"nomad", "blocked_evals", "total_blocked"}, float32(stats.TotalBlocked))
|
|
metrics.SetGauge([]string{"nomad", "blocked_evals", "total_escaped"}, float32(stats.TotalEscaped))
|
|
|
|
for k, v := range stats.BlockedResources.ByJob {
|
|
labels := []metrics.Label{
|
|
{Name: "namespace", Value: k.Namespace},
|
|
{Name: "job", Value: k.ID},
|
|
}
|
|
metrics.SetGaugeWithLabels([]string{"nomad", "blocked_evals", "job", "cpu"}, float32(v.CPU), labels)
|
|
metrics.SetGaugeWithLabels([]string{"nomad", "blocked_evals", "job", "memory"}, float32(v.MemoryMB), labels)
|
|
}
|
|
|
|
for k, v := range stats.BlockedResources.ByClassInDC {
|
|
labels := []metrics.Label{
|
|
{Name: "datacenter", Value: k.dc},
|
|
{Name: "node_class", Value: k.class},
|
|
}
|
|
metrics.SetGaugeWithLabels([]string{"nomad", "blocked_evals", "cpu"}, float32(v.CPU), labels)
|
|
metrics.SetGaugeWithLabels([]string{"nomad", "blocked_evals", "memory"}, float32(v.MemoryMB), labels)
|
|
}
|
|
case <-stopCh:
|
|
return
|
|
}
|
|
}
|
|
}
|
|
|
|
// prune is a long lived function that prunes unnecessary objects on a timer.
|
|
func (b *BlockedEvals) prune(stopCh <-chan struct{}) {
|
|
ticker := time.NewTicker(pruneInterval)
|
|
defer ticker.Stop()
|
|
|
|
for {
|
|
select {
|
|
case <-stopCh:
|
|
return
|
|
case t := <-ticker.C:
|
|
cutoff := t.UTC().Add(-1 * pruneThreshold)
|
|
b.pruneUnblockIndexes(cutoff)
|
|
b.pruneStats(cutoff)
|
|
}
|
|
}
|
|
}
|
|
|
|
// pruneUnblockIndexes is used to prune any tracked entry that is excessively
|
|
// old. This protects againsts unbounded growth of the map.
|
|
func (b *BlockedEvals) pruneUnblockIndexes(cutoff time.Time) {
|
|
b.l.Lock()
|
|
defer b.l.Unlock()
|
|
|
|
if b.timetable == nil {
|
|
return
|
|
}
|
|
|
|
oldThreshold := b.timetable.NearestIndex(cutoff)
|
|
for key, index := range b.unblockIndexes {
|
|
if index < oldThreshold {
|
|
delete(b.unblockIndexes, key)
|
|
}
|
|
}
|
|
}
|
|
|
|
// pruneStats is used to prune any zero value stats that are excessively old.
|
|
func (b *BlockedEvals) pruneStats(cutoff time.Time) {
|
|
b.l.Lock()
|
|
defer b.l.Unlock()
|
|
|
|
b.stats.prune(cutoff)
|
|
}
|