2023-04-10 15:36:59 +00:00
|
|
|
// Copyright (c) HashiCorp, Inc.
|
|
|
|
// SPDX-License-Identifier: MPL-2.0
|
|
|
|
|
2015-07-28 22:12:08 +00:00
|
|
|
package nomad
|
|
|
|
|
|
|
|
import (
|
2019-03-12 21:25:14 +00:00
|
|
|
"context"
|
2022-01-06 16:56:13 +00:00
|
|
|
"encoding/json"
|
core: backoff considerably when worker is behind raft (#15523)
Upon dequeuing an evaluation workers snapshot their state store at the
eval's wait index or later. This ensures we process an eval at a point
in time after it was created or updated. Processing an eval on an old
snapshot could cause any number of problems such as:
1. Since job registration atomically updates an eval and job in a single
raft entry, scheduling against indexes before that may not have the
eval's job or may have an older version.
2. The older the scheduler's snapshot, the higher the likelihood
something has changed in the cluster state which will cause the plan
applier to reject the scheduler's plan. This could waste work or
even cause eval's to be failed needlessly.
However, the workers run in parallel with a new server pulling the
cluster state from a peer. During this time, which may be many minutes
long, the state store is likely far behind the minimum index required
to process evaluations.
This PR addresses this by adding an additional long backoff period after
an eval is nacked. If the scheduler's indexes catches up within the
additional backoff, it will unblock early to dequeue the next eval.
When the server shuts down we'll get a `context.Canceled` error from the state
store method. We need to bubble this error up so that other callers can detect
it. Handle this case separately when waiting after dequeue so that we can warn
on shutdown instead of throwing an ambiguous error message with just the text
"canceled."
While there may be more precise ways to block scheduling until the
server catches up, this approach adds little risk and covers additional
cases where a server may be temporarily behind due to a spike in load or
a saturated network.
For testing, we make the `raftSyncLimit` into a parameter on the worker's `run` method
so that we can run backoff tests without waiting 30+ seconds. We haven't followed thru
and made all the worker globals into worker parameters, because there isn't much
use outside of testing, but we can consider that in the future.
Co-authored-by: Tim Gross <tgross@hashicorp.com>
2023-01-24 13:56:35 +00:00
|
|
|
"errors"
|
2015-07-28 22:12:08 +00:00
|
|
|
"fmt"
|
2015-08-05 23:23:37 +00:00
|
|
|
"strings"
|
2015-08-23 17:52:31 +00:00
|
|
|
"sync"
|
2015-07-28 22:12:08 +00:00
|
|
|
"time"
|
|
|
|
|
2018-09-15 23:23:13 +00:00
|
|
|
log "github.com/hashicorp/go-hclog"
|
2017-02-08 04:31:23 +00:00
|
|
|
memdb "github.com/hashicorp/go-memdb"
|
2023-11-28 18:52:25 +00:00
|
|
|
metrics "github.com/hashicorp/go-metrics"
|
2022-03-07 18:40:57 +00:00
|
|
|
"github.com/hashicorp/go-version"
|
2023-08-15 18:38:18 +00:00
|
|
|
"github.com/hashicorp/nomad/helper"
|
2022-01-06 16:56:13 +00:00
|
|
|
"github.com/hashicorp/nomad/helper/uuid"
|
2019-03-12 21:25:14 +00:00
|
|
|
"github.com/hashicorp/nomad/nomad/state"
|
2015-07-28 22:12:08 +00:00
|
|
|
"github.com/hashicorp/nomad/nomad/structs"
|
2015-07-28 23:15:32 +00:00
|
|
|
"github.com/hashicorp/nomad/scheduler"
|
2015-07-28 22:12:08 +00:00
|
|
|
)
|
|
|
|
|
|
|
|
const (
|
2015-08-24 00:39:49 +00:00
|
|
|
// backoffBaselineFast is the baseline time for exponential backoff
|
|
|
|
backoffBaselineFast = 20 * time.Millisecond
|
2015-07-28 22:12:08 +00:00
|
|
|
|
2015-08-24 00:39:49 +00:00
|
|
|
// backoffBaselineSlow is the baseline time for exponential backoff
|
|
|
|
// but that is much slower than backoffBaselineFast
|
|
|
|
backoffBaselineSlow = 500 * time.Millisecond
|
|
|
|
|
|
|
|
// backoffLimitSlow is the limit of the exponential backoff for
|
|
|
|
// the slower backoff
|
|
|
|
backoffLimitSlow = 10 * time.Second
|
2015-07-28 22:12:08 +00:00
|
|
|
|
2016-10-27 18:41:43 +00:00
|
|
|
// backoffSchedulerVersionMismatch is the backoff between retries when the
|
|
|
|
// scheduler version mismatches that of the leader.
|
|
|
|
backoffSchedulerVersionMismatch = 30 * time.Second
|
|
|
|
|
2015-07-28 22:12:08 +00:00
|
|
|
// dequeueTimeout is used to timeout an evaluation dequeue so that
|
|
|
|
// we can check if there is a shutdown event
|
|
|
|
dequeueTimeout = 500 * time.Millisecond
|
|
|
|
|
|
|
|
// raftSyncLimit is the limit of time we will wait for Raft replication
|
|
|
|
// to catch up to the evaluation. This is used to fast Nack and
|
|
|
|
// allow another scheduler to pick it up.
|
|
|
|
raftSyncLimit = 5 * time.Second
|
2015-09-07 03:18:47 +00:00
|
|
|
|
|
|
|
// dequeueErrGrace is the grace period where we don't log about
|
|
|
|
// dequeue errors after start. This is to improve the user experience
|
|
|
|
// in dev mode where the leader isn't elected for a few seconds.
|
|
|
|
dequeueErrGrace = 10 * time.Second
|
2015-07-28 22:12:08 +00:00
|
|
|
)
|
|
|
|
|
2022-01-06 16:56:13 +00:00
|
|
|
type WorkerStatus int
|
|
|
|
|
|
|
|
//go:generate stringer -trimprefix=Worker -output worker_string_workerstatus.go -linecomment -type=WorkerStatus
|
|
|
|
const (
|
|
|
|
WorkerUnknownStatus WorkerStatus = iota // Unknown
|
|
|
|
WorkerStarting
|
|
|
|
WorkerStarted
|
|
|
|
WorkerPausing
|
|
|
|
WorkerPaused
|
|
|
|
WorkerResuming
|
|
|
|
WorkerStopping
|
|
|
|
WorkerStopped
|
|
|
|
)
|
|
|
|
|
|
|
|
type SchedulerWorkerStatus int
|
|
|
|
|
|
|
|
//go:generate stringer -trimprefix=Workload -output worker_string_schedulerworkerstatus.go -linecomment -type=SchedulerWorkerStatus
|
|
|
|
const (
|
|
|
|
WorkloadUnknownStatus SchedulerWorkerStatus = iota
|
|
|
|
WorkloadRunning
|
|
|
|
WorkloadWaitingToDequeue
|
|
|
|
WorkloadWaitingForRaft
|
|
|
|
WorkloadScheduling
|
|
|
|
WorkloadSubmitting
|
|
|
|
WorkloadBackoff
|
|
|
|
WorkloadStopped
|
|
|
|
WorkloadPaused
|
|
|
|
)
|
|
|
|
|
2015-07-28 22:12:08 +00:00
|
|
|
// Worker is a single threaded scheduling worker. There may be multiple
|
|
|
|
// running per server (leader or follower). They are responsible for dequeuing
|
|
|
|
// pending evaluations, invoking schedulers, plan submission and the
|
|
|
|
// lifecycle around making task allocations. They bridge the business logic
|
|
|
|
// of the scheduler with the plumbing required to make it all work.
|
|
|
|
type Worker struct {
|
|
|
|
srv *Server
|
2018-09-15 23:23:13 +00:00
|
|
|
logger log.Logger
|
2015-09-07 03:18:47 +00:00
|
|
|
start time.Time
|
2022-01-06 16:56:13 +00:00
|
|
|
id string
|
|
|
|
|
|
|
|
status WorkerStatus
|
|
|
|
workloadStatus SchedulerWorkerStatus
|
|
|
|
statusLock sync.RWMutex
|
2015-07-28 22:12:08 +00:00
|
|
|
|
2022-01-06 16:56:13 +00:00
|
|
|
pauseFlag bool
|
2015-08-23 17:52:31 +00:00
|
|
|
pauseLock sync.Mutex
|
|
|
|
pauseCond *sync.Cond
|
2022-01-06 16:56:13 +00:00
|
|
|
ctx context.Context
|
|
|
|
cancelFn context.CancelFunc
|
2015-08-23 17:52:31 +00:00
|
|
|
|
2022-01-06 16:56:13 +00:00
|
|
|
// the Server.Config.EnabledSchedulers value is not safe for concurrent access, so
|
|
|
|
// the worker needs a cached copy of it. Workers are stopped if this value changes.
|
|
|
|
enabledSchedulers []string
|
2015-08-12 22:44:36 +00:00
|
|
|
|
2022-01-06 16:56:13 +00:00
|
|
|
// failures is the count of errors encountered while dequeueing evaluations
|
|
|
|
// and is used to calculate backoff.
|
2023-08-15 18:38:18 +00:00
|
|
|
failures uint64
|
|
|
|
failureBackoff time.Duration
|
|
|
|
evalToken string
|
2016-05-23 22:24:31 +00:00
|
|
|
|
|
|
|
// snapshotIndex is the index of the snapshot in which the scheduler was
|
2018-03-11 18:00:32 +00:00
|
|
|
// first invoked. It is used to mark the SnapshotIndex of evaluations
|
2016-05-23 22:24:31 +00:00
|
|
|
// Created, Updated or Reblocked.
|
2016-05-21 01:07:10 +00:00
|
|
|
snapshotIndex uint64
|
2015-07-28 22:12:08 +00:00
|
|
|
}
|
|
|
|
|
2022-01-06 16:56:13 +00:00
|
|
|
// NewWorker starts a new scheduler worker associated with the given server
|
|
|
|
func NewWorker(ctx context.Context, srv *Server, args SchedulerWorkerPoolArgs) (*Worker, error) {
|
|
|
|
w := newWorker(ctx, srv, args)
|
|
|
|
w.Start()
|
|
|
|
return w, nil
|
|
|
|
}
|
|
|
|
|
|
|
|
// _newWorker creates a worker without calling its Start func. This is useful for testing.
|
|
|
|
func newWorker(ctx context.Context, srv *Server, args SchedulerWorkerPoolArgs) *Worker {
|
2015-07-28 22:12:08 +00:00
|
|
|
w := &Worker{
|
2022-01-06 16:56:13 +00:00
|
|
|
id: uuid.Generate(),
|
|
|
|
srv: srv,
|
|
|
|
start: time.Now(),
|
|
|
|
status: WorkerStarting,
|
|
|
|
enabledSchedulers: make([]string, len(args.EnabledSchedulers)),
|
2023-08-15 18:38:18 +00:00
|
|
|
failureBackoff: time.Duration(0),
|
2015-07-28 22:12:08 +00:00
|
|
|
}
|
2022-01-06 16:56:13 +00:00
|
|
|
copy(w.enabledSchedulers, args.EnabledSchedulers)
|
|
|
|
|
|
|
|
w.logger = srv.logger.ResetNamed("worker").With("worker_id", w.id)
|
2015-08-23 17:52:31 +00:00
|
|
|
w.pauseCond = sync.NewCond(&w.pauseLock)
|
2022-01-06 16:56:13 +00:00
|
|
|
w.ctx, w.cancelFn = context.WithCancel(ctx)
|
|
|
|
|
|
|
|
return w
|
|
|
|
}
|
|
|
|
|
|
|
|
// ID returns a string ID for the worker.
|
|
|
|
func (w *Worker) ID() string {
|
|
|
|
return w.id
|
|
|
|
}
|
|
|
|
|
|
|
|
// Start transitions a worker to the starting state. Check
|
|
|
|
// to see if it paused using IsStarted()
|
|
|
|
func (w *Worker) Start() {
|
|
|
|
w.setStatus(WorkerStarting)
|
core: backoff considerably when worker is behind raft (#15523)
Upon dequeuing an evaluation workers snapshot their state store at the
eval's wait index or later. This ensures we process an eval at a point
in time after it was created or updated. Processing an eval on an old
snapshot could cause any number of problems such as:
1. Since job registration atomically updates an eval and job in a single
raft entry, scheduling against indexes before that may not have the
eval's job or may have an older version.
2. The older the scheduler's snapshot, the higher the likelihood
something has changed in the cluster state which will cause the plan
applier to reject the scheduler's plan. This could waste work or
even cause eval's to be failed needlessly.
However, the workers run in parallel with a new server pulling the
cluster state from a peer. During this time, which may be many minutes
long, the state store is likely far behind the minimum index required
to process evaluations.
This PR addresses this by adding an additional long backoff period after
an eval is nacked. If the scheduler's indexes catches up within the
additional backoff, it will unblock early to dequeue the next eval.
When the server shuts down we'll get a `context.Canceled` error from the state
store method. We need to bubble this error up so that other callers can detect
it. Handle this case separately when waiting after dequeue so that we can warn
on shutdown instead of throwing an ambiguous error message with just the text
"canceled."
While there may be more precise ways to block scheduling until the
server catches up, this approach adds little risk and covers additional
cases where a server may be temporarily behind due to a spike in load or
a saturated network.
For testing, we make the `raftSyncLimit` into a parameter on the worker's `run` method
so that we can run backoff tests without waiting 30+ seconds. We haven't followed thru
and made all the worker globals into worker parameters, because there isn't much
use outside of testing, but we can consider that in the future.
Co-authored-by: Tim Gross <tgross@hashicorp.com>
2023-01-24 13:56:35 +00:00
|
|
|
go w.run(raftSyncLimit)
|
2015-07-28 22:12:08 +00:00
|
|
|
}
|
|
|
|
|
2022-01-06 16:56:13 +00:00
|
|
|
// Pause transitions a worker to the pausing state. Check
|
|
|
|
// to see if it paused using IsPaused()
|
|
|
|
func (w *Worker) Pause() {
|
|
|
|
if w.isPausable() {
|
|
|
|
w.setStatus(WorkerPausing)
|
|
|
|
w.setPauseFlag(true)
|
|
|
|
}
|
|
|
|
}
|
|
|
|
|
|
|
|
// Resume transitions a worker to the resuming state. Check
|
|
|
|
// to see if the worker restarted by calling IsStarted()
|
|
|
|
func (w *Worker) Resume() {
|
|
|
|
if w.IsPaused() {
|
|
|
|
w.setStatus(WorkerResuming)
|
|
|
|
w.setPauseFlag(false)
|
2015-08-23 17:52:31 +00:00
|
|
|
w.pauseCond.Broadcast()
|
|
|
|
}
|
|
|
|
}
|
|
|
|
|
2022-01-06 16:56:13 +00:00
|
|
|
// Resume transitions a worker to the stopping state. Check
|
|
|
|
// to see if the worker stopped by calling IsStopped()
|
|
|
|
func (w *Worker) Stop() {
|
|
|
|
w.setStatus(WorkerStopping)
|
|
|
|
w.shutdown()
|
|
|
|
}
|
|
|
|
|
|
|
|
// IsStarted returns a boolean indicating if this worker has been started.
|
|
|
|
func (w *Worker) IsStarted() bool {
|
|
|
|
return w.GetStatus() == WorkerStarted
|
|
|
|
}
|
|
|
|
|
|
|
|
// IsPaused returns a boolean indicating if this worker has been paused.
|
|
|
|
func (w *Worker) IsPaused() bool {
|
|
|
|
return w.GetStatus() == WorkerPaused
|
|
|
|
}
|
|
|
|
|
|
|
|
// IsStopped returns a boolean indicating if this worker has been stopped.
|
|
|
|
func (w *Worker) IsStopped() bool {
|
|
|
|
return w.GetStatus() == WorkerStopped
|
|
|
|
}
|
|
|
|
|
|
|
|
func (w *Worker) isPausable() bool {
|
|
|
|
w.statusLock.RLock()
|
|
|
|
defer w.statusLock.RUnlock()
|
|
|
|
switch w.status {
|
|
|
|
case WorkerPausing, WorkerPaused, WorkerStopping, WorkerStopped:
|
|
|
|
return false
|
|
|
|
default:
|
|
|
|
return true
|
|
|
|
}
|
|
|
|
}
|
|
|
|
|
|
|
|
// GetStatus returns the status of the Worker
|
|
|
|
func (w *Worker) GetStatus() WorkerStatus {
|
|
|
|
w.statusLock.RLock()
|
|
|
|
defer w.statusLock.RUnlock()
|
|
|
|
return w.status
|
|
|
|
}
|
|
|
|
|
|
|
|
// setStatuses is used internally to the worker to update the
|
|
|
|
// status of the worker and workload at one time, since some
|
|
|
|
// transitions need to update both values using the same lock.
|
|
|
|
func (w *Worker) setStatuses(newWorkerStatus WorkerStatus, newWorkloadStatus SchedulerWorkerStatus) {
|
|
|
|
w.statusLock.Lock()
|
|
|
|
defer w.statusLock.Unlock()
|
|
|
|
w.setWorkerStatusLocked(newWorkerStatus)
|
|
|
|
w.setWorkloadStatusLocked(newWorkloadStatus)
|
|
|
|
}
|
|
|
|
|
|
|
|
// setStatus is used internally to the worker to update the
|
|
|
|
// status of the worker based on calls to the Worker API. For
|
|
|
|
// atomically updating the scheduler status and the workload
|
|
|
|
// status, use `setStatuses`.
|
|
|
|
func (w *Worker) setStatus(newStatus WorkerStatus) {
|
|
|
|
w.statusLock.Lock()
|
|
|
|
defer w.statusLock.Unlock()
|
|
|
|
w.setWorkerStatusLocked(newStatus)
|
|
|
|
}
|
|
|
|
|
|
|
|
func (w *Worker) setWorkerStatusLocked(newStatus WorkerStatus) {
|
|
|
|
if newStatus == w.status {
|
|
|
|
return
|
|
|
|
}
|
|
|
|
w.logger.Trace("changed worker status", "from", w.status, "to", newStatus)
|
|
|
|
w.status = newStatus
|
|
|
|
}
|
|
|
|
|
|
|
|
// GetStatus returns the status of the Worker's Workload.
|
|
|
|
func (w *Worker) GetWorkloadStatus() SchedulerWorkerStatus {
|
|
|
|
w.statusLock.RLock()
|
|
|
|
defer w.statusLock.RUnlock()
|
|
|
|
return w.workloadStatus
|
|
|
|
}
|
|
|
|
|
|
|
|
// setWorkloadStatus is used internally to the worker to update the
|
|
|
|
// status of the worker based updates from the workload.
|
|
|
|
func (w *Worker) setWorkloadStatus(newStatus SchedulerWorkerStatus) {
|
|
|
|
w.statusLock.Lock()
|
|
|
|
defer w.statusLock.Unlock()
|
|
|
|
w.setWorkloadStatusLocked(newStatus)
|
|
|
|
}
|
|
|
|
|
|
|
|
func (w *Worker) setWorkloadStatusLocked(newStatus SchedulerWorkerStatus) {
|
|
|
|
if newStatus == w.workloadStatus {
|
|
|
|
return
|
|
|
|
}
|
|
|
|
w.logger.Trace("changed workload status", "from", w.workloadStatus, "to", newStatus)
|
|
|
|
w.workloadStatus = newStatus
|
|
|
|
}
|
|
|
|
|
|
|
|
type WorkerInfo struct {
|
|
|
|
ID string `json:"id"`
|
|
|
|
EnabledSchedulers []string `json:"enabled_schedulers"`
|
|
|
|
Started time.Time `json:"started"`
|
|
|
|
Status string `json:"status"`
|
|
|
|
WorkloadStatus string `json:"workload_status"`
|
|
|
|
}
|
|
|
|
|
|
|
|
func (w WorkerInfo) Copy() WorkerInfo {
|
|
|
|
out := WorkerInfo{
|
|
|
|
ID: w.ID,
|
|
|
|
EnabledSchedulers: make([]string, len(w.EnabledSchedulers)),
|
|
|
|
Started: w.Started,
|
|
|
|
Status: w.Status,
|
|
|
|
WorkloadStatus: w.WorkloadStatus,
|
|
|
|
}
|
|
|
|
copy(out.EnabledSchedulers, w.EnabledSchedulers)
|
|
|
|
return out
|
|
|
|
}
|
|
|
|
|
|
|
|
func (w WorkerInfo) String() string {
|
|
|
|
// lazy implementation of WorkerInfo to string
|
|
|
|
out, _ := json.Marshal(w)
|
|
|
|
return string(out)
|
|
|
|
}
|
|
|
|
|
|
|
|
func (w *Worker) Info() WorkerInfo {
|
2015-08-23 17:52:31 +00:00
|
|
|
w.pauseLock.Lock()
|
2022-01-06 16:56:13 +00:00
|
|
|
defer w.pauseLock.Unlock()
|
|
|
|
out := WorkerInfo{
|
|
|
|
ID: w.id,
|
|
|
|
Status: w.status.String(),
|
|
|
|
WorkloadStatus: w.workloadStatus.String(),
|
|
|
|
EnabledSchedulers: make([]string, len(w.enabledSchedulers)),
|
|
|
|
}
|
|
|
|
out.Started = w.start
|
|
|
|
copy(out.EnabledSchedulers, w.enabledSchedulers)
|
|
|
|
return out
|
|
|
|
}
|
|
|
|
|
|
|
|
// ----------------------------------
|
|
|
|
// Pause Implementation
|
|
|
|
// These functions are used to support the worker's pause behaviors.
|
|
|
|
// ----------------------------------
|
|
|
|
|
|
|
|
func (w *Worker) setPauseFlag(pause bool) {
|
|
|
|
w.pauseLock.Lock()
|
|
|
|
defer w.pauseLock.Unlock()
|
|
|
|
w.pauseFlag = pause
|
|
|
|
}
|
|
|
|
|
|
|
|
// maybeWait is responsible for making the transition from `pausing`
|
|
|
|
// to `paused`, waiting, and then transitioning back to the running
|
|
|
|
// values.
|
|
|
|
func (w *Worker) maybeWait() {
|
|
|
|
w.pauseLock.Lock()
|
|
|
|
defer w.pauseLock.Unlock()
|
|
|
|
|
|
|
|
if !w.pauseFlag {
|
|
|
|
return
|
|
|
|
}
|
|
|
|
|
|
|
|
w.statusLock.Lock()
|
|
|
|
w.status = WorkerPaused
|
|
|
|
originalWorkloadStatus := w.workloadStatus
|
|
|
|
w.workloadStatus = WorkloadPaused
|
|
|
|
w.logger.Trace("changed workload status", "from", originalWorkloadStatus, "to", w.workloadStatus)
|
|
|
|
|
|
|
|
w.statusLock.Unlock()
|
|
|
|
|
|
|
|
for w.pauseFlag {
|
2015-08-23 17:52:31 +00:00
|
|
|
w.pauseCond.Wait()
|
|
|
|
}
|
2022-01-06 16:56:13 +00:00
|
|
|
|
|
|
|
w.statusLock.Lock()
|
|
|
|
|
|
|
|
w.logger.Trace("changed workload status", "from", w.workloadStatus, "to", originalWorkloadStatus)
|
|
|
|
w.workloadStatus = originalWorkloadStatus
|
|
|
|
|
|
|
|
// only reset the worker status if the worker is not resuming to stop the paused workload.
|
|
|
|
if w.status != WorkerStopping {
|
|
|
|
w.logger.Trace("changed worker status", "from", w.status, "to", WorkerStarted)
|
|
|
|
w.status = WorkerStarted
|
|
|
|
}
|
|
|
|
w.statusLock.Unlock()
|
|
|
|
}
|
|
|
|
|
|
|
|
// Shutdown is used to signal that the worker should shutdown.
|
|
|
|
func (w *Worker) shutdown() {
|
|
|
|
w.pauseLock.Lock()
|
|
|
|
wasPaused := w.pauseFlag
|
|
|
|
w.pauseFlag = false
|
2015-08-23 17:52:31 +00:00
|
|
|
w.pauseLock.Unlock()
|
2022-01-06 16:56:13 +00:00
|
|
|
|
|
|
|
w.logger.Trace("shutdown request received")
|
|
|
|
w.cancelFn()
|
|
|
|
if wasPaused {
|
|
|
|
w.pauseCond.Broadcast()
|
|
|
|
}
|
|
|
|
}
|
|
|
|
|
|
|
|
// markStopped is used to mark the worker and workload as stopped. It should be called in a
|
|
|
|
// defer immediately upon entering the run() function.
|
|
|
|
func (w *Worker) markStopped() {
|
|
|
|
w.setStatuses(WorkerStopped, WorkloadStopped)
|
|
|
|
w.logger.Debug("stopped")
|
2015-08-23 17:52:31 +00:00
|
|
|
}
|
|
|
|
|
2022-01-06 16:56:13 +00:00
|
|
|
func (w *Worker) workerShuttingDown() bool {
|
|
|
|
select {
|
|
|
|
case <-w.ctx.Done():
|
|
|
|
return true
|
|
|
|
default:
|
|
|
|
return false
|
|
|
|
}
|
|
|
|
}
|
|
|
|
|
|
|
|
// ----------------------------------
|
|
|
|
// Workload behavior code
|
|
|
|
// ----------------------------------
|
|
|
|
|
2015-07-28 22:12:08 +00:00
|
|
|
// run is the long-lived goroutine which is used to run the worker
|
core: backoff considerably when worker is behind raft (#15523)
Upon dequeuing an evaluation workers snapshot their state store at the
eval's wait index or later. This ensures we process an eval at a point
in time after it was created or updated. Processing an eval on an old
snapshot could cause any number of problems such as:
1. Since job registration atomically updates an eval and job in a single
raft entry, scheduling against indexes before that may not have the
eval's job or may have an older version.
2. The older the scheduler's snapshot, the higher the likelihood
something has changed in the cluster state which will cause the plan
applier to reject the scheduler's plan. This could waste work or
even cause eval's to be failed needlessly.
However, the workers run in parallel with a new server pulling the
cluster state from a peer. During this time, which may be many minutes
long, the state store is likely far behind the minimum index required
to process evaluations.
This PR addresses this by adding an additional long backoff period after
an eval is nacked. If the scheduler's indexes catches up within the
additional backoff, it will unblock early to dequeue the next eval.
When the server shuts down we'll get a `context.Canceled` error from the state
store method. We need to bubble this error up so that other callers can detect
it. Handle this case separately when waiting after dequeue so that we can warn
on shutdown instead of throwing an ambiguous error message with just the text
"canceled."
While there may be more precise ways to block scheduling until the
server catches up, this approach adds little risk and covers additional
cases where a server may be temporarily behind due to a spike in load or
a saturated network.
For testing, we make the `raftSyncLimit` into a parameter on the worker's `run` method
so that we can run backoff tests without waiting 30+ seconds. We haven't followed thru
and made all the worker globals into worker parameters, because there isn't much
use outside of testing, but we can consider that in the future.
Co-authored-by: Tim Gross <tgross@hashicorp.com>
2023-01-24 13:56:35 +00:00
|
|
|
func (w *Worker) run(raftSyncLimit time.Duration) {
|
2022-01-06 16:56:13 +00:00
|
|
|
defer func() {
|
|
|
|
w.markStopped()
|
|
|
|
}()
|
|
|
|
w.setStatuses(WorkerStarted, WorkloadRunning)
|
|
|
|
w.logger.Debug("running")
|
2015-07-28 22:12:08 +00:00
|
|
|
for {
|
2022-01-06 16:56:13 +00:00
|
|
|
// Check to see if the context has been cancelled. Server shutdown and Shutdown()
|
|
|
|
// should do this.
|
|
|
|
if w.workerShuttingDown() {
|
|
|
|
return
|
|
|
|
}
|
2015-07-28 22:12:08 +00:00
|
|
|
// Dequeue a pending evaluation
|
2017-09-13 20:47:01 +00:00
|
|
|
eval, token, waitIndex, shutdown := w.dequeueEvaluation(dequeueTimeout)
|
2015-07-28 22:12:08 +00:00
|
|
|
if shutdown {
|
|
|
|
return
|
|
|
|
}
|
|
|
|
|
core: backoff considerably when worker is behind raft (#15523)
Upon dequeuing an evaluation workers snapshot their state store at the
eval's wait index or later. This ensures we process an eval at a point
in time after it was created or updated. Processing an eval on an old
snapshot could cause any number of problems such as:
1. Since job registration atomically updates an eval and job in a single
raft entry, scheduling against indexes before that may not have the
eval's job or may have an older version.
2. The older the scheduler's snapshot, the higher the likelihood
something has changed in the cluster state which will cause the plan
applier to reject the scheduler's plan. This could waste work or
even cause eval's to be failed needlessly.
However, the workers run in parallel with a new server pulling the
cluster state from a peer. During this time, which may be many minutes
long, the state store is likely far behind the minimum index required
to process evaluations.
This PR addresses this by adding an additional long backoff period after
an eval is nacked. If the scheduler's indexes catches up within the
additional backoff, it will unblock early to dequeue the next eval.
When the server shuts down we'll get a `context.Canceled` error from the state
store method. We need to bubble this error up so that other callers can detect
it. Handle this case separately when waiting after dequeue so that we can warn
on shutdown instead of throwing an ambiguous error message with just the text
"canceled."
While there may be more precise ways to block scheduling until the
server catches up, this approach adds little risk and covers additional
cases where a server may be temporarily behind due to a spike in load or
a saturated network.
For testing, we make the `raftSyncLimit` into a parameter on the worker's `run` method
so that we can run backoff tests without waiting 30+ seconds. We haven't followed thru
and made all the worker globals into worker parameters, because there isn't much
use outside of testing, but we can consider that in the future.
Co-authored-by: Tim Gross <tgross@hashicorp.com>
2023-01-24 13:56:35 +00:00
|
|
|
// since dequeue takes time, we could have shutdown the server after
|
|
|
|
// getting an eval that needs to be nacked before we exit. Explicitly
|
|
|
|
// check the server whether to allow this eval to be processed.
|
2015-07-28 22:12:08 +00:00
|
|
|
if w.srv.IsShutdown() {
|
core: backoff considerably when worker is behind raft (#15523)
Upon dequeuing an evaluation workers snapshot their state store at the
eval's wait index or later. This ensures we process an eval at a point
in time after it was created or updated. Processing an eval on an old
snapshot could cause any number of problems such as:
1. Since job registration atomically updates an eval and job in a single
raft entry, scheduling against indexes before that may not have the
eval's job or may have an older version.
2. The older the scheduler's snapshot, the higher the likelihood
something has changed in the cluster state which will cause the plan
applier to reject the scheduler's plan. This could waste work or
even cause eval's to be failed needlessly.
However, the workers run in parallel with a new server pulling the
cluster state from a peer. During this time, which may be many minutes
long, the state store is likely far behind the minimum index required
to process evaluations.
This PR addresses this by adding an additional long backoff period after
an eval is nacked. If the scheduler's indexes catches up within the
additional backoff, it will unblock early to dequeue the next eval.
When the server shuts down we'll get a `context.Canceled` error from the state
store method. We need to bubble this error up so that other callers can detect
it. Handle this case separately when waiting after dequeue so that we can warn
on shutdown instead of throwing an ambiguous error message with just the text
"canceled."
While there may be more precise ways to block scheduling until the
server catches up, this approach adds little risk and covers additional
cases where a server may be temporarily behind due to a spike in load or
a saturated network.
For testing, we make the `raftSyncLimit` into a parameter on the worker's `run` method
so that we can run backoff tests without waiting 30+ seconds. We haven't followed thru
and made all the worker globals into worker parameters, because there isn't much
use outside of testing, but we can consider that in the future.
Co-authored-by: Tim Gross <tgross@hashicorp.com>
2023-01-24 13:56:35 +00:00
|
|
|
w.logger.Warn("nacking eval because the server is shutting down",
|
|
|
|
"eval", log.Fmt("%#v", eval))
|
2022-01-06 16:56:13 +00:00
|
|
|
w.sendNack(eval, token)
|
2015-07-28 22:12:08 +00:00
|
|
|
return
|
|
|
|
}
|
|
|
|
|
2016-10-11 19:31:40 +00:00
|
|
|
// Wait for the raft log to catchup to the evaluation
|
2022-01-06 16:56:13 +00:00
|
|
|
w.setWorkloadStatus(WorkloadWaitingForRaft)
|
2019-06-24 18:59:44 +00:00
|
|
|
snap, err := w.snapshotMinIndex(waitIndex, raftSyncLimit)
|
2019-03-12 21:25:14 +00:00
|
|
|
if err != nil {
|
core: backoff considerably when worker is behind raft (#15523)
Upon dequeuing an evaluation workers snapshot their state store at the
eval's wait index or later. This ensures we process an eval at a point
in time after it was created or updated. Processing an eval on an old
snapshot could cause any number of problems such as:
1. Since job registration atomically updates an eval and job in a single
raft entry, scheduling against indexes before that may not have the
eval's job or may have an older version.
2. The older the scheduler's snapshot, the higher the likelihood
something has changed in the cluster state which will cause the plan
applier to reject the scheduler's plan. This could waste work or
even cause eval's to be failed needlessly.
However, the workers run in parallel with a new server pulling the
cluster state from a peer. During this time, which may be many minutes
long, the state store is likely far behind the minimum index required
to process evaluations.
This PR addresses this by adding an additional long backoff period after
an eval is nacked. If the scheduler's indexes catches up within the
additional backoff, it will unblock early to dequeue the next eval.
When the server shuts down we'll get a `context.Canceled` error from the state
store method. We need to bubble this error up so that other callers can detect
it. Handle this case separately when waiting after dequeue so that we can warn
on shutdown instead of throwing an ambiguous error message with just the text
"canceled."
While there may be more precise ways to block scheduling until the
server catches up, this approach adds little risk and covers additional
cases where a server may be temporarily behind due to a spike in load or
a saturated network.
For testing, we make the `raftSyncLimit` into a parameter on the worker's `run` method
so that we can run backoff tests without waiting 30+ seconds. We haven't followed thru
and made all the worker globals into worker parameters, because there isn't much
use outside of testing, but we can consider that in the future.
Co-authored-by: Tim Gross <tgross@hashicorp.com>
2023-01-24 13:56:35 +00:00
|
|
|
var timeoutErr ErrMinIndexDeadlineExceeded
|
|
|
|
if errors.As(err, &timeoutErr) {
|
|
|
|
w.logger.Warn("timeout waiting for Raft index required by eval",
|
|
|
|
"eval", eval.ID, "index", waitIndex, "timeout", raftSyncLimit)
|
|
|
|
w.sendNack(eval, token)
|
|
|
|
|
|
|
|
// Timing out above means this server is woefully behind the
|
|
|
|
// leader's index. This can happen when a new server is added to
|
|
|
|
// a cluster and must initially sync the cluster state.
|
|
|
|
// Backoff dequeuing another eval until there's some indication
|
|
|
|
// this server would be up to date enough to process it.
|
|
|
|
slowServerSyncLimit := 10 * raftSyncLimit
|
|
|
|
if _, err := w.snapshotMinIndex(waitIndex, slowServerSyncLimit); err != nil {
|
|
|
|
w.logger.Warn("server is unable to catch up to last eval's index", "error", err)
|
|
|
|
}
|
|
|
|
|
|
|
|
} else if errors.Is(err, context.Canceled) {
|
|
|
|
// If the server has shutdown while we're waiting, we'll get the
|
|
|
|
// Canceled error from the worker's context. We need to nack any
|
|
|
|
// dequeued evals before we exit.
|
|
|
|
w.logger.Warn("nacking eval because the server is shutting down", "eval", eval.ID)
|
|
|
|
w.sendNack(eval, token)
|
|
|
|
return
|
|
|
|
} else {
|
|
|
|
w.logger.Error("error waiting for Raft index", "error", err, "index", waitIndex)
|
|
|
|
w.sendNack(eval, token)
|
|
|
|
}
|
|
|
|
|
2015-07-28 22:12:08 +00:00
|
|
|
continue
|
|
|
|
}
|
|
|
|
|
|
|
|
// Invoke the scheduler to determine placements
|
2022-01-06 16:56:13 +00:00
|
|
|
w.setWorkloadStatus(WorkloadScheduling)
|
2019-03-12 21:25:14 +00:00
|
|
|
if err := w.invokeScheduler(snap, eval, token); err != nil {
|
2018-09-25 17:47:45 +00:00
|
|
|
w.logger.Error("error invoking scheduler", "error", err)
|
2022-01-06 16:56:13 +00:00
|
|
|
w.sendNack(eval, token)
|
2015-07-28 22:12:08 +00:00
|
|
|
continue
|
|
|
|
}
|
|
|
|
|
|
|
|
// Complete the evaluation
|
2022-01-06 16:56:13 +00:00
|
|
|
w.sendAck(eval, token)
|
2015-07-28 22:12:08 +00:00
|
|
|
}
|
|
|
|
}
|
|
|
|
|
|
|
|
// dequeueEvaluation is used to fetch the next ready evaluation.
|
|
|
|
// This blocks until an evaluation is available or a timeout is reached.
|
2017-09-13 20:47:01 +00:00
|
|
|
func (w *Worker) dequeueEvaluation(timeout time.Duration) (
|
|
|
|
eval *structs.Evaluation, token string, waitIndex uint64, shutdown bool) {
|
2015-07-28 22:12:08 +00:00
|
|
|
// Setup the request
|
|
|
|
req := structs.EvalDequeueRequest{
|
2022-01-06 16:56:13 +00:00
|
|
|
Schedulers: w.enabledSchedulers,
|
2016-10-26 21:52:48 +00:00
|
|
|
Timeout: timeout,
|
|
|
|
SchedulerVersion: scheduler.SchedulerVersion,
|
2015-07-28 22:12:08 +00:00
|
|
|
WriteRequest: structs.WriteRequest{
|
|
|
|
Region: w.srv.config.Region,
|
|
|
|
},
|
|
|
|
}
|
2015-08-12 22:25:31 +00:00
|
|
|
var resp structs.EvalDequeueResponse
|
2015-07-28 22:12:08 +00:00
|
|
|
|
|
|
|
REQ:
|
2022-01-06 16:56:13 +00:00
|
|
|
// Wait inside this function if the worker is paused.
|
|
|
|
w.maybeWait()
|
|
|
|
// Immediately check to see if the worker has been shutdown.
|
|
|
|
if w.workerShuttingDown() {
|
|
|
|
return nil, "", 0, true
|
|
|
|
}
|
2015-08-23 17:52:31 +00:00
|
|
|
|
2015-07-28 22:12:08 +00:00
|
|
|
// Make a blocking RPC
|
2015-07-28 23:36:15 +00:00
|
|
|
start := time.Now()
|
2022-01-06 16:56:13 +00:00
|
|
|
w.setWorkloadStatus(WorkloadWaitingToDequeue)
|
2015-07-28 22:12:08 +00:00
|
|
|
err := w.srv.RPC("Eval.Dequeue", &req, &resp)
|
2015-07-28 23:36:15 +00:00
|
|
|
metrics.MeasureSince([]string{"nomad", "worker", "dequeue_eval"}, start)
|
2015-07-28 22:12:08 +00:00
|
|
|
if err != nil {
|
2022-01-06 16:56:13 +00:00
|
|
|
if time.Since(w.start) > dequeueErrGrace && !w.workerShuttingDown() {
|
2018-09-15 23:23:13 +00:00
|
|
|
w.logger.Error("failed to dequeue evaluation", "error", err)
|
2015-09-07 03:18:47 +00:00
|
|
|
}
|
2016-10-27 18:41:43 +00:00
|
|
|
|
|
|
|
// Adjust the backoff based on the error. If it is a scheduler version
|
|
|
|
// mismatch we increase the baseline.
|
|
|
|
base, limit := backoffBaselineFast, backoffLimitSlow
|
|
|
|
if strings.Contains(err.Error(), "calling scheduler version") {
|
|
|
|
base = backoffSchedulerVersionMismatch
|
|
|
|
limit = backoffSchedulerVersionMismatch
|
|
|
|
}
|
|
|
|
|
|
|
|
if w.backoffErr(base, limit) {
|
2017-09-13 20:47:01 +00:00
|
|
|
return nil, "", 0, true
|
2015-07-28 22:12:08 +00:00
|
|
|
}
|
|
|
|
goto REQ
|
|
|
|
}
|
|
|
|
w.backoffReset()
|
|
|
|
|
|
|
|
// Check if we got a response
|
|
|
|
if resp.Eval != nil {
|
2022-01-06 16:56:13 +00:00
|
|
|
w.logger.Debug("dequeued evaluation", "eval_id", resp.Eval.ID, "type", resp.Eval.Type, "namespace", resp.Eval.Namespace, "job_id", resp.Eval.JobID, "node_id", resp.Eval.NodeID, "triggered_by", resp.Eval.TriggeredBy)
|
2017-09-13 20:47:01 +00:00
|
|
|
return resp.Eval, resp.Token, resp.GetWaitIndex(), false
|
2015-07-28 22:12:08 +00:00
|
|
|
}
|
|
|
|
|
|
|
|
goto REQ
|
|
|
|
}
|
|
|
|
|
2021-11-17 15:49:55 +00:00
|
|
|
// sendAcknowledgement should not be called directly. Call `sendAck` or `sendNack` instead.
|
|
|
|
// This function implements `ack`ing or `nack`ing the evaluation generally.
|
2015-07-28 22:12:08 +00:00
|
|
|
// Any errors are logged but swallowed.
|
2022-01-06 16:56:13 +00:00
|
|
|
func (w *Worker) sendAcknowledgement(eval *structs.Evaluation, token string, ack bool) {
|
2015-07-28 23:36:15 +00:00
|
|
|
defer metrics.MeasureSince([]string{"nomad", "worker", "send_ack"}, time.Now())
|
2015-07-28 22:12:08 +00:00
|
|
|
// Setup the request
|
2015-08-12 22:25:31 +00:00
|
|
|
req := structs.EvalAckRequest{
|
2022-01-06 16:56:13 +00:00
|
|
|
EvalID: eval.ID,
|
2015-08-12 22:25:31 +00:00
|
|
|
Token: token,
|
2015-07-28 22:12:08 +00:00
|
|
|
WriteRequest: structs.WriteRequest{
|
|
|
|
Region: w.srv.config.Region,
|
|
|
|
},
|
|
|
|
}
|
|
|
|
var resp structs.GenericResponse
|
|
|
|
|
|
|
|
// Determine if this is an Ack or Nack
|
|
|
|
verb := "ack"
|
|
|
|
endpoint := "Eval.Ack"
|
|
|
|
if !ack {
|
|
|
|
verb = "nack"
|
|
|
|
endpoint = "Eval.Nack"
|
|
|
|
}
|
|
|
|
|
|
|
|
// Make the RPC call
|
|
|
|
err := w.srv.RPC(endpoint, &req, &resp)
|
|
|
|
if err != nil {
|
2022-01-06 16:56:13 +00:00
|
|
|
w.logger.Error(fmt.Sprintf("failed to %s evaluation", verb), "eval_id", eval.ID, "error", err)
|
2015-07-28 23:36:15 +00:00
|
|
|
} else {
|
2022-01-06 16:56:13 +00:00
|
|
|
w.logger.Debug(fmt.Sprintf("%s evaluation", verb), "eval_id", eval.ID, "type", eval.Type, "namespace", eval.Namespace, "job_id", eval.JobID, "node_id", eval.NodeID, "triggered_by", eval.TriggeredBy)
|
2015-07-28 22:12:08 +00:00
|
|
|
}
|
|
|
|
}
|
|
|
|
|
2021-11-17 15:49:55 +00:00
|
|
|
// sendNack makes a best effort to nack the evaluation.
|
|
|
|
// Any errors are logged but swallowed.
|
2022-01-06 16:56:13 +00:00
|
|
|
func (w *Worker) sendNack(eval *structs.Evaluation, token string) {
|
|
|
|
w.sendAcknowledgement(eval, token, false)
|
2021-11-17 15:49:55 +00:00
|
|
|
}
|
|
|
|
|
|
|
|
// sendAck makes a best effort to ack the evaluation.
|
|
|
|
// Any errors are logged but swallowed.
|
2022-01-06 16:56:13 +00:00
|
|
|
func (w *Worker) sendAck(eval *structs.Evaluation, token string) {
|
|
|
|
w.sendAcknowledgement(eval, token, true)
|
2021-11-17 15:49:55 +00:00
|
|
|
}
|
|
|
|
|
core: backoff considerably when worker is behind raft (#15523)
Upon dequeuing an evaluation workers snapshot their state store at the
eval's wait index or later. This ensures we process an eval at a point
in time after it was created or updated. Processing an eval on an old
snapshot could cause any number of problems such as:
1. Since job registration atomically updates an eval and job in a single
raft entry, scheduling against indexes before that may not have the
eval's job or may have an older version.
2. The older the scheduler's snapshot, the higher the likelihood
something has changed in the cluster state which will cause the plan
applier to reject the scheduler's plan. This could waste work or
even cause eval's to be failed needlessly.
However, the workers run in parallel with a new server pulling the
cluster state from a peer. During this time, which may be many minutes
long, the state store is likely far behind the minimum index required
to process evaluations.
This PR addresses this by adding an additional long backoff period after
an eval is nacked. If the scheduler's indexes catches up within the
additional backoff, it will unblock early to dequeue the next eval.
When the server shuts down we'll get a `context.Canceled` error from the state
store method. We need to bubble this error up so that other callers can detect
it. Handle this case separately when waiting after dequeue so that we can warn
on shutdown instead of throwing an ambiguous error message with just the text
"canceled."
While there may be more precise ways to block scheduling until the
server catches up, this approach adds little risk and covers additional
cases where a server may be temporarily behind due to a spike in load or
a saturated network.
For testing, we make the `raftSyncLimit` into a parameter on the worker's `run` method
so that we can run backoff tests without waiting 30+ seconds. We haven't followed thru
and made all the worker globals into worker parameters, because there isn't much
use outside of testing, but we can consider that in the future.
Co-authored-by: Tim Gross <tgross@hashicorp.com>
2023-01-24 13:56:35 +00:00
|
|
|
type ErrMinIndexDeadlineExceeded struct {
|
|
|
|
waitIndex uint64
|
|
|
|
timeout time.Duration
|
|
|
|
}
|
|
|
|
|
|
|
|
// Unwrapping an ErrMinIndexDeadlineExceeded always return
|
|
|
|
// context.DeadlineExceeded
|
|
|
|
func (ErrMinIndexDeadlineExceeded) Unwrap() error {
|
|
|
|
return context.DeadlineExceeded
|
|
|
|
}
|
|
|
|
|
|
|
|
func (e ErrMinIndexDeadlineExceeded) Error() string {
|
|
|
|
return fmt.Sprintf("timed out after %s waiting for index=%d", e.timeout, e.waitIndex)
|
|
|
|
}
|
|
|
|
|
2019-06-24 18:59:44 +00:00
|
|
|
// snapshotMinIndex times calls to StateStore.SnapshotAfter which may block.
|
|
|
|
func (w *Worker) snapshotMinIndex(waitIndex uint64, timeout time.Duration) (*state.StateSnapshot, error) {
|
core: backoff considerably when worker is behind raft (#15523)
Upon dequeuing an evaluation workers snapshot their state store at the
eval's wait index or later. This ensures we process an eval at a point
in time after it was created or updated. Processing an eval on an old
snapshot could cause any number of problems such as:
1. Since job registration atomically updates an eval and job in a single
raft entry, scheduling against indexes before that may not have the
eval's job or may have an older version.
2. The older the scheduler's snapshot, the higher the likelihood
something has changed in the cluster state which will cause the plan
applier to reject the scheduler's plan. This could waste work or
even cause eval's to be failed needlessly.
However, the workers run in parallel with a new server pulling the
cluster state from a peer. During this time, which may be many minutes
long, the state store is likely far behind the minimum index required
to process evaluations.
This PR addresses this by adding an additional long backoff period after
an eval is nacked. If the scheduler's indexes catches up within the
additional backoff, it will unblock early to dequeue the next eval.
When the server shuts down we'll get a `context.Canceled` error from the state
store method. We need to bubble this error up so that other callers can detect
it. Handle this case separately when waiting after dequeue so that we can warn
on shutdown instead of throwing an ambiguous error message with just the text
"canceled."
While there may be more precise ways to block scheduling until the
server catches up, this approach adds little risk and covers additional
cases where a server may be temporarily behind due to a spike in load or
a saturated network.
For testing, we make the `raftSyncLimit` into a parameter on the worker's `run` method
so that we can run backoff tests without waiting 30+ seconds. We haven't followed thru
and made all the worker globals into worker parameters, because there isn't much
use outside of testing, but we can consider that in the future.
Co-authored-by: Tim Gross <tgross@hashicorp.com>
2023-01-24 13:56:35 +00:00
|
|
|
defer metrics.MeasureSince([]string{"nomad", "worker", "wait_for_index"}, time.Now())
|
|
|
|
|
2022-01-06 16:56:13 +00:00
|
|
|
ctx, cancel := context.WithTimeout(w.ctx, timeout)
|
2019-06-24 18:59:44 +00:00
|
|
|
snap, err := w.srv.fsm.State().SnapshotMinIndex(ctx, waitIndex)
|
2019-03-12 21:25:14 +00:00
|
|
|
cancel()
|
2019-05-17 21:37:42 +00:00
|
|
|
|
core: backoff considerably when worker is behind raft (#15523)
Upon dequeuing an evaluation workers snapshot their state store at the
eval's wait index or later. This ensures we process an eval at a point
in time after it was created or updated. Processing an eval on an old
snapshot could cause any number of problems such as:
1. Since job registration atomically updates an eval and job in a single
raft entry, scheduling against indexes before that may not have the
eval's job or may have an older version.
2. The older the scheduler's snapshot, the higher the likelihood
something has changed in the cluster state which will cause the plan
applier to reject the scheduler's plan. This could waste work or
even cause eval's to be failed needlessly.
However, the workers run in parallel with a new server pulling the
cluster state from a peer. During this time, which may be many minutes
long, the state store is likely far behind the minimum index required
to process evaluations.
This PR addresses this by adding an additional long backoff period after
an eval is nacked. If the scheduler's indexes catches up within the
additional backoff, it will unblock early to dequeue the next eval.
When the server shuts down we'll get a `context.Canceled` error from the state
store method. We need to bubble this error up so that other callers can detect
it. Handle this case separately when waiting after dequeue so that we can warn
on shutdown instead of throwing an ambiguous error message with just the text
"canceled."
While there may be more precise ways to block scheduling until the
server catches up, this approach adds little risk and covers additional
cases where a server may be temporarily behind due to a spike in load or
a saturated network.
For testing, we make the `raftSyncLimit` into a parameter on the worker's `run` method
so that we can run backoff tests without waiting 30+ seconds. We haven't followed thru
and made all the worker globals into worker parameters, because there isn't much
use outside of testing, but we can consider that in the future.
Co-authored-by: Tim Gross <tgross@hashicorp.com>
2023-01-24 13:56:35 +00:00
|
|
|
// Wrap error to ensure callers can detect timeouts.
|
|
|
|
if errors.Is(err, context.DeadlineExceeded) {
|
|
|
|
return nil, ErrMinIndexDeadlineExceeded{
|
|
|
|
waitIndex: waitIndex,
|
|
|
|
timeout: timeout,
|
|
|
|
}
|
2019-05-17 21:37:42 +00:00
|
|
|
}
|
|
|
|
|
2019-03-12 21:25:14 +00:00
|
|
|
return snap, err
|
2015-07-28 22:12:08 +00:00
|
|
|
}
|
|
|
|
|
|
|
|
// invokeScheduler is used to invoke the business logic of the scheduler
|
2019-03-12 21:25:14 +00:00
|
|
|
func (w *Worker) invokeScheduler(snap *state.StateSnapshot, eval *structs.Evaluation, token string) error {
|
2015-08-07 00:45:27 +00:00
|
|
|
defer metrics.MeasureSince([]string{"nomad", "worker", "invoke_scheduler", eval.Type}, time.Now())
|
2015-08-12 22:44:36 +00:00
|
|
|
// Store the evaluation token
|
|
|
|
w.evalToken = token
|
|
|
|
|
2016-05-21 01:07:10 +00:00
|
|
|
// Store the snapshot's index
|
2019-03-12 21:25:14 +00:00
|
|
|
var err error
|
2016-05-21 01:07:10 +00:00
|
|
|
w.snapshotIndex, err = snap.LatestIndex()
|
|
|
|
if err != nil {
|
|
|
|
return fmt.Errorf("failed to determine snapshot's index: %v", err)
|
|
|
|
}
|
|
|
|
|
2019-07-17 19:37:25 +00:00
|
|
|
// Create the scheduler, or use the special core scheduler
|
2015-08-07 00:04:35 +00:00
|
|
|
var sched scheduler.Scheduler
|
2015-08-15 19:38:58 +00:00
|
|
|
if eval.Type == structs.JobTypeCore {
|
|
|
|
sched = NewCoreScheduler(w.srv, snap)
|
2015-08-07 00:04:35 +00:00
|
|
|
} else {
|
2022-01-15 01:09:14 +00:00
|
|
|
sched, err = scheduler.NewScheduler(eval.Type, w.logger, w.srv.workersEventCh, snap, w)
|
2015-08-07 00:04:35 +00:00
|
|
|
if err != nil {
|
|
|
|
return fmt.Errorf("failed to instantiate scheduler: %v", err)
|
|
|
|
}
|
2015-07-28 23:15:32 +00:00
|
|
|
}
|
|
|
|
|
|
|
|
// Process the evaluation
|
|
|
|
err = sched.Process(eval)
|
|
|
|
if err != nil {
|
|
|
|
return fmt.Errorf("failed to process evaluation: %v", err)
|
|
|
|
}
|
2015-07-28 22:12:08 +00:00
|
|
|
return nil
|
|
|
|
}
|
|
|
|
|
2022-03-07 18:40:57 +00:00
|
|
|
// ServersMeetMinimumVersion allows implementations of the Scheduler interface in
|
|
|
|
// other packages to perform server version checks without direct references to
|
|
|
|
// the Nomad server.
|
|
|
|
func (w *Worker) ServersMeetMinimumVersion(minVersion *version.Version, checkFailedServers bool) bool {
|
2022-10-17 20:23:51 +00:00
|
|
|
return ServersMeetMinimumVersion(w.srv.Members(), w.srv.Region(), minVersion, checkFailedServers)
|
2022-03-07 18:40:57 +00:00
|
|
|
}
|
|
|
|
|
2015-07-28 23:15:32 +00:00
|
|
|
// SubmitPlan is used to submit a plan for consideration. This allows
|
|
|
|
// the worker to act as the planner for the scheduler.
|
2015-07-28 23:36:15 +00:00
|
|
|
func (w *Worker) SubmitPlan(plan *structs.Plan) (*structs.PlanResult, scheduler.State, error) {
|
2022-01-06 16:56:13 +00:00
|
|
|
// Check for a shutdown before plan submission. Checking server state rather than
|
|
|
|
// worker state to allow work in flight to complete before stopping.
|
2015-07-28 23:52:13 +00:00
|
|
|
if w.srv.IsShutdown() {
|
|
|
|
return nil, nil, fmt.Errorf("shutdown while planning")
|
|
|
|
}
|
2015-07-28 23:36:15 +00:00
|
|
|
defer metrics.MeasureSince([]string{"nomad", "worker", "submit_plan"}, time.Now())
|
2015-07-28 23:52:13 +00:00
|
|
|
|
2015-08-12 22:44:36 +00:00
|
|
|
// Add the evaluation token to the plan
|
|
|
|
plan.EvalToken = w.evalToken
|
|
|
|
|
2019-06-06 22:44:47 +00:00
|
|
|
// Add SnapshotIndex to ensure leader's StateStore processes the Plan
|
|
|
|
// at or after the index it was created.
|
|
|
|
plan.SnapshotIndex = w.snapshotIndex
|
|
|
|
|
2019-03-04 09:49:32 +00:00
|
|
|
// Normalize stopped and preempted allocs before RPC
|
2022-10-17 20:23:51 +00:00
|
|
|
normalizePlan := ServersMeetMinimumVersion(w.srv.Members(), w.srv.Region(), MinVersionPlanNormalization, true)
|
2019-03-08 12:48:12 +00:00
|
|
|
if normalizePlan {
|
|
|
|
plan.NormalizeAllocations()
|
|
|
|
}
|
2019-03-04 09:49:32 +00:00
|
|
|
|
2015-07-28 23:36:15 +00:00
|
|
|
// Setup the request
|
|
|
|
req := structs.PlanRequest{
|
|
|
|
Plan: plan,
|
|
|
|
WriteRequest: structs.WriteRequest{
|
|
|
|
Region: w.srv.config.Region,
|
|
|
|
},
|
|
|
|
}
|
|
|
|
var resp structs.PlanResponse
|
|
|
|
|
2015-08-05 23:23:37 +00:00
|
|
|
SUBMIT:
|
2015-07-28 23:36:15 +00:00
|
|
|
// Make the RPC call
|
|
|
|
if err := w.srv.RPC("Plan.Submit", &req, &resp); err != nil {
|
2018-09-15 23:23:13 +00:00
|
|
|
w.logger.Error("failed to submit plan for evaluation", "eval_id", plan.EvalID, "error", err)
|
2015-08-24 00:39:49 +00:00
|
|
|
if w.shouldResubmit(err) && !w.backoffErr(backoffBaselineSlow, backoffLimitSlow) {
|
2015-08-05 23:23:37 +00:00
|
|
|
goto SUBMIT
|
|
|
|
}
|
2015-07-28 23:36:15 +00:00
|
|
|
return nil, nil, err
|
|
|
|
} else {
|
2018-09-21 23:55:35 +00:00
|
|
|
w.logger.Debug("submitted plan for evaluation", "eval_id", plan.EvalID)
|
2015-08-05 23:23:37 +00:00
|
|
|
w.backoffReset()
|
2015-07-28 23:36:15 +00:00
|
|
|
}
|
|
|
|
|
|
|
|
// Look for a result
|
|
|
|
result := resp.Result
|
|
|
|
if result == nil {
|
|
|
|
return nil, nil, fmt.Errorf("missing result")
|
|
|
|
}
|
|
|
|
|
|
|
|
// Check if a state update is required. This could be required if we
|
2019-06-06 22:44:47 +00:00
|
|
|
// planned based on stale data, which is causing issues. For example, a
|
2015-07-28 23:36:15 +00:00
|
|
|
// node failure since the time we've started planning or conflicting task
|
|
|
|
// allocations.
|
|
|
|
var state scheduler.State
|
|
|
|
if result.RefreshIndex != 0 {
|
2016-10-11 19:31:40 +00:00
|
|
|
// Wait for the raft log to catchup to the evaluation
|
2018-09-15 23:23:13 +00:00
|
|
|
w.logger.Debug("refreshing state", "refresh_index", result.RefreshIndex, "eval_id", plan.EvalID)
|
2015-07-28 23:36:15 +00:00
|
|
|
|
2019-03-12 21:25:14 +00:00
|
|
|
var err error
|
2019-06-24 18:59:44 +00:00
|
|
|
state, err = w.snapshotMinIndex(result.RefreshIndex, raftSyncLimit)
|
2015-07-28 23:36:15 +00:00
|
|
|
if err != nil {
|
2019-03-12 21:25:14 +00:00
|
|
|
return nil, nil, err
|
2015-07-28 23:36:15 +00:00
|
|
|
}
|
|
|
|
}
|
|
|
|
|
|
|
|
// Return the result and potential state update
|
|
|
|
return result, state, nil
|
2015-07-28 23:15:32 +00:00
|
|
|
}
|
|
|
|
|
2015-08-15 21:25:00 +00:00
|
|
|
// UpdateEval is used to submit an updated evaluation. This allows
|
|
|
|
// the worker to act as the planner for the scheduler.
|
|
|
|
func (w *Worker) UpdateEval(eval *structs.Evaluation) error {
|
2022-01-06 16:56:13 +00:00
|
|
|
// Check for a shutdown before plan submission. Checking server state rather than
|
|
|
|
// worker state to allow a workers work in flight to complete before stopping.
|
2015-08-15 21:25:00 +00:00
|
|
|
if w.srv.IsShutdown() {
|
|
|
|
return fmt.Errorf("shutdown while planning")
|
|
|
|
}
|
|
|
|
defer metrics.MeasureSince([]string{"nomad", "worker", "update_eval"}, time.Now())
|
|
|
|
|
2016-05-21 01:07:10 +00:00
|
|
|
// Store the snapshot index in the eval
|
|
|
|
eval.SnapshotIndex = w.snapshotIndex
|
2019-08-07 16:50:35 +00:00
|
|
|
eval.UpdateModifyTime()
|
2016-05-21 01:07:10 +00:00
|
|
|
|
2015-08-15 21:25:00 +00:00
|
|
|
// Setup the request
|
|
|
|
req := structs.EvalUpdateRequest{
|
|
|
|
Evals: []*structs.Evaluation{eval},
|
|
|
|
EvalToken: w.evalToken,
|
|
|
|
WriteRequest: structs.WriteRequest{
|
|
|
|
Region: w.srv.config.Region,
|
|
|
|
},
|
|
|
|
}
|
|
|
|
var resp structs.GenericResponse
|
|
|
|
|
|
|
|
SUBMIT:
|
|
|
|
// Make the RPC call
|
|
|
|
if err := w.srv.RPC("Eval.Update", &req, &resp); err != nil {
|
2018-09-15 23:23:13 +00:00
|
|
|
w.logger.Error("failed to update evaluation", "eval", log.Fmt("%#v", eval), "error", err)
|
2015-08-24 00:39:49 +00:00
|
|
|
if w.shouldResubmit(err) && !w.backoffErr(backoffBaselineSlow, backoffLimitSlow) {
|
2015-08-15 21:25:00 +00:00
|
|
|
goto SUBMIT
|
|
|
|
}
|
|
|
|
return err
|
|
|
|
} else {
|
2018-09-15 23:23:13 +00:00
|
|
|
w.logger.Debug("updated evaluation", "eval", log.Fmt("%#v", eval))
|
2015-08-15 21:25:00 +00:00
|
|
|
w.backoffReset()
|
|
|
|
}
|
|
|
|
return nil
|
|
|
|
}
|
|
|
|
|
2015-09-07 21:23:48 +00:00
|
|
|
// CreateEval is used to create a new evaluation. This allows
|
|
|
|
// the worker to act as the planner for the scheduler.
|
|
|
|
func (w *Worker) CreateEval(eval *structs.Evaluation) error {
|
2022-01-06 16:56:13 +00:00
|
|
|
// Check for a shutdown before plan submission. This consults the server Shutdown state
|
|
|
|
// instead of the worker's to prevent aborting work in flight.
|
2015-09-07 21:23:48 +00:00
|
|
|
if w.srv.IsShutdown() {
|
|
|
|
return fmt.Errorf("shutdown while planning")
|
|
|
|
}
|
|
|
|
defer metrics.MeasureSince([]string{"nomad", "worker", "create_eval"}, time.Now())
|
|
|
|
|
2016-05-21 01:07:10 +00:00
|
|
|
// Store the snapshot index in the eval
|
|
|
|
eval.SnapshotIndex = w.snapshotIndex
|
|
|
|
|
2019-08-07 16:50:35 +00:00
|
|
|
now := time.Now().UTC().UnixNano()
|
|
|
|
eval.CreateTime = now
|
|
|
|
eval.ModifyTime = now
|
|
|
|
|
2015-09-07 21:23:48 +00:00
|
|
|
// Setup the request
|
|
|
|
req := structs.EvalUpdateRequest{
|
|
|
|
Evals: []*structs.Evaluation{eval},
|
|
|
|
EvalToken: w.evalToken,
|
|
|
|
WriteRequest: structs.WriteRequest{
|
|
|
|
Region: w.srv.config.Region,
|
|
|
|
},
|
|
|
|
}
|
|
|
|
var resp structs.GenericResponse
|
|
|
|
|
|
|
|
SUBMIT:
|
|
|
|
// Make the RPC call
|
|
|
|
if err := w.srv.RPC("Eval.Create", &req, &resp); err != nil {
|
2018-09-15 23:23:13 +00:00
|
|
|
w.logger.Error("failed to create evaluation", "eval", log.Fmt("%#v", eval), "error", err)
|
2015-09-07 21:23:48 +00:00
|
|
|
if w.shouldResubmit(err) && !w.backoffErr(backoffBaselineSlow, backoffLimitSlow) {
|
|
|
|
goto SUBMIT
|
|
|
|
}
|
|
|
|
return err
|
|
|
|
} else {
|
2023-10-27 15:20:53 +00:00
|
|
|
w.logger.Debug("created evaluation", "eval", log.Fmt("%#v", eval), "waitUntil", log.Fmt("%#v", eval.WaitUntil.String()))
|
2015-09-07 21:23:48 +00:00
|
|
|
w.backoffReset()
|
2016-05-20 23:03:53 +00:00
|
|
|
}
|
|
|
|
return nil
|
|
|
|
}
|
|
|
|
|
|
|
|
// ReblockEval is used to reinsert a blocked evaluation into the blocked eval
|
|
|
|
// tracker. This allows the worker to act as the planner for the scheduler.
|
|
|
|
func (w *Worker) ReblockEval(eval *structs.Evaluation) error {
|
2022-01-06 16:56:13 +00:00
|
|
|
// Check for a shutdown before plan submission. This checks the server state rather than
|
|
|
|
// the worker's to prevent erroring on work in flight that would complete otherwise.
|
2016-05-20 23:03:53 +00:00
|
|
|
if w.srv.IsShutdown() {
|
|
|
|
return fmt.Errorf("shutdown while planning")
|
|
|
|
}
|
|
|
|
defer metrics.MeasureSince([]string{"nomad", "worker", "reblock_eval"}, time.Now())
|
|
|
|
|
2016-07-20 21:09:03 +00:00
|
|
|
// Update the evaluation if the queued jobs is not same as what is
|
|
|
|
// recorded in the job summary
|
2017-02-08 04:31:23 +00:00
|
|
|
ws := memdb.NewWatchSet()
|
2017-09-07 23:56:15 +00:00
|
|
|
summary, err := w.srv.fsm.state.JobSummaryByID(ws, eval.Namespace, eval.JobID)
|
2016-07-20 21:09:03 +00:00
|
|
|
if err != nil {
|
2017-08-07 21:13:05 +00:00
|
|
|
return fmt.Errorf("couldn't retrieve job summary: %v", err)
|
2016-07-20 21:09:03 +00:00
|
|
|
}
|
|
|
|
if summary != nil {
|
|
|
|
var hasChanged bool
|
|
|
|
for tg, summary := range summary.Summary {
|
|
|
|
if queued, ok := eval.QueuedAllocations[tg]; ok {
|
2016-07-22 06:13:07 +00:00
|
|
|
if queued != summary.Queued {
|
|
|
|
hasChanged = true
|
|
|
|
break
|
|
|
|
}
|
2016-07-20 21:09:03 +00:00
|
|
|
}
|
|
|
|
}
|
|
|
|
if hasChanged {
|
|
|
|
if err := w.UpdateEval(eval); err != nil {
|
|
|
|
return err
|
|
|
|
}
|
|
|
|
}
|
|
|
|
}
|
|
|
|
|
2016-05-21 01:07:10 +00:00
|
|
|
// Store the snapshot index in the eval
|
|
|
|
eval.SnapshotIndex = w.snapshotIndex
|
2019-08-07 16:50:35 +00:00
|
|
|
eval.UpdateModifyTime()
|
2016-05-20 23:03:53 +00:00
|
|
|
|
|
|
|
// Setup the request
|
|
|
|
req := structs.EvalUpdateRequest{
|
|
|
|
Evals: []*structs.Evaluation{eval},
|
|
|
|
EvalToken: w.evalToken,
|
|
|
|
WriteRequest: structs.WriteRequest{
|
|
|
|
Region: w.srv.config.Region,
|
|
|
|
},
|
|
|
|
}
|
|
|
|
var resp structs.GenericResponse
|
|
|
|
|
|
|
|
SUBMIT:
|
|
|
|
// Make the RPC call
|
|
|
|
if err := w.srv.RPC("Eval.Reblock", &req, &resp); err != nil {
|
2018-09-15 23:23:13 +00:00
|
|
|
w.logger.Error("failed to reblock evaluation", "eval", log.Fmt("%#v", eval), "error", err)
|
2016-05-20 23:03:53 +00:00
|
|
|
if w.shouldResubmit(err) && !w.backoffErr(backoffBaselineSlow, backoffLimitSlow) {
|
|
|
|
goto SUBMIT
|
|
|
|
}
|
|
|
|
return err
|
|
|
|
} else {
|
2018-09-15 23:23:13 +00:00
|
|
|
w.logger.Debug("reblocked evaluation", "eval", log.Fmt("%#v", eval))
|
2016-05-20 23:03:53 +00:00
|
|
|
w.backoffReset()
|
2015-09-07 21:23:48 +00:00
|
|
|
}
|
|
|
|
return nil
|
|
|
|
}
|
|
|
|
|
2015-08-05 23:23:37 +00:00
|
|
|
// shouldResubmit checks if a given error should be swallowed and the plan
|
|
|
|
// resubmitted after a backoff. Usually these are transient errors that
|
|
|
|
// the cluster should heal from quickly.
|
|
|
|
func (w *Worker) shouldResubmit(err error) bool {
|
|
|
|
s := err.Error()
|
|
|
|
switch {
|
|
|
|
case strings.Contains(s, "No cluster leader"):
|
|
|
|
return true
|
|
|
|
case strings.Contains(s, "plan queue is disabled"):
|
|
|
|
return true
|
|
|
|
default:
|
|
|
|
return false
|
|
|
|
}
|
|
|
|
}
|
|
|
|
|
2015-07-28 22:12:08 +00:00
|
|
|
// backoffErr is used to do an exponential back off on error. This is
|
|
|
|
// maintained statefully for the worker. Returns if attempts should be
|
2018-03-11 17:34:16 +00:00
|
|
|
// abandoned due to shutdown.
|
2022-01-06 16:56:13 +00:00
|
|
|
// This uses the worker's context in order to immediately stop the
|
|
|
|
// backoff if the server or the worker is shutdown.
|
2015-08-24 00:39:49 +00:00
|
|
|
func (w *Worker) backoffErr(base, limit time.Duration) bool {
|
2022-01-06 16:56:13 +00:00
|
|
|
w.setWorkloadStatus(WorkloadBackoff)
|
2023-08-15 18:38:18 +00:00
|
|
|
|
|
|
|
backoff := helper.Backoff(base, limit, w.failures)
|
|
|
|
w.failures++
|
|
|
|
|
2015-07-28 22:12:08 +00:00
|
|
|
select {
|
|
|
|
case <-time.After(backoff):
|
|
|
|
return false
|
2022-01-06 16:56:13 +00:00
|
|
|
case <-w.ctx.Done():
|
2015-07-28 22:12:08 +00:00
|
|
|
return true
|
|
|
|
}
|
|
|
|
}
|
|
|
|
|
|
|
|
// backoffReset is used to reset the failure count for
|
|
|
|
// exponential backoff
|
|
|
|
func (w *Worker) backoffReset() {
|
|
|
|
w.failures = 0
|
|
|
|
}
|