open-nomad/client/allocrunner/alloc_runner.go

1172 lines
35 KiB
Go
Raw Normal View History

2018-06-11 20:33:18 +00:00
package allocrunner
2015-08-23 22:06:47 +00:00
2015-08-23 22:15:48 +00:00
import (
2017-07-03 22:03:42 +00:00
"context"
2015-08-29 22:46:10 +00:00
"fmt"
2015-08-23 22:15:48 +00:00
"log"
"path/filepath"
2015-08-23 22:15:48 +00:00
"sync"
2015-08-29 22:46:10 +00:00
"time"
2015-08-23 22:15:48 +00:00
metrics "github.com/armon/go-metrics"
2017-04-29 22:43:23 +00:00
"github.com/boltdb/bolt"
2016-02-20 00:31:04 +00:00
"github.com/hashicorp/go-multierror"
"github.com/hashicorp/nomad/client/allocdir"
2018-06-11 20:33:18 +00:00
"github.com/hashicorp/nomad/client/allocrunner/taskrunner"
"github.com/hashicorp/nomad/client/allocwatcher"
"github.com/hashicorp/nomad/client/config"
2018-06-11 20:33:18 +00:00
consulApi "github.com/hashicorp/nomad/client/consul"
2016-09-14 20:30:01 +00:00
"github.com/hashicorp/nomad/client/vaultclient"
2017-07-03 04:49:56 +00:00
"github.com/hashicorp/nomad/helper"
2015-08-23 22:15:48 +00:00
"github.com/hashicorp/nomad/nomad/structs"
cstructs "github.com/hashicorp/nomad/client/structs"
2015-08-23 22:15:48 +00:00
)
2015-08-23 22:06:47 +00:00
2017-04-29 22:43:23 +00:00
var (
// The following are the key paths written to the state database
allocRunnerStateAllocKey = []byte("alloc")
2017-04-29 22:43:23 +00:00
allocRunnerStateImmutableKey = []byte("immutable")
allocRunnerStateMutableKey = []byte("mutable")
allocRunnerStateAllocDirKey = []byte("alloc-dir")
)
// AllocStateUpdater is used to update the status of an allocation
type AllocStateUpdater func(alloc *structs.Allocation)
type AllocStatsReporter interface {
LatestAllocStats(taskFilter string) (*cstructs.AllocResourceUsage, error)
}
// AllocRunner is used to wrap an allocation and provide the execution context.
type AllocRunner struct {
config *config.Config
updater AllocStateUpdater
logger *log.Logger
2015-08-23 22:15:48 +00:00
// allocID is the ID of this runner's allocation. Since it does not
// change for the lifetime of the AllocRunner it is safe to read
// without acquiring a lock (unlike alloc).
allocID string
2016-02-10 21:44:53 +00:00
alloc *structs.Allocation
allocClientStatus string // Explicit status of allocation. Set when there are failures
allocClientDescription string
2018-04-10 19:49:59 +00:00
allocHealth *bool // Whether the allocation is healthy
allocHealthTime time.Time // Time at which allocation health has been set
2017-07-03 04:49:56 +00:00
allocBroadcast *cstructs.AllocBroadcaster
2016-02-10 21:44:53 +00:00
allocLock sync.Mutex
2015-08-29 22:46:10 +00:00
dirtyCh chan struct{}
allocDir *allocdir.AllocDir
allocDirLock sync.Mutex
2018-06-11 20:33:18 +00:00
tasks map[string]*taskrunner.TaskRunner
taskStates map[string]*structs.TaskState
restored map[string]struct{}
taskLock sync.RWMutex
2015-08-23 23:49:48 +00:00
2016-03-22 02:59:58 +00:00
taskStatusLock sync.RWMutex
2015-08-29 22:46:10 +00:00
2015-08-23 22:15:48 +00:00
updateCh chan *structs.Allocation
vaultClient vaultclient.VaultClient
2018-06-11 20:33:18 +00:00
consulClient consulApi.ConsulServiceAPI
2016-09-14 20:30:01 +00:00
// prevAlloc allows for Waiting until a previous allocation exits and
// the migrates it data. If sticky volumes aren't used and there's no
// previous allocation a noop implementation is used so it always safe
// to call.
prevAlloc allocwatcher.PrevAllocWatcher
// ctx is cancelled with exitFn to cause the alloc to be destroyed
// (stopped and GC'd).
2017-07-03 22:03:42 +00:00
ctx context.Context
exitFn context.CancelFunc
// waitCh is closed when the Run method exits. At that point the alloc
// has stopped and been GC'd.
2017-07-03 22:03:42 +00:00
waitCh chan struct{}
2017-04-29 22:43:23 +00:00
// State related fields
// stateDB is used to store the alloc runners state
stateDB *bolt.DB
allocStateLock sync.Mutex
2017-04-29 22:43:23 +00:00
2017-06-23 00:32:20 +00:00
// persistedEval is the last persisted evaluation ID. Since evaluation
// IDs change on every allocation update we only need to persist the
// allocation when its eval ID != the last persisted eval ID.
persistedEvalLock sync.Mutex
persistedEval string
2017-04-29 22:43:23 +00:00
// immutablePersisted and allocDirPersisted are used to track whether the
// immutable data and the alloc dir have been persisted. Once persisted we
// can lower write volume by not re-writing these values
immutablePersisted bool
allocDirPersisted bool
2017-11-02 17:05:38 +00:00
// baseLabels are used when emitting tagged metrics. All alloc runner metrics
// will have these tags, and optionally more.
baseLabels []metrics.Label
2015-08-23 22:06:47 +00:00
}
// allocRunnerAllocState is state that only has to be written when the alloc
// changes.
type allocRunnerAllocState struct {
Alloc *structs.Allocation
}
// allocRunnerImmutableState is state that only has to be written once.
2017-04-29 22:43:23 +00:00
type allocRunnerImmutableState struct {
Version string
}
// allocRunnerMutableState is state that has to be written on each save as it
// changes over the life-cycle of the alloc_runner.
type allocRunnerMutableState struct {
AllocClientStatus string
AllocClientDescription string
TaskStates map[string]*structs.TaskState
2017-07-18 19:31:05 +00:00
DeploymentStatus *structs.AllocDeploymentStatus
2017-04-29 22:43:23 +00:00
}
// NewAllocRunner is used to create a new allocation context
2017-04-29 22:43:23 +00:00
func NewAllocRunner(logger *log.Logger, config *config.Config, stateDB *bolt.DB, updater AllocStateUpdater,
2018-06-11 20:33:18 +00:00
alloc *structs.Allocation, vaultClient vaultclient.VaultClient, consulClient consulApi.ConsulServiceAPI,
prevAlloc allocwatcher.PrevAllocWatcher) *AllocRunner {
2015-08-30 02:14:47 +00:00
ar := &AllocRunner{
config: config,
stateDB: stateDB,
updater: updater,
logger: logger,
alloc: alloc,
allocID: alloc.ID,
//allocBroadcast: cstructs.NewAllocBroadcaster(8),
prevAlloc: prevAlloc,
dirtyCh: make(chan struct{}, 1),
allocDir: allocdir.NewAllocDir(logger, filepath.Join(config.AllocDir, alloc.ID)),
tasks: make(map[string]*taskrunner.TaskRunner),
taskStates: copyTaskStates(alloc.TaskStates),
restored: make(map[string]struct{}),
updateCh: make(chan *structs.Allocation, 64),
waitCh: make(chan struct{}),
vaultClient: vaultClient,
consulClient: consulClient,
2015-08-23 22:06:47 +00:00
}
2017-07-03 22:03:42 +00:00
// TODO Should be passed a context
ar.ctx, ar.exitFn = context.WithCancel(context.TODO())
2017-11-02 17:05:38 +00:00
return ar
}
// setBaseLabels creates the set of base labels. This should be called after
// Restore has been called so the allocation is guaranteed to be loaded
func (r *AllocRunner) setBaseLabels() {
r.baseLabels = make([]metrics.Label, 0, 3)
if r.alloc.Job != nil {
r.baseLabels = append(r.baseLabels, metrics.Label{
2017-11-02 17:05:38 +00:00
Name: "job",
Value: r.alloc.Job.Name,
})
}
if r.alloc.TaskGroup != "" {
r.baseLabels = append(r.baseLabels, metrics.Label{
2017-11-02 17:05:38 +00:00
Name: "task_group",
Value: r.alloc.TaskGroup,
})
}
if r.config != nil && r.config.Node != nil {
r.baseLabels = append(r.baseLabels, metrics.Label{
2017-11-02 20:29:10 +00:00
Name: "node_id",
Value: r.config.Node.ID,
})
2017-11-02 17:05:38 +00:00
}
2015-08-23 22:06:47 +00:00
}
2017-05-02 20:31:56 +00:00
// pre060StateFilePath returns the path to our state file that would have been
// written pre v0.6.0
// COMPAT: Remove in 0.7.0
func (r *AllocRunner) pre060StateFilePath() string {
r.allocLock.Lock()
defer r.allocLock.Unlock()
path := filepath.Join(r.config.StateDir, "alloc", r.allocID, "state.json")
return path
}
// RestoreState is used to restore the state of the alloc runner
func (r *AllocRunner) RestoreState() error {
//XXX Deprecated: see allocrunnerv2
//err := r.stateDB.View(func(tx *bolt.Tx) error {
// bkt, err := state.GetAllocationBucket(tx, r.allocID)
// if err != nil {
// return fmt.Errorf("failed to get allocation bucket: %v", err)
// }
// // Get the state objects
// var mutable allocRunnerMutableState
// var immutable allocRunnerImmutableState
// var allocState allocRunnerAllocState
// var allocDir allocdir.AllocDir
// if err := state.GetObject(bkt, allocRunnerStateAllocKey, &allocState); err != nil {
// return fmt.Errorf("failed to read alloc runner alloc state: %v", err)
// }
// if err := state.GetObject(bkt, allocRunnerStateImmutableKey, &immutable); err != nil {
// return fmt.Errorf("failed to read alloc runner immutable state: %v", err)
// }
// if err := state.GetObject(bkt, allocRunnerStateMutableKey, &mutable); err != nil {
// return fmt.Errorf("failed to read alloc runner mutable state: %v", err)
// }
// if err := state.GetObject(bkt, allocRunnerStateAllocDirKey, &allocDir); err != nil {
// return fmt.Errorf("failed to read alloc runner alloc_dir state: %v", err)
// }
// // Populate the fields
// r.alloc = allocState.Alloc
// r.allocDir = &allocDir
// r.allocClientStatus = mutable.AllocClientStatus
// r.allocClientDescription = mutable.AllocClientDescription
// r.taskStates = mutable.TaskStates
// r.alloc.ClientStatus = getClientStatus(r.taskStates)
// r.alloc.DeploymentStatus = mutable.DeploymentStatus
// return nil
//})
//if err != nil {
// return fmt.Errorf("failed to read allocation state: %v", err)
//}
2016-06-17 21:58:53 +00:00
var snapshotErrors multierror.Error
if r.alloc == nil {
snapshotErrors.Errors = append(snapshotErrors.Errors, fmt.Errorf("alloc_runner snapshot includes a nil allocation"))
}
if r.allocDir == nil {
snapshotErrors.Errors = append(snapshotErrors.Errors, fmt.Errorf("alloc_runner snapshot includes a nil alloc dir"))
}
2016-06-17 21:58:53 +00:00
if e := snapshotErrors.ErrorOrNil(); e != nil {
return e
}
2017-05-02 20:31:56 +00:00
tg := r.alloc.Job.LookupTaskGroup(r.alloc.TaskGroup)
if tg == nil {
return fmt.Errorf("restored allocation doesn't contain task group %q", r.alloc.TaskGroup)
}
// Restore the task runners
taskDestroyEvent := structs.NewTaskEvent(structs.TaskKilled)
var mErr multierror.Error
2017-05-02 20:31:56 +00:00
for _, task := range tg.Tasks {
name := task.Name
state := r.taskStates[name]
// Nomad exited before task could start, nothing to restore.
// AllocRunner.Run will start a new TaskRunner for this task
if state == nil {
continue
}
// Mark the task as restored.
r.restored[name] = struct{}{}
td, ok := r.allocDir.TaskDirs[name]
if !ok {
// Create the task dir metadata if it doesn't exist.
// Since task dirs are created during r.Run() the
// client may save state and exit before all task dirs
// are created
td = r.allocDir.NewTaskDir(name)
}
2015-11-09 23:55:31 +00:00
// Skip tasks in terminal states.
if state.State == structs.TaskStateDead {
2015-11-09 23:55:31 +00:00
continue
}
2018-06-11 20:33:18 +00:00
tr := taskrunner.NewTaskRunner(r.logger, r.config, r.stateDB, r.setTaskState, td, r.Alloc(), task, r.vaultClient, r.consulClient)
r.tasks[name] = tr
if restartReason, err := tr.RestoreState(); err != nil {
r.logger.Printf("[ERR] client: failed to restore state for alloc %s task %q: %v", r.allocID, name, err)
mErr.Errors = append(mErr.Errors, err)
2016-02-19 23:49:32 +00:00
} else if !r.alloc.TerminalStatus() {
// Only start if the alloc isn't in a terminal status.
go tr.Run()
2017-05-09 17:50:24 +00:00
// Restart task runner if RestoreState gave a reason
if restartReason != "" {
r.logger.Printf("[INFO] client: restarting alloc %s task %s: %v", r.allocID, name, restartReason)
const failure = false
tr.Restart("upgrade", restartReason, failure)
}
} else {
2018-06-12 19:16:20 +00:00
// XXX This does nothing and is broken since the task runner is not
// running yet, and there is nothing listening to the destroy ch.
// XXX When a single task is dead in the allocation we should kill
// all the task. This currently does NOT happen. Re-enable test:
// TestAllocRunner_TaskLeader_StopRestoredTG
tr.Destroy(taskDestroyEvent)
}
}
2016-09-14 20:30:01 +00:00
return mErr.ErrorOrNil()
}
2015-11-10 00:15:11 +00:00
// SaveState is used to snapshot the state of the alloc runner
// if the fullSync is marked as false only the state of the Alloc Runner
// is snapshotted. If fullSync is marked as true, we snapshot
// all the Task Runners associated with the Alloc
func (r *AllocRunner) SaveState() error {
if err := r.saveAllocRunnerState(); err != nil {
return err
}
2015-11-10 00:15:11 +00:00
// Save state for each task
2016-07-21 20:41:01 +00:00
runners := r.getTaskRunners()
var mErr multierror.Error
2016-07-21 20:41:01 +00:00
for _, tr := range runners {
if err := tr.SaveState(); err != nil {
mErr.Errors = append(mErr.Errors, fmt.Errorf("failed to save state for alloc %s task %q: %v",
2018-06-11 20:33:18 +00:00
r.allocID, tr.Name(), err))
2015-11-10 00:15:11 +00:00
}
}
return mErr.ErrorOrNil()
}
func (r *AllocRunner) saveAllocRunnerState() error {
r.allocStateLock.Lock()
defer r.allocStateLock.Unlock()
if r.ctx.Err() == context.Canceled {
return nil
}
//XXX Deprecated: see allocrunnerv2
return nil
//// Grab all the relevant data
//alloc := r.Alloc()
//r.allocLock.Lock()
//allocClientStatus := r.allocClientStatus
//allocClientDescription := r.allocClientDescription
//r.allocLock.Unlock()
//r.allocDirLock.Lock()
//allocDir := r.allocDir.Copy()
//r.allocDirLock.Unlock()
//// Start the transaction.
//return r.stateDB.Batch(func(tx *bolt.Tx) error {
// // Grab the allocation bucket
// allocBkt, err := state.GetAllocationBucket(tx, r.allocID)
// if err != nil {
// return fmt.Errorf("failed to retrieve allocation bucket: %v", err)
// }
// // Write the allocation if the eval has changed
// r.persistedEvalLock.Lock()
// lastPersisted := r.persistedEval
// r.persistedEvalLock.Unlock()
// if alloc.EvalID != lastPersisted {
// allocState := &allocRunnerAllocState{
// Alloc: alloc,
// }
// if err := state.PutObject(allocBkt, allocRunnerStateAllocKey, &allocState); err != nil {
// return fmt.Errorf("failed to write alloc_runner alloc state: %v", err)
// }
// tx.OnCommit(func() {
// r.persistedEvalLock.Lock()
// r.persistedEval = alloc.EvalID
// r.persistedEvalLock.Unlock()
// })
// }
// // Write immutable data iff it hasn't been written yet
// if !r.immutablePersisted {
// immutable := &allocRunnerImmutableState{
// Version: r.config.Version.VersionNumber(),
// }
// if err := state.PutObject(allocBkt, allocRunnerStateImmutableKey, &immutable); err != nil {
// return fmt.Errorf("failed to write alloc_runner immutable state: %v", err)
// }
// tx.OnCommit(func() {
// r.immutablePersisted = true
// })
// }
// // Write the alloc dir data if it hasn't been written before and it exists.
// if !r.allocDirPersisted && allocDir != nil {
// if err := state.PutObject(allocBkt, allocRunnerStateAllocDirKey, allocDir); err != nil {
// return fmt.Errorf("failed to write alloc_runner allocDir state: %v", err)
// }
// tx.OnCommit(func() {
// r.allocDirPersisted = true
// })
// }
// // Write the mutable state every time
// mutable := &allocRunnerMutableState{
// AllocClientStatus: allocClientStatus,
// AllocClientDescription: allocClientDescription,
// TaskStates: alloc.TaskStates,
// DeploymentStatus: alloc.DeploymentStatus,
// }
// if err := state.PutObject(allocBkt, allocRunnerStateMutableKey, &mutable); err != nil {
// return fmt.Errorf("failed to write alloc_runner mutable state: %v", err)
// }
// return nil
//})
}
// DestroyState is used to cleanup after ourselves
func (r *AllocRunner) DestroyState() error {
//r.allocStateLock.Lock()
//defer r.allocStateLock.Unlock()
//return r.stateDB.Update(func(tx *bolt.Tx) error {
// if err := state.DeleteAllocationBucket(tx, r.allocID); err != nil {
// return fmt.Errorf("failed to delete allocation bucket: %v", err)
// }
// return nil
//})
panic("deprecated: use allocrunnerv2")
}
2015-08-31 00:35:58 +00:00
// DestroyContext is used to destroy the context
func (r *AllocRunner) DestroyContext() error {
return r.allocDir.Destroy()
2015-08-31 00:35:58 +00:00
}
2017-05-02 20:31:56 +00:00
// GetAllocDir returns the alloc dir for the alloc runner
func (r *AllocRunner) GetAllocDir() *allocdir.AllocDir {
return r.allocDir
}
// GetListener returns a listener for updates broadcast by this alloc runner.
// Callers are responsible for calling Close on their Listener.
func (r *AllocRunner) GetListener() *cstructs.AllocListener {
return r.allocBroadcast.Listen()
}
2016-02-10 21:44:53 +00:00
// copyTaskStates returns a copy of the passed task states.
func copyTaskStates(states map[string]*structs.TaskState) map[string]*structs.TaskState {
copy := make(map[string]*structs.TaskState, len(states))
for task, state := range states {
copy[task] = state.Copy()
}
return copy
}
// finalizeTerminalAlloc sets any missing required fields like
// finishedAt in the alloc runner's task States. finishedAt is used
// to calculate reschedule time for failed allocs, so we make sure that
// it is set
func (r *AllocRunner) finalizeTerminalAlloc(alloc *structs.Allocation) {
if !alloc.ClientTerminalStatus() {
return
}
r.taskStatusLock.Lock()
defer r.taskStatusLock.Unlock()
group := alloc.Job.LookupTaskGroup(alloc.TaskGroup)
if r.taskStates == nil {
r.taskStates = make(map[string]*structs.TaskState)
}
now := time.Now()
for _, task := range group.Tasks {
ts, ok := r.taskStates[task.Name]
if !ok {
ts = &structs.TaskState{}
r.taskStates[task.Name] = ts
}
if ts.FinishedAt.IsZero() {
ts.FinishedAt = now
}
}
alloc.TaskStates = copyTaskStates(r.taskStates)
}
2015-08-23 22:06:47 +00:00
// Alloc returns the associated allocation
2015-08-23 22:36:06 +00:00
func (r *AllocRunner) Alloc() *structs.Allocation {
r.allocLock.Lock()
// Don't do a deep copy of the job
alloc := r.alloc.CopySkipJob()
2016-02-10 21:44:53 +00:00
2016-05-15 16:41:34 +00:00
// The status has explicitly been set.
2016-02-10 21:44:53 +00:00
if r.allocClientStatus != "" || r.allocClientDescription != "" {
alloc.ClientStatus = r.allocClientStatus
alloc.ClientDescription = r.allocClientDescription
// Copy over the task states so we don't lose them
r.taskStatusLock.RLock()
alloc.TaskStates = copyTaskStates(r.taskStates)
r.taskStatusLock.RUnlock()
2016-02-10 21:44:53 +00:00
r.allocLock.Unlock()
r.finalizeTerminalAlloc(alloc)
2016-02-10 21:44:53 +00:00
return alloc
}
2017-07-03 04:49:56 +00:00
// The health has been set
if r.allocHealth != nil {
if alloc.DeploymentStatus == nil {
2018-04-10 19:49:59 +00:00
alloc.DeploymentStatus = &structs.AllocDeploymentStatus{}
2017-07-03 04:49:56 +00:00
}
alloc.DeploymentStatus.Healthy = helper.BoolToPtr(*r.allocHealth)
2018-04-10 19:49:59 +00:00
alloc.DeploymentStatus.Timestamp = r.allocHealthTime
2017-07-03 04:49:56 +00:00
}
r.allocLock.Unlock()
// Scan the task states to determine the status of the alloc
r.taskStatusLock.RLock()
2016-02-10 21:44:53 +00:00
alloc.TaskStates = copyTaskStates(r.taskStates)
2017-05-09 17:50:24 +00:00
alloc.ClientStatus = getClientStatus(r.taskStates)
r.taskStatusLock.RUnlock()
// If the client status is failed and we are part of a deployment, mark the
// alloc as unhealthy. This guards against the watcher not be started.
r.allocLock.Lock()
if alloc.ClientStatus == structs.AllocClientStatusFailed &&
alloc.DeploymentID != "" && !alloc.DeploymentStatus.IsUnhealthy() {
alloc.DeploymentStatus = &structs.AllocDeploymentStatus{
Healthy: helper.BoolToPtr(false),
}
}
r.allocLock.Unlock()
r.finalizeTerminalAlloc(alloc)
2017-05-09 17:50:24 +00:00
return alloc
}
// getClientStatus takes in the task states for a given allocation and computes
// the client status
func getClientStatus(taskStates map[string]*structs.TaskState) string {
var pending, running, dead, failed bool
for _, state := range taskStates {
switch state.State {
case structs.TaskStateRunning:
running = true
case structs.TaskStatePending:
pending = true
case structs.TaskStateDead:
if state.Failed {
failed = true
} else {
dead = true
}
}
}
// Determine the alloc status
if failed {
2017-05-09 17:50:24 +00:00
return structs.AllocClientStatusFailed
} else if running {
2017-05-09 17:50:24 +00:00
return structs.AllocClientStatusRunning
} else if pending {
2017-05-09 17:50:24 +00:00
return structs.AllocClientStatusPending
} else if dead {
2017-05-09 17:50:24 +00:00
return structs.AllocClientStatusComplete
}
2017-05-09 17:50:24 +00:00
return ""
2015-08-23 22:06:47 +00:00
}
2015-08-31 00:10:17 +00:00
// dirtySyncState is used to watch for state being marked dirty to sync
func (r *AllocRunner) dirtySyncState() {
2015-08-29 22:46:10 +00:00
for {
select {
case <-r.dirtyCh:
if err := r.syncStatus(); err != nil {
// Only WARN instead of ERR because we continue on
r.logger.Printf("[WARN] client: error persisting alloc %q state: %v",
r.allocID, err)
}
2017-07-03 22:03:42 +00:00
case <-r.ctx.Done():
2015-08-29 22:46:10 +00:00
return
}
2015-08-31 00:10:17 +00:00
}
}
2015-08-29 22:46:10 +00:00
2015-08-31 00:10:17 +00:00
// syncStatus is used to run and sync the status when it changes
func (r *AllocRunner) syncStatus() error {
// Get a copy of our alloc, update status server side and sync to disk
alloc := r.Alloc()
r.updater(alloc)
r.sendBroadcast(alloc)
return r.saveAllocRunnerState()
2015-08-29 22:46:10 +00:00
}
// sendBroadcast broadcasts an alloc update.
func (r *AllocRunner) sendBroadcast(alloc *structs.Allocation) {
2017-07-18 18:05:38 +00:00
// Try to send the alloc up to three times with a delay to allow recovery.
sent := false
for i := 0; i < 3; i++ {
//if sent = r.allocBroadcast.Send(alloc); sent {
// break
//}
2017-07-18 18:05:38 +00:00
time.Sleep(500 * time.Millisecond)
}
if !sent {
2017-07-18 19:31:05 +00:00
r.logger.Printf("[WARN] client: failed to broadcast update to allocation %q", r.allocID)
2017-07-18 18:05:38 +00:00
}
2015-08-29 22:46:10 +00:00
}
// setStatus is used to update the allocation status
func (r *AllocRunner) setStatus(status, desc string) {
2016-02-10 21:44:53 +00:00
r.allocLock.Lock()
r.allocClientStatus = status
r.allocClientDescription = desc
r.allocLock.Unlock()
2015-08-29 22:46:10 +00:00
select {
case r.dirtyCh <- struct{}{}:
default:
}
}
// setTaskState is used to set the status of a task. If lazySync is set then the
// event is appended but not synced with the server. If state is omitted, the
// last known state is used.
2017-08-08 04:26:04 +00:00
func (r *AllocRunner) setTaskState(taskName, state string, event *structs.TaskEvent, lazySync bool) {
r.taskStatusLock.Lock()
defer r.taskStatusLock.Unlock()
taskState, ok := r.taskStates[taskName]
if !ok {
2016-03-01 22:09:25 +00:00
taskState = &structs.TaskState{}
r.taskStates[taskName] = taskState
}
// Set the tasks state.
2016-11-04 23:57:24 +00:00
if event != nil {
if event.FailsTask {
taskState.Failed = true
}
2017-07-03 04:49:56 +00:00
if event.Type == structs.TaskRestarting {
2017-11-02 17:05:38 +00:00
if !r.config.DisableTaggedMetrics {
metrics.IncrCounterWithLabels([]string{"client", "allocs", "restart"},
1, r.baseLabels)
}
if r.config.BackwardsCompatibleMetrics {
metrics.IncrCounter([]string{"client", "allocs", r.alloc.Job.Name, r.alloc.TaskGroup, taskName, "restart"}, 1)
}
2017-07-03 04:49:56 +00:00
taskState.Restarts++
taskState.LastRestart = time.Unix(0, event.Time)
2017-07-03 04:49:56 +00:00
}
2016-11-04 23:57:24 +00:00
r.appendTaskEvent(taskState, event)
}
2017-08-08 04:26:04 +00:00
if lazySync {
2016-11-04 23:57:24 +00:00
return
}
2017-08-08 04:26:04 +00:00
// If the state hasn't been set use the existing state.
if state == "" {
state = taskState.State
if taskState.State == "" {
state = structs.TaskStatePending
}
}
switch state {
case structs.TaskStateRunning:
// Capture the start time if it is just starting
if taskState.State != structs.TaskStateRunning {
taskState.StartedAt = time.Now().UTC()
2017-11-02 17:05:38 +00:00
if !r.config.DisableTaggedMetrics {
metrics.IncrCounterWithLabels([]string{"client", "allocs", "running"},
1, r.baseLabels)
}
if r.config.BackwardsCompatibleMetrics {
metrics.IncrCounter([]string{"client", "allocs", r.alloc.Job.Name, r.alloc.TaskGroup, taskName, "running"}, 1)
}
}
case structs.TaskStateDead:
// Capture the finished time if not already set
if taskState.FinishedAt.IsZero() {
taskState.FinishedAt = time.Now().UTC()
}
2017-02-11 01:55:19 +00:00
// Find all tasks that are not the one that is dead and check if the one
// that is dead is a leader
2018-06-11 20:33:18 +00:00
var otherTaskRunners []*taskrunner.TaskRunner
2017-02-11 01:55:19 +00:00
var otherTaskNames []string
leader := false
for task, tr := range r.tasks {
if task != taskName {
otherTaskRunners = append(otherTaskRunners, tr)
otherTaskNames = append(otherTaskNames, task)
2018-06-11 20:33:18 +00:00
} else if tr.IsLeader() {
2017-02-11 01:55:19 +00:00
leader = true
}
}
// Emitting metrics to indicate task complete and failures
if taskState.Failed {
2017-11-02 17:05:38 +00:00
if !r.config.DisableTaggedMetrics {
metrics.IncrCounterWithLabels([]string{"client", "allocs", "failed"},
1, r.baseLabels)
}
if r.config.BackwardsCompatibleMetrics {
metrics.IncrCounter([]string{"client", "allocs", r.alloc.Job.Name, r.alloc.TaskGroup, taskName, "failed"}, 1)
}
} else {
2017-11-02 17:05:38 +00:00
if !r.config.DisableTaggedMetrics {
metrics.IncrCounterWithLabels([]string{"client", "allocs", "complete"},
1, r.baseLabels)
}
if r.config.BackwardsCompatibleMetrics {
metrics.IncrCounter([]string{"client", "allocs", r.alloc.Job.Name, r.alloc.TaskGroup, taskName, "complete"}, 1)
}
}
2018-06-12 19:16:20 +00:00
2016-09-14 22:04:25 +00:00
// If the task failed, we should kill all the other tasks in the task group.
if taskState.Failed {
2017-02-11 01:55:19 +00:00
for _, tr := range otherTaskRunners {
tr.Destroy(structs.NewTaskEvent(structs.TaskSiblingFailed).SetFailedSibling(taskName))
2016-09-14 22:04:25 +00:00
}
2017-02-11 01:55:19 +00:00
if len(otherTaskRunners) > 0 {
r.logger.Printf("[DEBUG] client: task %q failed, destroying other tasks in task group: %v", taskName, otherTaskNames)
}
} else if leader {
// If the task was a leader task we should kill all the other tasks.
2017-02-11 01:55:19 +00:00
for _, tr := range otherTaskRunners {
tr.Destroy(structs.NewTaskEvent(structs.TaskLeaderDead))
}
if len(otherTaskRunners) > 0 {
r.logger.Printf("[DEBUG] client: leader task %q is dead, destroying other tasks in task group: %v", taskName, otherTaskNames)
2016-09-14 22:04:25 +00:00
}
}
2016-02-19 22:49:43 +00:00
}
// Store the new state
taskState.State = state
2015-08-29 22:46:10 +00:00
select {
case r.dirtyCh <- struct{}{}:
default:
}
}
// appendTaskEvent updates the task status by appending the new event.
func (r *AllocRunner) appendTaskEvent(state *structs.TaskState, event *structs.TaskEvent) {
capacity := 10
if state.Events == nil {
state.Events = make([]*structs.TaskEvent, 0, capacity)
}
// If we hit capacity, then shift it.
if len(state.Events) == capacity {
old := state.Events
state.Events = make([]*structs.TaskEvent, 0, capacity)
state.Events = append(state.Events, old[1:]...)
}
state.Events = append(state.Events, event)
}
2015-08-23 22:06:47 +00:00
// Run is a long running goroutine used to manage an allocation
2015-08-23 22:36:06 +00:00
func (r *AllocRunner) Run() {
2015-10-04 20:36:03 +00:00
defer close(r.waitCh)
r.setBaseLabels()
2015-08-31 00:10:17 +00:00
go r.dirtySyncState()
2015-08-23 22:15:48 +00:00
2015-08-29 22:46:10 +00:00
// Find the task group to run in the allocation
2017-05-02 20:31:56 +00:00
alloc := r.Alloc()
2015-08-30 02:14:47 +00:00
tg := alloc.Job.LookupTaskGroup(alloc.TaskGroup)
2015-08-23 23:49:48 +00:00
if tg == nil {
r.logger.Printf("[ERR] client: alloc %q for missing task group %q", r.allocID, alloc.TaskGroup)
2015-08-29 22:46:10 +00:00
r.setStatus(structs.AllocClientStatusFailed, fmt.Sprintf("missing task group '%s'", alloc.TaskGroup))
2015-08-23 23:49:48 +00:00
return
}
// Build allocation directory (idempotent)
r.allocDirLock.Lock()
err := r.allocDir.Build()
r.allocDirLock.Unlock()
if err != nil {
r.logger.Printf("[ERR] client: alloc %q failed to build task directories: %v", r.allocID, err)
r.setStatus(structs.AllocClientStatusFailed, fmt.Sprintf("failed to build task dirs for '%s'", alloc.TaskGroup))
return
}
// Wait for a previous alloc - if any - to terminate
if err := r.prevAlloc.Wait(r.ctx); err != nil {
if err == context.Canceled {
return
}
r.setStatus(structs.AllocClientStatusFailed, fmt.Sprintf("error while waiting for previous alloc to terminate: %v", err))
return
}
// Wait for data to be migrated from a previous alloc if applicable
if err := r.prevAlloc.Migrate(r.ctx, r.allocDir); err != nil {
if err == context.Canceled {
return
}
2017-08-11 23:50:30 +00:00
// Soft-fail on migration errors
r.logger.Printf("[WARN] client: alloc %q error while migrating data from previous alloc: %v", r.allocID, err)
// Recreate alloc dir to ensure a clean slate
r.allocDir.Destroy()
if err := r.allocDir.Build(); err != nil {
r.logger.Printf("[ERR] client: alloc %q failed to clean task directories after failed migration: %v", r.allocID, err)
r.setStatus(structs.AllocClientStatusFailed, fmt.Sprintf("failed to rebuild task dirs for '%s'", alloc.TaskGroup))
return
}
2015-08-30 02:14:47 +00:00
}
2015-08-23 23:49:48 +00:00
// Check if the allocation is in a terminal status. In this case, we don't
// start any of the task runners and directly wait for the destroy signal to
// clean up the allocation.
if alloc.TerminalStatus() {
r.logger.Printf("[DEBUG] client: alloc %q in terminal status, waiting for destroy", r.allocID)
// mark this allocation as completed if it is not already in a
// terminal state
if !alloc.Terminated() {
r.setStatus(structs.AllocClientStatusComplete, "canceled running tasks for allocation in terminal state")
}
r.handleDestroy()
r.logger.Printf("[DEBUG] client: terminating runner for alloc '%s'", r.allocID)
2016-02-04 22:19:27 +00:00
return
}
// Increment alloc runner start counter. Incr'd even when restoring existing tasks so 1 start != 1 task execution
2017-11-02 17:05:38 +00:00
if !r.config.DisableTaggedMetrics {
metrics.IncrCounterWithLabels([]string{"client", "allocs", "start"},
1, r.baseLabels)
}
if r.config.BackwardsCompatibleMetrics {
metrics.IncrCounter([]string{"client", "allocs", r.alloc.Job.Name, r.alloc.TaskGroup, "start"}, 1)
}
// Start the watcher
wCtx, watcherCancel := context.WithCancel(r.ctx)
go r.watchHealth(wCtx)
2015-08-23 23:49:48 +00:00
// Start the task runners
r.logger.Printf("[DEBUG] client: starting task runners for alloc '%s'", r.allocID)
r.taskLock.Lock()
2015-08-23 23:49:48 +00:00
for _, task := range tg.Tasks {
if _, ok := r.restored[task.Name]; ok {
2015-08-30 02:14:47 +00:00
continue
}
r.allocDirLock.Lock()
taskdir := r.allocDir.NewTaskDir(task.Name)
r.allocDirLock.Unlock()
2018-06-11 20:33:18 +00:00
tr := taskrunner.NewTaskRunner(r.logger, r.config, r.stateDB, r.setTaskState, taskdir, r.Alloc(), task.Copy(), r.vaultClient, r.consulClient)
2015-08-23 23:49:48 +00:00
r.tasks[task.Name] = tr
tr.MarkReceived()
2016-09-14 20:30:01 +00:00
2015-08-23 23:49:48 +00:00
go tr.Run()
}
r.taskLock.Unlock()
2015-08-23 23:49:48 +00:00
2018-03-11 17:54:39 +00:00
// taskDestroyEvent contains an event that caused the destruction of a task
// in the allocation.
var taskDestroyEvent *structs.TaskEvent
2015-08-29 22:46:10 +00:00
OUTER:
2015-08-23 23:49:48 +00:00
// Wait for updates
2015-08-23 22:06:47 +00:00
for {
2015-08-23 22:15:48 +00:00
select {
2015-08-23 22:36:06 +00:00
case update := <-r.updateCh:
// Store the updated allocation.
r.allocLock.Lock()
// If the deployment ids have changed clear the health
if r.alloc.DeploymentID != update.DeploymentID {
r.allocHealth = nil
2018-04-10 19:49:59 +00:00
r.allocHealthTime = time.Time{}
}
r.alloc = update
r.allocLock.Unlock()
// Create a new watcher
watcherCancel()
wCtx, watcherCancel = context.WithCancel(r.ctx)
go r.watchHealth(wCtx)
2015-08-29 22:46:10 +00:00
// Check if we're in a terminal status
if update.TerminalStatus() {
taskDestroyEvent = structs.NewTaskEvent(structs.TaskKilled)
2015-08-29 22:46:10 +00:00
break OUTER
}
// Update the task groups
2016-07-21 20:41:01 +00:00
runners := r.getTaskRunners()
for _, tr := range runners {
tr.Update(update)
2015-08-29 22:46:10 +00:00
}
2017-05-02 20:31:56 +00:00
if err := r.syncStatus(); err != nil {
r.logger.Printf("[WARN] client: failed to sync alloc %q status upon receiving alloc update: %v",
r.allocID, err)
2017-05-02 20:31:56 +00:00
}
2017-07-03 22:03:42 +00:00
case <-r.ctx.Done():
taskDestroyEvent = structs.NewTaskEvent(structs.TaskKilled)
2015-08-29 22:46:10 +00:00
break OUTER
2015-08-23 22:15:48 +00:00
}
2015-08-23 22:06:47 +00:00
}
2015-08-29 22:46:10 +00:00
2016-09-15 18:37:20 +00:00
// Kill the task runners
r.destroyTaskRunners(taskDestroyEvent)
// Block until we should destroy the state of the alloc
r.handleDestroy()
// Free up the context. It has likely exited already
watcherCancel()
r.logger.Printf("[DEBUG] client: terminating runner for alloc '%s'", r.allocID)
2016-09-15 18:37:20 +00:00
}
// destroyTaskRunners destroys the task runners, waits for them to terminate and
// then saves state.
func (r *AllocRunner) destroyTaskRunners(destroyEvent *structs.TaskEvent) {
// First destroy the leader if one exists
tg := r.alloc.Job.LookupTaskGroup(r.alloc.TaskGroup)
leader := ""
for _, task := range tg.Tasks {
if task.Leader {
leader = task.Name
break
}
}
if leader != "" {
r.taskLock.RLock()
tr := r.tasks[leader]
r.taskLock.RUnlock()
// Dead tasks don't have a task runner created so guard against
// the leader being dead when this AR was saved.
if tr == nil {
r.logger.Printf("[DEBUG] client: alloc %q leader task %q of task group %q already stopped",
r.allocID, leader, r.alloc.TaskGroup)
} else {
r.logger.Printf("[DEBUG] client: alloc %q destroying leader task %q of task group %q first",
r.allocID, leader, r.alloc.TaskGroup)
tr.Destroy(destroyEvent)
<-tr.WaitCh()
}
}
// Then destroy non-leader tasks concurrently
r.taskLock.RLock()
for name, tr := range r.tasks {
if name != leader {
tr.Destroy(destroyEvent)
}
2015-08-29 22:46:10 +00:00
}
r.taskLock.RUnlock()
2015-08-29 22:46:10 +00:00
// Wait for termination of the task runners
for _, tr := range r.getTaskRunners() {
2015-08-29 22:46:10 +00:00
<-tr.WaitCh()
}
}
2015-08-31 00:10:17 +00:00
// handleDestroy blocks till the AllocRunner should be destroyed and does the
// necessary cleanup.
func (r *AllocRunner) handleDestroy() {
// Final state sync. We do this to ensure that the server has the correct
// state as we wait for a destroy.
alloc := r.Alloc()
// Increment the destroy count for this alloc runner since this allocation is being removed from this client.
2017-11-02 17:05:38 +00:00
if !r.config.DisableTaggedMetrics {
metrics.IncrCounterWithLabels([]string{"client", "allocs", "destroy"},
1, r.baseLabels)
}
if r.config.BackwardsCompatibleMetrics {
metrics.IncrCounter([]string{"client", "allocs", r.alloc.Job.Name, r.alloc.TaskGroup, "destroy"}, 1)
}
// Broadcast and persist state synchronously
r.sendBroadcast(alloc)
if err := r.saveAllocRunnerState(); err != nil {
r.logger.Printf("[WARN] client: alloc %q unable to persist state but should be GC'd soon anyway:%v",
r.allocID, err)
}
// Unmount any mounted directories as no tasks are running and makes
// cleaning up Nomad's data directory simpler.
if err := r.allocDir.UnmountAll(); err != nil {
r.logger.Printf("[ERR] client: alloc %q unable unmount task directories: %v", r.allocID, err)
}
// Update the server with the alloc's status -- also marks the alloc as
// being eligible for GC, so from this point on the alloc can be gc'd
// at any time.
r.updater(alloc)
2017-03-10 05:05:34 +00:00
for {
select {
2017-07-03 22:03:42 +00:00
case <-r.ctx.Done():
2017-03-10 05:05:34 +00:00
if err := r.DestroyContext(); err != nil {
r.logger.Printf("[ERR] client: failed to destroy context for alloc '%s': %v",
r.allocID, err)
2017-03-10 05:05:34 +00:00
}
if err := r.DestroyState(); err != nil {
r.logger.Printf("[ERR] client: failed to destroy state for alloc '%s': %v",
r.allocID, err)
2017-03-10 05:05:34 +00:00
}
return
case <-r.updateCh:
r.logger.Printf("[DEBUG] client: dropping update to terminal alloc '%s'", r.allocID)
2015-08-31 00:35:58 +00:00
}
}
2015-08-23 22:06:47 +00:00
}
// IsWaiting returns true if this alloc is waiting on a previous allocation to
// terminate.
func (r *AllocRunner) IsWaiting() bool {
return r.prevAlloc.IsWaiting()
}
// IsMigrating returns true if this alloc is migrating data from a previous
// allocation.
func (r *AllocRunner) IsMigrating() bool {
return r.prevAlloc.IsMigrating()
}
2015-08-23 22:06:47 +00:00
// Update is used to update the allocation of the context
2015-08-23 22:36:06 +00:00
func (r *AllocRunner) Update(update *structs.Allocation) {
2015-08-23 22:15:48 +00:00
select {
2015-08-23 22:36:06 +00:00
case r.updateCh <- update:
2015-08-23 22:15:48 +00:00
default:
2015-08-23 22:36:06 +00:00
r.logger.Printf("[ERR] client: dropping update to alloc '%s'", update.ID)
2015-08-23 22:15:48 +00:00
}
2015-08-23 22:06:47 +00:00
}
// StatsReporter returns an interface to query resource usage statistics of an
// allocation
func (r *AllocRunner) StatsReporter() AllocStatsReporter {
return r
}
2016-07-21 20:41:01 +00:00
// getTaskRunners is a helper that returns a copy of the task runners list using
// the taskLock.
2018-06-11 20:33:18 +00:00
func (r *AllocRunner) getTaskRunners() []*taskrunner.TaskRunner {
2016-07-21 20:41:01 +00:00
// Get the task runners
r.taskLock.RLock()
defer r.taskLock.RUnlock()
2018-06-11 20:33:18 +00:00
runners := make([]*taskrunner.TaskRunner, 0, len(r.tasks))
2016-07-21 20:41:01 +00:00
for _, tr := range r.tasks {
runners = append(runners, tr)
}
return runners
}
// LatestAllocStats returns the latest allocation stats. If the optional taskFilter is set
// the allocation stats will only include the given task.
func (r *AllocRunner) LatestAllocStats(taskFilter string) (*cstructs.AllocResourceUsage, error) {
astat := &cstructs.AllocResourceUsage{
Tasks: make(map[string]*cstructs.TaskResourceUsage),
}
var flat []*cstructs.TaskResourceUsage
if taskFilter != "" {
2016-06-20 17:19:06 +00:00
r.taskLock.RLock()
tr, ok := r.tasks[taskFilter]
2016-06-20 17:19:06 +00:00
r.taskLock.RUnlock()
if !ok {
return nil, fmt.Errorf("allocation %q has no task %q", r.allocID, taskFilter)
}
l := tr.LatestResourceUsage()
if l != nil {
astat.Tasks[taskFilter] = l
flat = []*cstructs.TaskResourceUsage{l}
astat.Timestamp = l.Timestamp
}
} else {
2016-06-20 17:19:06 +00:00
// Get the task runners
2016-07-21 20:41:01 +00:00
runners := r.getTaskRunners()
2016-06-20 17:19:06 +00:00
for _, tr := range runners {
l := tr.LatestResourceUsage()
if l != nil {
2018-06-11 20:33:18 +00:00
astat.Tasks[tr.Name()] = l
flat = append(flat, l)
if l.Timestamp > astat.Timestamp {
astat.Timestamp = l.Timestamp
}
}
}
}
astat.ResourceUsage = sumTaskResourceUsage(flat)
return astat, nil
}
// sumTaskResourceUsage takes a set of task resources and sums their resources
func sumTaskResourceUsage(usages []*cstructs.TaskResourceUsage) *cstructs.ResourceUsage {
summed := &cstructs.ResourceUsage{
MemoryStats: &cstructs.MemoryStats{},
CpuStats: &cstructs.CpuStats{},
}
for _, usage := range usages {
summed.Add(usage.ResourceUsage)
}
return summed
}
2018-06-11 20:33:18 +00:00
// ShouldUpdate takes the AllocModifyIndex of an allocation sent from the server and
// checks if the current running allocation is behind and should be updated.
2018-06-11 20:33:18 +00:00
func (r *AllocRunner) ShouldUpdate(serverIndex uint64) bool {
r.allocLock.Lock()
defer r.allocLock.Unlock()
return r.alloc.AllocModifyIndex < serverIndex
}
2015-08-23 22:06:47 +00:00
// Destroy is used to indicate that the allocation context should be destroyed
2015-08-23 22:36:06 +00:00
func (r *AllocRunner) Destroy() {
// Lock when closing the context as that gives the save state code
// serialization.
r.allocStateLock.Lock()
defer r.allocStateLock.Unlock()
2017-07-03 22:03:42 +00:00
r.exitFn()
2017-07-03 04:49:56 +00:00
r.allocBroadcast.Close()
2015-08-23 22:06:47 +00:00
}
2015-10-04 20:36:03 +00:00
// IsDestroyed returns true if the AllocRunner is not running and has been
// destroyed (GC'd).
func (r *AllocRunner) IsDestroyed() bool {
select {
case <-r.waitCh:
return true
default:
return false
}
}
2015-10-04 20:36:03 +00:00
// WaitCh returns a channel to wait for termination
func (r *AllocRunner) WaitCh() <-chan struct{} {
return r.waitCh
}
2018-06-11 20:33:18 +00:00
// AllocID returns the allocation ID of the allocation being run
func (r *AllocRunner) AllocID() string {
if r == nil {
return ""
}
return r.allocID
}