357 lines
9.6 KiB
Go
357 lines
9.6 KiB
Go
package client
|
|
|
|
import (
|
|
"crypto/md5"
|
|
"encoding/hex"
|
|
"fmt"
|
|
"log"
|
|
"os"
|
|
"path/filepath"
|
|
"sync"
|
|
"time"
|
|
|
|
"github.com/hashicorp/nomad/client/config"
|
|
"github.com/hashicorp/nomad/client/driver"
|
|
"github.com/hashicorp/nomad/nomad/structs"
|
|
|
|
cstructs "github.com/hashicorp/nomad/client/driver/structs"
|
|
)
|
|
|
|
// TaskRunner is used to wrap a task within an allocation and provide the execution context.
|
|
type TaskRunner struct {
|
|
config *config.Config
|
|
updater TaskStateUpdater
|
|
logger *log.Logger
|
|
ctx *driver.ExecContext
|
|
alloc *structs.Allocation
|
|
restartTracker *RestartTracker
|
|
consulService *ConsulService
|
|
|
|
task *structs.Task
|
|
state *structs.TaskState
|
|
updateCh chan *structs.Task
|
|
handle driver.DriverHandle
|
|
|
|
destroy bool
|
|
destroyCh chan struct{}
|
|
destroyLock sync.Mutex
|
|
waitCh chan struct{}
|
|
|
|
snapshotLock sync.Mutex
|
|
}
|
|
|
|
// taskRunnerState is used to snapshot the state of the task runner
|
|
type taskRunnerState struct {
|
|
Task *structs.Task
|
|
HandleID string
|
|
}
|
|
|
|
// TaskStateUpdater is used to signal that tasks state has changed.
|
|
type TaskStateUpdater func(taskName string)
|
|
|
|
// NewTaskRunner is used to create a new task context
|
|
func NewTaskRunner(logger *log.Logger, config *config.Config,
|
|
updater TaskStateUpdater, ctx *driver.ExecContext,
|
|
alloc *structs.Allocation, task *structs.Task, state *structs.TaskState,
|
|
restartTracker *RestartTracker, consulService *ConsulService) *TaskRunner {
|
|
|
|
tc := &TaskRunner{
|
|
config: config,
|
|
updater: updater,
|
|
logger: logger,
|
|
restartTracker: restartTracker,
|
|
consulService: consulService,
|
|
ctx: ctx,
|
|
alloc: alloc,
|
|
task: task,
|
|
state: state,
|
|
updateCh: make(chan *structs.Task, 8),
|
|
destroyCh: make(chan struct{}),
|
|
waitCh: make(chan struct{}),
|
|
}
|
|
return tc
|
|
}
|
|
|
|
// WaitCh returns a channel to wait for termination
|
|
func (r *TaskRunner) WaitCh() <-chan struct{} {
|
|
return r.waitCh
|
|
}
|
|
|
|
// stateFilePath returns the path to our state file
|
|
func (r *TaskRunner) stateFilePath() string {
|
|
// Get the MD5 of the task name
|
|
hashVal := md5.Sum([]byte(r.task.Name))
|
|
hashHex := hex.EncodeToString(hashVal[:])
|
|
dirName := fmt.Sprintf("task-%s", hashHex)
|
|
|
|
// Generate the path
|
|
path := filepath.Join(r.config.StateDir, "alloc", r.alloc.ID,
|
|
dirName, "state.json")
|
|
return path
|
|
}
|
|
|
|
// RestoreState is used to restore our state
|
|
func (r *TaskRunner) RestoreState() error {
|
|
// Load the snapshot
|
|
var snap taskRunnerState
|
|
if err := restoreState(r.stateFilePath(), &snap); err != nil {
|
|
return err
|
|
}
|
|
|
|
// Restore fields
|
|
r.task = snap.Task
|
|
|
|
// Restore the driver
|
|
if snap.HandleID != "" {
|
|
driver, err := r.createDriver()
|
|
if err != nil {
|
|
return err
|
|
}
|
|
|
|
handle, err := driver.Open(r.ctx, snap.HandleID)
|
|
|
|
// In the case it fails, we relaunch the task in the Run() method.
|
|
if err != nil {
|
|
r.logger.Printf("[ERR] client: failed to open handle to task '%s' for alloc '%s': %v",
|
|
r.task.Name, r.alloc.ID, err)
|
|
return nil
|
|
}
|
|
r.handle = handle
|
|
}
|
|
return nil
|
|
}
|
|
|
|
// SaveState is used to snapshot our state
|
|
func (r *TaskRunner) SaveState() error {
|
|
r.snapshotLock.Lock()
|
|
defer r.snapshotLock.Unlock()
|
|
snap := taskRunnerState{
|
|
Task: r.task,
|
|
}
|
|
if r.handle != nil {
|
|
snap.HandleID = r.handle.ID()
|
|
}
|
|
return persistState(r.stateFilePath(), &snap)
|
|
}
|
|
|
|
// DestroyState is used to cleanup after ourselves
|
|
func (r *TaskRunner) DestroyState() error {
|
|
return os.RemoveAll(r.stateFilePath())
|
|
}
|
|
|
|
func (r *TaskRunner) appendEvent(event *structs.TaskEvent) {
|
|
capacity := 10
|
|
if r.state.Events == nil {
|
|
r.state.Events = make([]*structs.TaskEvent, 0, capacity)
|
|
}
|
|
|
|
// If we hit capacity, then shift it.
|
|
if len(r.state.Events) == capacity {
|
|
old := r.state.Events
|
|
r.state.Events = make([]*structs.TaskEvent, 0, capacity)
|
|
r.state.Events = append(r.state.Events, old[1:]...)
|
|
}
|
|
|
|
r.state.Events = append(r.state.Events, event)
|
|
}
|
|
|
|
// setState is used to update the state of the task runner
|
|
func (r *TaskRunner) setState(state string, event *structs.TaskEvent) {
|
|
// Update the task.
|
|
r.state.State = state
|
|
r.appendEvent(event)
|
|
|
|
// Persist our state to disk.
|
|
if err := r.SaveState(); err != nil {
|
|
r.logger.Printf("[ERR] client: failed to save state of Task Runner: %v", r.task.Name)
|
|
}
|
|
|
|
// Indicate the task has been updated.
|
|
r.updater(r.task.Name)
|
|
}
|
|
|
|
// createDriver makes a driver for the task
|
|
func (r *TaskRunner) createDriver() (driver.Driver, error) {
|
|
taskEnv, err := driver.GetTaskEnv(r.ctx.AllocDir, r.config.Node, r.task)
|
|
if err != nil {
|
|
err = fmt.Errorf("failed to create driver '%s' for alloc %s: %v",
|
|
r.task.Driver, r.alloc.ID, err)
|
|
r.logger.Printf("[ERR] client: %s", err)
|
|
return nil, err
|
|
|
|
}
|
|
|
|
driverCtx := driver.NewDriverContext(r.task.Name, r.config, r.config.Node, r.logger, taskEnv)
|
|
driver, err := driver.NewDriver(r.task.Driver, driverCtx)
|
|
if err != nil {
|
|
err = fmt.Errorf("failed to create driver '%s' for alloc %s: %v",
|
|
r.task.Driver, r.alloc.ID, err)
|
|
r.logger.Printf("[ERR] client: %s", err)
|
|
return nil, err
|
|
}
|
|
return driver, err
|
|
}
|
|
|
|
// startTask is used to start the task if there is no handle
|
|
func (r *TaskRunner) startTask() error {
|
|
// Create a driver
|
|
driver, err := r.createDriver()
|
|
if err != nil {
|
|
e := structs.NewTaskEvent(structs.TaskDriverFailure).SetDriverError(err)
|
|
r.setState(structs.TaskStateDead, e)
|
|
return err
|
|
}
|
|
|
|
// Start the job
|
|
handle, err := driver.Start(r.ctx, r.task)
|
|
if err != nil {
|
|
r.logger.Printf("[ERR] client: failed to start task '%s' for alloc '%s': %v",
|
|
r.task.Name, r.alloc.ID, err)
|
|
e := structs.NewTaskEvent(structs.TaskDriverFailure).
|
|
SetDriverError(fmt.Errorf("failed to start: %v", err))
|
|
r.setState(structs.TaskStateDead, e)
|
|
return err
|
|
}
|
|
r.handle = handle
|
|
r.setState(structs.TaskStateRunning, structs.NewTaskEvent(structs.TaskStarted))
|
|
return nil
|
|
}
|
|
|
|
// Run is a long running routine used to manage the task
|
|
func (r *TaskRunner) Run() {
|
|
defer close(r.waitCh)
|
|
r.logger.Printf("[DEBUG] client: starting task context for '%s' (alloc '%s')",
|
|
r.task.Name, r.alloc.ID)
|
|
|
|
r.run()
|
|
return
|
|
}
|
|
|
|
func (r *TaskRunner) run() {
|
|
var forceStart bool
|
|
for {
|
|
// Start the task if not yet started or it is being forced.
|
|
if r.handle == nil || forceStart {
|
|
forceStart = false
|
|
if err := r.startTask(); err != nil {
|
|
return
|
|
}
|
|
}
|
|
|
|
// Store the errors that caused use to stop waiting for updates.
|
|
var waitRes *cstructs.WaitResult
|
|
var destroyErr error
|
|
destroyed := false
|
|
|
|
// Register the services defined by the task with Consil
|
|
r.consulService.Register(r.task, r.alloc)
|
|
|
|
OUTER:
|
|
// Wait for updates
|
|
for {
|
|
select {
|
|
case waitRes = <-r.handle.WaitCh():
|
|
break OUTER
|
|
case update := <-r.updateCh:
|
|
// Update
|
|
r.task = update
|
|
if err := r.handle.Update(update); err != nil {
|
|
r.logger.Printf("[ERR] client: failed to update task '%s' for alloc '%s': %v", r.task.Name, r.alloc.ID, err)
|
|
}
|
|
case <-r.destroyCh:
|
|
// Avoid destroying twice
|
|
if destroyed {
|
|
continue
|
|
}
|
|
|
|
// Send the kill signal, and use the WaitCh to block until complete
|
|
if err := r.handle.Kill(); err != nil {
|
|
r.logger.Printf("[ERR] client: failed to kill task '%s' for alloc '%s': %v", r.task.Name, r.alloc.ID, err)
|
|
destroyErr = err
|
|
}
|
|
destroyed = true
|
|
}
|
|
}
|
|
|
|
// De-Register the services belonging to the task from consul
|
|
r.consulService.Deregister(r.task, r.alloc)
|
|
|
|
// If the user destroyed the task, we do not attempt to do any restarts.
|
|
if destroyed {
|
|
r.setState(structs.TaskStateDead, structs.NewTaskEvent(structs.TaskKilled).SetKillError(destroyErr))
|
|
return
|
|
}
|
|
|
|
// Log whether the task was successful or not.
|
|
if !waitRes.Successful() {
|
|
r.logger.Printf("[ERR] client: failed to complete task '%s' for alloc '%s': %v", r.task.Name, r.alloc.ID, waitRes)
|
|
} else {
|
|
r.logger.Printf("[INFO] client: completed task '%s' for alloc '%s'", r.task.Name, r.alloc.ID)
|
|
}
|
|
|
|
// Check if we should restart. If not mark task as dead and exit.
|
|
shouldRestart, when := r.restartTracker.NextRestart(waitRes.ExitCode)
|
|
waitEvent := r.waitErrorToEvent(waitRes)
|
|
if !shouldRestart {
|
|
r.logger.Printf("[INFO] client: Not restarting task: %v for alloc: %v ", r.task.Name, r.alloc.ID)
|
|
r.setState(structs.TaskStateDead, waitEvent)
|
|
return
|
|
}
|
|
|
|
r.logger.Printf("[INFO] client: Restarting Task: %v", r.task.Name)
|
|
r.logger.Printf("[DEBUG] client: Sleeping for %v before restarting Task %v", when, r.task.Name)
|
|
r.setState(structs.TaskStatePending, waitEvent)
|
|
|
|
// Sleep but watch for destroy events.
|
|
select {
|
|
case <-time.After(when):
|
|
case <-r.destroyCh:
|
|
}
|
|
|
|
// Destroyed while we were waiting to restart, so abort.
|
|
r.destroyLock.Lock()
|
|
destroyed = r.destroy
|
|
r.destroyLock.Unlock()
|
|
if destroyed {
|
|
r.logger.Printf("[DEBUG] client: Not restarting task: %v because it's destroyed by user", r.task.Name)
|
|
r.setState(structs.TaskStateDead, structs.NewTaskEvent(structs.TaskKilled))
|
|
return
|
|
}
|
|
|
|
// Set force start because we are restarting the task.
|
|
forceStart = true
|
|
}
|
|
|
|
return
|
|
}
|
|
|
|
// Helper function for converting a WaitResult into a TaskTerminated event.
|
|
func (r *TaskRunner) waitErrorToEvent(res *cstructs.WaitResult) *structs.TaskEvent {
|
|
return structs.NewTaskEvent(structs.TaskTerminated).
|
|
SetExitCode(res.ExitCode).
|
|
SetSignal(res.Signal).
|
|
SetExitMessage(res.Err)
|
|
}
|
|
|
|
// Update is used to update the task of the context
|
|
func (r *TaskRunner) Update(update *structs.Task) {
|
|
select {
|
|
case r.updateCh <- update:
|
|
default:
|
|
r.logger.Printf("[ERR] client: dropping task update '%s' (alloc '%s')",
|
|
update.Name, r.alloc.ID)
|
|
}
|
|
}
|
|
|
|
// Destroy is used to indicate that the task context should be destroyed
|
|
func (r *TaskRunner) Destroy() {
|
|
r.destroyLock.Lock()
|
|
defer r.destroyLock.Unlock()
|
|
|
|
if r.destroy {
|
|
return
|
|
}
|
|
r.destroy = true
|
|
close(r.destroyCh)
|
|
}
|