open-nomad/client/allochealth/tracker.go
Seth Hoenig b3ea68948b build: run gofmt on all go source files
Go 1.19 will forecefully format all your doc strings. To get this
out of the way, here is one big commit with all the changes gofmt
wants to make.
2022-08-16 11:14:11 -05:00

698 lines
19 KiB
Go

package allochealth
import (
"context"
"fmt"
"strings"
"sync"
"time"
"github.com/hashicorp/consul/api"
"github.com/hashicorp/go-hclog"
"github.com/hashicorp/nomad/client/serviceregistration"
"github.com/hashicorp/nomad/client/serviceregistration/checks/checkstore"
cstructs "github.com/hashicorp/nomad/client/structs"
"github.com/hashicorp/nomad/helper"
"github.com/hashicorp/nomad/nomad/structs"
)
const (
// AllocHealthEventSource is the source used for emitting task events
AllocHealthEventSource = "Alloc Unhealthy"
// checkLookupInterval is the pace at which we check if the Consul or Nomad
// checks for an allocation are healthy or unhealthy.
checkLookupInterval = 500 * time.Millisecond
)
// Tracker tracks the health of an allocation and makes health events watchable
// via channels.
type Tracker struct {
// ctx and cancelFn is used to shutdown the tracker
ctx context.Context
cancelFn context.CancelFunc
// alloc is the alloc we are tracking
alloc *structs.Allocation
// tg is the task group we are tracking
tg *structs.TaskGroup
// minHealthyTime is the duration an alloc must remain healthy to be
// considered healthy
minHealthyTime time.Duration
// checkLookupInterval is the repeated interval after which which we check
// if the Consul checks are healthy or unhealthy.
checkLookupInterval time.Duration
// useChecks specifies whether to consider Consul and Nomad service checks.
useChecks bool
// consulCheckCount is the total number of Consul service checks in the task
// group including task level checks.
consulCheckCount int
// nomadCheckCount is the total the number of Nomad service checks in the task
// group including task level checks.
nomadCheckCount int
// allocUpdates is a listener for retrieving new alloc updates
allocUpdates *cstructs.AllocListener
// consulClient is used to look up the status of Consul service checks
consulClient serviceregistration.Handler
// checkStore is used to lookup the status of Nomad service checks
checkStore checkstore.Shim
// healthy is used to signal whether we have determined the allocation to be
// healthy or unhealthy
healthy chan bool
// allocStopped is triggered when the allocation is stopped and tracking is
// not needed
allocStopped chan struct{}
// lifecycleTasks is a map of ephemeral tasks and their lifecycle hooks.
// These tasks may terminate without affecting alloc health
lifecycleTasks map[string]string
// lock is used to lock shared fields listed below
lock sync.Mutex
// tasksHealthy marks whether all the tasks have met their health check
// (disregards Consul and Nomad checks)
tasksHealthy bool
// allocFailed marks whether the allocation failed
allocFailed bool
// checksHealthy marks whether all the task's Consul checks are healthy
checksHealthy bool
// taskHealth contains the health state for each task in the allocation
// name -> state
taskHealth map[string]*taskHealthState
// logger is for logging things
logger hclog.Logger
}
// NewTracker returns a health tracker for the given allocation.
//
// Depending on job configuration, an allocation's health takes into consideration
// - An alloc listener
// - Consul checks (via consul API)
// - Nomad checks (via client state)
func NewTracker(
parentCtx context.Context,
logger hclog.Logger,
alloc *structs.Allocation,
allocUpdates *cstructs.AllocListener,
consulClient serviceregistration.Handler,
checkStore checkstore.Shim,
minHealthyTime time.Duration,
useChecks bool,
) *Tracker {
t := &Tracker{
healthy: make(chan bool, 1),
allocStopped: make(chan struct{}),
alloc: alloc,
tg: alloc.Job.LookupTaskGroup(alloc.TaskGroup),
minHealthyTime: minHealthyTime,
useChecks: useChecks,
allocUpdates: allocUpdates,
consulClient: consulClient,
checkStore: checkStore,
checkLookupInterval: checkLookupInterval,
logger: logger,
lifecycleTasks: map[string]string{},
}
t.taskHealth = make(map[string]*taskHealthState, len(t.tg.Tasks))
for _, task := range t.tg.Tasks {
t.taskHealth[task.Name] = &taskHealthState{task: task}
if task.Lifecycle != nil && !task.Lifecycle.Sidecar {
t.lifecycleTasks[task.Name] = task.Lifecycle.Hook
}
c, n := countChecks(task.Services)
t.consulCheckCount += c
t.nomadCheckCount += n
}
c, n := countChecks(t.tg.Services)
t.consulCheckCount += c
t.nomadCheckCount += n
t.ctx, t.cancelFn = context.WithCancel(parentCtx)
return t
}
func countChecks(services []*structs.Service) (consul, nomad int) {
for _, service := range services {
switch service.Provider {
case structs.ServiceProviderNomad:
nomad += len(service.Checks)
default:
consul += len(service.Checks)
}
}
return
}
// Start starts the watcher.
func (t *Tracker) Start() {
go t.watchTaskEvents()
switch {
case !t.useChecks:
return
case t.consulCheckCount > 0:
go t.watchConsulEvents()
case t.nomadCheckCount > 0:
go t.watchNomadEvents()
}
}
// HealthyCh returns a channel that will emit a boolean indicating the health of
// the allocation.
func (t *Tracker) HealthyCh() <-chan bool {
return t.healthy
}
// AllocStoppedCh returns a channel that will be fired if the allocation is
// stopped. This means that health will not be set.
func (t *Tracker) AllocStoppedCh() <-chan struct{} {
return t.allocStopped
}
// TaskEvents returns a map of events by task. This should only be called after
// health has been determined. Only tasks that have contributed to the
// allocation being unhealthy will have an event.
func (t *Tracker) TaskEvents() map[string]*structs.TaskEvent {
t.lock.Lock()
defer t.lock.Unlock()
// Nothing to do since the failure wasn't task related
if t.allocFailed {
return nil
}
deadline, _ := t.ctx.Deadline()
events := make(map[string]*structs.TaskEvent, len(t.tg.Tasks))
// Go through are task information and build the event map
for task, state := range t.taskHealth {
useChecks := t.tg.Update.HealthCheck == structs.UpdateStrategyHealthCheck_Checks
if e, ok := state.event(deadline, t.tg.Update.HealthyDeadline, t.tg.Update.MinHealthyTime, useChecks); ok {
events[task] = structs.NewTaskEvent(AllocHealthEventSource).SetMessage(e)
}
}
return events
}
// setTaskHealth is used to set the tasks health as healthy or unhealthy. If the
// allocation is terminal, health is immediately broadcast.
func (t *Tracker) setTaskHealth(healthy, terminal bool) {
t.lock.Lock()
defer t.lock.Unlock()
t.tasksHealthy = healthy
// if unhealthy, force waiting for new checks health status
if !terminal && !healthy {
t.checksHealthy = false
return
}
// If we are marked healthy but we also require Consul checks to be healthy
// and they are not yet, return, unless the task is terminal.
usesConsulChecks := t.useChecks && t.consulCheckCount > 0
if !terminal && healthy && usesConsulChecks && !t.checksHealthy {
return
}
// If we are marked healthy but also require Nomad checks to be healthy and
// they are not yet, return, unless the task is terminal.
usesNomadChecks := t.useChecks && t.nomadCheckCount > 0
if !terminal && healthy && usesNomadChecks && !t.checksHealthy {
return
}
select {
case t.healthy <- healthy:
// nothing
default:
}
// Shutdown the tracker
t.cancelFn()
}
// setCheckHealth is used to mark the checks as either healthy or unhealthy.
// returns true if health is propagated and no more health monitoring is needed
//
// todo: this is currently being shared by watchConsulEvents and watchNomadEvents,
//
// and must be split up if/when we support registering services (and thus checks)
// of different providers.
func (t *Tracker) setCheckHealth(healthy bool) bool {
t.lock.Lock()
defer t.lock.Unlock()
// check health should always be false if tasks are unhealthy
// as checks might be missing from unhealthy tasks
t.checksHealthy = healthy && t.tasksHealthy
// Only signal if we are healthy and so is the tasks
if !t.checksHealthy {
return false
}
select {
case t.healthy <- healthy:
// nothing
default:
}
// Shutdown the tracker, things are healthy so nothing to do
t.cancelFn()
return true
}
// markAllocStopped is used to mark the allocation as having stopped.
func (t *Tracker) markAllocStopped() {
close(t.allocStopped)
t.cancelFn()
}
// watchTaskEvents is a long lived watcher that watches for the health of the
// allocation's tasks.
func (t *Tracker) watchTaskEvents() {
alloc := t.alloc
allStartedTime := time.Time{}
waiter := newHealthyFuture()
for {
// If the alloc is being stopped by the server just exit
switch alloc.DesiredStatus {
case structs.AllocDesiredStatusStop, structs.AllocDesiredStatusEvict:
t.logger.Trace("desired status is terminal for alloc", "alloc_id", alloc.ID, "desired_status", alloc.DesiredStatus)
t.markAllocStopped()
return
}
// Store the task states
t.lock.Lock()
for task, state := range alloc.TaskStates {
//TODO(schmichael) for now skip unknown tasks as
//they're task group services which don't currently
//support checks anyway
if v, ok := t.taskHealth[task]; ok {
v.state = state
}
}
t.lock.Unlock()
// Detect if the alloc is unhealthy or if all tasks have started yet
latestStartTime := time.Time{}
for taskName, state := range alloc.TaskStates {
// If the task is a poststop task we do not want to evaluate it
// since it will remain pending until the main task has finished
// or exited.
if t.lifecycleTasks[taskName] == structs.TaskLifecycleHookPoststop {
continue
}
// If this is a poststart task which has already succeeded, we
// should skip evaluation.
if t.lifecycleTasks[taskName] == structs.TaskLifecycleHookPoststart && state.Successful() {
continue
}
// One of the tasks has failed so we can exit watching
if state.Failed || (!state.FinishedAt.IsZero() && t.lifecycleTasks[taskName] != structs.TaskLifecycleHookPrestart) {
t.setTaskHealth(false, true)
return
}
if state.State == structs.TaskStatePending {
latestStartTime = time.Time{}
break
} else if state.StartedAt.After(latestStartTime) {
// task is either running or exited successfully
latestStartTime = state.StartedAt
}
}
// If the alloc is marked as failed by the client but none of the
// individual tasks failed, that means something failed at the alloc
// level.
if alloc.ClientStatus == structs.AllocClientStatusFailed {
t.lock.Lock()
t.allocFailed = true
t.lock.Unlock()
t.setTaskHealth(false, true)
return
}
if !latestStartTime.Equal(allStartedTime) {
// reset task health
t.setTaskHealth(false, false)
// Prevent the timer from firing at the old start time
waiter.disable()
// Set the timer since all tasks are started
if !latestStartTime.IsZero() {
allStartedTime = latestStartTime
waiter.wait(t.minHealthyTime)
}
}
select {
case <-t.ctx.Done():
return
case newAlloc, ok := <-t.allocUpdates.Ch():
if !ok {
return
}
alloc = newAlloc
case <-waiter.C():
t.setTaskHealth(true, false)
}
}
}
// healthyFuture is used to fire after checks have been healthy for MinHealthyTime
type healthyFuture struct {
timer *time.Timer
}
// newHealthyFuture will create a healthyFuture in a disabled state, and
// will do nothing until a call to wait takes place
func newHealthyFuture() *healthyFuture {
timer := time.NewTimer(0)
ht := &healthyFuture{timer: timer}
ht.disable()
return ht
}
// disable the healthyFuture from triggering
func (h *healthyFuture) disable() {
if !h.timer.Stop() {
// must ensure channel is clear
// https://pkg.go.dev/time#Timer.Stop
select {
case <-h.timer.C:
default:
}
}
}
// wait will reset the healthyFuture to trigger after dur passes.
func (h *healthyFuture) wait(dur time.Duration) {
// must ensure timer is stopped
// https://pkg.go.dev/time#Timer.Reset
h.disable()
h.timer.Reset(dur)
}
// C returns a channel on which the future will send when ready.
func (h *healthyFuture) C() <-chan time.Time {
return h.timer.C
}
// watchConsulEvents is a watcher for the health of the allocation's Consul
// checks. If all checks report healthy the watcher will exit after the
// MinHealthyTime has been reached, otherwise the watcher will continue to
// check unhealthy checks until the ctx is cancelled.
//
// Does not watch Nomad service checks; see watchNomadEvents for those.
func (t *Tracker) watchConsulEvents() {
// checkTicker is the ticker that triggers us to look at the checks in Consul
checkTicker := time.NewTicker(t.checkLookupInterval)
defer checkTicker.Stop()
// waiter is used to fire when the checks have been healthy for the MinHealthyTime
waiter := newHealthyFuture()
// primed marks whether the healthy waiter has been set
primed := false
// Store whether the last Consul checks call was successful or not
consulChecksErr := false
// allocReg are the registered objects in Consul for the allocation
var allocReg *serviceregistration.AllocRegistration
OUTER:
for {
select {
// we are shutting down
case <-t.ctx.Done():
return
// it is time to check the checks
case <-checkTicker.C:
newAllocReg, err := t.consulClient.AllocRegistrations(t.alloc.ID)
if err != nil {
if !consulChecksErr {
consulChecksErr = true
t.logger.Warn("error looking up Consul registrations for allocation", "error", err, "alloc_id", t.alloc.ID)
}
continue OUTER
} else {
consulChecksErr = false
allocReg = newAllocReg
}
// enough time has passed with healthy checks
case <-waiter.C():
if t.setCheckHealth(true) {
// final health set and propagated
return
}
// checks are healthy but tasks are unhealthy,
// reset and wait until all is healthy
primed = false
}
if allocReg == nil {
continue
}
// Store the task registrations
t.lock.Lock()
for task, reg := range allocReg.Tasks {
if v, ok := t.taskHealth[task]; ok {
v.taskRegistrations = reg
}
}
t.lock.Unlock()
// Detect if all the checks are passing
passed := true
CHECKS:
for _, treg := range allocReg.Tasks {
for _, sreg := range treg.Services {
for _, check := range sreg.Checks {
onUpdate := sreg.CheckOnUpdate[check.CheckID]
switch check.Status {
case api.HealthPassing:
continue
case api.HealthWarning:
if onUpdate == structs.OnUpdateIgnoreWarn || onUpdate == structs.OnUpdateIgnore {
continue
}
case api.HealthCritical:
if onUpdate == structs.OnUpdateIgnore {
continue
}
default:
}
passed = false
t.setCheckHealth(false)
break CHECKS
}
}
}
if !passed {
// Reset the timer since we have transitioned back to unhealthy
if primed {
primed = false
waiter.disable()
}
} else if !primed {
// Reset the timer to fire after MinHealthyTime
primed = true
waiter.disable()
waiter.wait(t.minHealthyTime)
}
}
}
// watchNomadEvents is a watcher for the health of the allocation's Nomad checks.
// If all checks report healthy the watcher will exit after the MinHealthyTime has
// been reached, otherwise the watcher will continue to check unhealthy checks until
// the ctx is cancelled.
//
// Does not watch Consul service checks; see watchConsulEvents for those.
func (t *Tracker) watchNomadEvents() {
// checkTicker is the ticker that triggers us to look at the checks in Nomad
checkTicker, cancel := helper.NewSafeTimer(t.checkLookupInterval)
defer cancel()
// waiter is used to fire when the checks have been healthy for the MinHealthyTime
waiter := newHealthyFuture()
// allocID of the allocation we are watching checks for
allocID := t.alloc.ID
// primed marks whether the healthy waiter has been set
primed := false
// latest set of nomad check results
var results map[structs.CheckID]*structs.CheckQueryResult
for {
select {
// tracker has been canceled, so stop waiting
case <-t.ctx.Done():
return
// it is time to check the checks
case <-checkTicker.C:
results = t.checkStore.List(allocID)
checkTicker.Reset(t.checkLookupInterval)
// enough time has passed with healthy checks
case <-waiter.C():
if t.setCheckHealth(true) { // todo(shoenig) this needs to be split between Consul and Nomad
return // final health set and propagated
}
// checks are healthy but tasks are unhealthy, reset and wait
primed = false
}
// scan to see if any checks are failing
passing := true
for _, result := range results {
switch result.Status {
case structs.CheckSuccess:
continue
case structs.CheckFailure:
if result.Mode == structs.Readiness {
continue
}
passing = false
default:
// i.e. pending check; do not consider healthy or ready
passing = false
}
if !passing {
break // 1+ check is failing; no need to continue
}
}
if !passing {
// at least one check is failing, transition to unhealthy
t.setCheckHealth(false)
primed = false
waiter.disable()
}
if passing && !primed {
// healthy but not yet primed, set timer to wait
primed = true
waiter.wait(t.minHealthyTime)
}
}
}
// taskHealthState captures all known health information about a task. It is
// largely used to determine if the task has contributed to the allocation being
// unhealthy.
type taskHealthState struct {
task *structs.Task
state *structs.TaskState
taskRegistrations *serviceregistration.ServiceRegistrations
}
// event takes the deadline time for the allocation to be healthy and the update
// strategy of the group. It returns true if the task has contributed to the
// allocation being unhealthy and if so, an event description of why.
func (t *taskHealthState) event(deadline time.Time, healthyDeadline, minHealthyTime time.Duration, useChecks bool) (string, bool) {
desiredChecks := 0
for _, s := range t.task.Services {
if nc := len(s.Checks); nc > 0 {
desiredChecks += nc
}
}
requireChecks := (desiredChecks > 0) && useChecks
if t.state != nil {
if t.state.Failed {
return "Unhealthy because of failed task", true
}
switch t.state.State {
case structs.TaskStatePending:
return fmt.Sprintf("Task not running by healthy_deadline of %v", healthyDeadline), true
case structs.TaskStateDead:
// non-sidecar hook lifecycle tasks are healthy if they exit with success
if t.task.Lifecycle == nil || t.task.Lifecycle.Sidecar {
return "Unhealthy because of dead task", true
}
case structs.TaskStateRunning:
// We are running so check if we have been running long enough
if t.state.StartedAt.Add(minHealthyTime).After(deadline) {
return fmt.Sprintf("Task not running for min_healthy_time of %v by healthy_deadline of %v", minHealthyTime, healthyDeadline), true
}
}
}
if t.taskRegistrations != nil {
var notPassing []string
passing := 0
OUTER:
for _, sreg := range t.taskRegistrations.Services {
for _, check := range sreg.Checks {
if check.Status != api.HealthPassing {
notPassing = append(notPassing, sreg.Service.Service)
continue OUTER
} else {
passing++
}
}
}
if len(notPassing) != 0 {
return fmt.Sprintf("Services not healthy by deadline: %s", strings.Join(notPassing, ", ")), true
}
if passing != desiredChecks {
return fmt.Sprintf("Only %d out of %d checks registered and passing", passing, desiredChecks), true
}
} else if requireChecks {
return "Service checks not registered", true
}
return "", false
}