e552e1726f
Update `go-metrics` to v0.5.3 to pick https://github.com/hashicorp/go-metrics/pull/146.
265 lines
7.8 KiB
Go
265 lines
7.8 KiB
Go
// Copyright (c) HashiCorp, Inc.
|
|
// SPDX-License-Identifier: MPL-2.0
|
|
|
|
package nomad
|
|
|
|
import (
|
|
"errors"
|
|
"sync"
|
|
"time"
|
|
|
|
log "github.com/hashicorp/go-hclog"
|
|
memdb "github.com/hashicorp/go-memdb"
|
|
metrics "github.com/hashicorp/go-metrics"
|
|
|
|
"github.com/hashicorp/nomad/helper"
|
|
"github.com/hashicorp/nomad/nomad/structs"
|
|
)
|
|
|
|
const (
|
|
// heartbeatNotLeader is the error string returned when the heartbeat request
|
|
// couldn't be completed since the server is not the leader.
|
|
heartbeatNotLeader = "failed to reset heartbeat since server is not leader"
|
|
|
|
// NodeHeartbeatEventMissed is the event used when the Nodes heartbeat is
|
|
// missed.
|
|
NodeHeartbeatEventMissed = "Node heartbeat missed"
|
|
)
|
|
|
|
var (
|
|
// heartbeatNotLeaderErr is the error returned when the heartbeat request
|
|
// couldn't be completed since the server is not the leader.
|
|
heartbeatNotLeaderErr = errors.New(heartbeatNotLeader)
|
|
)
|
|
|
|
// nodeHeartbeater is used to track expiration times of node heartbeats. If it
|
|
// detects an expired node, the node status is updated to be 'down'.
|
|
type nodeHeartbeater struct {
|
|
srv *Server
|
|
logger log.Logger
|
|
|
|
// heartbeatTimers track the expiration time of each heartbeat that has
|
|
// a TTL. On expiration, the node status is updated to be 'down'.
|
|
heartbeatTimers map[string]*time.Timer
|
|
heartbeatTimersLock sync.Mutex
|
|
}
|
|
|
|
// newNodeHeartbeater returns a new node heartbeater used to detect and act on
|
|
// failed node heartbeats.
|
|
func newNodeHeartbeater(s *Server) *nodeHeartbeater {
|
|
return &nodeHeartbeater{
|
|
srv: s,
|
|
logger: s.logger.Named("heartbeat"),
|
|
}
|
|
}
|
|
|
|
// initializeHeartbeatTimers is used when a leader is newly elected to create
|
|
// a new map to track heartbeat expiration and to reset all the timers from
|
|
// the previously known set of timers.
|
|
func (h *nodeHeartbeater) initializeHeartbeatTimers() error {
|
|
// Scan all nodes and reset their timer
|
|
snap, err := h.srv.fsm.State().Snapshot()
|
|
if err != nil {
|
|
return err
|
|
}
|
|
|
|
// Get an iterator over nodes
|
|
ws := memdb.NewWatchSet()
|
|
iter, err := snap.Nodes(ws)
|
|
if err != nil {
|
|
return err
|
|
}
|
|
|
|
h.heartbeatTimersLock.Lock()
|
|
defer h.heartbeatTimersLock.Unlock()
|
|
|
|
// Handle each node
|
|
for {
|
|
raw := iter.Next()
|
|
if raw == nil {
|
|
break
|
|
}
|
|
node := raw.(*structs.Node)
|
|
if node.TerminalStatus() {
|
|
continue
|
|
}
|
|
h.resetHeartbeatTimerLocked(node.ID, h.srv.config.FailoverHeartbeatTTL)
|
|
}
|
|
return nil
|
|
}
|
|
|
|
// resetHeartbeatTimer is used to reset the TTL of a heartbeat.
|
|
// This can be used for new heartbeats and existing ones.
|
|
func (h *nodeHeartbeater) resetHeartbeatTimer(id string) (time.Duration, error) {
|
|
h.heartbeatTimersLock.Lock()
|
|
defer h.heartbeatTimersLock.Unlock()
|
|
|
|
// Do not create a timer for the node since we are not the leader. This
|
|
// check avoids the race in which leadership is lost but a timer is created
|
|
// on this server since it was servicing an RPC during a leadership loss.
|
|
if !h.srv.IsLeader() {
|
|
h.logger.Debug("ignoring resetting node TTL since this server is not the leader", "node_id", id)
|
|
return 0, heartbeatNotLeaderErr
|
|
}
|
|
|
|
// Compute the target TTL value
|
|
n := len(h.heartbeatTimers)
|
|
ttl := helper.RateScaledInterval(h.srv.config.MaxHeartbeatsPerSecond, h.srv.config.MinHeartbeatTTL, n)
|
|
ttl += helper.RandomStagger(ttl)
|
|
|
|
// Reset the TTL
|
|
h.resetHeartbeatTimerLocked(id, ttl+h.srv.config.HeartbeatGrace)
|
|
return ttl, nil
|
|
}
|
|
|
|
// resetHeartbeatTimerLocked is used to reset a heartbeat timer
|
|
// assuming the heartbeatTimerLock is already held
|
|
func (h *nodeHeartbeater) resetHeartbeatTimerLocked(id string, ttl time.Duration) {
|
|
// Ensure a timer map exists
|
|
if h.heartbeatTimers == nil {
|
|
h.heartbeatTimers = make(map[string]*time.Timer)
|
|
}
|
|
|
|
// Renew the heartbeat timer if it exists
|
|
if timer, ok := h.heartbeatTimers[id]; ok {
|
|
timer.Reset(ttl)
|
|
return
|
|
}
|
|
|
|
// Create a new timer to track expiration of this heartbeat
|
|
timer := time.AfterFunc(ttl, func() {
|
|
h.invalidateHeartbeat(id)
|
|
})
|
|
h.heartbeatTimers[id] = timer
|
|
}
|
|
|
|
// invalidateHeartbeat is invoked when a heartbeat TTL is reached and we
|
|
// need to invalidate the heartbeat.
|
|
func (h *nodeHeartbeater) invalidateHeartbeat(id string) {
|
|
defer metrics.MeasureSince([]string{"nomad", "heartbeat", "invalidate"}, time.Now())
|
|
// Clear the heartbeat timer
|
|
h.heartbeatTimersLock.Lock()
|
|
if timer, ok := h.heartbeatTimers[id]; ok {
|
|
timer.Stop()
|
|
delete(h.heartbeatTimers, id)
|
|
}
|
|
h.heartbeatTimersLock.Unlock()
|
|
|
|
// Do not invalidate the node since we are not the leader. This check avoids
|
|
// the race in which leadership is lost but a timer is created on this
|
|
// server since it was servicing an RPC during a leadership loss.
|
|
if !h.srv.IsLeader() {
|
|
h.logger.Debug("ignoring node TTL since this server is not the leader", "node_id", id)
|
|
return
|
|
}
|
|
|
|
h.logger.Warn("node TTL expired", "node_id", id)
|
|
|
|
canDisconnect, hasPendingReconnects := h.disconnectState(id)
|
|
|
|
// Make a request to update the node status
|
|
req := structs.NodeUpdateStatusRequest{
|
|
NodeID: id,
|
|
Status: structs.NodeStatusDown,
|
|
NodeEvent: structs.NewNodeEvent().SetSubsystem(structs.NodeEventSubsystemCluster).SetMessage(NodeHeartbeatEventMissed),
|
|
WriteRequest: structs.WriteRequest{
|
|
Region: h.srv.config.Region,
|
|
},
|
|
}
|
|
|
|
if canDisconnect && hasPendingReconnects {
|
|
req.Status = structs.NodeStatusDisconnected
|
|
}
|
|
var resp structs.NodeUpdateResponse
|
|
|
|
if err := h.srv.RPC("Node.UpdateStatus", &req, &resp); err != nil {
|
|
h.logger.Error("update node status failed", "error", err)
|
|
}
|
|
}
|
|
|
|
func (h *nodeHeartbeater) disconnectState(id string) (bool, bool) {
|
|
node, err := h.srv.State().NodeByID(nil, id)
|
|
if err != nil {
|
|
h.logger.Error("error retrieving node by id", "error", err)
|
|
return false, false
|
|
}
|
|
|
|
// Exit if the node is already down or just initializing.
|
|
if node.Status == structs.NodeStatusDown || node.Status == structs.NodeStatusInit {
|
|
return false, false
|
|
}
|
|
|
|
allocs, err := h.srv.State().AllocsByNode(nil, id)
|
|
if err != nil {
|
|
h.logger.Error("error retrieving allocs by node", "error", err)
|
|
return false, false
|
|
}
|
|
|
|
now := time.Now().UTC()
|
|
// Check if the node has any allocs that are configured with max_client_disconnect,
|
|
// that are past the disconnect window, and if so, whether it has at least one
|
|
// alloc that isn't yet expired.
|
|
nodeCanDisconnect := false
|
|
for _, alloc := range allocs {
|
|
allocCanDisconnect := alloc.DisconnectTimeout(now).After(now)
|
|
// Only process this until we find that at least one alloc is configured
|
|
// with max_client_disconnect.
|
|
if !nodeCanDisconnect && allocCanDisconnect {
|
|
nodeCanDisconnect = true
|
|
}
|
|
// Only process this until we find one that we want to run and has not
|
|
// yet expired.
|
|
if allocCanDisconnect &&
|
|
alloc.DesiredStatus == structs.AllocDesiredStatusRun &&
|
|
!alloc.Expired(now) {
|
|
return true, true
|
|
}
|
|
}
|
|
|
|
return nodeCanDisconnect, false
|
|
}
|
|
|
|
// clearHeartbeatTimer is used to clear the heartbeat time for
|
|
// a single heartbeat. This is used when a heartbeat is destroyed
|
|
// explicitly and no longer needed.
|
|
func (h *nodeHeartbeater) clearHeartbeatTimer(id string) error {
|
|
h.heartbeatTimersLock.Lock()
|
|
defer h.heartbeatTimersLock.Unlock()
|
|
|
|
if timer, ok := h.heartbeatTimers[id]; ok {
|
|
timer.Stop()
|
|
delete(h.heartbeatTimers, id)
|
|
}
|
|
return nil
|
|
}
|
|
|
|
// clearAllHeartbeatTimers is used when a leader is stepping
|
|
// down and we no longer need to track any heartbeat timers.
|
|
func (h *nodeHeartbeater) clearAllHeartbeatTimers() error {
|
|
h.heartbeatTimersLock.Lock()
|
|
defer h.heartbeatTimersLock.Unlock()
|
|
|
|
for _, t := range h.heartbeatTimers {
|
|
t.Stop()
|
|
}
|
|
h.heartbeatTimers = nil
|
|
return nil
|
|
}
|
|
|
|
// heartbeatStats is a long running routine used to capture
|
|
// the number of active heartbeats being tracked
|
|
func (h *nodeHeartbeater) heartbeatStats() {
|
|
for {
|
|
select {
|
|
case <-time.After(5 * time.Second):
|
|
h.heartbeatTimersLock.Lock()
|
|
num := len(h.heartbeatTimers)
|
|
h.heartbeatTimersLock.Unlock()
|
|
metrics.SetGauge([]string{"nomad", "heartbeat", "active"}, float32(num))
|
|
|
|
case <-h.srv.shutdownCh:
|
|
return
|
|
}
|
|
}
|
|
}
|