49b7d23cea
When summing up the capability of your Nomad fleet for scaling purposes, it's important to exclude draining nodes, as they won't accept new jobs.
2910 lines
88 KiB
Go
2910 lines
88 KiB
Go
package client
|
|
|
|
import (
|
|
"errors"
|
|
"fmt"
|
|
"io/ioutil"
|
|
"net"
|
|
"net/rpc"
|
|
"os"
|
|
"path/filepath"
|
|
"sort"
|
|
"strconv"
|
|
"strings"
|
|
"sync"
|
|
"time"
|
|
|
|
metrics "github.com/armon/go-metrics"
|
|
consulapi "github.com/hashicorp/consul/api"
|
|
"github.com/hashicorp/consul/lib"
|
|
hclog "github.com/hashicorp/go-hclog"
|
|
multierror "github.com/hashicorp/go-multierror"
|
|
"github.com/hashicorp/nomad/client/allocdir"
|
|
"github.com/hashicorp/nomad/client/allocrunner"
|
|
"github.com/hashicorp/nomad/client/allocrunner/interfaces"
|
|
arstate "github.com/hashicorp/nomad/client/allocrunner/state"
|
|
"github.com/hashicorp/nomad/client/allocwatcher"
|
|
"github.com/hashicorp/nomad/client/config"
|
|
consulApi "github.com/hashicorp/nomad/client/consul"
|
|
"github.com/hashicorp/nomad/client/devicemanager"
|
|
"github.com/hashicorp/nomad/client/fingerprint"
|
|
"github.com/hashicorp/nomad/client/pluginmanager"
|
|
"github.com/hashicorp/nomad/client/pluginmanager/drivermanager"
|
|
"github.com/hashicorp/nomad/client/servers"
|
|
"github.com/hashicorp/nomad/client/state"
|
|
"github.com/hashicorp/nomad/client/stats"
|
|
cstructs "github.com/hashicorp/nomad/client/structs"
|
|
"github.com/hashicorp/nomad/client/vaultclient"
|
|
"github.com/hashicorp/nomad/command/agent/consul"
|
|
"github.com/hashicorp/nomad/helper"
|
|
"github.com/hashicorp/nomad/helper/pool"
|
|
hstats "github.com/hashicorp/nomad/helper/stats"
|
|
"github.com/hashicorp/nomad/helper/tlsutil"
|
|
"github.com/hashicorp/nomad/helper/uuid"
|
|
"github.com/hashicorp/nomad/nomad/structs"
|
|
nconfig "github.com/hashicorp/nomad/nomad/structs/config"
|
|
"github.com/hashicorp/nomad/plugins/device"
|
|
"github.com/hashicorp/nomad/plugins/drivers"
|
|
vaultapi "github.com/hashicorp/vault/api"
|
|
"github.com/shirou/gopsutil/host"
|
|
)
|
|
|
|
const (
|
|
// clientRPCCache controls how long we keep an idle connection
|
|
// open to a server
|
|
clientRPCCache = 5 * time.Minute
|
|
|
|
// clientMaxStreams controls how many idle streams we keep
|
|
// open to a server
|
|
clientMaxStreams = 2
|
|
|
|
// datacenterQueryLimit searches through up to this many adjacent
|
|
// datacenters looking for the Nomad server service.
|
|
datacenterQueryLimit = 9
|
|
|
|
// registerRetryIntv is minimum interval on which we retry
|
|
// registration. We pick a value between this and 2x this.
|
|
registerRetryIntv = 15 * time.Second
|
|
|
|
// getAllocRetryIntv is minimum interval on which we retry
|
|
// to fetch allocations. We pick a value between this and 2x this.
|
|
getAllocRetryIntv = 30 * time.Second
|
|
|
|
// devModeRetryIntv is the retry interval used for development
|
|
devModeRetryIntv = time.Second
|
|
|
|
// stateSnapshotIntv is how often the client snapshots state
|
|
stateSnapshotIntv = 60 * time.Second
|
|
|
|
// initialHeartbeatStagger is used to stagger the interval between
|
|
// starting and the initial heartbeat. After the initial heartbeat,
|
|
// we switch to using the TTL specified by the servers.
|
|
initialHeartbeatStagger = 10 * time.Second
|
|
|
|
// nodeUpdateRetryIntv is how often the client checks for updates to the
|
|
// node attributes or meta map.
|
|
nodeUpdateRetryIntv = 5 * time.Second
|
|
|
|
// allocSyncIntv is the batching period of allocation updates before they
|
|
// are synced with the server.
|
|
allocSyncIntv = 200 * time.Millisecond
|
|
|
|
// allocSyncRetryIntv is the interval on which we retry updating
|
|
// the status of the allocation
|
|
allocSyncRetryIntv = 5 * time.Second
|
|
|
|
// defaultConnectSidecarImage is the image set in the node meta by default
|
|
// to be used by Consul Connect sidecar tasks
|
|
defaultConnectSidecarImage = "envoyproxy/envoy:v1.11.1"
|
|
|
|
// defaultConnectLogLevel is the log level set in the node meta by default
|
|
// to be used by Consul Connect sidecar tasks
|
|
defaultConnectLogLevel = "info"
|
|
)
|
|
|
|
var (
|
|
// grace period to allow for batch fingerprint processing
|
|
batchFirstFingerprintsProcessingGrace = batchFirstFingerprintsTimeout + 5*time.Second
|
|
)
|
|
|
|
// ClientStatsReporter exposes all the APIs related to resource usage of a Nomad
|
|
// Client
|
|
type ClientStatsReporter interface {
|
|
// GetAllocStats returns the AllocStatsReporter for the passed allocation.
|
|
// If it does not exist an error is reported.
|
|
GetAllocStats(allocID string) (interfaces.AllocStatsReporter, error)
|
|
|
|
// LatestHostStats returns the latest resource usage stats for the host
|
|
LatestHostStats() *stats.HostStats
|
|
}
|
|
|
|
// AllocRunner is the interface implemented by the core alloc runner.
|
|
//TODO Create via factory to allow testing Client with mock AllocRunners.
|
|
type AllocRunner interface {
|
|
Alloc() *structs.Allocation
|
|
AllocState() *arstate.State
|
|
Destroy()
|
|
Shutdown()
|
|
GetAllocDir() *allocdir.AllocDir
|
|
IsDestroyed() bool
|
|
IsMigrating() bool
|
|
IsWaiting() bool
|
|
Listener() *cstructs.AllocListener
|
|
Restore() error
|
|
Run()
|
|
StatsReporter() interfaces.AllocStatsReporter
|
|
Update(*structs.Allocation)
|
|
WaitCh() <-chan struct{}
|
|
DestroyCh() <-chan struct{}
|
|
ShutdownCh() <-chan struct{}
|
|
Signal(taskName, signal string) error
|
|
GetTaskEventHandler(taskName string) drivermanager.EventHandler
|
|
PersistState() error
|
|
|
|
RestartTask(taskName string, taskEvent *structs.TaskEvent) error
|
|
RestartAll(taskEvent *structs.TaskEvent) error
|
|
|
|
GetTaskExecHandler(taskName string) drivermanager.TaskExecHandler
|
|
GetTaskDriverCapabilities(taskName string) (*drivers.Capabilities, error)
|
|
}
|
|
|
|
// Client is used to implement the client interaction with Nomad. Clients
|
|
// are expected to register as a schedulable node to the servers, and to
|
|
// run allocations as determined by the servers.
|
|
type Client struct {
|
|
config *config.Config
|
|
start time.Time
|
|
|
|
// stateDB is used to efficiently store client state.
|
|
stateDB state.StateDB
|
|
|
|
// configCopy is a copy that should be passed to alloc-runners.
|
|
configCopy *config.Config
|
|
configLock sync.RWMutex
|
|
|
|
logger hclog.Logger
|
|
rpcLogger hclog.Logger
|
|
|
|
connPool *pool.ConnPool
|
|
|
|
// tlsWrap is used to wrap outbound connections using TLS. It should be
|
|
// accessed using the lock.
|
|
tlsWrap tlsutil.RegionWrapper
|
|
tlsWrapLock sync.RWMutex
|
|
|
|
// servers is the list of nomad servers
|
|
servers *servers.Manager
|
|
|
|
// heartbeat related times for tracking how often to heartbeat
|
|
lastHeartbeat time.Time
|
|
heartbeatTTL time.Duration
|
|
haveHeartbeated bool
|
|
heartbeatLock sync.Mutex
|
|
|
|
// triggerDiscoveryCh triggers Consul discovery; see triggerDiscovery
|
|
triggerDiscoveryCh chan struct{}
|
|
|
|
// triggerNodeUpdate triggers the client to mark the Node as changed and
|
|
// update it.
|
|
triggerNodeUpdate chan struct{}
|
|
|
|
// triggerEmitNodeEvent sends an event and triggers the client to update the
|
|
// server for the node event
|
|
triggerEmitNodeEvent chan *structs.NodeEvent
|
|
|
|
// rpcRetryCh is closed when there an event such as server discovery or a
|
|
// successful RPC occurring happens such that a retry should happen. Access
|
|
// should only occur via the getter method
|
|
rpcRetryCh chan struct{}
|
|
rpcRetryLock sync.Mutex
|
|
|
|
// allocs maps alloc IDs to their AllocRunner. This map includes all
|
|
// AllocRunners - running and GC'd - until the server GCs them.
|
|
allocs map[string]AllocRunner
|
|
allocLock sync.RWMutex
|
|
|
|
// invalidAllocs is a map that tracks allocations that failed because
|
|
// the client couldn't initialize alloc or task runners for it. This can
|
|
// happen due to driver errors
|
|
invalidAllocs map[string]struct{}
|
|
invalidAllocsLock sync.Mutex
|
|
|
|
// allocUpdates stores allocations that need to be synced to the server.
|
|
allocUpdates chan *structs.Allocation
|
|
|
|
// consulService is Nomad's custom Consul client for managing services
|
|
// and checks.
|
|
consulService consulApi.ConsulServiceAPI
|
|
|
|
// consulCatalog is the subset of Consul's Catalog API Nomad uses.
|
|
consulCatalog consul.CatalogAPI
|
|
|
|
// HostStatsCollector collects host resource usage stats
|
|
hostStatsCollector *stats.HostStatsCollector
|
|
|
|
// shutdown is true when the Client has been shutdown. Must hold
|
|
// shutdownLock to access.
|
|
shutdown bool
|
|
|
|
// shutdownCh is closed to signal the Client is shutting down.
|
|
shutdownCh chan struct{}
|
|
|
|
shutdownLock sync.Mutex
|
|
|
|
// shutdownGroup are goroutines that exit when shutdownCh is closed.
|
|
// Shutdown() blocks on Wait() after closing shutdownCh.
|
|
shutdownGroup group
|
|
|
|
// vaultClient is used to interact with Vault for token and secret renewals
|
|
vaultClient vaultclient.VaultClient
|
|
|
|
// garbageCollector is used to garbage collect terminal allocations present
|
|
// in the node automatically
|
|
garbageCollector *AllocGarbageCollector
|
|
|
|
// clientACLResolver holds the ACL resolution state
|
|
clientACLResolver
|
|
|
|
// rpcServer is used to serve RPCs by the local agent.
|
|
rpcServer *rpc.Server
|
|
endpoints rpcEndpoints
|
|
streamingRpcs *structs.StreamingRpcRegistry
|
|
|
|
// pluginManagers is the set of PluginManagers registered by the client
|
|
pluginManagers *pluginmanager.PluginGroup
|
|
|
|
// devicemanger is responsible for managing device plugins.
|
|
devicemanager devicemanager.Manager
|
|
|
|
// drivermanager is responsible for managing driver plugins
|
|
drivermanager drivermanager.Manager
|
|
|
|
// baseLabels are used when emitting tagged metrics. All client metrics will
|
|
// have these tags, and optionally more.
|
|
baseLabels []metrics.Label
|
|
|
|
// batchNodeUpdates is used to batch initial updates to the node
|
|
batchNodeUpdates *batchNodeUpdates
|
|
|
|
// fpInitialized chan is closed when the first batch of fingerprints are
|
|
// applied to the node and the server is updated
|
|
fpInitialized chan struct{}
|
|
|
|
// serversContactedCh is closed when GetClientAllocs and runAllocs have
|
|
// successfully run once.
|
|
serversContactedCh chan struct{}
|
|
serversContactedOnce sync.Once
|
|
}
|
|
|
|
var (
|
|
// noServersErr is returned by the RPC method when the client has no
|
|
// configured servers. This is used to trigger Consul discovery if
|
|
// enabled.
|
|
noServersErr = errors.New("no servers")
|
|
)
|
|
|
|
// NewClient is used to create a new client from the given configuration
|
|
func NewClient(cfg *config.Config, consulCatalog consul.CatalogAPI, consulService consulApi.ConsulServiceAPI) (*Client, error) {
|
|
// Create the tls wrapper
|
|
var tlsWrap tlsutil.RegionWrapper
|
|
if cfg.TLSConfig.EnableRPC {
|
|
tw, err := tlsutil.NewTLSConfiguration(cfg.TLSConfig, true, true)
|
|
if err != nil {
|
|
return nil, err
|
|
}
|
|
tlsWrap, err = tw.OutgoingTLSWrapper()
|
|
if err != nil {
|
|
return nil, err
|
|
}
|
|
}
|
|
|
|
if cfg.StateDBFactory == nil {
|
|
cfg.StateDBFactory = state.GetStateDBFactory(cfg.DevMode)
|
|
}
|
|
|
|
// Create the logger
|
|
logger := cfg.Logger.ResetNamed("client")
|
|
|
|
// Create the client
|
|
c := &Client{
|
|
config: cfg,
|
|
consulCatalog: consulCatalog,
|
|
consulService: consulService,
|
|
start: time.Now(),
|
|
connPool: pool.NewPool(logger, clientRPCCache, clientMaxStreams, tlsWrap),
|
|
tlsWrap: tlsWrap,
|
|
streamingRpcs: structs.NewStreamingRpcRegistry(),
|
|
logger: logger,
|
|
rpcLogger: logger.Named("rpc"),
|
|
allocs: make(map[string]AllocRunner),
|
|
allocUpdates: make(chan *structs.Allocation, 64),
|
|
shutdownCh: make(chan struct{}),
|
|
triggerDiscoveryCh: make(chan struct{}),
|
|
triggerNodeUpdate: make(chan struct{}, 8),
|
|
triggerEmitNodeEvent: make(chan *structs.NodeEvent, 8),
|
|
fpInitialized: make(chan struct{}),
|
|
invalidAllocs: make(map[string]struct{}),
|
|
serversContactedCh: make(chan struct{}),
|
|
serversContactedOnce: sync.Once{},
|
|
}
|
|
|
|
c.batchNodeUpdates = newBatchNodeUpdates(
|
|
c.updateNodeFromDriver,
|
|
c.updateNodeFromDevices,
|
|
)
|
|
|
|
// Initialize the server manager
|
|
c.servers = servers.New(c.logger, c.shutdownCh, c)
|
|
|
|
// Start server manager rebalancing go routine
|
|
go c.servers.Start()
|
|
|
|
// Initialize the client
|
|
if err := c.init(); err != nil {
|
|
return nil, fmt.Errorf("failed to initialize client: %v", err)
|
|
}
|
|
|
|
// Setup the clients RPC server
|
|
c.setupClientRpc()
|
|
|
|
// Initialize the ACL state
|
|
if err := c.clientACLResolver.init(); err != nil {
|
|
return nil, fmt.Errorf("failed to initialize ACL state: %v", err)
|
|
}
|
|
|
|
// Setup the node
|
|
if err := c.setupNode(); err != nil {
|
|
return nil, fmt.Errorf("node setup failed: %v", err)
|
|
}
|
|
|
|
// Store the config copy before restoring state but after it has been
|
|
// initialized.
|
|
c.configLock.Lock()
|
|
c.configCopy = c.config.Copy()
|
|
c.configLock.Unlock()
|
|
|
|
fingerprintManager := NewFingerprintManager(
|
|
c.configCopy.PluginSingletonLoader, c.GetConfig, c.configCopy.Node,
|
|
c.shutdownCh, c.updateNodeFromFingerprint, c.logger)
|
|
|
|
c.pluginManagers = pluginmanager.New(c.logger)
|
|
|
|
// Fingerprint the node and scan for drivers
|
|
if err := fingerprintManager.Run(); err != nil {
|
|
return nil, fmt.Errorf("fingerprinting failed: %v", err)
|
|
}
|
|
|
|
// Build the white/blacklists of drivers.
|
|
allowlistDrivers := cfg.ReadStringListToMap("driver.whitelist")
|
|
blocklistDrivers := cfg.ReadStringListToMap("driver.blacklist")
|
|
|
|
// Setup the driver manager
|
|
driverConfig := &drivermanager.Config{
|
|
Logger: c.logger,
|
|
Loader: c.configCopy.PluginSingletonLoader,
|
|
PluginConfig: c.configCopy.NomadPluginConfig(),
|
|
Updater: c.batchNodeUpdates.updateNodeFromDriver,
|
|
EventHandlerFactory: c.GetTaskEventHandler,
|
|
State: c.stateDB,
|
|
AllowedDrivers: allowlistDrivers,
|
|
BlockedDrivers: blocklistDrivers,
|
|
}
|
|
drvManager := drivermanager.New(driverConfig)
|
|
c.drivermanager = drvManager
|
|
c.pluginManagers.RegisterAndRun(drvManager)
|
|
|
|
// Setup the device manager
|
|
devConfig := &devicemanager.Config{
|
|
Logger: c.logger,
|
|
Loader: c.configCopy.PluginSingletonLoader,
|
|
PluginConfig: c.configCopy.NomadPluginConfig(),
|
|
Updater: c.batchNodeUpdates.updateNodeFromDevices,
|
|
StatsInterval: c.configCopy.StatsCollectionInterval,
|
|
State: c.stateDB,
|
|
}
|
|
devManager := devicemanager.New(devConfig)
|
|
c.devicemanager = devManager
|
|
c.pluginManagers.RegisterAndRun(devManager)
|
|
|
|
// Batching of initial fingerprints is done to reduce the number of node
|
|
// updates sent to the server on startup.
|
|
go c.batchFirstFingerprints()
|
|
|
|
// Add the stats collector
|
|
statsCollector := stats.NewHostStatsCollector(c.logger, c.config.AllocDir, c.devicemanager.AllStats)
|
|
c.hostStatsCollector = statsCollector
|
|
|
|
// Add the garbage collector
|
|
gcConfig := &GCConfig{
|
|
MaxAllocs: cfg.GCMaxAllocs,
|
|
DiskUsageThreshold: cfg.GCDiskUsageThreshold,
|
|
InodeUsageThreshold: cfg.GCInodeUsageThreshold,
|
|
Interval: cfg.GCInterval,
|
|
ParallelDestroys: cfg.GCParallelDestroys,
|
|
ReservedDiskMB: cfg.Node.Reserved.DiskMB,
|
|
}
|
|
c.garbageCollector = NewAllocGarbageCollector(c.logger, statsCollector, c, gcConfig)
|
|
go c.garbageCollector.Run()
|
|
|
|
// Set the preconfigured list of static servers
|
|
c.configLock.RLock()
|
|
if len(c.configCopy.Servers) > 0 {
|
|
if _, err := c.setServersImpl(c.configCopy.Servers, true); err != nil {
|
|
logger.Warn("none of the configured servers are valid", "error", err)
|
|
}
|
|
}
|
|
c.configLock.RUnlock()
|
|
|
|
// Setup Consul discovery if enabled
|
|
if c.configCopy.ConsulConfig.ClientAutoJoin != nil && *c.configCopy.ConsulConfig.ClientAutoJoin {
|
|
c.shutdownGroup.Go(c.consulDiscovery)
|
|
if c.servers.NumServers() == 0 {
|
|
// No configured servers; trigger discovery manually
|
|
c.triggerDiscoveryCh <- struct{}{}
|
|
}
|
|
}
|
|
|
|
// Setup the vault client for token and secret renewals
|
|
if err := c.setupVaultClient(); err != nil {
|
|
return nil, fmt.Errorf("failed to setup vault client: %v", err)
|
|
}
|
|
|
|
// wait until drivers are healthy before restoring or registering with servers
|
|
select {
|
|
case <-c.Ready():
|
|
case <-time.After(batchFirstFingerprintsProcessingGrace):
|
|
logger.Warn("batch fingerprint operation timed out; proceeding to register with fingerprinted plugins so far")
|
|
}
|
|
|
|
// Register and then start heartbeating to the servers.
|
|
c.shutdownGroup.Go(c.registerAndHeartbeat)
|
|
|
|
// Restore the state
|
|
if err := c.restoreState(); err != nil {
|
|
logger.Error("failed to restore state", "error", err)
|
|
logger.Error("Nomad is unable to start due to corrupt state. "+
|
|
"The safest way to proceed is to manually stop running task processes "+
|
|
"and remove Nomad's state and alloc directories before "+
|
|
"restarting. Lost allocations will be rescheduled.",
|
|
"state_dir", c.config.StateDir, "alloc_dir", c.config.AllocDir)
|
|
logger.Error("Corrupt state is often caused by a bug. Please " +
|
|
"report as much information as possible to " +
|
|
"https://github.com/hashicorp/nomad/issues")
|
|
return nil, fmt.Errorf("failed to restore state")
|
|
}
|
|
|
|
// Begin periodic snapshotting of state.
|
|
c.shutdownGroup.Go(c.periodicSnapshot)
|
|
|
|
// Begin syncing allocations to the server
|
|
c.shutdownGroup.Go(c.allocSync)
|
|
|
|
// Start the client! Don't use the shutdownGroup as run handles
|
|
// shutdowns manually to prevent updates from being applied during
|
|
// shutdown.
|
|
go c.run()
|
|
|
|
// Start collecting stats
|
|
c.shutdownGroup.Go(c.emitStats)
|
|
|
|
c.logger.Info("started client", "node_id", c.NodeID())
|
|
return c, nil
|
|
}
|
|
|
|
// Ready returns a chan that is closed when the client is fully initialized
|
|
func (c *Client) Ready() <-chan struct{} {
|
|
return c.fpInitialized
|
|
}
|
|
|
|
// init is used to initialize the client and perform any setup
|
|
// needed before we begin starting its various components.
|
|
func (c *Client) init() error {
|
|
// Ensure the state dir exists if we have one
|
|
if c.config.StateDir != "" {
|
|
if err := os.MkdirAll(c.config.StateDir, 0700); err != nil {
|
|
return fmt.Errorf("failed creating state dir: %s", err)
|
|
}
|
|
|
|
} else {
|
|
// Otherwise make a temp directory to use.
|
|
p, err := ioutil.TempDir("", "NomadClient")
|
|
if err != nil {
|
|
return fmt.Errorf("failed creating temporary directory for the StateDir: %v", err)
|
|
}
|
|
|
|
p, err = filepath.EvalSymlinks(p)
|
|
if err != nil {
|
|
return fmt.Errorf("failed to find temporary directory for the StateDir: %v", err)
|
|
}
|
|
|
|
c.config.StateDir = p
|
|
}
|
|
c.logger.Info("using state directory", "state_dir", c.config.StateDir)
|
|
|
|
// Open the state database
|
|
db, err := c.config.StateDBFactory(c.logger, c.config.StateDir)
|
|
if err != nil {
|
|
return fmt.Errorf("failed to open state database: %v", err)
|
|
}
|
|
|
|
// Upgrade the state database
|
|
if err := db.Upgrade(); err != nil {
|
|
// Upgrade only returns an error on critical persistence
|
|
// failures in which an operator should intervene before the
|
|
// node is accessible. Upgrade drops and logs corrupt state it
|
|
// encounters, so failing to start the agent should be extremely
|
|
// rare.
|
|
return fmt.Errorf("failed to upgrade state database: %v", err)
|
|
}
|
|
|
|
c.stateDB = db
|
|
|
|
// Ensure the alloc dir exists if we have one
|
|
if c.config.AllocDir != "" {
|
|
if err := os.MkdirAll(c.config.AllocDir, 0711); err != nil {
|
|
return fmt.Errorf("failed creating alloc dir: %s", err)
|
|
}
|
|
} else {
|
|
// Otherwise make a temp directory to use.
|
|
p, err := ioutil.TempDir("", "NomadClient")
|
|
if err != nil {
|
|
return fmt.Errorf("failed creating temporary directory for the AllocDir: %v", err)
|
|
}
|
|
|
|
p, err = filepath.EvalSymlinks(p)
|
|
if err != nil {
|
|
return fmt.Errorf("failed to find temporary directory for the AllocDir: %v", err)
|
|
}
|
|
|
|
// Change the permissions to have the execute bit
|
|
if err := os.Chmod(p, 0711); err != nil {
|
|
return fmt.Errorf("failed to change directory permissions for the AllocDir: %v", err)
|
|
}
|
|
|
|
c.config.AllocDir = p
|
|
}
|
|
|
|
c.logger.Info("using alloc directory", "alloc_dir", c.config.AllocDir)
|
|
return nil
|
|
}
|
|
|
|
// reloadTLSConnections allows a client to reload its TLS configuration on the
|
|
// fly
|
|
func (c *Client) reloadTLSConnections(newConfig *nconfig.TLSConfig) error {
|
|
var tlsWrap tlsutil.RegionWrapper
|
|
if newConfig != nil && newConfig.EnableRPC {
|
|
tw, err := tlsutil.NewTLSConfiguration(newConfig, true, true)
|
|
if err != nil {
|
|
return err
|
|
}
|
|
|
|
twWrap, err := tw.OutgoingTLSWrapper()
|
|
if err != nil {
|
|
return err
|
|
}
|
|
tlsWrap = twWrap
|
|
}
|
|
|
|
// Store the new tls wrapper.
|
|
c.tlsWrapLock.Lock()
|
|
c.tlsWrap = tlsWrap
|
|
c.tlsWrapLock.Unlock()
|
|
|
|
// Keep the client configuration up to date as we use configuration values to
|
|
// decide on what type of connections to accept
|
|
c.configLock.Lock()
|
|
c.config.TLSConfig = newConfig
|
|
c.configLock.Unlock()
|
|
|
|
c.connPool.ReloadTLS(tlsWrap)
|
|
|
|
return nil
|
|
}
|
|
|
|
// Reload allows a client to reload its configuration on the fly
|
|
func (c *Client) Reload(newConfig *config.Config) error {
|
|
shouldReloadTLS, err := tlsutil.ShouldReloadRPCConnections(c.config.TLSConfig, newConfig.TLSConfig)
|
|
if err != nil {
|
|
c.logger.Error("error parsing TLS configuration", "error", err)
|
|
return err
|
|
}
|
|
|
|
if shouldReloadTLS {
|
|
return c.reloadTLSConnections(newConfig.TLSConfig)
|
|
}
|
|
|
|
return nil
|
|
}
|
|
|
|
// Leave is used to prepare the client to leave the cluster
|
|
func (c *Client) Leave() error {
|
|
// TODO
|
|
return nil
|
|
}
|
|
|
|
// GetConfig returns the config of the client
|
|
func (c *Client) GetConfig() *config.Config {
|
|
c.configLock.Lock()
|
|
defer c.configLock.Unlock()
|
|
return c.configCopy
|
|
}
|
|
|
|
// Datacenter returns the datacenter for the given client
|
|
func (c *Client) Datacenter() string {
|
|
return c.config.Node.Datacenter
|
|
}
|
|
|
|
// Region returns the region for the given client
|
|
func (c *Client) Region() string {
|
|
return c.config.Region
|
|
}
|
|
|
|
// NodeID returns the node ID for the given client
|
|
func (c *Client) NodeID() string {
|
|
return c.config.Node.ID
|
|
}
|
|
|
|
// secretNodeID returns the secret node ID for the given client
|
|
func (c *Client) secretNodeID() string {
|
|
return c.config.Node.SecretID
|
|
}
|
|
|
|
// RPCMajorVersion returns the structs.ApiMajorVersion supported by the
|
|
// client.
|
|
func (c *Client) RPCMajorVersion() int {
|
|
return structs.ApiMajorVersion
|
|
}
|
|
|
|
// RPCMinorVersion returns the structs.ApiMinorVersion supported by the
|
|
// client.
|
|
func (c *Client) RPCMinorVersion() int {
|
|
return structs.ApiMinorVersion
|
|
}
|
|
|
|
// Shutdown is used to tear down the client
|
|
func (c *Client) Shutdown() error {
|
|
c.shutdownLock.Lock()
|
|
defer c.shutdownLock.Unlock()
|
|
|
|
if c.shutdown {
|
|
c.logger.Info("already shutdown")
|
|
return nil
|
|
}
|
|
c.logger.Info("shutting down")
|
|
|
|
// Stop renewing tokens and secrets
|
|
if c.vaultClient != nil {
|
|
c.vaultClient.Stop()
|
|
}
|
|
|
|
// Stop Garbage collector
|
|
c.garbageCollector.Stop()
|
|
|
|
arGroup := group{}
|
|
if c.config.DevMode {
|
|
// In DevMode destroy all the running allocations.
|
|
for _, ar := range c.getAllocRunners() {
|
|
ar.Destroy()
|
|
arGroup.AddCh(ar.DestroyCh())
|
|
}
|
|
} else {
|
|
// In normal mode call shutdown
|
|
for _, ar := range c.getAllocRunners() {
|
|
ar.Shutdown()
|
|
arGroup.AddCh(ar.ShutdownCh())
|
|
}
|
|
}
|
|
arGroup.Wait()
|
|
|
|
// Shutdown the plugin managers
|
|
c.pluginManagers.Shutdown()
|
|
|
|
c.shutdown = true
|
|
close(c.shutdownCh)
|
|
|
|
// Must close connection pool to unblock alloc watcher
|
|
c.connPool.Shutdown()
|
|
|
|
// Wait for goroutines to stop
|
|
c.shutdownGroup.Wait()
|
|
|
|
// One final save state
|
|
c.saveState()
|
|
return c.stateDB.Close()
|
|
}
|
|
|
|
// Stats is used to return statistics for debugging and insight
|
|
// for various sub-systems
|
|
func (c *Client) Stats() map[string]map[string]string {
|
|
c.heartbeatLock.Lock()
|
|
defer c.heartbeatLock.Unlock()
|
|
stats := map[string]map[string]string{
|
|
"client": {
|
|
"node_id": c.NodeID(),
|
|
"known_servers": strings.Join(c.GetServers(), ","),
|
|
"num_allocations": strconv.Itoa(c.NumAllocs()),
|
|
"last_heartbeat": fmt.Sprintf("%v", time.Since(c.lastHeartbeat)),
|
|
"heartbeat_ttl": fmt.Sprintf("%v", c.heartbeatTTL),
|
|
},
|
|
"runtime": hstats.RuntimeStats(),
|
|
}
|
|
return stats
|
|
}
|
|
|
|
// SignalAllocation sends a signal to the tasks within an allocation.
|
|
// If the provided task is empty, then every allocation will be signalled.
|
|
// If a task is provided, then only an exactly matching task will be signalled.
|
|
func (c *Client) SignalAllocation(allocID, task, signal string) error {
|
|
ar, err := c.getAllocRunner(allocID)
|
|
if err != nil {
|
|
return err
|
|
}
|
|
|
|
return ar.Signal(task, signal)
|
|
}
|
|
|
|
// CollectAllocation garbage collects a single allocation on a node. Returns
|
|
// true if alloc was found and garbage collected; otherwise false.
|
|
func (c *Client) CollectAllocation(allocID string) bool {
|
|
return c.garbageCollector.Collect(allocID)
|
|
}
|
|
|
|
// CollectAllAllocs garbage collects all allocations on a node in the terminal
|
|
// state
|
|
func (c *Client) CollectAllAllocs() {
|
|
c.garbageCollector.CollectAll()
|
|
}
|
|
|
|
func (c *Client) RestartAllocation(allocID, taskName string) error {
|
|
ar, err := c.getAllocRunner(allocID)
|
|
if err != nil {
|
|
return err
|
|
}
|
|
|
|
event := structs.NewTaskEvent(structs.TaskRestartSignal).
|
|
SetRestartReason("User requested restart")
|
|
|
|
if taskName != "" {
|
|
return ar.RestartTask(taskName, event)
|
|
}
|
|
|
|
return ar.RestartAll(event)
|
|
}
|
|
|
|
// Node returns the locally registered node
|
|
func (c *Client) Node() *structs.Node {
|
|
c.configLock.RLock()
|
|
defer c.configLock.RUnlock()
|
|
return c.configCopy.Node
|
|
}
|
|
|
|
func (c *Client) getAllocRunner(allocID string) (AllocRunner, error) {
|
|
c.allocLock.RLock()
|
|
defer c.allocLock.RUnlock()
|
|
|
|
ar, ok := c.allocs[allocID]
|
|
if !ok {
|
|
return nil, structs.NewErrUnknownAllocation(allocID)
|
|
}
|
|
|
|
return ar, nil
|
|
}
|
|
|
|
// StatsReporter exposes the various APIs related resource usage of a Nomad
|
|
// client
|
|
func (c *Client) StatsReporter() ClientStatsReporter {
|
|
return c
|
|
}
|
|
|
|
func (c *Client) GetAllocStats(allocID string) (interfaces.AllocStatsReporter, error) {
|
|
ar, err := c.getAllocRunner(allocID)
|
|
if err != nil {
|
|
return nil, err
|
|
}
|
|
return ar.StatsReporter(), nil
|
|
}
|
|
|
|
// HostStats returns all the stats related to a Nomad client
|
|
func (c *Client) LatestHostStats() *stats.HostStats {
|
|
return c.hostStatsCollector.Stats()
|
|
}
|
|
|
|
func (c *Client) LatestDeviceResourceStats(devices []*structs.AllocatedDeviceResource) []*device.DeviceGroupStats {
|
|
return c.computeAllocatedDeviceGroupStats(devices, c.LatestHostStats().DeviceStats)
|
|
}
|
|
|
|
func (c *Client) computeAllocatedDeviceGroupStats(devices []*structs.AllocatedDeviceResource, hostDeviceGroupStats []*device.DeviceGroupStats) []*device.DeviceGroupStats {
|
|
// basic optimization for the usual case
|
|
if len(devices) == 0 || len(hostDeviceGroupStats) == 0 {
|
|
return nil
|
|
}
|
|
|
|
// Build an index of allocated devices
|
|
adIdx := map[structs.DeviceIdTuple][]string{}
|
|
|
|
total := 0
|
|
for _, ds := range devices {
|
|
adIdx[*ds.ID()] = ds.DeviceIDs
|
|
total += len(ds.DeviceIDs)
|
|
}
|
|
|
|
// Collect allocated device stats from host stats
|
|
result := make([]*device.DeviceGroupStats, 0, len(adIdx))
|
|
|
|
for _, dg := range hostDeviceGroupStats {
|
|
k := structs.DeviceIdTuple{
|
|
Vendor: dg.Vendor,
|
|
Type: dg.Type,
|
|
Name: dg.Name,
|
|
}
|
|
|
|
allocatedDeviceIDs, ok := adIdx[k]
|
|
if !ok {
|
|
continue
|
|
}
|
|
|
|
rdgStats := &device.DeviceGroupStats{
|
|
Vendor: dg.Vendor,
|
|
Type: dg.Type,
|
|
Name: dg.Name,
|
|
InstanceStats: map[string]*device.DeviceStats{},
|
|
}
|
|
|
|
for _, adID := range allocatedDeviceIDs {
|
|
deviceStats, ok := dg.InstanceStats[adID]
|
|
if !ok || deviceStats == nil {
|
|
c.logger.Warn("device not found in stats", "device_id", adID, "device_group_id", k)
|
|
continue
|
|
}
|
|
|
|
rdgStats.InstanceStats[adID] = deviceStats
|
|
}
|
|
result = append(result, rdgStats)
|
|
}
|
|
|
|
return result
|
|
}
|
|
|
|
// ValidateMigrateToken verifies that a token is for a specific client and
|
|
// allocation, and has been created by a trusted party that has privileged
|
|
// knowledge of the client's secret identifier
|
|
func (c *Client) ValidateMigrateToken(allocID, migrateToken string) bool {
|
|
if !c.config.ACLEnabled {
|
|
return true
|
|
}
|
|
|
|
return structs.CompareMigrateToken(allocID, c.secretNodeID(), migrateToken)
|
|
}
|
|
|
|
// GetAllocFS returns the AllocFS interface for the alloc dir of an allocation
|
|
func (c *Client) GetAllocFS(allocID string) (allocdir.AllocDirFS, error) {
|
|
ar, err := c.getAllocRunner(allocID)
|
|
if err != nil {
|
|
return nil, err
|
|
}
|
|
|
|
return ar.GetAllocDir(), nil
|
|
}
|
|
|
|
// GetAllocState returns a copy of an allocation's state on this client. It
|
|
// returns either an AllocState or an unknown allocation error.
|
|
func (c *Client) GetAllocState(allocID string) (*arstate.State, error) {
|
|
ar, err := c.getAllocRunner(allocID)
|
|
if err != nil {
|
|
return nil, err
|
|
}
|
|
|
|
return ar.AllocState(), nil
|
|
}
|
|
|
|
// GetServers returns the list of nomad servers this client is aware of.
|
|
func (c *Client) GetServers() []string {
|
|
endpoints := c.servers.GetServers()
|
|
res := make([]string, len(endpoints))
|
|
for i := range endpoints {
|
|
res[i] = endpoints[i].String()
|
|
}
|
|
sort.Strings(res)
|
|
return res
|
|
}
|
|
|
|
// SetServers sets a new list of nomad servers to connect to. As long as one
|
|
// server is resolvable no error is returned.
|
|
func (c *Client) SetServers(in []string) (int, error) {
|
|
return c.setServersImpl(in, false)
|
|
}
|
|
|
|
// setServersImpl sets a new list of nomad servers to connect to. If force is
|
|
// set, we add the server to the internal serverlist even if the server could not
|
|
// be pinged. An error is returned if no endpoints were valid when non-forcing.
|
|
//
|
|
// Force should be used when setting the servers from the initial configuration
|
|
// since the server may be starting up in parallel and initial pings may fail.
|
|
func (c *Client) setServersImpl(in []string, force bool) (int, error) {
|
|
var mu sync.Mutex
|
|
var wg sync.WaitGroup
|
|
var merr multierror.Error
|
|
|
|
endpoints := make([]*servers.Server, 0, len(in))
|
|
wg.Add(len(in))
|
|
|
|
for _, s := range in {
|
|
go func(srv string) {
|
|
defer wg.Done()
|
|
addr, err := resolveServer(srv)
|
|
if err != nil {
|
|
mu.Lock()
|
|
c.logger.Debug("ignoring server due to resolution error", "error", err, "server", srv)
|
|
merr.Errors = append(merr.Errors, err)
|
|
mu.Unlock()
|
|
return
|
|
}
|
|
|
|
// Try to ping to check if it is a real server
|
|
if err := c.Ping(addr); err != nil {
|
|
mu.Lock()
|
|
merr.Errors = append(merr.Errors, fmt.Errorf("Server at address %s failed ping: %v", addr, err))
|
|
mu.Unlock()
|
|
|
|
// If we are forcing the setting of the servers, inject it to
|
|
// the serverlist even if we can't ping immediately.
|
|
if !force {
|
|
return
|
|
}
|
|
}
|
|
|
|
mu.Lock()
|
|
endpoints = append(endpoints, &servers.Server{Addr: addr})
|
|
mu.Unlock()
|
|
}(s)
|
|
}
|
|
|
|
wg.Wait()
|
|
|
|
// Only return errors if no servers are valid
|
|
if len(endpoints) == 0 {
|
|
if len(merr.Errors) > 0 {
|
|
return 0, merr.ErrorOrNil()
|
|
}
|
|
return 0, noServersErr
|
|
}
|
|
|
|
c.servers.SetServers(endpoints)
|
|
return len(endpoints), nil
|
|
}
|
|
|
|
// restoreState is used to restore our state from the data dir
|
|
// If there are errors restoring a specific allocation it is marked
|
|
// as failed whenever possible.
|
|
func (c *Client) restoreState() error {
|
|
if c.config.DevMode {
|
|
return nil
|
|
}
|
|
|
|
//XXX REMOVED! make a note in backward compat / upgrading doc
|
|
// COMPAT: Remove in 0.7.0
|
|
// 0.6.0 transitioned from individual state files to a single bolt-db.
|
|
// The upgrade path is to:
|
|
// Check if old state exists
|
|
// If so, restore from that and delete old state
|
|
// Restore using state database
|
|
|
|
// Restore allocations
|
|
allocs, allocErrs, err := c.stateDB.GetAllAllocations()
|
|
if err != nil {
|
|
return err
|
|
}
|
|
|
|
for allocID, err := range allocErrs {
|
|
c.logger.Error("error restoring alloc", "error", err, "alloc_id", allocID)
|
|
//TODO Cleanup
|
|
// Try to clean up alloc dir
|
|
// Remove boltdb entries?
|
|
// Send to server with clientstatus=failed
|
|
}
|
|
|
|
// Load each alloc back
|
|
for _, alloc := range allocs {
|
|
|
|
// COMPAT(0.12): remove once upgrading from 0.9.5 is no longer supported
|
|
// See hasLocalState for details. Skipping suspicious allocs
|
|
// now. If allocs should be run, they will be started when the client
|
|
// gets allocs from servers.
|
|
if !c.hasLocalState(alloc) {
|
|
c.logger.Warn("found a alloc without any local state, skipping restore", "alloc_id", alloc.ID)
|
|
continue
|
|
}
|
|
|
|
//XXX On Restore we give up on watching previous allocs because
|
|
// we need the local AllocRunners initialized first. We could
|
|
// add a second loop to initialize just the alloc watcher.
|
|
prevAllocWatcher := allocwatcher.NoopPrevAlloc{}
|
|
prevAllocMigrator := allocwatcher.NoopPrevAlloc{}
|
|
|
|
c.configLock.RLock()
|
|
arConf := &allocrunner.Config{
|
|
Alloc: alloc,
|
|
Logger: c.logger,
|
|
ClientConfig: c.configCopy,
|
|
StateDB: c.stateDB,
|
|
StateUpdater: c,
|
|
DeviceStatsReporter: c,
|
|
Consul: c.consulService,
|
|
Vault: c.vaultClient,
|
|
PrevAllocWatcher: prevAllocWatcher,
|
|
PrevAllocMigrator: prevAllocMigrator,
|
|
DeviceManager: c.devicemanager,
|
|
DriverManager: c.drivermanager,
|
|
ServersContactedCh: c.serversContactedCh,
|
|
}
|
|
c.configLock.RUnlock()
|
|
|
|
ar, err := allocrunner.NewAllocRunner(arConf)
|
|
if err != nil {
|
|
c.logger.Error("error running alloc", "error", err, "alloc_id", alloc.ID)
|
|
c.handleInvalidAllocs(alloc, err)
|
|
continue
|
|
}
|
|
|
|
// Restore state
|
|
if err := ar.Restore(); err != nil {
|
|
c.logger.Error("error restoring alloc", "error", err, "alloc_id", alloc.ID)
|
|
// Override the status of the alloc to failed
|
|
ar.SetClientStatus(structs.AllocClientStatusFailed)
|
|
// Destroy the alloc runner since this is a failed restore
|
|
ar.Destroy()
|
|
continue
|
|
}
|
|
|
|
//XXX is this locking necessary?
|
|
c.allocLock.Lock()
|
|
c.allocs[alloc.ID] = ar
|
|
c.allocLock.Unlock()
|
|
}
|
|
|
|
// All allocs restored successfully, run them!
|
|
c.allocLock.Lock()
|
|
for _, ar := range c.allocs {
|
|
go ar.Run()
|
|
}
|
|
c.allocLock.Unlock()
|
|
return nil
|
|
}
|
|
|
|
// hasLocalState returns true if we have any other associated state
|
|
// with alloc beyond the task itself
|
|
//
|
|
// Useful for detecting if a potentially completed alloc got resurrected
|
|
// after AR was destroyed. In such cases, re-running the alloc lead to
|
|
// unexpected reruns and may lead to process and task exhaustion on node.
|
|
//
|
|
// The heuristic used here is an alloc is suspect if we see no other information
|
|
// and no other task/status info is found.
|
|
//
|
|
// Also, an alloc without any client state will not be restored correctly; there will
|
|
// be no tasks processes to reattach to, etc. In such cases, client should
|
|
// wait until it gets allocs from server to launch them.
|
|
//
|
|
// See:
|
|
// * https://github.com/hashicorp/nomad/pull/6207
|
|
// * https://github.com/hashicorp/nomad/issues/5984
|
|
//
|
|
// COMPAT(0.12): remove once upgrading from 0.9.5 is no longer supported
|
|
func (c *Client) hasLocalState(alloc *structs.Allocation) bool {
|
|
tg := alloc.Job.LookupTaskGroup(alloc.TaskGroup)
|
|
if tg == nil {
|
|
// corrupt alloc?!
|
|
return false
|
|
}
|
|
|
|
for _, task := range tg.Tasks {
|
|
ls, tr, _ := c.stateDB.GetTaskRunnerState(alloc.ID, task.Name)
|
|
if ls != nil || tr != nil {
|
|
return true
|
|
}
|
|
}
|
|
|
|
return false
|
|
}
|
|
|
|
func (c *Client) handleInvalidAllocs(alloc *structs.Allocation, err error) {
|
|
c.invalidAllocsLock.Lock()
|
|
c.invalidAllocs[alloc.ID] = struct{}{}
|
|
c.invalidAllocsLock.Unlock()
|
|
|
|
// Mark alloc as failed so server can handle this
|
|
failed := makeFailedAlloc(alloc, err)
|
|
select {
|
|
case c.allocUpdates <- failed:
|
|
case <-c.shutdownCh:
|
|
}
|
|
}
|
|
|
|
// saveState is used to snapshot our state into the data dir.
|
|
func (c *Client) saveState() error {
|
|
var wg sync.WaitGroup
|
|
var l sync.Mutex
|
|
var mErr multierror.Error
|
|
runners := c.getAllocRunners()
|
|
wg.Add(len(runners))
|
|
|
|
for id, ar := range runners {
|
|
go func(id string, ar AllocRunner) {
|
|
err := ar.PersistState()
|
|
if err != nil {
|
|
c.logger.Error("error saving alloc state", "error", err, "alloc_id", id)
|
|
l.Lock()
|
|
multierror.Append(&mErr, err)
|
|
l.Unlock()
|
|
}
|
|
wg.Done()
|
|
}(id, ar)
|
|
}
|
|
|
|
wg.Wait()
|
|
return mErr.ErrorOrNil()
|
|
}
|
|
|
|
// getAllocRunners returns a snapshot of the current set of alloc runners.
|
|
func (c *Client) getAllocRunners() map[string]AllocRunner {
|
|
c.allocLock.RLock()
|
|
defer c.allocLock.RUnlock()
|
|
runners := make(map[string]AllocRunner, len(c.allocs))
|
|
for id, ar := range c.allocs {
|
|
runners[id] = ar
|
|
}
|
|
return runners
|
|
}
|
|
|
|
// NumAllocs returns the number of un-GC'd allocs this client has. Used to
|
|
// fulfill the AllocCounter interface for the GC.
|
|
func (c *Client) NumAllocs() int {
|
|
n := 0
|
|
c.allocLock.RLock()
|
|
for _, a := range c.allocs {
|
|
if !a.IsDestroyed() {
|
|
n++
|
|
}
|
|
}
|
|
c.allocLock.RUnlock()
|
|
return n
|
|
}
|
|
|
|
// nodeID restores, or generates if necessary, a unique node ID and SecretID.
|
|
// The node ID is, if available, a persistent unique ID. The secret ID is a
|
|
// high-entropy random UUID.
|
|
func (c *Client) nodeID() (id, secret string, err error) {
|
|
var hostID string
|
|
hostInfo, err := host.Info()
|
|
if !c.config.NoHostUUID && err == nil {
|
|
if hashed, ok := helper.HashUUID(hostInfo.HostID); ok {
|
|
hostID = hashed
|
|
}
|
|
}
|
|
|
|
if hostID == "" {
|
|
// Generate a random hostID if no constant ID is available on
|
|
// this platform.
|
|
hostID = uuid.Generate()
|
|
}
|
|
|
|
// Do not persist in dev mode
|
|
if c.config.DevMode {
|
|
return hostID, uuid.Generate(), nil
|
|
}
|
|
|
|
// Attempt to read existing ID
|
|
idPath := filepath.Join(c.config.StateDir, "client-id")
|
|
idBuf, err := ioutil.ReadFile(idPath)
|
|
if err != nil && !os.IsNotExist(err) {
|
|
return "", "", err
|
|
}
|
|
|
|
// Attempt to read existing secret ID
|
|
secretPath := filepath.Join(c.config.StateDir, "secret-id")
|
|
secretBuf, err := ioutil.ReadFile(secretPath)
|
|
if err != nil && !os.IsNotExist(err) {
|
|
return "", "", err
|
|
}
|
|
|
|
// Use existing ID if any
|
|
if len(idBuf) != 0 {
|
|
id = strings.ToLower(string(idBuf))
|
|
} else {
|
|
id = hostID
|
|
|
|
// Persist the ID
|
|
if err := ioutil.WriteFile(idPath, []byte(id), 0700); err != nil {
|
|
return "", "", err
|
|
}
|
|
}
|
|
|
|
if len(secretBuf) != 0 {
|
|
secret = string(secretBuf)
|
|
} else {
|
|
// Generate new ID
|
|
secret = uuid.Generate()
|
|
|
|
// Persist the ID
|
|
if err := ioutil.WriteFile(secretPath, []byte(secret), 0700); err != nil {
|
|
return "", "", err
|
|
}
|
|
}
|
|
|
|
return id, secret, nil
|
|
}
|
|
|
|
// setupNode is used to setup the initial node
|
|
func (c *Client) setupNode() error {
|
|
node := c.config.Node
|
|
if node == nil {
|
|
node = &structs.Node{}
|
|
c.config.Node = node
|
|
}
|
|
// Generate an ID and secret for the node
|
|
id, secretID, err := c.nodeID()
|
|
if err != nil {
|
|
return fmt.Errorf("node ID setup failed: %v", err)
|
|
}
|
|
|
|
node.ID = id
|
|
node.SecretID = secretID
|
|
if node.Attributes == nil {
|
|
node.Attributes = make(map[string]string)
|
|
}
|
|
if node.Links == nil {
|
|
node.Links = make(map[string]string)
|
|
}
|
|
if node.Drivers == nil {
|
|
node.Drivers = make(map[string]*structs.DriverInfo)
|
|
}
|
|
if node.Meta == nil {
|
|
node.Meta = make(map[string]string)
|
|
}
|
|
if node.NodeResources == nil {
|
|
node.NodeResources = &structs.NodeResources{}
|
|
}
|
|
if node.ReservedResources == nil {
|
|
node.ReservedResources = &structs.NodeReservedResources{}
|
|
}
|
|
if node.Resources == nil {
|
|
node.Resources = &structs.Resources{}
|
|
}
|
|
if node.Reserved == nil {
|
|
node.Reserved = &structs.Resources{}
|
|
}
|
|
if node.Datacenter == "" {
|
|
node.Datacenter = "dc1"
|
|
}
|
|
if node.Name == "" {
|
|
node.Name, _ = os.Hostname()
|
|
}
|
|
// TODO(dani): Fingerprint these to handle volumes that don't exist/have bad perms.
|
|
if node.HostVolumes == nil {
|
|
if l := len(c.config.HostVolumes); l != 0 {
|
|
node.HostVolumes = make(map[string]*structs.ClientHostVolumeConfig, l)
|
|
for k, v := range c.config.HostVolumes {
|
|
node.HostVolumes[k] = v.Copy()
|
|
}
|
|
}
|
|
}
|
|
|
|
if node.Name == "" {
|
|
node.Name = node.ID
|
|
}
|
|
node.Status = structs.NodeStatusInit
|
|
|
|
// Setup default meta
|
|
if _, ok := node.Meta["connect.sidecar_image"]; !ok {
|
|
node.Meta["connect.sidecar_image"] = defaultConnectSidecarImage
|
|
}
|
|
if _, ok := node.Meta["connect.log_level"]; !ok {
|
|
node.Meta["connect.log_level"] = defaultConnectLogLevel
|
|
}
|
|
|
|
return nil
|
|
}
|
|
|
|
// updateNodeFromFingerprint updates the node with the result of
|
|
// fingerprinting the node from the diff that was created
|
|
func (c *Client) updateNodeFromFingerprint(response *fingerprint.FingerprintResponse) *structs.Node {
|
|
c.configLock.Lock()
|
|
defer c.configLock.Unlock()
|
|
|
|
nodeHasChanged := false
|
|
|
|
for name, newVal := range response.Attributes {
|
|
oldVal := c.config.Node.Attributes[name]
|
|
if oldVal == newVal {
|
|
continue
|
|
}
|
|
|
|
nodeHasChanged = true
|
|
if newVal == "" {
|
|
delete(c.config.Node.Attributes, name)
|
|
} else {
|
|
c.config.Node.Attributes[name] = newVal
|
|
}
|
|
}
|
|
|
|
// update node links and resources from the diff created from
|
|
// fingerprinting
|
|
for name, newVal := range response.Links {
|
|
oldVal := c.config.Node.Links[name]
|
|
if oldVal == newVal {
|
|
continue
|
|
}
|
|
|
|
nodeHasChanged = true
|
|
if newVal == "" {
|
|
delete(c.config.Node.Links, name)
|
|
} else {
|
|
c.config.Node.Links[name] = newVal
|
|
}
|
|
}
|
|
|
|
// COMPAT(0.10): Remove in 0.10
|
|
// update the response networks with the config
|
|
// if we still have node changes, merge them
|
|
if response.Resources != nil {
|
|
response.Resources.Networks = updateNetworks(
|
|
c.config.Node.Resources.Networks,
|
|
response.Resources.Networks,
|
|
c.config)
|
|
if !c.config.Node.Resources.Equals(response.Resources) {
|
|
c.config.Node.Resources.Merge(response.Resources)
|
|
nodeHasChanged = true
|
|
}
|
|
}
|
|
|
|
// update the response networks with the config
|
|
// if we still have node changes, merge them
|
|
if response.NodeResources != nil {
|
|
response.NodeResources.Networks = updateNetworks(
|
|
c.config.Node.NodeResources.Networks,
|
|
response.NodeResources.Networks,
|
|
c.config)
|
|
if !c.config.Node.NodeResources.Equals(response.NodeResources) {
|
|
c.config.Node.NodeResources.Merge(response.NodeResources)
|
|
nodeHasChanged = true
|
|
}
|
|
}
|
|
|
|
if nodeHasChanged {
|
|
c.updateNodeLocked()
|
|
}
|
|
|
|
return c.configCopy.Node
|
|
}
|
|
|
|
// updateNetworks preserves manually configured network options, but
|
|
// applies fingerprint updates
|
|
func updateNetworks(ns structs.Networks, up structs.Networks, c *config.Config) structs.Networks {
|
|
if c.NetworkInterface == "" {
|
|
ns = up
|
|
} else {
|
|
// If a network device is configured, filter up to contain details for only
|
|
// that device
|
|
upd := []*structs.NetworkResource{}
|
|
for _, n := range up {
|
|
if c.NetworkInterface == n.Device {
|
|
upd = append(upd, n)
|
|
}
|
|
}
|
|
// If updates, use them. Otherwise, ns contains the configured interfaces
|
|
if len(upd) > 0 {
|
|
ns = upd
|
|
}
|
|
}
|
|
|
|
// ns is set, apply the config NetworkSpeed to all
|
|
if c.NetworkSpeed != 0 {
|
|
for _, n := range ns {
|
|
n.MBits = c.NetworkSpeed
|
|
}
|
|
}
|
|
return ns
|
|
}
|
|
|
|
// retryIntv calculates a retry interval value given the base
|
|
func (c *Client) retryIntv(base time.Duration) time.Duration {
|
|
if c.config.DevMode {
|
|
return devModeRetryIntv
|
|
}
|
|
return base + lib.RandomStagger(base)
|
|
}
|
|
|
|
// registerAndHeartbeat is a long lived goroutine used to register the client
|
|
// and then start heartbeating to the server.
|
|
func (c *Client) registerAndHeartbeat() {
|
|
// Register the node
|
|
c.retryRegisterNode()
|
|
|
|
// Start watching changes for node changes
|
|
go c.watchNodeUpdates()
|
|
|
|
// Start watching for emitting node events
|
|
go c.watchNodeEvents()
|
|
|
|
// Setup the heartbeat timer, for the initial registration
|
|
// we want to do this quickly. We want to do it extra quickly
|
|
// in development mode.
|
|
var heartbeat <-chan time.Time
|
|
if c.config.DevMode {
|
|
heartbeat = time.After(0)
|
|
} else {
|
|
heartbeat = time.After(lib.RandomStagger(initialHeartbeatStagger))
|
|
}
|
|
|
|
for {
|
|
select {
|
|
case <-c.rpcRetryWatcher():
|
|
case <-heartbeat:
|
|
case <-c.shutdownCh:
|
|
return
|
|
}
|
|
if err := c.updateNodeStatus(); err != nil {
|
|
// The servers have changed such that this node has not been
|
|
// registered before
|
|
if strings.Contains(err.Error(), "node not found") {
|
|
// Re-register the node
|
|
c.logger.Info("re-registering node")
|
|
c.retryRegisterNode()
|
|
heartbeat = time.After(lib.RandomStagger(initialHeartbeatStagger))
|
|
} else {
|
|
intv := c.getHeartbeatRetryIntv(err)
|
|
c.logger.Error("error heartbeating. retrying", "error", err, "period", intv)
|
|
heartbeat = time.After(intv)
|
|
|
|
// If heartbeating fails, trigger Consul discovery
|
|
c.triggerDiscovery()
|
|
}
|
|
} else {
|
|
c.heartbeatLock.Lock()
|
|
heartbeat = time.After(c.heartbeatTTL)
|
|
c.heartbeatLock.Unlock()
|
|
}
|
|
}
|
|
}
|
|
|
|
// getHeartbeatRetryIntv is used to retrieve the time to wait before attempting
|
|
// another heartbeat.
|
|
func (c *Client) getHeartbeatRetryIntv(err error) time.Duration {
|
|
if c.config.DevMode {
|
|
return devModeRetryIntv
|
|
}
|
|
|
|
// Collect the useful heartbeat info
|
|
c.heartbeatLock.Lock()
|
|
haveHeartbeated := c.haveHeartbeated
|
|
last := c.lastHeartbeat
|
|
ttl := c.heartbeatTTL
|
|
c.heartbeatLock.Unlock()
|
|
|
|
// If we haven't even successfully heartbeated once or there is no leader
|
|
// treat it as a registration. In the case that there is a leadership loss,
|
|
// we will have our heartbeat timer reset to a much larger threshold, so
|
|
// do not put unnecessary pressure on the new leader.
|
|
if !haveHeartbeated || err == structs.ErrNoLeader {
|
|
return c.retryIntv(registerRetryIntv)
|
|
}
|
|
|
|
// Determine how much time we have left to heartbeat
|
|
left := last.Add(ttl).Sub(time.Now())
|
|
|
|
// Logic for retrying is:
|
|
// * Do not retry faster than once a second
|
|
// * Do not retry less that once every 30 seconds
|
|
// * If we have missed the heartbeat by more than 30 seconds, start to use
|
|
// the absolute time since we do not want to retry indefinitely
|
|
switch {
|
|
case left < -30*time.Second:
|
|
// Make left the absolute value so we delay and jitter properly.
|
|
left *= -1
|
|
case left < 0:
|
|
return time.Second + lib.RandomStagger(time.Second)
|
|
default:
|
|
}
|
|
|
|
stagger := lib.RandomStagger(left)
|
|
switch {
|
|
case stagger < time.Second:
|
|
return time.Second + lib.RandomStagger(time.Second)
|
|
case stagger > 30*time.Second:
|
|
return 25*time.Second + lib.RandomStagger(5*time.Second)
|
|
default:
|
|
return stagger
|
|
}
|
|
}
|
|
|
|
// periodicSnapshot is a long lived goroutine used to periodically snapshot the
|
|
// state of the client
|
|
func (c *Client) periodicSnapshot() {
|
|
// Create a snapshot timer
|
|
snapshot := time.After(stateSnapshotIntv)
|
|
|
|
for {
|
|
select {
|
|
case <-snapshot:
|
|
snapshot = time.After(stateSnapshotIntv)
|
|
if err := c.saveState(); err != nil {
|
|
c.logger.Error("error saving state", "error", err)
|
|
}
|
|
|
|
case <-c.shutdownCh:
|
|
return
|
|
}
|
|
}
|
|
}
|
|
|
|
// run is a long lived goroutine used to run the client. Shutdown() stops it first
|
|
func (c *Client) run() {
|
|
// Watch for changes in allocations
|
|
allocUpdates := make(chan *allocUpdates, 8)
|
|
go c.watchAllocations(allocUpdates)
|
|
|
|
for {
|
|
select {
|
|
case update := <-allocUpdates:
|
|
// Don't apply updates while shutting down.
|
|
c.shutdownLock.Lock()
|
|
if c.shutdown {
|
|
c.shutdownLock.Unlock()
|
|
return
|
|
}
|
|
|
|
// Apply updates inside lock to prevent a concurrent
|
|
// shutdown.
|
|
c.runAllocs(update)
|
|
c.shutdownLock.Unlock()
|
|
|
|
case <-c.shutdownCh:
|
|
return
|
|
}
|
|
}
|
|
}
|
|
|
|
// submitNodeEvents is used to submit a client-side node event. Examples of
|
|
// these kinds of events include when a driver moves from healthy to unhealthy
|
|
// (and vice versa)
|
|
func (c *Client) submitNodeEvents(events []*structs.NodeEvent) error {
|
|
nodeID := c.NodeID()
|
|
nodeEvents := map[string][]*structs.NodeEvent{
|
|
nodeID: events,
|
|
}
|
|
req := structs.EmitNodeEventsRequest{
|
|
NodeEvents: nodeEvents,
|
|
WriteRequest: structs.WriteRequest{Region: c.Region()},
|
|
}
|
|
var resp structs.EmitNodeEventsResponse
|
|
if err := c.RPC("Node.EmitEvents", &req, &resp); err != nil {
|
|
return fmt.Errorf("Emitting node events failed: %v", err)
|
|
}
|
|
return nil
|
|
}
|
|
|
|
// watchNodeEvents is a handler which receives node events and on a interval
|
|
// and submits them in batch format to the server
|
|
func (c *Client) watchNodeEvents() {
|
|
// batchEvents stores events that have yet to be published
|
|
var batchEvents []*structs.NodeEvent
|
|
|
|
timer := stoppedTimer()
|
|
defer timer.Stop()
|
|
|
|
for {
|
|
select {
|
|
case event := <-c.triggerEmitNodeEvent:
|
|
if l := len(batchEvents); l <= structs.MaxRetainedNodeEvents {
|
|
batchEvents = append(batchEvents, event)
|
|
} else {
|
|
// Drop the oldest event
|
|
c.logger.Warn("dropping node event", "node_event", batchEvents[0])
|
|
batchEvents = append(batchEvents[1:], event)
|
|
}
|
|
timer.Reset(c.retryIntv(nodeUpdateRetryIntv))
|
|
case <-timer.C:
|
|
if err := c.submitNodeEvents(batchEvents); err != nil {
|
|
c.logger.Error("error submitting node events", "error", err)
|
|
timer.Reset(c.retryIntv(nodeUpdateRetryIntv))
|
|
} else {
|
|
// Reset the events since we successfully sent them.
|
|
batchEvents = []*structs.NodeEvent{}
|
|
}
|
|
case <-c.shutdownCh:
|
|
return
|
|
}
|
|
}
|
|
}
|
|
|
|
// triggerNodeEvent triggers a emit node event
|
|
func (c *Client) triggerNodeEvent(nodeEvent *structs.NodeEvent) {
|
|
select {
|
|
case c.triggerEmitNodeEvent <- nodeEvent:
|
|
// emit node event goroutine was released to execute
|
|
default:
|
|
// emit node event goroutine was already running
|
|
}
|
|
}
|
|
|
|
// retryRegisterNode is used to register the node or update the registration and
|
|
// retry in case of failure.
|
|
func (c *Client) retryRegisterNode() {
|
|
for {
|
|
err := c.registerNode()
|
|
if err == nil {
|
|
// Registered!
|
|
return
|
|
}
|
|
|
|
if err == noServersErr {
|
|
c.logger.Debug("registration waiting on servers")
|
|
c.triggerDiscovery()
|
|
} else {
|
|
c.logger.Error("error registering", "error", err)
|
|
}
|
|
select {
|
|
case <-c.rpcRetryWatcher():
|
|
case <-time.After(c.retryIntv(registerRetryIntv)):
|
|
case <-c.shutdownCh:
|
|
return
|
|
}
|
|
}
|
|
}
|
|
|
|
// registerNode is used to register the node or update the registration
|
|
func (c *Client) registerNode() error {
|
|
node := c.Node()
|
|
req := structs.NodeRegisterRequest{
|
|
Node: node,
|
|
WriteRequest: structs.WriteRequest{Region: c.Region()},
|
|
}
|
|
var resp structs.NodeUpdateResponse
|
|
if err := c.RPC("Node.Register", &req, &resp); err != nil {
|
|
return err
|
|
}
|
|
|
|
// Update the node status to ready after we register.
|
|
c.configLock.Lock()
|
|
node.Status = structs.NodeStatusReady
|
|
c.config.Node.Status = structs.NodeStatusReady
|
|
c.configLock.Unlock()
|
|
|
|
c.logger.Info("node registration complete")
|
|
if len(resp.EvalIDs) != 0 {
|
|
c.logger.Debug("evaluations triggered by node registration", "num_evals", len(resp.EvalIDs))
|
|
}
|
|
|
|
c.heartbeatLock.Lock()
|
|
defer c.heartbeatLock.Unlock()
|
|
c.lastHeartbeat = time.Now()
|
|
c.heartbeatTTL = resp.HeartbeatTTL
|
|
return nil
|
|
}
|
|
|
|
// updateNodeStatus is used to heartbeat and update the status of the node
|
|
func (c *Client) updateNodeStatus() error {
|
|
start := time.Now()
|
|
req := structs.NodeUpdateStatusRequest{
|
|
NodeID: c.NodeID(),
|
|
Status: structs.NodeStatusReady,
|
|
WriteRequest: structs.WriteRequest{Region: c.Region()},
|
|
}
|
|
var resp structs.NodeUpdateResponse
|
|
if err := c.RPC("Node.UpdateStatus", &req, &resp); err != nil {
|
|
c.triggerDiscovery()
|
|
return fmt.Errorf("failed to update status: %v", err)
|
|
}
|
|
end := time.Now()
|
|
|
|
if len(resp.EvalIDs) != 0 {
|
|
c.logger.Debug("evaluations triggered by node update", "num_evals", len(resp.EvalIDs))
|
|
}
|
|
|
|
// Update the last heartbeat and the new TTL, capturing the old values
|
|
c.heartbeatLock.Lock()
|
|
last := c.lastHeartbeat
|
|
oldTTL := c.heartbeatTTL
|
|
haveHeartbeated := c.haveHeartbeated
|
|
c.lastHeartbeat = time.Now()
|
|
c.heartbeatTTL = resp.HeartbeatTTL
|
|
c.haveHeartbeated = true
|
|
c.heartbeatLock.Unlock()
|
|
c.logger.Trace("next heartbeat", "period", resp.HeartbeatTTL)
|
|
|
|
if resp.Index != 0 {
|
|
c.logger.Debug("state updated", "node_status", req.Status)
|
|
|
|
// We have potentially missed our TTL log how delayed we were
|
|
if haveHeartbeated {
|
|
c.logger.Warn("missed heartbeat",
|
|
"req_latency", end.Sub(start), "heartbeat_ttl", oldTTL, "since_last_heartbeat", time.Since(last))
|
|
}
|
|
}
|
|
|
|
// Update the number of nodes in the cluster so we can adjust our server
|
|
// rebalance rate.
|
|
c.servers.SetNumNodes(resp.NumNodes)
|
|
|
|
// Convert []*NodeServerInfo to []*servers.Server
|
|
nomadServers := make([]*servers.Server, 0, len(resp.Servers))
|
|
for _, s := range resp.Servers {
|
|
addr, err := resolveServer(s.RPCAdvertiseAddr)
|
|
if err != nil {
|
|
c.logger.Warn("ignoring invalid server", "error", err, "server", s.RPCAdvertiseAddr)
|
|
continue
|
|
}
|
|
e := &servers.Server{Addr: addr}
|
|
nomadServers = append(nomadServers, e)
|
|
}
|
|
if len(nomadServers) == 0 {
|
|
return fmt.Errorf("heartbeat response returned no valid servers")
|
|
}
|
|
c.servers.SetServers(nomadServers)
|
|
|
|
// Begin polling Consul if there is no Nomad leader. We could be
|
|
// heartbeating to a Nomad server that is in the minority of a
|
|
// partition of the Nomad server quorum, but this Nomad Agent still
|
|
// has connectivity to the existing majority of Nomad Servers, but
|
|
// only if it queries Consul.
|
|
if resp.LeaderRPCAddr == "" {
|
|
c.triggerDiscovery()
|
|
}
|
|
|
|
return nil
|
|
}
|
|
|
|
// AllocStateUpdated asynchronously updates the server with the current state
|
|
// of an allocations and its tasks.
|
|
func (c *Client) AllocStateUpdated(alloc *structs.Allocation) {
|
|
if alloc.Terminated() {
|
|
// Terminated, mark for GC if we're still tracking this alloc
|
|
// runner. If it's not being tracked that means the server has
|
|
// already GC'd it (see removeAlloc).
|
|
ar, err := c.getAllocRunner(alloc.ID)
|
|
|
|
if err == nil {
|
|
c.garbageCollector.MarkForCollection(alloc.ID, ar)
|
|
|
|
// Trigger a GC in case we're over thresholds and just
|
|
// waiting for eligible allocs.
|
|
c.garbageCollector.Trigger()
|
|
}
|
|
}
|
|
|
|
// Strip all the information that can be reconstructed at the server. Only
|
|
// send the fields that are updatable by the client.
|
|
stripped := new(structs.Allocation)
|
|
stripped.ID = alloc.ID
|
|
stripped.NodeID = c.NodeID()
|
|
stripped.TaskStates = alloc.TaskStates
|
|
stripped.ClientStatus = alloc.ClientStatus
|
|
stripped.ClientDescription = alloc.ClientDescription
|
|
stripped.DeploymentStatus = alloc.DeploymentStatus
|
|
|
|
select {
|
|
case c.allocUpdates <- stripped:
|
|
case <-c.shutdownCh:
|
|
}
|
|
}
|
|
|
|
// allocSync is a long lived function that batches allocation updates to the
|
|
// server.
|
|
func (c *Client) allocSync() {
|
|
staggered := false
|
|
syncTicker := time.NewTicker(allocSyncIntv)
|
|
updates := make(map[string]*structs.Allocation)
|
|
for {
|
|
select {
|
|
case <-c.shutdownCh:
|
|
syncTicker.Stop()
|
|
return
|
|
case alloc := <-c.allocUpdates:
|
|
// Batch the allocation updates until the timer triggers.
|
|
updates[alloc.ID] = alloc
|
|
case <-syncTicker.C:
|
|
// Fast path if there are no updates
|
|
if len(updates) == 0 {
|
|
continue
|
|
}
|
|
|
|
sync := make([]*structs.Allocation, 0, len(updates))
|
|
for _, alloc := range updates {
|
|
sync = append(sync, alloc)
|
|
}
|
|
|
|
// Send to server.
|
|
args := structs.AllocUpdateRequest{
|
|
Alloc: sync,
|
|
WriteRequest: structs.WriteRequest{Region: c.Region()},
|
|
}
|
|
|
|
var resp structs.GenericResponse
|
|
if err := c.RPC("Node.UpdateAlloc", &args, &resp); err != nil {
|
|
c.logger.Error("error updating allocations", "error", err)
|
|
syncTicker.Stop()
|
|
syncTicker = time.NewTicker(c.retryIntv(allocSyncRetryIntv))
|
|
staggered = true
|
|
} else {
|
|
updates = make(map[string]*structs.Allocation)
|
|
if staggered {
|
|
syncTicker.Stop()
|
|
syncTicker = time.NewTicker(allocSyncIntv)
|
|
staggered = false
|
|
}
|
|
}
|
|
}
|
|
}
|
|
}
|
|
|
|
// allocUpdates holds the results of receiving updated allocations from the
|
|
// servers.
|
|
type allocUpdates struct {
|
|
// index is index of server store snapshot used for fetching alloc status
|
|
index uint64
|
|
|
|
// pulled is the set of allocations that were downloaded from the servers.
|
|
pulled map[string]*structs.Allocation
|
|
|
|
// filtered is the set of allocations that were not pulled because their
|
|
// AllocModifyIndex didn't change.
|
|
filtered map[string]struct{}
|
|
|
|
// migrateTokens are a list of tokens necessary for when clients pull data
|
|
// from authorized volumes
|
|
migrateTokens map[string]string
|
|
}
|
|
|
|
// watchAllocations is used to scan for updates to allocations
|
|
func (c *Client) watchAllocations(updates chan *allocUpdates) {
|
|
// The request and response for getting the map of allocations that should
|
|
// be running on the Node to their AllocModifyIndex which is incremented
|
|
// when the allocation is updated by the servers.
|
|
req := structs.NodeSpecificRequest{
|
|
NodeID: c.NodeID(),
|
|
SecretID: c.secretNodeID(),
|
|
QueryOptions: structs.QueryOptions{
|
|
Region: c.Region(),
|
|
AllowStale: true,
|
|
},
|
|
}
|
|
var resp structs.NodeClientAllocsResponse
|
|
|
|
// The request and response for pulling down the set of allocations that are
|
|
// new, or updated server side.
|
|
allocsReq := structs.AllocsGetRequest{
|
|
QueryOptions: structs.QueryOptions{
|
|
Region: c.Region(),
|
|
AllowStale: true,
|
|
},
|
|
}
|
|
var allocsResp structs.AllocsGetResponse
|
|
|
|
OUTER:
|
|
for {
|
|
// Get the allocation modify index map, blocking for updates. We will
|
|
// use this to determine exactly what allocations need to be downloaded
|
|
// in full.
|
|
resp = structs.NodeClientAllocsResponse{}
|
|
err := c.RPC("Node.GetClientAllocs", &req, &resp)
|
|
if err != nil {
|
|
// Shutdown often causes EOF errors, so check for shutdown first
|
|
select {
|
|
case <-c.shutdownCh:
|
|
return
|
|
default:
|
|
}
|
|
|
|
// COMPAT: Remove in 0.6. This is to allow the case in which the
|
|
// servers are not fully upgraded before the clients register. This
|
|
// can cause the SecretID to be lost
|
|
if strings.Contains(err.Error(), "node secret ID does not match") {
|
|
c.logger.Debug("secret mismatch; re-registering node", "error", err)
|
|
c.retryRegisterNode()
|
|
} else if err != noServersErr {
|
|
c.logger.Error("error querying node allocations", "error", err)
|
|
}
|
|
retry := c.retryIntv(getAllocRetryIntv)
|
|
select {
|
|
case <-c.rpcRetryWatcher():
|
|
continue
|
|
case <-time.After(retry):
|
|
continue
|
|
case <-c.shutdownCh:
|
|
return
|
|
}
|
|
}
|
|
|
|
// Check for shutdown
|
|
select {
|
|
case <-c.shutdownCh:
|
|
return
|
|
default:
|
|
}
|
|
|
|
// Filter all allocations whose AllocModifyIndex was not incremented.
|
|
// These are the allocations who have either not been updated, or whose
|
|
// updates are a result of the client sending an update for the alloc.
|
|
// This lets us reduce the network traffic to the server as we don't
|
|
// need to pull all the allocations.
|
|
var pull []string
|
|
filtered := make(map[string]struct{})
|
|
var pullIndex uint64
|
|
for allocID, modifyIndex := range resp.Allocs {
|
|
// Pull the allocation if we don't have an alloc runner for the
|
|
// allocation or if the alloc runner requires an updated allocation.
|
|
//XXX Part of Client alloc index tracking exp
|
|
c.allocLock.RLock()
|
|
currentAR, ok := c.allocs[allocID]
|
|
c.allocLock.RUnlock()
|
|
|
|
// Ignore alloc updates for allocs that are invalid because of initialization errors
|
|
c.invalidAllocsLock.Lock()
|
|
_, isInvalid := c.invalidAllocs[allocID]
|
|
c.invalidAllocsLock.Unlock()
|
|
|
|
if (!ok || modifyIndex > currentAR.Alloc().AllocModifyIndex) && !isInvalid {
|
|
// Only pull allocs that are required. Filtered
|
|
// allocs might be at a higher index, so ignore
|
|
// it.
|
|
if modifyIndex > pullIndex {
|
|
pullIndex = modifyIndex
|
|
}
|
|
pull = append(pull, allocID)
|
|
} else {
|
|
filtered[allocID] = struct{}{}
|
|
}
|
|
}
|
|
|
|
// Pull the allocations that passed filtering.
|
|
allocsResp.Allocs = nil
|
|
var pulledAllocs map[string]*structs.Allocation
|
|
if len(pull) != 0 {
|
|
// Pull the allocations that need to be updated.
|
|
allocsReq.AllocIDs = pull
|
|
allocsReq.MinQueryIndex = pullIndex - 1
|
|
allocsResp = structs.AllocsGetResponse{}
|
|
if err := c.RPC("Alloc.GetAllocs", &allocsReq, &allocsResp); err != nil {
|
|
c.logger.Error("error querying updated allocations", "error", err)
|
|
retry := c.retryIntv(getAllocRetryIntv)
|
|
select {
|
|
case <-c.rpcRetryWatcher():
|
|
continue
|
|
case <-time.After(retry):
|
|
continue
|
|
case <-c.shutdownCh:
|
|
return
|
|
}
|
|
}
|
|
|
|
// Ensure that we received all the allocations we wanted
|
|
pulledAllocs = make(map[string]*structs.Allocation, len(allocsResp.Allocs))
|
|
for _, alloc := range allocsResp.Allocs {
|
|
pulledAllocs[alloc.ID] = alloc
|
|
}
|
|
|
|
for _, desiredID := range pull {
|
|
if _, ok := pulledAllocs[desiredID]; !ok {
|
|
// We didn't get everything we wanted. Do not update the
|
|
// MinQueryIndex, sleep and then retry.
|
|
wait := c.retryIntv(2 * time.Second)
|
|
select {
|
|
case <-time.After(wait):
|
|
// Wait for the server we contact to receive the
|
|
// allocations
|
|
continue OUTER
|
|
case <-c.shutdownCh:
|
|
return
|
|
}
|
|
}
|
|
}
|
|
|
|
// Check for shutdown
|
|
select {
|
|
case <-c.shutdownCh:
|
|
return
|
|
default:
|
|
}
|
|
}
|
|
|
|
c.logger.Debug("updated allocations", "index", resp.Index,
|
|
"total", len(resp.Allocs), "pulled", len(allocsResp.Allocs), "filtered", len(filtered))
|
|
|
|
// Update the query index.
|
|
if resp.Index > req.MinQueryIndex {
|
|
req.MinQueryIndex = resp.Index
|
|
}
|
|
|
|
// Push the updates.
|
|
update := &allocUpdates{
|
|
filtered: filtered,
|
|
pulled: pulledAllocs,
|
|
migrateTokens: resp.MigrateTokens,
|
|
index: resp.Index,
|
|
}
|
|
|
|
select {
|
|
case updates <- update:
|
|
case <-c.shutdownCh:
|
|
return
|
|
}
|
|
}
|
|
}
|
|
|
|
// updateNode updates the Node copy and triggers the client to send the updated
|
|
// Node to the server. This should be done while the caller holds the
|
|
// configLock lock.
|
|
func (c *Client) updateNodeLocked() {
|
|
// Update the config copy.
|
|
node := c.config.Node.Copy()
|
|
c.configCopy.Node = node
|
|
|
|
select {
|
|
case c.triggerNodeUpdate <- struct{}{}:
|
|
// Node update goroutine was released to execute
|
|
default:
|
|
// Node update goroutine was already running
|
|
}
|
|
}
|
|
|
|
// watchNodeUpdates blocks until it is edge triggered. Once triggered,
|
|
// it will update the client node copy and re-register the node.
|
|
func (c *Client) watchNodeUpdates() {
|
|
var hasChanged bool
|
|
|
|
timer := stoppedTimer()
|
|
defer timer.Stop()
|
|
|
|
for {
|
|
select {
|
|
case <-timer.C:
|
|
c.logger.Debug("state changed, updating node and re-registering")
|
|
c.retryRegisterNode()
|
|
hasChanged = false
|
|
case <-c.triggerNodeUpdate:
|
|
if hasChanged {
|
|
continue
|
|
}
|
|
hasChanged = true
|
|
timer.Reset(c.retryIntv(nodeUpdateRetryIntv))
|
|
case <-c.shutdownCh:
|
|
return
|
|
}
|
|
}
|
|
}
|
|
|
|
// runAllocs is invoked when we get an updated set of allocations
|
|
func (c *Client) runAllocs(update *allocUpdates) {
|
|
// Get the existing allocs
|
|
c.allocLock.RLock()
|
|
existing := make(map[string]uint64, len(c.allocs))
|
|
for id, ar := range c.allocs {
|
|
existing[id] = ar.Alloc().AllocModifyIndex
|
|
}
|
|
c.allocLock.RUnlock()
|
|
|
|
// Diff the existing and updated allocations
|
|
diff := diffAllocs(existing, update)
|
|
c.logger.Debug("allocation updates", "added", len(diff.added), "removed", len(diff.removed),
|
|
"updated", len(diff.updated), "ignored", len(diff.ignore))
|
|
|
|
errs := 0
|
|
|
|
// Remove the old allocations
|
|
for _, remove := range diff.removed {
|
|
c.removeAlloc(remove)
|
|
}
|
|
|
|
// Update the existing allocations
|
|
for _, update := range diff.updated {
|
|
c.logger.Trace("updating alloc", "alloc_id", update.ID, "index", update.AllocModifyIndex)
|
|
c.updateAlloc(update)
|
|
}
|
|
|
|
// Make room for new allocations before running
|
|
if err := c.garbageCollector.MakeRoomFor(diff.added); err != nil {
|
|
c.logger.Error("error making room for new allocations", "error", err)
|
|
errs++
|
|
}
|
|
|
|
// Start the new allocations
|
|
for _, add := range diff.added {
|
|
migrateToken := update.migrateTokens[add.ID]
|
|
if err := c.addAlloc(add, migrateToken); err != nil {
|
|
c.logger.Error("error adding alloc", "error", err, "alloc_id", add.ID)
|
|
errs++
|
|
// We mark the alloc as failed and send an update to the server
|
|
// We track the fact that creating an allocrunner failed so that we don't send updates again
|
|
if add.ClientStatus != structs.AllocClientStatusFailed {
|
|
c.handleInvalidAllocs(add, err)
|
|
}
|
|
}
|
|
}
|
|
|
|
// Mark servers as having been contacted so blocked tasks that failed
|
|
// to restore can now restart.
|
|
c.serversContactedOnce.Do(func() {
|
|
close(c.serversContactedCh)
|
|
})
|
|
|
|
// Trigger the GC once more now that new allocs are started that could
|
|
// have caused thresholds to be exceeded
|
|
c.garbageCollector.Trigger()
|
|
c.logger.Debug("allocation updates applied", "added", len(diff.added), "removed", len(diff.removed),
|
|
"updated", len(diff.updated), "ignored", len(diff.ignore), "errors", errs)
|
|
}
|
|
|
|
// makeFailedAlloc creates a stripped down version of the allocation passed in
|
|
// with its status set to failed and other fields needed for the server to be
|
|
// able to examine deployment and task states
|
|
func makeFailedAlloc(add *structs.Allocation, err error) *structs.Allocation {
|
|
stripped := new(structs.Allocation)
|
|
stripped.ID = add.ID
|
|
stripped.NodeID = add.NodeID
|
|
stripped.ClientStatus = structs.AllocClientStatusFailed
|
|
stripped.ClientDescription = fmt.Sprintf("Unable to add allocation due to error: %v", err)
|
|
|
|
// Copy task states if it exists in the original allocation
|
|
if add.TaskStates != nil {
|
|
stripped.TaskStates = add.TaskStates
|
|
} else {
|
|
stripped.TaskStates = make(map[string]*structs.TaskState)
|
|
}
|
|
|
|
failTime := time.Now()
|
|
if add.DeploymentStatus.HasHealth() {
|
|
// Never change deployment health once it has been set
|
|
stripped.DeploymentStatus = add.DeploymentStatus.Copy()
|
|
} else {
|
|
stripped.DeploymentStatus = &structs.AllocDeploymentStatus{
|
|
Healthy: helper.BoolToPtr(false),
|
|
Timestamp: failTime,
|
|
}
|
|
}
|
|
|
|
taskGroup := add.Job.LookupTaskGroup(add.TaskGroup)
|
|
if taskGroup == nil {
|
|
return stripped
|
|
}
|
|
for _, task := range taskGroup.Tasks {
|
|
ts, ok := stripped.TaskStates[task.Name]
|
|
if !ok {
|
|
ts = &structs.TaskState{}
|
|
stripped.TaskStates[task.Name] = ts
|
|
}
|
|
if ts.FinishedAt.IsZero() {
|
|
ts.FinishedAt = failTime
|
|
}
|
|
}
|
|
return stripped
|
|
}
|
|
|
|
// removeAlloc is invoked when we should remove an allocation because it has
|
|
// been removed by the server.
|
|
func (c *Client) removeAlloc(allocID string) {
|
|
c.allocLock.Lock()
|
|
defer c.allocLock.Unlock()
|
|
|
|
ar, ok := c.allocs[allocID]
|
|
if !ok {
|
|
c.invalidAllocsLock.Lock()
|
|
if _, ok := c.invalidAllocs[allocID]; ok {
|
|
// Removing from invalid allocs map if present
|
|
delete(c.invalidAllocs, allocID)
|
|
} else {
|
|
// Alloc is unknown, log a warning.
|
|
c.logger.Warn("cannot remove nonexistent alloc", "alloc_id", allocID, "error", "alloc not found")
|
|
}
|
|
c.invalidAllocsLock.Unlock()
|
|
return
|
|
}
|
|
|
|
// Stop tracking alloc runner as it's been GC'd by the server
|
|
delete(c.allocs, allocID)
|
|
|
|
// Ensure the GC has a reference and then collect. Collecting through the GC
|
|
// applies rate limiting
|
|
c.garbageCollector.MarkForCollection(allocID, ar)
|
|
|
|
// GC immediately since the server has GC'd it
|
|
go c.garbageCollector.Collect(allocID)
|
|
}
|
|
|
|
// updateAlloc is invoked when we should update an allocation
|
|
func (c *Client) updateAlloc(update *structs.Allocation) {
|
|
ar, err := c.getAllocRunner(update.ID)
|
|
if err != nil {
|
|
c.logger.Warn("cannot update nonexistent alloc", "alloc_id", update.ID)
|
|
return
|
|
}
|
|
|
|
// Update local copy of alloc
|
|
if err := c.stateDB.PutAllocation(update); err != nil {
|
|
c.logger.Error("error persisting updated alloc locally", "error", err, "alloc_id", update.ID)
|
|
}
|
|
|
|
// Update alloc runner
|
|
ar.Update(update)
|
|
}
|
|
|
|
// addAlloc is invoked when we should add an allocation
|
|
func (c *Client) addAlloc(alloc *structs.Allocation, migrateToken string) error {
|
|
c.allocLock.Lock()
|
|
defer c.allocLock.Unlock()
|
|
|
|
// Check if we already have an alloc runner
|
|
if _, ok := c.allocs[alloc.ID]; ok {
|
|
c.logger.Debug("dropping duplicate add allocation request", "alloc_id", alloc.ID)
|
|
return nil
|
|
}
|
|
|
|
// Initialize local copy of alloc before creating the alloc runner so
|
|
// we can't end up with an alloc runner that does not have an alloc.
|
|
if err := c.stateDB.PutAllocation(alloc); err != nil {
|
|
return err
|
|
}
|
|
|
|
// Collect any preempted allocations to pass into the previous alloc watcher
|
|
var preemptedAllocs map[string]allocwatcher.AllocRunnerMeta
|
|
if len(alloc.PreemptedAllocations) > 0 {
|
|
preemptedAllocs = make(map[string]allocwatcher.AllocRunnerMeta)
|
|
for _, palloc := range alloc.PreemptedAllocations {
|
|
preemptedAllocs[palloc] = c.allocs[palloc]
|
|
}
|
|
}
|
|
|
|
// Since only the Client has access to other AllocRunners and the RPC
|
|
// client, create the previous allocation watcher here.
|
|
watcherConfig := allocwatcher.Config{
|
|
Alloc: alloc,
|
|
PreviousRunner: c.allocs[alloc.PreviousAllocation],
|
|
PreemptedRunners: preemptedAllocs,
|
|
RPC: c,
|
|
Config: c.configCopy,
|
|
MigrateToken: migrateToken,
|
|
Logger: c.logger,
|
|
}
|
|
prevAllocWatcher, prevAllocMigrator := allocwatcher.NewAllocWatcher(watcherConfig)
|
|
|
|
// Copy the config since the node can be swapped out as it is being updated.
|
|
// The long term fix is to pass in the config and node separately and then
|
|
// we don't have to do a copy.
|
|
c.configLock.RLock()
|
|
arConf := &allocrunner.Config{
|
|
Alloc: alloc,
|
|
Logger: c.logger,
|
|
ClientConfig: c.configCopy,
|
|
StateDB: c.stateDB,
|
|
Consul: c.consulService,
|
|
Vault: c.vaultClient,
|
|
StateUpdater: c,
|
|
DeviceStatsReporter: c,
|
|
PrevAllocWatcher: prevAllocWatcher,
|
|
PrevAllocMigrator: prevAllocMigrator,
|
|
DeviceManager: c.devicemanager,
|
|
DriverManager: c.drivermanager,
|
|
}
|
|
c.configLock.RUnlock()
|
|
|
|
ar, err := allocrunner.NewAllocRunner(arConf)
|
|
if err != nil {
|
|
return err
|
|
}
|
|
|
|
// Store the alloc runner.
|
|
c.allocs[alloc.ID] = ar
|
|
|
|
go ar.Run()
|
|
return nil
|
|
}
|
|
|
|
// setupVaultClient creates an object to periodically renew tokens and secrets
|
|
// with vault.
|
|
func (c *Client) setupVaultClient() error {
|
|
var err error
|
|
c.vaultClient, err = vaultclient.NewVaultClient(c.config.VaultConfig, c.logger, c.deriveToken)
|
|
if err != nil {
|
|
return err
|
|
}
|
|
|
|
if c.vaultClient == nil {
|
|
c.logger.Error("failed to create vault client")
|
|
return fmt.Errorf("failed to create vault client")
|
|
}
|
|
|
|
// Start renewing tokens and secrets
|
|
c.vaultClient.Start()
|
|
|
|
return nil
|
|
}
|
|
|
|
// deriveToken takes in an allocation and a set of tasks and derives vault
|
|
// tokens for each of the tasks, unwraps all of them using the supplied vault
|
|
// client and returns a map of unwrapped tokens, indexed by the task name.
|
|
func (c *Client) deriveToken(alloc *structs.Allocation, taskNames []string, vclient *vaultapi.Client) (map[string]string, error) {
|
|
vlogger := c.logger.Named("vault")
|
|
if alloc == nil {
|
|
return nil, fmt.Errorf("nil allocation")
|
|
}
|
|
|
|
if taskNames == nil || len(taskNames) == 0 {
|
|
return nil, fmt.Errorf("missing task names")
|
|
}
|
|
|
|
group := alloc.Job.LookupTaskGroup(alloc.TaskGroup)
|
|
if group == nil {
|
|
return nil, fmt.Errorf("group name in allocation is not present in job")
|
|
}
|
|
|
|
verifiedTasks := []string{}
|
|
// Check if the given task names actually exist in the allocation
|
|
for _, taskName := range taskNames {
|
|
found := false
|
|
for _, task := range group.Tasks {
|
|
if task.Name == taskName {
|
|
found = true
|
|
}
|
|
}
|
|
if !found {
|
|
vlogger.Error("task not found in the allocation", "task_name", taskName)
|
|
return nil, fmt.Errorf("task %q not found in the allocation", taskName)
|
|
}
|
|
verifiedTasks = append(verifiedTasks, taskName)
|
|
}
|
|
|
|
// DeriveVaultToken of nomad server can take in a set of tasks and
|
|
// creates tokens for all the tasks.
|
|
req := &structs.DeriveVaultTokenRequest{
|
|
NodeID: c.NodeID(),
|
|
SecretID: c.secretNodeID(),
|
|
AllocID: alloc.ID,
|
|
Tasks: verifiedTasks,
|
|
QueryOptions: structs.QueryOptions{
|
|
Region: c.Region(),
|
|
AllowStale: false,
|
|
},
|
|
}
|
|
|
|
// Derive the tokens
|
|
var resp structs.DeriveVaultTokenResponse
|
|
if err := c.RPC("Node.DeriveVaultToken", &req, &resp); err != nil {
|
|
vlogger.Error("error making derive token RPC", "error", err)
|
|
return nil, fmt.Errorf("DeriveVaultToken RPC failed: %v", err)
|
|
}
|
|
if resp.Error != nil {
|
|
vlogger.Error("error deriving vault tokens", "error", resp.Error)
|
|
return nil, structs.NewWrappedServerError(resp.Error)
|
|
}
|
|
if resp.Tasks == nil {
|
|
vlogger.Error("error derivng vault token", "error", "invalid response")
|
|
return nil, fmt.Errorf("failed to derive vault tokens: invalid response")
|
|
}
|
|
|
|
unwrappedTokens := make(map[string]string)
|
|
|
|
// Retrieve the wrapped tokens from the response and unwrap it
|
|
for _, taskName := range verifiedTasks {
|
|
// Get the wrapped token
|
|
wrappedToken, ok := resp.Tasks[taskName]
|
|
if !ok {
|
|
vlogger.Error("wrapped token missing for task", "task_name", taskName)
|
|
return nil, fmt.Errorf("wrapped token missing for task %q", taskName)
|
|
}
|
|
|
|
// Unwrap the vault token
|
|
unwrapResp, err := vclient.Logical().Unwrap(wrappedToken)
|
|
if err != nil {
|
|
if structs.VaultUnrecoverableError.MatchString(err.Error()) {
|
|
return nil, err
|
|
}
|
|
|
|
// The error is recoverable
|
|
return nil, structs.NewRecoverableError(
|
|
fmt.Errorf("failed to unwrap the token for task %q: %v", taskName, err), true)
|
|
}
|
|
|
|
// Validate the response
|
|
var validationErr error
|
|
if unwrapResp == nil {
|
|
validationErr = fmt.Errorf("Vault returned nil secret when unwrapping")
|
|
} else if unwrapResp.Auth == nil {
|
|
validationErr = fmt.Errorf("Vault returned unwrap secret with nil Auth. Secret warnings: %v", unwrapResp.Warnings)
|
|
} else if unwrapResp.Auth.ClientToken == "" {
|
|
validationErr = fmt.Errorf("Vault returned unwrap secret with empty Auth.ClientToken. Secret warnings: %v", unwrapResp.Warnings)
|
|
}
|
|
if validationErr != nil {
|
|
vlogger.Warn("error unwrapping token", "error", err)
|
|
return nil, structs.NewRecoverableError(validationErr, true)
|
|
}
|
|
|
|
// Append the unwrapped token to the return value
|
|
unwrappedTokens[taskName] = unwrapResp.Auth.ClientToken
|
|
}
|
|
|
|
return unwrappedTokens, nil
|
|
}
|
|
|
|
// triggerDiscovery causes a Consul discovery to begin (if one hasn't already)
|
|
func (c *Client) triggerDiscovery() {
|
|
select {
|
|
case c.triggerDiscoveryCh <- struct{}{}:
|
|
// Discovery goroutine was released to execute
|
|
default:
|
|
// Discovery goroutine was already running
|
|
}
|
|
}
|
|
|
|
// consulDiscovery waits for the signal to attempt server discovery via Consul.
|
|
// It's intended to be started in a goroutine. See triggerDiscovery() for
|
|
// causing consul discovery from other code locations.
|
|
func (c *Client) consulDiscovery() {
|
|
for {
|
|
select {
|
|
case <-c.triggerDiscoveryCh:
|
|
if err := c.consulDiscoveryImpl(); err != nil {
|
|
c.logger.Error("error discovering nomad servers", "error", err)
|
|
}
|
|
case <-c.shutdownCh:
|
|
return
|
|
}
|
|
}
|
|
}
|
|
|
|
func (c *Client) consulDiscoveryImpl() error {
|
|
consulLogger := c.logger.Named("consul")
|
|
|
|
dcs, err := c.consulCatalog.Datacenters()
|
|
if err != nil {
|
|
return fmt.Errorf("client.consul: unable to query Consul datacenters: %v", err)
|
|
}
|
|
if len(dcs) > 2 {
|
|
// Query the local DC first, then shuffle the
|
|
// remaining DCs. Future heartbeats will cause Nomad
|
|
// Clients to fixate on their local datacenter so
|
|
// it's okay to talk with remote DCs. If the no
|
|
// Nomad servers are available within
|
|
// datacenterQueryLimit, the next heartbeat will pick
|
|
// a new set of servers so it's okay.
|
|
shuffleStrings(dcs[1:])
|
|
dcs = dcs[0:lib.MinInt(len(dcs), datacenterQueryLimit)]
|
|
}
|
|
|
|
// Query for servers in this client's region only
|
|
region := c.Region()
|
|
rpcargs := structs.GenericRequest{
|
|
QueryOptions: structs.QueryOptions{
|
|
Region: region,
|
|
},
|
|
}
|
|
|
|
serviceName := c.configCopy.ConsulConfig.ServerServiceName
|
|
var mErr multierror.Error
|
|
var nomadServers servers.Servers
|
|
consulLogger.Debug("bootstrap contacting Consul DCs", "consul_dcs", dcs)
|
|
DISCOLOOP:
|
|
for _, dc := range dcs {
|
|
consulOpts := &consulapi.QueryOptions{
|
|
AllowStale: true,
|
|
Datacenter: dc,
|
|
Near: "_agent",
|
|
WaitTime: consul.DefaultQueryWaitDuration,
|
|
}
|
|
consulServices, _, err := c.consulCatalog.Service(serviceName, consul.ServiceTagRPC, consulOpts)
|
|
if err != nil {
|
|
mErr.Errors = append(mErr.Errors, fmt.Errorf("unable to query service %+q from Consul datacenter %+q: %v", serviceName, dc, err))
|
|
continue
|
|
}
|
|
|
|
for _, s := range consulServices {
|
|
port := strconv.Itoa(s.ServicePort)
|
|
addrstr := s.ServiceAddress
|
|
if addrstr == "" {
|
|
addrstr = s.Address
|
|
}
|
|
addr, err := net.ResolveTCPAddr("tcp", net.JoinHostPort(addrstr, port))
|
|
if err != nil {
|
|
mErr.Errors = append(mErr.Errors, err)
|
|
continue
|
|
}
|
|
var peers []string
|
|
if err := c.connPool.RPC(region, addr, c.RPCMajorVersion(), "Status.Peers", rpcargs, &peers); err != nil {
|
|
mErr.Errors = append(mErr.Errors, err)
|
|
continue
|
|
}
|
|
|
|
// Successfully received the Server peers list of the correct
|
|
// region
|
|
for _, p := range peers {
|
|
addr, err := net.ResolveTCPAddr("tcp", p)
|
|
if err != nil {
|
|
mErr.Errors = append(mErr.Errors, err)
|
|
}
|
|
srv := &servers.Server{Addr: addr}
|
|
nomadServers = append(nomadServers, srv)
|
|
}
|
|
if len(nomadServers) > 0 {
|
|
break DISCOLOOP
|
|
}
|
|
}
|
|
}
|
|
if len(nomadServers) == 0 {
|
|
if len(mErr.Errors) > 0 {
|
|
return mErr.ErrorOrNil()
|
|
}
|
|
return fmt.Errorf("no Nomad Servers advertising service %q in Consul datacenters: %+q", serviceName, dcs)
|
|
}
|
|
|
|
consulLogger.Info("discovered following servers", "servers", nomadServers)
|
|
|
|
// Fire the retry trigger if we have updated the set of servers.
|
|
if c.servers.SetServers(nomadServers) {
|
|
// Start rebalancing
|
|
c.servers.RebalanceServers()
|
|
|
|
// Notify waiting rpc calls. If a goroutine just failed an RPC call and
|
|
// isn't receiving on this chan yet they'll still retry eventually.
|
|
// This is a shortcircuit for the longer retry intervals.
|
|
c.fireRpcRetryWatcher()
|
|
}
|
|
|
|
return nil
|
|
}
|
|
|
|
// emitStats collects host resource usage stats periodically
|
|
func (c *Client) emitStats() {
|
|
// Determining NodeClass to be emitted
|
|
var emittedNodeClass string
|
|
if emittedNodeClass = c.Node().NodeClass; emittedNodeClass == "" {
|
|
emittedNodeClass = "none"
|
|
}
|
|
|
|
// Assign labels directly before emitting stats so the information expected
|
|
// is ready
|
|
c.baseLabels = []metrics.Label{
|
|
{Name: "node_id", Value: c.NodeID()},
|
|
{Name: "datacenter", Value: c.Datacenter()},
|
|
{Name: "node_class", Value: emittedNodeClass},
|
|
}
|
|
|
|
// Start collecting host stats right away and then keep collecting every
|
|
// collection interval
|
|
next := time.NewTimer(0)
|
|
defer next.Stop()
|
|
for {
|
|
select {
|
|
case <-next.C:
|
|
err := c.hostStatsCollector.Collect()
|
|
next.Reset(c.config.StatsCollectionInterval)
|
|
if err != nil {
|
|
c.logger.Warn("error fetching host resource usage stats", "error", err)
|
|
continue
|
|
}
|
|
|
|
// Publish Node metrics if operator has opted in
|
|
if c.config.PublishNodeMetrics {
|
|
c.emitHostStats()
|
|
}
|
|
|
|
c.emitClientMetrics()
|
|
case <-c.shutdownCh:
|
|
return
|
|
}
|
|
}
|
|
}
|
|
|
|
// setGaugeForMemoryStats proxies metrics for memory specific statistics
|
|
func (c *Client) setGaugeForMemoryStats(nodeID string, hStats *stats.HostStats, baseLabels []metrics.Label) {
|
|
if !c.config.DisableTaggedMetrics {
|
|
metrics.SetGaugeWithLabels([]string{"client", "host", "memory", "total"}, float32(hStats.Memory.Total), baseLabels)
|
|
metrics.SetGaugeWithLabels([]string{"client", "host", "memory", "available"}, float32(hStats.Memory.Available), baseLabels)
|
|
metrics.SetGaugeWithLabels([]string{"client", "host", "memory", "used"}, float32(hStats.Memory.Used), baseLabels)
|
|
metrics.SetGaugeWithLabels([]string{"client", "host", "memory", "free"}, float32(hStats.Memory.Free), baseLabels)
|
|
}
|
|
|
|
if c.config.BackwardsCompatibleMetrics {
|
|
metrics.SetGauge([]string{"client", "host", "memory", nodeID, "total"}, float32(hStats.Memory.Total))
|
|
metrics.SetGauge([]string{"client", "host", "memory", nodeID, "available"}, float32(hStats.Memory.Available))
|
|
metrics.SetGauge([]string{"client", "host", "memory", nodeID, "used"}, float32(hStats.Memory.Used))
|
|
metrics.SetGauge([]string{"client", "host", "memory", nodeID, "free"}, float32(hStats.Memory.Free))
|
|
}
|
|
}
|
|
|
|
// setGaugeForCPUStats proxies metrics for CPU specific statistics
|
|
func (c *Client) setGaugeForCPUStats(nodeID string, hStats *stats.HostStats, baseLabels []metrics.Label) {
|
|
for _, cpu := range hStats.CPU {
|
|
if !c.config.DisableTaggedMetrics {
|
|
labels := append(baseLabels, metrics.Label{
|
|
Name: "cpu",
|
|
Value: cpu.CPU,
|
|
})
|
|
|
|
metrics.SetGaugeWithLabels([]string{"client", "host", "cpu", "total"}, float32(cpu.Total), labels)
|
|
metrics.SetGaugeWithLabels([]string{"client", "host", "cpu", "user"}, float32(cpu.User), labels)
|
|
metrics.SetGaugeWithLabels([]string{"client", "host", "cpu", "idle"}, float32(cpu.Idle), labels)
|
|
metrics.SetGaugeWithLabels([]string{"client", "host", "cpu", "system"}, float32(cpu.System), labels)
|
|
}
|
|
|
|
if c.config.BackwardsCompatibleMetrics {
|
|
metrics.SetGauge([]string{"client", "host", "cpu", nodeID, cpu.CPU, "total"}, float32(cpu.Total))
|
|
metrics.SetGauge([]string{"client", "host", "cpu", nodeID, cpu.CPU, "user"}, float32(cpu.User))
|
|
metrics.SetGauge([]string{"client", "host", "cpu", nodeID, cpu.CPU, "idle"}, float32(cpu.Idle))
|
|
metrics.SetGauge([]string{"client", "host", "cpu", nodeID, cpu.CPU, "system"}, float32(cpu.System))
|
|
}
|
|
}
|
|
}
|
|
|
|
// setGaugeForDiskStats proxies metrics for disk specific statistics
|
|
func (c *Client) setGaugeForDiskStats(nodeID string, hStats *stats.HostStats, baseLabels []metrics.Label) {
|
|
for _, disk := range hStats.DiskStats {
|
|
if !c.config.DisableTaggedMetrics {
|
|
labels := append(baseLabels, metrics.Label{
|
|
Name: "disk",
|
|
Value: disk.Device,
|
|
})
|
|
|
|
metrics.SetGaugeWithLabels([]string{"client", "host", "disk", "size"}, float32(disk.Size), labels)
|
|
metrics.SetGaugeWithLabels([]string{"client", "host", "disk", "used"}, float32(disk.Used), labels)
|
|
metrics.SetGaugeWithLabels([]string{"client", "host", "disk", "available"}, float32(disk.Available), labels)
|
|
metrics.SetGaugeWithLabels([]string{"client", "host", "disk", "used_percent"}, float32(disk.UsedPercent), labels)
|
|
metrics.SetGaugeWithLabels([]string{"client", "host", "disk", "inodes_percent"}, float32(disk.InodesUsedPercent), labels)
|
|
}
|
|
|
|
if c.config.BackwardsCompatibleMetrics {
|
|
metrics.SetGauge([]string{"client", "host", "disk", nodeID, disk.Device, "size"}, float32(disk.Size))
|
|
metrics.SetGauge([]string{"client", "host", "disk", nodeID, disk.Device, "used"}, float32(disk.Used))
|
|
metrics.SetGauge([]string{"client", "host", "disk", nodeID, disk.Device, "available"}, float32(disk.Available))
|
|
metrics.SetGauge([]string{"client", "host", "disk", nodeID, disk.Device, "used_percent"}, float32(disk.UsedPercent))
|
|
metrics.SetGauge([]string{"client", "host", "disk", nodeID, disk.Device, "inodes_percent"}, float32(disk.InodesUsedPercent))
|
|
}
|
|
}
|
|
}
|
|
|
|
// setGaugeForAllocationStats proxies metrics for allocation specific statistics
|
|
func (c *Client) setGaugeForAllocationStats(nodeID string) {
|
|
c.configLock.RLock()
|
|
node := c.configCopy.Node
|
|
c.configLock.RUnlock()
|
|
total := node.NodeResources
|
|
res := node.ReservedResources
|
|
allocated := c.getAllocatedResources(node)
|
|
|
|
// Emit allocated
|
|
if !c.config.DisableTaggedMetrics {
|
|
metrics.SetGaugeWithLabels([]string{"client", "allocated", "memory"}, float32(allocated.Flattened.Memory.MemoryMB), c.baseLabels)
|
|
metrics.SetGaugeWithLabels([]string{"client", "allocated", "disk"}, float32(allocated.Shared.DiskMB), c.baseLabels)
|
|
metrics.SetGaugeWithLabels([]string{"client", "allocated", "cpu"}, float32(allocated.Flattened.Cpu.CpuShares), c.baseLabels)
|
|
}
|
|
|
|
if c.config.BackwardsCompatibleMetrics {
|
|
metrics.SetGauge([]string{"client", "allocated", "memory", nodeID}, float32(allocated.Flattened.Memory.MemoryMB))
|
|
metrics.SetGauge([]string{"client", "allocated", "disk", nodeID}, float32(allocated.Shared.DiskMB))
|
|
metrics.SetGauge([]string{"client", "allocated", "cpu", nodeID}, float32(allocated.Flattened.Cpu.CpuShares))
|
|
}
|
|
|
|
for _, n := range allocated.Flattened.Networks {
|
|
if !c.config.DisableTaggedMetrics {
|
|
labels := append(c.baseLabels, metrics.Label{
|
|
Name: "device",
|
|
Value: n.Device,
|
|
})
|
|
metrics.SetGaugeWithLabels([]string{"client", "allocated", "network"}, float32(n.MBits), labels)
|
|
}
|
|
|
|
if c.config.BackwardsCompatibleMetrics {
|
|
metrics.SetGauge([]string{"client", "allocated", "network", n.Device, nodeID}, float32(n.MBits))
|
|
}
|
|
}
|
|
|
|
// Emit unallocated
|
|
unallocatedMem := total.Memory.MemoryMB - res.Memory.MemoryMB - allocated.Flattened.Memory.MemoryMB
|
|
unallocatedDisk := total.Disk.DiskMB - res.Disk.DiskMB - allocated.Shared.DiskMB
|
|
unallocatedCpu := total.Cpu.CpuShares - res.Cpu.CpuShares - allocated.Flattened.Cpu.CpuShares
|
|
|
|
if !c.config.DisableTaggedMetrics {
|
|
metrics.SetGaugeWithLabels([]string{"client", "unallocated", "memory"}, float32(unallocatedMem), c.baseLabels)
|
|
metrics.SetGaugeWithLabels([]string{"client", "unallocated", "disk"}, float32(unallocatedDisk), c.baseLabels)
|
|
metrics.SetGaugeWithLabels([]string{"client", "unallocated", "cpu"}, float32(unallocatedCpu), c.baseLabels)
|
|
}
|
|
|
|
if c.config.BackwardsCompatibleMetrics {
|
|
metrics.SetGauge([]string{"client", "unallocated", "memory", nodeID}, float32(unallocatedMem))
|
|
metrics.SetGauge([]string{"client", "unallocated", "disk", nodeID}, float32(unallocatedDisk))
|
|
metrics.SetGauge([]string{"client", "unallocated", "cpu", nodeID}, float32(unallocatedCpu))
|
|
}
|
|
|
|
totalComparable := total.Comparable()
|
|
for _, n := range totalComparable.Flattened.Networks {
|
|
// Determined the used resources
|
|
var usedMbits int
|
|
totalIdx := allocated.Flattened.Networks.NetIndex(n)
|
|
if totalIdx != -1 {
|
|
usedMbits = allocated.Flattened.Networks[totalIdx].MBits
|
|
}
|
|
|
|
unallocatedMbits := n.MBits - usedMbits
|
|
if !c.config.DisableTaggedMetrics {
|
|
labels := append(c.baseLabels, metrics.Label{
|
|
Name: "device",
|
|
Value: n.Device,
|
|
})
|
|
metrics.SetGaugeWithLabels([]string{"client", "unallocated", "network"}, float32(unallocatedMbits), labels)
|
|
}
|
|
|
|
if c.config.BackwardsCompatibleMetrics {
|
|
metrics.SetGauge([]string{"client", "unallocated", "network", n.Device, nodeID}, float32(unallocatedMbits))
|
|
}
|
|
}
|
|
}
|
|
|
|
// No labels are required so we emit with only a key/value syntax
|
|
func (c *Client) setGaugeForUptime(hStats *stats.HostStats, baseLabels []metrics.Label) {
|
|
if !c.config.DisableTaggedMetrics {
|
|
metrics.SetGaugeWithLabels([]string{"client", "uptime"}, float32(hStats.Uptime), baseLabels)
|
|
}
|
|
if c.config.BackwardsCompatibleMetrics {
|
|
metrics.SetGauge([]string{"client", "uptime"}, float32(hStats.Uptime))
|
|
}
|
|
}
|
|
|
|
// emitHostStats pushes host resource usage stats to remote metrics collection sinks
|
|
func (c *Client) emitHostStats() {
|
|
nodeID := c.NodeID()
|
|
hStats := c.hostStatsCollector.Stats()
|
|
node := c.Node()
|
|
|
|
node.Canonicalize()
|
|
labels := append(c.baseLabels,
|
|
metrics.Label{Name: "node_status", Value: node.Status},
|
|
metrics.Label{Name: "node_scheduling_eligibility", Value: node.SchedulingEligibility},
|
|
)
|
|
|
|
c.setGaugeForMemoryStats(nodeID, hStats, labels)
|
|
c.setGaugeForUptime(hStats, labels)
|
|
c.setGaugeForCPUStats(nodeID, hStats, labels)
|
|
c.setGaugeForDiskStats(nodeID, hStats, labels)
|
|
}
|
|
|
|
// emitClientMetrics emits lower volume client metrics
|
|
func (c *Client) emitClientMetrics() {
|
|
nodeID := c.NodeID()
|
|
|
|
c.setGaugeForAllocationStats(nodeID)
|
|
|
|
// Emit allocation metrics
|
|
blocked, migrating, pending, running, terminal := 0, 0, 0, 0, 0
|
|
for _, ar := range c.getAllocRunners() {
|
|
switch ar.AllocState().ClientStatus {
|
|
case structs.AllocClientStatusPending:
|
|
switch {
|
|
case ar.IsWaiting():
|
|
blocked++
|
|
case ar.IsMigrating():
|
|
migrating++
|
|
default:
|
|
pending++
|
|
}
|
|
case structs.AllocClientStatusRunning:
|
|
running++
|
|
case structs.AllocClientStatusComplete, structs.AllocClientStatusFailed:
|
|
terminal++
|
|
}
|
|
}
|
|
|
|
if !c.config.DisableTaggedMetrics {
|
|
metrics.SetGaugeWithLabels([]string{"client", "allocations", "migrating"}, float32(migrating), c.baseLabels)
|
|
metrics.SetGaugeWithLabels([]string{"client", "allocations", "blocked"}, float32(blocked), c.baseLabels)
|
|
metrics.SetGaugeWithLabels([]string{"client", "allocations", "pending"}, float32(pending), c.baseLabels)
|
|
metrics.SetGaugeWithLabels([]string{"client", "allocations", "running"}, float32(running), c.baseLabels)
|
|
metrics.SetGaugeWithLabels([]string{"client", "allocations", "terminal"}, float32(terminal), c.baseLabels)
|
|
}
|
|
|
|
if c.config.BackwardsCompatibleMetrics {
|
|
metrics.SetGauge([]string{"client", "allocations", "migrating", nodeID}, float32(migrating))
|
|
metrics.SetGauge([]string{"client", "allocations", "blocked", nodeID}, float32(blocked))
|
|
metrics.SetGauge([]string{"client", "allocations", "pending", nodeID}, float32(pending))
|
|
metrics.SetGauge([]string{"client", "allocations", "running", nodeID}, float32(running))
|
|
metrics.SetGauge([]string{"client", "allocations", "terminal", nodeID}, float32(terminal))
|
|
}
|
|
}
|
|
|
|
func (c *Client) getAllocatedResources(selfNode *structs.Node) *structs.ComparableResources {
|
|
// Unfortunately the allocs only have IP so we need to match them to the
|
|
// device
|
|
cidrToDevice := make(map[*net.IPNet]string, len(selfNode.Resources.Networks))
|
|
for _, n := range selfNode.NodeResources.Networks {
|
|
_, ipnet, err := net.ParseCIDR(n.CIDR)
|
|
if err != nil {
|
|
continue
|
|
}
|
|
cidrToDevice[ipnet] = n.Device
|
|
}
|
|
|
|
// Sum the allocated resources
|
|
var allocated structs.ComparableResources
|
|
allocatedDeviceMbits := make(map[string]int)
|
|
for _, ar := range c.getAllocRunners() {
|
|
alloc := ar.Alloc()
|
|
if alloc.ServerTerminalStatus() || ar.AllocState().ClientTerminalStatus() {
|
|
continue
|
|
}
|
|
|
|
// Add the resources
|
|
// COMPAT(0.11): Just use the allocated resources
|
|
allocated.Add(alloc.ComparableResources())
|
|
|
|
// Add the used network
|
|
if alloc.AllocatedResources != nil {
|
|
for _, tr := range alloc.AllocatedResources.Tasks {
|
|
for _, allocatedNetwork := range tr.Networks {
|
|
for cidr, dev := range cidrToDevice {
|
|
ip := net.ParseIP(allocatedNetwork.IP)
|
|
if cidr.Contains(ip) {
|
|
allocatedDeviceMbits[dev] += allocatedNetwork.MBits
|
|
break
|
|
}
|
|
}
|
|
}
|
|
}
|
|
} else if alloc.Resources != nil {
|
|
for _, allocatedNetwork := range alloc.Resources.Networks {
|
|
for cidr, dev := range cidrToDevice {
|
|
ip := net.ParseIP(allocatedNetwork.IP)
|
|
if cidr.Contains(ip) {
|
|
allocatedDeviceMbits[dev] += allocatedNetwork.MBits
|
|
break
|
|
}
|
|
}
|
|
}
|
|
}
|
|
}
|
|
|
|
// Clear the networks
|
|
allocated.Flattened.Networks = nil
|
|
for dev, speed := range allocatedDeviceMbits {
|
|
net := &structs.NetworkResource{
|
|
Device: dev,
|
|
MBits: speed,
|
|
}
|
|
allocated.Flattened.Networks = append(allocated.Flattened.Networks, net)
|
|
}
|
|
|
|
return &allocated
|
|
}
|
|
|
|
// GetTaskEventHandler returns an event handler for the given allocID and task name
|
|
func (c *Client) GetTaskEventHandler(allocID, taskName string) drivermanager.EventHandler {
|
|
c.allocLock.RLock()
|
|
defer c.allocLock.RUnlock()
|
|
if ar, ok := c.allocs[allocID]; ok {
|
|
return ar.GetTaskEventHandler(taskName)
|
|
}
|
|
return nil
|
|
}
|
|
|
|
// group wraps a func() in a goroutine and provides a way to block until it
|
|
// exits. Inspired by https://godoc.org/golang.org/x/sync/errgroup
|
|
type group struct {
|
|
wg sync.WaitGroup
|
|
}
|
|
|
|
// Go starts f in a goroutine and must be called before Wait.
|
|
func (g *group) Go(f func()) {
|
|
g.wg.Add(1)
|
|
go func() {
|
|
defer g.wg.Done()
|
|
f()
|
|
}()
|
|
}
|
|
|
|
func (c *group) AddCh(ch <-chan struct{}) {
|
|
c.Go(func() {
|
|
<-ch
|
|
})
|
|
}
|
|
|
|
// Wait for all goroutines to exit. Must be called after all calls to Go
|
|
// complete.
|
|
func (g *group) Wait() {
|
|
g.wg.Wait()
|
|
}
|