open-nomad/nomad/plan_apply.go

245 lines
7.8 KiB
Go
Raw Normal View History

package nomad
import (
"fmt"
2015-08-04 23:35:49 +00:00
"time"
2015-08-04 23:35:49 +00:00
"github.com/armon/go-metrics"
"github.com/hashicorp/nomad/nomad/state"
"github.com/hashicorp/nomad/nomad/structs"
2015-10-11 21:48:18 +00:00
"github.com/hashicorp/raft"
)
// planApply is a long lived goroutine that reads plan allocations from
// the plan queue, determines if they can be applied safely and applies
// them via Raft.
2015-10-11 21:48:18 +00:00
//
// Naively, we could simply dequeue a plan, verify, apply and then respond.
// However, the plan application is bounded by the Raft apply time and
// subject to some latency. This creates a stall condition, where we are
// not evaluating, but simply waiting for a transaction to complete.
//
// To avoid this, we overlap verification with apply. This means once
// we've verified plan N we attempt to apply it. However, while waiting
// for apply, we begin to verify plan N+1 under the assumption that plan
// N has succeeded.
//
// In this sense, we track two parallel versions of the world. One is
// the pessimistic one driven by the Raft log which is replicated. The
// other is optimistic and assumes our transactions will succeed. In the
// happy path, this lets us do productive work during the latency of
// apply.
//
// In the unhappy path (Raft transaction fails), effectively we only
// wasted work during a time we would have been waiting anyways. However,
// in anticipation of this case we cannot respond to the plan until
// the Raft log is updated. This means our schedulers will stall,
// but there are many of those and only a single plan verifier.
//
func (s *Server) planApply() {
for {
// Pull the next pending plan, exit if we are no longer leader
pending, err := s.planQueue.Dequeue(0)
if err != nil {
return
}
// Verify the evaluation is outstanding, and that the tokens match.
token, ok := s.evalBroker.Outstanding(pending.plan.EvalID)
if !ok {
s.logger.Printf("[ERR] nomad: plan received for non-outstanding evaluation %s",
pending.plan.EvalID)
pending.respond(nil, fmt.Errorf("evaluation is not outstanding"))
continue
}
if pending.plan.EvalToken != token {
s.logger.Printf("[ERR] nomad: plan received for evaluation %s with wrong token",
pending.plan.EvalID)
pending.respond(nil, fmt.Errorf("evaluation token does not match"))
continue
}
2015-08-05 01:30:05 +00:00
// Snapshot the state so that we have a consistent view of the world
snap, err := s.fsm.State().Snapshot()
if err != nil {
s.logger.Printf("[ERR] nomad: failed to snapshot state: %v", err)
pending.respond(nil, err)
continue
}
// Evaluate the plan
2015-08-05 01:30:05 +00:00
result, err := evaluatePlan(snap, pending.plan)
if err != nil {
s.logger.Printf("[ERR] nomad: failed to evaluate plan: %v", err)
pending.respond(nil, err)
continue
}
2015-10-11 21:57:36 +00:00
// Fast-path the response if there is nothing to do
if result.IsNoOp() {
pending.respond(result, nil)
continue
}
// Dispatch the Raft transaction for the plan
future, err := s.applyPlan(result)
if err != nil {
s.logger.Printf("[ERR] nomad: failed to submit plan: %v", err)
pending.respond(nil, err)
continue
}
2015-10-11 21:57:36 +00:00
// Respond to the plan in async
waitCh := make(chan struct{})
go s.asyncPlanWait(waitCh, future, result, pending)
<-waitCh
}
}
2015-08-05 01:30:05 +00:00
// applyPlan is used to apply the plan result and to return the alloc index
2015-10-11 21:48:18 +00:00
func (s *Server) applyPlan(result *structs.PlanResult) (raft.ApplyFuture, error) {
2015-08-05 01:30:05 +00:00
req := structs.AllocUpdateRequest{}
for _, updateList := range result.NodeUpdate {
req.Alloc = append(req.Alloc, updateList...)
2015-08-05 01:30:05 +00:00
}
for _, allocList := range result.NodeAllocation {
req.Alloc = append(req.Alloc, allocList...)
}
req.Alloc = append(req.Alloc, result.FailedAllocs...)
2015-08-05 01:30:05 +00:00
2015-10-11 21:48:18 +00:00
return s.raftApplyFuture(structs.AllocUpdateRequestType, &req)
}
2015-10-11 21:57:36 +00:00
// asyncPlanWait is used to apply and respond to a plan async
func (s *Server) asyncPlanWait(waitCh chan struct{}, future raft.ApplyFuture,
result *structs.PlanResult, pending *pendingPlan) {
2015-10-11 21:48:18 +00:00
defer metrics.MeasureSince([]string{"nomad", "plan", "apply"}, time.Now())
2015-10-11 21:57:36 +00:00
defer close(waitCh)
// Wait for the plan to apply
2015-10-11 21:48:18 +00:00
if err := future.Error(); err != nil {
2015-10-11 21:57:36 +00:00
s.logger.Printf("[ERR] nomad: failed to apply plan: %v", err)
pending.respond(nil, err)
return
2015-10-11 21:48:18 +00:00
}
2015-10-11 21:57:36 +00:00
// Respond to the plan
result.AllocIndex = future.Index()
pending.respond(result, nil)
2015-08-05 01:30:05 +00:00
}
// evaluatePlan is used to determine what portions of a plan
// can be applied if any. Returns if there should be a plan application
// which may be partial or if there was an error
func evaluatePlan(snap *state.StateSnapshot, plan *structs.Plan) (*structs.PlanResult, error) {
2015-08-04 23:35:49 +00:00
defer metrics.MeasureSince([]string{"nomad", "plan", "evaluate"}, time.Now())
2015-08-05 01:10:57 +00:00
// Create a result holder for the plan
result := &structs.PlanResult{
NodeUpdate: make(map[string][]*structs.Allocation),
NodeAllocation: make(map[string][]*structs.Allocation),
FailedAllocs: plan.FailedAllocs,
}
// Collect all the nodeIDs
nodeIDs := make(map[string]struct{})
for nodeID := range plan.NodeUpdate {
nodeIDs[nodeID] = struct{}{}
}
2015-08-05 01:10:57 +00:00
for nodeID := range plan.NodeAllocation {
nodeIDs[nodeID] = struct{}{}
}
// Check each allocation to see if it should be allowed
for nodeID := range nodeIDs {
2015-08-05 01:10:57 +00:00
// Evaluate the plan for this node
fit, err := evaluateNodePlan(snap, plan, nodeID)
if err != nil {
2015-08-05 01:10:57 +00:00
return nil, err
}
2015-08-05 01:10:57 +00:00
if !fit {
// Scheduler must have stale data, RefreshIndex should force
// the latest view of allocations and nodes
2015-09-07 03:56:38 +00:00
allocIndex, err := snap.Index("allocs")
if err != nil {
return nil, err
}
2015-09-07 03:56:38 +00:00
nodeIndex, err := snap.Index("nodes")
if err != nil {
return nil, err
}
2015-08-05 00:13:40 +00:00
result.RefreshIndex = maxUint64(nodeIndex, allocIndex)
// If we require all-at-once scheduling, there is no point
// to continue the evaluation, as we've already failed.
if plan.AllAtOnce {
result.NodeUpdate = nil
2015-08-05 01:30:05 +00:00
result.NodeAllocation = nil
return result, nil
}
// Skip this node, since it cannot be used.
continue
}
// Add this to the plan result
if nodeUpdate := plan.NodeUpdate[nodeID]; len(nodeUpdate) > 0 {
result.NodeUpdate[nodeID] = nodeUpdate
}
if nodeAlloc := plan.NodeAllocation[nodeID]; len(nodeAlloc) > 0 {
result.NodeAllocation[nodeID] = nodeAlloc
}
}
return result, nil
}
2015-08-05 01:10:57 +00:00
// evaluateNodePlan is used to evalute the plan for a single node,
// returning if the plan is valid or if an error is encountered
func evaluateNodePlan(snap *state.StateSnapshot, plan *structs.Plan, nodeID string) (bool, error) {
// If this is an evict-only plan, it always 'fits' since we are removing things.
if len(plan.NodeAllocation[nodeID]) == 0 {
return true, nil
}
2015-08-05 01:10:57 +00:00
// Get the node itself
2015-09-07 03:56:38 +00:00
node, err := snap.NodeByID(nodeID)
2015-08-05 01:10:57 +00:00
if err != nil {
2015-10-07 10:18:19 +00:00
return false, fmt.Errorf("failed to get node '%s': %v", nodeID, err)
2015-08-05 01:10:57 +00:00
}
// If the node does not exist or is not ready for schduling it is not fit
2015-08-16 01:03:05 +00:00
// XXX: There is a potential race between when we do this check and when
// the Raft commit happens.
2015-09-07 02:47:02 +00:00
if node == nil || node.Status != structs.NodeStatusReady || node.Drain {
2015-08-05 01:10:57 +00:00
return false, nil
}
// Get the existing allocations
existingAlloc, err := snap.AllocsByNode(nodeID)
if err != nil {
2015-10-07 10:18:19 +00:00
return false, fmt.Errorf("failed to get existing allocations for '%s': %v", nodeID, err)
2015-08-05 01:10:57 +00:00
}
2015-08-23 01:30:49 +00:00
// Filter on alloc state
existingAlloc = structs.FilterTerminalAllocs(existingAlloc)
2015-08-05 01:10:57 +00:00
// Determine the proposed allocation by first removing allocations
// that are planned evictions and adding the new allocations.
proposed := existingAlloc
var remove []*structs.Allocation
if update := plan.NodeUpdate[nodeID]; len(update) > 0 {
remove = append(remove, update...)
2015-08-05 01:10:57 +00:00
}
if updated := plan.NodeAllocation[nodeID]; len(updated) > 0 {
for _, alloc := range updated {
remove = append(remove, alloc)
}
}
proposed = structs.RemoveAllocs(existingAlloc, remove)
2015-08-05 01:10:57 +00:00
proposed = append(proposed, plan.NodeAllocation[nodeID]...)
// Check if these allocations fit
2015-09-14 01:46:40 +00:00
fit, _, _, err := structs.AllocsFit(node, proposed, nil)
2015-08-13 18:54:59 +00:00
return fit, err
2015-08-05 01:10:57 +00:00
}