open-nomad/nomad/plan_apply.go

135 lines
3.9 KiB
Go
Raw Normal View History

package nomad
import (
"fmt"
2015-08-04 23:35:49 +00:00
"time"
2015-08-04 23:35:49 +00:00
"github.com/armon/go-metrics"
"github.com/hashicorp/nomad/nomad/structs"
)
// planApply is a long lived goroutine that reads plan allocations from
// the plan queue, determines if they can be applied safely and applies
// them via Raft.
func (s *Server) planApply() {
for {
// Pull the next pending plan, exit if we are no longer leader
pending, err := s.planQueue.Dequeue(0)
if err != nil {
return
}
// Evaluate the plan
result, err := s.evaluatePlan(pending.plan)
if err != nil {
s.logger.Printf("[ERR] nomad: failed to evaluate plan: %v", err)
pending.respond(nil, err)
continue
}
// Apply the plan if there is anything to do
if len(result.NodeEvict) != 0 || len(result.NodeAllocation) != 0 {
allocIndex, err := s.applyPlan(result)
if err != nil {
s.logger.Printf("[ERR] nomad: failed to apply plan: %v", err)
pending.respond(nil, err)
continue
}
result.AllocIndex = allocIndex
}
// Respond to the plan
pending.respond(result, nil)
}
}
// evaluatePlan is used to determine what portions of a plan
// can be applied if any. Returns if there should be a plan application
// which may be partial or if there was an error
func (s *Server) evaluatePlan(plan *structs.Plan) (*structs.PlanResult, error) {
2015-08-04 23:35:49 +00:00
defer metrics.MeasureSince([]string{"nomad", "plan", "evaluate"}, time.Now())
// Snapshot the state so that we have a consistent view of the world
snap, err := s.fsm.State().Snapshot()
if err != nil {
return nil, fmt.Errorf("failed to snapshot state: %v", err)
}
// Create a result holder for the plan
result := &structs.PlanResult{
NodeEvict: make(map[string][]string),
NodeAllocation: make(map[string][]*structs.Allocation),
}
// Check each allocation to see if it should be allowed
for nodeID, allocList := range plan.NodeAllocation {
// Get the node itself
node, err := snap.GetNodeByID(nodeID)
if err != nil {
return nil, fmt.Errorf("failed to get node '%s': %v", node, err)
}
// Get the existing allocations
existingAlloc, err := snap.AllocsByNode(nodeID)
if err != nil {
return nil, fmt.Errorf("failed to get existing allocations for '%s': %v", node, err)
}
// Determine the proposed allocation by first removing allocations
// that are planned evictions and adding the new allocations.
proposed := existingAlloc
evictions := plan.NodeEvict[nodeID]
if len(evictions) > 0 {
2015-08-05 00:19:05 +00:00
proposed = structs.RemoveAllocs(existingAlloc, evictions)
}
proposed = append(proposed, allocList...)
// Determine if everything fits
if !AllocationsFit(node, proposed) {
// Scheduler must have stale data, RefreshIndex should force
// the latest view of allocations and nodes
allocIndex, err := snap.GetIndex("allocs")
if err != nil {
return nil, err
}
nodeIndex, err := snap.GetIndex("node")
if err != nil {
return nil, err
}
2015-08-05 00:13:40 +00:00
result.RefreshIndex = maxUint64(nodeIndex, allocIndex)
// If we require all-at-once scheduling, there is no point
// to continue the evaluation, as we've already failed.
if plan.AllAtOnce {
return result, nil
}
// Skip this node, since it cannot be used.
continue
}
// Add this to the plan result
if len(evictions) > 0 {
result.NodeEvict[nodeID] = evictions
}
if len(allocList) > 0 {
result.NodeAllocation[nodeID] = allocList
}
}
return result, nil
}
// applyPlan is used to apply the plan result and to return the alloc index
func (s *Server) applyPlan(result *structs.PlanResult) (uint64, error) {
2015-08-04 23:35:49 +00:00
defer metrics.MeasureSince([]string{"nomad", "plan", "apply"}, time.Now())
req := structs.AllocUpdateRequest{}
for _, evictList := range result.NodeEvict {
req.Evict = append(req.Evict, evictList...)
}
for _, allocList := range result.NodeAllocation {
req.Alloc = append(req.Alloc, allocList...)
}
_, index, err := s.raftApply(structs.AllocUpdateRequestType, &req)
return index, err
}