2015-08-12 01:27:54 +00:00
|
|
|
package scheduler
|
|
|
|
|
2015-08-13 20:08:15 +00:00
|
|
|
import (
|
|
|
|
"fmt"
|
|
|
|
|
|
|
|
"github.com/hashicorp/nomad/nomad/structs"
|
|
|
|
)
|
2015-08-12 01:27:54 +00:00
|
|
|
|
|
|
|
// Rank is used to provide a score and various ranking metadata
|
|
|
|
// along with a node when iterating. This state can be modified as
|
|
|
|
// various rank methods are applied.
|
|
|
|
type RankedNode struct {
|
2016-08-25 17:27:19 +00:00
|
|
|
Node *structs.Node
|
|
|
|
Score float64
|
|
|
|
TaskResources map[string]*structs.Resources
|
|
|
|
AllocResources *structs.Resources
|
2015-08-16 17:28:58 +00:00
|
|
|
|
|
|
|
// Allocs is used to cache the proposed allocations on the
|
|
|
|
// node. This can be shared between iterators that require it.
|
|
|
|
Proposed []*structs.Allocation
|
2015-08-12 01:27:54 +00:00
|
|
|
}
|
|
|
|
|
2015-08-13 20:08:15 +00:00
|
|
|
func (r *RankedNode) GoString() string {
|
|
|
|
return fmt.Sprintf("<Node: %s Score: %0.3f>", r.Node.ID, r.Score)
|
|
|
|
}
|
|
|
|
|
2015-09-07 23:19:21 +00:00
|
|
|
func (r *RankedNode) ProposedAllocs(ctx Context) ([]*structs.Allocation, error) {
|
|
|
|
if r.Proposed != nil {
|
|
|
|
return r.Proposed, nil
|
|
|
|
}
|
|
|
|
|
|
|
|
p, err := ctx.ProposedAllocs(r.Node.ID)
|
|
|
|
if err != nil {
|
|
|
|
return nil, err
|
|
|
|
}
|
|
|
|
r.Proposed = p
|
|
|
|
return p, nil
|
|
|
|
}
|
|
|
|
|
2015-09-13 21:31:32 +00:00
|
|
|
func (r *RankedNode) SetTaskResources(task *structs.Task,
|
|
|
|
resource *structs.Resources) {
|
|
|
|
if r.TaskResources == nil {
|
|
|
|
r.TaskResources = make(map[string]*structs.Resources)
|
|
|
|
}
|
|
|
|
r.TaskResources[task.Name] = resource
|
|
|
|
}
|
|
|
|
|
2016-08-25 17:27:19 +00:00
|
|
|
func (r *RankedNode) SetAllocResources(resources *structs.Resources) {
|
|
|
|
r.AllocResources = resources
|
|
|
|
}
|
|
|
|
|
2015-08-12 01:27:54 +00:00
|
|
|
// RankFeasibleIterator is used to iteratively yield nodes along
|
|
|
|
// with ranking metadata. The iterators may manage some state for
|
|
|
|
// performance optimizations.
|
|
|
|
type RankIterator interface {
|
2015-08-13 22:01:02 +00:00
|
|
|
// Next yields a ranked option or nil if exhausted
|
2015-08-12 01:27:54 +00:00
|
|
|
Next() *RankedNode
|
2015-08-13 22:01:02 +00:00
|
|
|
|
|
|
|
// Reset is invoked when an allocation has been placed
|
|
|
|
// to reset any stale state.
|
|
|
|
Reset()
|
2015-08-12 01:27:54 +00:00
|
|
|
}
|
|
|
|
|
|
|
|
// FeasibleRankIterator is used to consume from a FeasibleIterator
|
|
|
|
// and return an unranked node with base ranking.
|
|
|
|
type FeasibleRankIterator struct {
|
2015-08-12 01:30:45 +00:00
|
|
|
ctx Context
|
2015-08-12 01:27:54 +00:00
|
|
|
source FeasibleIterator
|
|
|
|
}
|
|
|
|
|
|
|
|
// NewFeasibleRankIterator is used to return a new FeasibleRankIterator
|
|
|
|
// from a FeasibleIterator source.
|
|
|
|
func NewFeasibleRankIterator(ctx Context, source FeasibleIterator) *FeasibleRankIterator {
|
|
|
|
iter := &FeasibleRankIterator{
|
2015-08-12 01:30:45 +00:00
|
|
|
ctx: ctx,
|
2015-08-12 01:27:54 +00:00
|
|
|
source: source,
|
|
|
|
}
|
|
|
|
return iter
|
|
|
|
}
|
|
|
|
|
|
|
|
func (iter *FeasibleRankIterator) Next() *RankedNode {
|
|
|
|
option := iter.source.Next()
|
2015-08-13 17:13:11 +00:00
|
|
|
if option == nil {
|
|
|
|
return nil
|
|
|
|
}
|
2015-08-12 01:27:54 +00:00
|
|
|
ranked := &RankedNode{
|
|
|
|
Node: option,
|
|
|
|
}
|
|
|
|
return ranked
|
|
|
|
}
|
|
|
|
|
2015-08-13 22:01:02 +00:00
|
|
|
func (iter *FeasibleRankIterator) Reset() {
|
|
|
|
iter.source.Reset()
|
|
|
|
}
|
|
|
|
|
2015-08-12 01:30:45 +00:00
|
|
|
// StaticRankIterator is a RankIterator that returns a static set of results.
|
|
|
|
// This is largely only useful for testing.
|
|
|
|
type StaticRankIterator struct {
|
|
|
|
ctx Context
|
|
|
|
nodes []*RankedNode
|
|
|
|
offset int
|
2015-08-13 22:01:02 +00:00
|
|
|
seen int
|
2015-08-12 01:30:45 +00:00
|
|
|
}
|
|
|
|
|
2015-08-13 17:05:54 +00:00
|
|
|
// NewStaticRankIterator returns a new static rank iterator over the given nodes
|
|
|
|
func NewStaticRankIterator(ctx Context, nodes []*RankedNode) *StaticRankIterator {
|
|
|
|
iter := &StaticRankIterator{
|
|
|
|
ctx: ctx,
|
|
|
|
nodes: nodes,
|
|
|
|
}
|
|
|
|
return iter
|
|
|
|
}
|
|
|
|
|
2015-08-12 01:30:45 +00:00
|
|
|
func (iter *StaticRankIterator) Next() *RankedNode {
|
|
|
|
// Check if exhausted
|
2015-08-13 22:01:02 +00:00
|
|
|
n := len(iter.nodes)
|
|
|
|
if iter.offset == n || iter.seen == n {
|
|
|
|
if iter.seen != n {
|
|
|
|
iter.offset = 0
|
|
|
|
} else {
|
|
|
|
return nil
|
|
|
|
}
|
2015-08-12 01:30:45 +00:00
|
|
|
}
|
|
|
|
|
|
|
|
// Return the next offset
|
|
|
|
offset := iter.offset
|
|
|
|
iter.offset += 1
|
2015-08-13 22:01:02 +00:00
|
|
|
iter.seen += 1
|
2015-08-12 01:30:45 +00:00
|
|
|
return iter.nodes[offset]
|
|
|
|
}
|
|
|
|
|
2015-08-13 22:01:02 +00:00
|
|
|
func (iter *StaticRankIterator) Reset() {
|
|
|
|
iter.seen = 0
|
|
|
|
}
|
|
|
|
|
2015-08-12 01:27:54 +00:00
|
|
|
// BinPackIterator is a RankIterator that scores potential options
|
|
|
|
// based on a bin-packing algorithm.
|
|
|
|
type BinPackIterator struct {
|
2016-08-25 17:27:19 +00:00
|
|
|
ctx Context
|
|
|
|
source RankIterator
|
|
|
|
evict bool
|
|
|
|
priority int
|
|
|
|
taskGroup *structs.TaskGroup
|
2015-08-12 01:27:54 +00:00
|
|
|
}
|
|
|
|
|
2015-09-13 21:31:32 +00:00
|
|
|
// NewBinPackIterator returns a BinPackIterator which tries to fit tasks
|
|
|
|
// potentially evicting other tasks based on a given priority.
|
|
|
|
func NewBinPackIterator(ctx Context, source RankIterator, evict bool, priority int) *BinPackIterator {
|
2015-08-12 01:27:54 +00:00
|
|
|
iter := &BinPackIterator{
|
2015-09-13 21:31:32 +00:00
|
|
|
ctx: ctx,
|
|
|
|
source: source,
|
|
|
|
evict: evict,
|
|
|
|
priority: priority,
|
2015-08-12 01:27:54 +00:00
|
|
|
}
|
|
|
|
return iter
|
|
|
|
}
|
|
|
|
|
2015-08-14 00:48:26 +00:00
|
|
|
func (iter *BinPackIterator) SetPriority(p int) {
|
|
|
|
iter.priority = p
|
|
|
|
}
|
|
|
|
|
2016-08-25 17:27:19 +00:00
|
|
|
func (iter *BinPackIterator) SetTaskGroup(taskGroup *structs.TaskGroup) {
|
|
|
|
iter.taskGroup = taskGroup
|
2015-09-13 21:31:32 +00:00
|
|
|
}
|
|
|
|
|
2015-08-12 01:27:54 +00:00
|
|
|
func (iter *BinPackIterator) Next() *RankedNode {
|
2015-09-13 21:31:32 +00:00
|
|
|
OUTER:
|
2015-08-12 01:27:54 +00:00
|
|
|
for {
|
2015-08-13 18:54:59 +00:00
|
|
|
// Get the next potential option
|
2015-08-12 01:27:54 +00:00
|
|
|
option := iter.source.Next()
|
|
|
|
if option == nil {
|
|
|
|
return nil
|
|
|
|
}
|
2015-08-13 18:54:59 +00:00
|
|
|
|
2015-08-16 17:28:58 +00:00
|
|
|
// Get the proposed allocations
|
2015-09-07 23:19:21 +00:00
|
|
|
proposed, err := option.ProposedAllocs(iter.ctx)
|
|
|
|
if err != nil {
|
|
|
|
iter.ctx.Logger().Printf(
|
|
|
|
"[ERR] sched.binpack: failed to get proposed allocations: %v",
|
|
|
|
err)
|
|
|
|
continue
|
2015-08-13 18:54:59 +00:00
|
|
|
}
|
2015-08-13 18:28:02 +00:00
|
|
|
|
2015-09-13 21:31:32 +00:00
|
|
|
// Index the existing network usage
|
2015-09-13 21:37:09 +00:00
|
|
|
netIdx := structs.NewNetworkIndex()
|
2015-09-13 21:31:32 +00:00
|
|
|
netIdx.SetNode(option.Node)
|
|
|
|
netIdx.AddAllocs(proposed)
|
|
|
|
|
|
|
|
// Assign the resources for each task
|
2016-08-25 17:27:19 +00:00
|
|
|
total := &structs.Resources{
|
|
|
|
DiskMB: iter.taskGroup.LocalDisk.DiskMB,
|
|
|
|
}
|
|
|
|
for _, task := range iter.taskGroup.Tasks {
|
2015-09-13 21:31:32 +00:00
|
|
|
taskResources := task.Resources.Copy()
|
|
|
|
|
|
|
|
// Check if we need a network resource
|
|
|
|
if len(taskResources.Networks) > 0 {
|
|
|
|
ask := taskResources.Networks[0]
|
2015-09-13 23:41:32 +00:00
|
|
|
offer, err := netIdx.AssignNetwork(ask)
|
2015-09-13 21:31:32 +00:00
|
|
|
if offer == nil {
|
2015-09-13 23:48:01 +00:00
|
|
|
iter.ctx.Metrics().ExhaustedNode(option.Node,
|
|
|
|
fmt.Sprintf("network: %s", err))
|
2016-02-20 20:18:22 +00:00
|
|
|
netIdx.Release()
|
2015-09-13 21:31:32 +00:00
|
|
|
continue OUTER
|
|
|
|
}
|
|
|
|
|
|
|
|
// Reserve this to prevent another task from colliding
|
|
|
|
netIdx.AddReserved(offer)
|
|
|
|
|
|
|
|
// Update the network ask to the offer
|
|
|
|
taskResources.Networks = []*structs.NetworkResource{offer}
|
|
|
|
}
|
|
|
|
|
|
|
|
// Store the task resource
|
|
|
|
option.SetTaskResources(task, taskResources)
|
|
|
|
|
|
|
|
// Accumulate the total resource requirement
|
|
|
|
total.Add(taskResources)
|
|
|
|
}
|
2016-08-25 17:27:19 +00:00
|
|
|
option.AllocResources = total
|
2015-09-13 21:31:32 +00:00
|
|
|
|
2015-08-13 18:54:59 +00:00
|
|
|
// Add the resources we are trying to fit
|
2015-09-13 21:31:32 +00:00
|
|
|
proposed = append(proposed, &structs.Allocation{Resources: total})
|
2015-08-13 18:28:02 +00:00
|
|
|
|
2015-08-13 19:02:42 +00:00
|
|
|
// Check if these allocations fit, if they do not, simply skip this node
|
2015-09-14 01:38:26 +00:00
|
|
|
fit, dim, util, _ := structs.AllocsFit(option.Node, proposed, netIdx)
|
2016-02-20 20:18:22 +00:00
|
|
|
netIdx.Release()
|
2015-08-13 18:54:59 +00:00
|
|
|
if !fit {
|
2015-09-14 01:38:26 +00:00
|
|
|
iter.ctx.Metrics().ExhaustedNode(option.Node, dim)
|
2015-08-13 19:02:42 +00:00
|
|
|
continue
|
2015-08-13 18:54:59 +00:00
|
|
|
}
|
|
|
|
|
2015-08-13 19:02:42 +00:00
|
|
|
// XXX: For now we completely ignore evictions. We should use that flag
|
|
|
|
// to determine if its possible to evict other lower priority allocations
|
|
|
|
// to make room. This explodes the search space, so it must be done
|
|
|
|
// carefully.
|
|
|
|
|
2015-08-13 18:54:59 +00:00
|
|
|
// Score the fit normally otherwise
|
2015-08-16 17:28:58 +00:00
|
|
|
fitness := structs.ScoreFit(option.Node, util)
|
|
|
|
option.Score += fitness
|
|
|
|
iter.ctx.Metrics().ScoreNode(option.Node, "binpack", fitness)
|
2015-08-13 18:54:59 +00:00
|
|
|
return option
|
2015-08-13 18:28:02 +00:00
|
|
|
}
|
|
|
|
}
|
2015-08-13 22:01:02 +00:00
|
|
|
|
|
|
|
func (iter *BinPackIterator) Reset() {
|
|
|
|
iter.source.Reset()
|
|
|
|
}
|
2015-08-16 17:32:25 +00:00
|
|
|
|
|
|
|
// JobAntiAffinityIterator is used to apply an anti-affinity to allocating
|
|
|
|
// along side other allocations from this job. This is used to help distribute
|
|
|
|
// load across the cluster.
|
|
|
|
type JobAntiAffinityIterator struct {
|
|
|
|
ctx Context
|
|
|
|
source RankIterator
|
|
|
|
penalty float64
|
|
|
|
jobID string
|
|
|
|
}
|
|
|
|
|
|
|
|
// NewJobAntiAffinityIterator is used to create a JobAntiAffinityIterator that
|
|
|
|
// applies the given penalty for co-placement with allocs from this job.
|
|
|
|
func NewJobAntiAffinityIterator(ctx Context, source RankIterator, penalty float64, jobID string) *JobAntiAffinityIterator {
|
|
|
|
iter := &JobAntiAffinityIterator{
|
|
|
|
ctx: ctx,
|
|
|
|
source: source,
|
|
|
|
penalty: penalty,
|
|
|
|
jobID: jobID,
|
|
|
|
}
|
|
|
|
return iter
|
|
|
|
}
|
|
|
|
|
|
|
|
func (iter *JobAntiAffinityIterator) SetJob(jobID string) {
|
|
|
|
iter.jobID = jobID
|
|
|
|
}
|
|
|
|
|
|
|
|
func (iter *JobAntiAffinityIterator) Next() *RankedNode {
|
|
|
|
for {
|
|
|
|
option := iter.source.Next()
|
|
|
|
if option == nil {
|
|
|
|
return nil
|
|
|
|
}
|
|
|
|
|
|
|
|
// Get the proposed allocations
|
2015-09-07 23:19:21 +00:00
|
|
|
proposed, err := option.ProposedAllocs(iter.ctx)
|
|
|
|
if err != nil {
|
|
|
|
iter.ctx.Logger().Printf(
|
|
|
|
"[ERR] sched.job-anti-aff: failed to get proposed allocations: %v",
|
|
|
|
err)
|
|
|
|
continue
|
2015-08-16 17:32:25 +00:00
|
|
|
}
|
|
|
|
|
|
|
|
// Determine the number of collisions
|
|
|
|
collisions := 0
|
|
|
|
for _, alloc := range proposed {
|
|
|
|
if alloc.JobID == iter.jobID {
|
|
|
|
collisions += 1
|
|
|
|
}
|
|
|
|
}
|
|
|
|
|
|
|
|
// Apply a penalty if there are collisions
|
|
|
|
if collisions > 0 {
|
2015-09-23 05:24:07 +00:00
|
|
|
scorePenalty := -1 * float64(collisions) * iter.penalty
|
|
|
|
option.Score += scorePenalty
|
2015-08-16 17:32:25 +00:00
|
|
|
iter.ctx.Metrics().ScoreNode(option.Node, "job-anti-affinity", scorePenalty)
|
|
|
|
}
|
|
|
|
return option
|
|
|
|
}
|
|
|
|
}
|
|
|
|
|
|
|
|
func (iter *JobAntiAffinityIterator) Reset() {
|
|
|
|
iter.source.Reset()
|
|
|
|
}
|