3131 lines
70 KiB
Go
3131 lines
70 KiB
Go
// Copyright (c) HashiCorp, Inc.
|
|
// SPDX-License-Identifier: MPL-2.0
|
|
|
|
package scheduler
|
|
|
|
import (
|
|
"fmt"
|
|
"reflect"
|
|
"testing"
|
|
"time"
|
|
|
|
"github.com/hashicorp/nomad/ci"
|
|
"github.com/hashicorp/nomad/helper/uuid"
|
|
"github.com/hashicorp/nomad/nomad/mock"
|
|
"github.com/hashicorp/nomad/nomad/structs"
|
|
psstructs "github.com/hashicorp/nomad/plugins/shared/structs"
|
|
"github.com/shoenig/test/must"
|
|
"github.com/stretchr/testify/assert"
|
|
"github.com/stretchr/testify/require"
|
|
)
|
|
|
|
func TestStaticIterator_Reset(t *testing.T) {
|
|
ci.Parallel(t)
|
|
|
|
_, ctx := testContext(t)
|
|
var nodes []*structs.Node
|
|
for i := 0; i < 3; i++ {
|
|
nodes = append(nodes, mock.Node())
|
|
}
|
|
static := NewStaticIterator(ctx, nodes)
|
|
|
|
for i := 0; i < 6; i++ {
|
|
static.Reset()
|
|
for j := 0; j < i; j++ {
|
|
static.Next()
|
|
}
|
|
static.Reset()
|
|
|
|
out := collectFeasible(static)
|
|
if len(out) != len(nodes) {
|
|
t.Fatalf("out: %#v missing nodes %d %#v", out, i, static)
|
|
}
|
|
|
|
ids := make(map[string]struct{})
|
|
for _, o := range out {
|
|
if _, ok := ids[o.ID]; ok {
|
|
t.Fatalf("duplicate")
|
|
}
|
|
ids[o.ID] = struct{}{}
|
|
}
|
|
}
|
|
}
|
|
|
|
func TestStaticIterator_SetNodes(t *testing.T) {
|
|
ci.Parallel(t)
|
|
|
|
_, ctx := testContext(t)
|
|
var nodes []*structs.Node
|
|
for i := 0; i < 3; i++ {
|
|
nodes = append(nodes, mock.Node())
|
|
}
|
|
static := NewStaticIterator(ctx, nodes)
|
|
|
|
newNodes := []*structs.Node{mock.Node()}
|
|
static.SetNodes(newNodes)
|
|
|
|
out := collectFeasible(static)
|
|
if !reflect.DeepEqual(out, newNodes) {
|
|
t.Fatalf("bad: %#v", out)
|
|
}
|
|
}
|
|
|
|
func TestRandomIterator(t *testing.T) {
|
|
ci.Parallel(t)
|
|
|
|
_, ctx := testContext(t)
|
|
var nodes []*structs.Node
|
|
for i := 0; i < 10; i++ {
|
|
nodes = append(nodes, mock.Node())
|
|
}
|
|
|
|
nc := make([]*structs.Node, len(nodes))
|
|
copy(nc, nodes)
|
|
rand := NewRandomIterator(ctx, nc)
|
|
|
|
out := collectFeasible(rand)
|
|
if len(out) != len(nodes) {
|
|
t.Fatalf("missing nodes")
|
|
}
|
|
if reflect.DeepEqual(out, nodes) {
|
|
t.Fatalf("same order")
|
|
}
|
|
}
|
|
|
|
func TestHostVolumeChecker(t *testing.T) {
|
|
ci.Parallel(t)
|
|
|
|
_, ctx := testContext(t)
|
|
nodes := []*structs.Node{
|
|
mock.Node(),
|
|
mock.Node(),
|
|
mock.Node(),
|
|
mock.Node(),
|
|
mock.Node(),
|
|
mock.Node(),
|
|
}
|
|
nodes[1].HostVolumes = map[string]*structs.ClientHostVolumeConfig{"foo": {Name: "foo"}}
|
|
nodes[2].HostVolumes = map[string]*structs.ClientHostVolumeConfig{
|
|
"foo": {},
|
|
"bar": {},
|
|
"unique-volume[0]": {},
|
|
}
|
|
nodes[3].HostVolumes = map[string]*structs.ClientHostVolumeConfig{
|
|
"foo": {},
|
|
"bar": {},
|
|
}
|
|
nodes[4].HostVolumes = map[string]*structs.ClientHostVolumeConfig{
|
|
"foo": {},
|
|
"baz": {},
|
|
}
|
|
|
|
noVolumes := map[string]*structs.VolumeRequest{}
|
|
|
|
volumes := map[string]*structs.VolumeRequest{
|
|
"foo": {
|
|
Type: "host",
|
|
Source: "foo",
|
|
},
|
|
"bar": {
|
|
Type: "host",
|
|
Source: "bar",
|
|
},
|
|
"baz": {
|
|
Type: "nothost",
|
|
Source: "baz",
|
|
},
|
|
"unique": {
|
|
Type: "host",
|
|
Source: "unique-volume[0]",
|
|
PerAlloc: true,
|
|
},
|
|
}
|
|
|
|
checker := NewHostVolumeChecker(ctx)
|
|
cases := []struct {
|
|
Node *structs.Node
|
|
RequestedVolumes map[string]*structs.VolumeRequest
|
|
Result bool
|
|
}{
|
|
{ // Nil Volumes, some requested
|
|
Node: nodes[0],
|
|
RequestedVolumes: volumes,
|
|
Result: false,
|
|
},
|
|
{ // Mismatched set of volumes
|
|
Node: nodes[1],
|
|
RequestedVolumes: volumes,
|
|
Result: false,
|
|
},
|
|
{ // Happy Path
|
|
Node: nodes[2],
|
|
RequestedVolumes: volumes,
|
|
Result: true,
|
|
},
|
|
{ // No Volumes requested or available
|
|
Node: nodes[3],
|
|
RequestedVolumes: noVolumes,
|
|
Result: true,
|
|
},
|
|
{ // No Volumes requested, some available
|
|
Node: nodes[4],
|
|
RequestedVolumes: noVolumes,
|
|
Result: true,
|
|
},
|
|
}
|
|
|
|
alloc := mock.Alloc()
|
|
alloc.NodeID = nodes[2].ID
|
|
|
|
for i, c := range cases {
|
|
checker.SetVolumes(alloc.Name, c.RequestedVolumes)
|
|
if act := checker.Feasible(c.Node); act != c.Result {
|
|
t.Fatalf("case(%d) failed: got %v; want %v", i, act, c.Result)
|
|
}
|
|
}
|
|
}
|
|
|
|
func TestHostVolumeChecker_ReadOnly(t *testing.T) {
|
|
ci.Parallel(t)
|
|
|
|
_, ctx := testContext(t)
|
|
nodes := []*structs.Node{
|
|
mock.Node(),
|
|
mock.Node(),
|
|
}
|
|
|
|
nodes[0].HostVolumes = map[string]*structs.ClientHostVolumeConfig{
|
|
"foo": {
|
|
ReadOnly: true,
|
|
},
|
|
}
|
|
nodes[1].HostVolumes = map[string]*structs.ClientHostVolumeConfig{
|
|
"foo": {
|
|
ReadOnly: false,
|
|
},
|
|
}
|
|
|
|
readwriteRequest := map[string]*structs.VolumeRequest{
|
|
"foo": {
|
|
Type: "host",
|
|
Source: "foo",
|
|
},
|
|
}
|
|
|
|
readonlyRequest := map[string]*structs.VolumeRequest{
|
|
"foo": {
|
|
Type: "host",
|
|
Source: "foo",
|
|
ReadOnly: true,
|
|
},
|
|
}
|
|
|
|
checker := NewHostVolumeChecker(ctx)
|
|
cases := []struct {
|
|
Node *structs.Node
|
|
RequestedVolumes map[string]*structs.VolumeRequest
|
|
Result bool
|
|
}{
|
|
{ // ReadWrite Request, ReadOnly Host
|
|
Node: nodes[0],
|
|
RequestedVolumes: readwriteRequest,
|
|
Result: false,
|
|
},
|
|
{ // ReadOnly Request, ReadOnly Host
|
|
Node: nodes[0],
|
|
RequestedVolumes: readonlyRequest,
|
|
Result: true,
|
|
},
|
|
{ // ReadOnly Request, ReadWrite Host
|
|
Node: nodes[1],
|
|
RequestedVolumes: readonlyRequest,
|
|
Result: true,
|
|
},
|
|
{ // ReadWrite Request, ReadWrite Host
|
|
Node: nodes[1],
|
|
RequestedVolumes: readwriteRequest,
|
|
Result: true,
|
|
},
|
|
}
|
|
|
|
alloc := mock.Alloc()
|
|
alloc.NodeID = nodes[1].ID
|
|
|
|
for i, c := range cases {
|
|
checker.SetVolumes(alloc.Name, c.RequestedVolumes)
|
|
if act := checker.Feasible(c.Node); act != c.Result {
|
|
t.Fatalf("case(%d) failed: got %v; want %v", i, act, c.Result)
|
|
}
|
|
}
|
|
}
|
|
|
|
func TestCSIVolumeChecker(t *testing.T) {
|
|
ci.Parallel(t)
|
|
state, ctx := testContext(t)
|
|
nodes := []*structs.Node{
|
|
mock.Node(),
|
|
mock.Node(),
|
|
mock.Node(),
|
|
mock.Node(),
|
|
mock.Node(),
|
|
mock.Node(),
|
|
mock.Node(),
|
|
}
|
|
|
|
// Register running plugins on some nodes
|
|
nodes[0].CSIControllerPlugins = map[string]*structs.CSIInfo{
|
|
"foo": {
|
|
PluginID: "foo",
|
|
Healthy: true,
|
|
ControllerInfo: &structs.CSIControllerInfo{},
|
|
},
|
|
}
|
|
nodes[0].CSINodePlugins = map[string]*structs.CSIInfo{
|
|
"foo": {
|
|
PluginID: "foo",
|
|
Healthy: true,
|
|
NodeInfo: &structs.CSINodeInfo{
|
|
MaxVolumes: 1,
|
|
AccessibleTopology: &structs.CSITopology{
|
|
Segments: map[string]string{"rack": "R1"},
|
|
},
|
|
},
|
|
},
|
|
}
|
|
nodes[1].CSINodePlugins = map[string]*structs.CSIInfo{
|
|
"foo": {
|
|
PluginID: "foo",
|
|
Healthy: false,
|
|
NodeInfo: &structs.CSINodeInfo{
|
|
MaxVolumes: 1,
|
|
AccessibleTopology: &structs.CSITopology{
|
|
Segments: map[string]string{"rack": "R1"},
|
|
},
|
|
},
|
|
},
|
|
}
|
|
nodes[2].CSINodePlugins = map[string]*structs.CSIInfo{
|
|
"bar": {
|
|
PluginID: "bar",
|
|
Healthy: true,
|
|
NodeInfo: &structs.CSINodeInfo{
|
|
MaxVolumes: 1,
|
|
AccessibleTopology: &structs.CSITopology{
|
|
Segments: map[string]string{"rack": "R1"},
|
|
},
|
|
},
|
|
},
|
|
}
|
|
nodes[4].CSINodePlugins = map[string]*structs.CSIInfo{
|
|
"foo": {
|
|
PluginID: "foo",
|
|
Healthy: true,
|
|
NodeInfo: &structs.CSINodeInfo{
|
|
MaxVolumes: 1,
|
|
AccessibleTopology: &structs.CSITopology{
|
|
Segments: map[string]string{"rack": "R1"},
|
|
},
|
|
},
|
|
},
|
|
}
|
|
nodes[5].CSINodePlugins = map[string]*structs.CSIInfo{
|
|
"foo": {
|
|
PluginID: "foo",
|
|
Healthy: true,
|
|
NodeInfo: &structs.CSINodeInfo{
|
|
MaxVolumes: 1,
|
|
AccessibleTopology: &structs.CSITopology{
|
|
Segments: map[string]string{"rack": "R4"},
|
|
},
|
|
},
|
|
},
|
|
}
|
|
nodes[6].CSINodePlugins = map[string]*structs.CSIInfo{
|
|
"foo": {
|
|
PluginID: "foo",
|
|
Healthy: true,
|
|
NodeInfo: &structs.CSINodeInfo{
|
|
MaxVolumes: 1,
|
|
},
|
|
},
|
|
}
|
|
|
|
// Create the plugins in the state store
|
|
index := uint64(999)
|
|
for _, node := range nodes {
|
|
err := state.UpsertNode(structs.MsgTypeTestSetup, index, node)
|
|
require.NoError(t, err)
|
|
index++
|
|
}
|
|
|
|
// Create the volume in the state store
|
|
vid := "volume-id"
|
|
vol := structs.NewCSIVolume(vid, index)
|
|
vol.PluginID = "foo"
|
|
vol.Namespace = structs.DefaultNamespace
|
|
vol.AccessMode = structs.CSIVolumeAccessModeMultiNodeMultiWriter
|
|
vol.AttachmentMode = structs.CSIVolumeAttachmentModeFilesystem
|
|
vol.Topologies = []*structs.CSITopology{
|
|
{Segments: map[string]string{"rack": "R1"}},
|
|
{Segments: map[string]string{"rack": "R2"}},
|
|
}
|
|
err := state.UpsertCSIVolume(index, []*structs.CSIVolume{vol})
|
|
require.NoError(t, err)
|
|
index++
|
|
|
|
// Create some other volumes in use on nodes[3] to trip MaxVolumes
|
|
vid2 := uuid.Generate()
|
|
vol2 := structs.NewCSIVolume(vid2, index)
|
|
vol2.PluginID = "foo"
|
|
vol2.Namespace = structs.DefaultNamespace
|
|
vol2.AccessMode = structs.CSIVolumeAccessModeMultiNodeSingleWriter
|
|
vol2.AttachmentMode = structs.CSIVolumeAttachmentModeFilesystem
|
|
err = state.UpsertCSIVolume(index, []*structs.CSIVolume{vol2})
|
|
require.NoError(t, err)
|
|
index++
|
|
|
|
vid3 := "volume-id[0]"
|
|
vol3 := vol.Copy()
|
|
vol3.ID = vid3
|
|
err = state.UpsertCSIVolume(index, []*structs.CSIVolume{vol3})
|
|
require.NoError(t, err)
|
|
index++
|
|
|
|
alloc := mock.Alloc()
|
|
alloc.NodeID = nodes[4].ID
|
|
alloc.Job.TaskGroups[0].Volumes = map[string]*structs.VolumeRequest{
|
|
vid2: {
|
|
Name: vid2,
|
|
Type: "csi",
|
|
Source: vid2,
|
|
},
|
|
}
|
|
err = state.UpsertJob(structs.MsgTypeTestSetup, index, nil, alloc.Job)
|
|
require.NoError(t, err)
|
|
index++
|
|
summary := mock.JobSummary(alloc.JobID)
|
|
require.NoError(t, state.UpsertJobSummary(index, summary))
|
|
index++
|
|
err = state.UpsertAllocs(structs.MsgTypeTestSetup, index, []*structs.Allocation{alloc})
|
|
require.NoError(t, err)
|
|
index++
|
|
|
|
// Create volume requests
|
|
noVolumes := map[string]*structs.VolumeRequest{}
|
|
|
|
volumes := map[string]*structs.VolumeRequest{
|
|
"shared": {
|
|
Type: "csi",
|
|
Name: "baz",
|
|
Source: "volume-id",
|
|
},
|
|
"unique": {
|
|
Type: "csi",
|
|
Name: "baz",
|
|
Source: "volume-id[0]",
|
|
PerAlloc: true,
|
|
},
|
|
"nonsense": {
|
|
Type: "host",
|
|
Name: "nonsense",
|
|
Source: "my-host-volume",
|
|
},
|
|
}
|
|
|
|
checker := NewCSIVolumeChecker(ctx)
|
|
checker.SetNamespace(structs.DefaultNamespace)
|
|
|
|
cases := []struct {
|
|
Name string
|
|
Node *structs.Node
|
|
RequestedVolumes map[string]*structs.VolumeRequest
|
|
Result bool
|
|
}{
|
|
{
|
|
Name: "ok",
|
|
Node: nodes[0],
|
|
RequestedVolumes: volumes,
|
|
Result: true,
|
|
},
|
|
{
|
|
Name: "unhealthy node",
|
|
Node: nodes[1],
|
|
RequestedVolumes: volumes,
|
|
Result: false,
|
|
},
|
|
{
|
|
Name: "wrong id",
|
|
Node: nodes[2],
|
|
RequestedVolumes: volumes,
|
|
Result: false,
|
|
},
|
|
{
|
|
Name: "no volumes requested or available",
|
|
Node: nodes[3],
|
|
RequestedVolumes: noVolumes,
|
|
Result: true,
|
|
},
|
|
{
|
|
Name: "no volumes requested, some available",
|
|
Node: nodes[0],
|
|
RequestedVolumes: noVolumes,
|
|
Result: true,
|
|
},
|
|
{
|
|
Name: "volumes requested, none available",
|
|
Node: nodes[3],
|
|
RequestedVolumes: volumes,
|
|
Result: false,
|
|
},
|
|
{
|
|
Name: "volumes requested, max volumes exceeded",
|
|
Node: nodes[4],
|
|
RequestedVolumes: volumes,
|
|
Result: false,
|
|
},
|
|
{
|
|
Name: "no matching topology",
|
|
Node: nodes[5],
|
|
RequestedVolumes: volumes,
|
|
Result: false,
|
|
},
|
|
{
|
|
Name: "nil topology",
|
|
Node: nodes[6],
|
|
RequestedVolumes: volumes,
|
|
Result: false,
|
|
},
|
|
}
|
|
|
|
for _, c := range cases {
|
|
checker.SetVolumes(alloc.Name, c.RequestedVolumes)
|
|
assert.Equal(t, c.Result, checker.Feasible(c.Node), c.Name)
|
|
}
|
|
|
|
// add a missing volume
|
|
volumes["missing"] = &structs.VolumeRequest{
|
|
Type: "csi",
|
|
Name: "bar",
|
|
Source: "does-not-exist",
|
|
}
|
|
|
|
checker = NewCSIVolumeChecker(ctx)
|
|
checker.SetNamespace(structs.DefaultNamespace)
|
|
|
|
for _, node := range nodes {
|
|
checker.SetVolumes(alloc.Name, volumes)
|
|
act := checker.Feasible(node)
|
|
require.False(t, act, "request with missing volume should never be feasible")
|
|
}
|
|
|
|
}
|
|
|
|
func TestNetworkChecker(t *testing.T) {
|
|
ci.Parallel(t)
|
|
|
|
_, ctx := testContext(t)
|
|
|
|
node := func(mode string) *structs.Node {
|
|
n := mock.Node()
|
|
n.NodeResources.Networks = append(n.NodeResources.Networks, &structs.NetworkResource{Mode: mode})
|
|
if mode == "bridge" {
|
|
n.NodeResources.NodeNetworks = []*structs.NodeNetworkResource{
|
|
{
|
|
Addresses: []structs.NodeNetworkAddress{
|
|
{
|
|
Alias: "public",
|
|
}, {
|
|
Alias: "private",
|
|
},
|
|
},
|
|
},
|
|
}
|
|
}
|
|
n.Attributes["nomad.version"] = "0.12.0" // mock version is 0.5.0
|
|
n.Meta["public_network"] = "public"
|
|
n.Meta["private_network"] = "private"
|
|
n.Meta["wrong_network"] = "empty"
|
|
return n
|
|
}
|
|
|
|
nodes := []*structs.Node{
|
|
node("bridge"),
|
|
node("bridge"),
|
|
node("cni/mynet"),
|
|
}
|
|
|
|
checker := NewNetworkChecker(ctx)
|
|
cases := []struct {
|
|
network *structs.NetworkResource
|
|
results []bool
|
|
}{
|
|
{
|
|
network: &structs.NetworkResource{Mode: "host"},
|
|
results: []bool{true, true, true},
|
|
},
|
|
{
|
|
network: &structs.NetworkResource{Mode: "bridge"},
|
|
results: []bool{true, true, false},
|
|
},
|
|
{
|
|
network: &structs.NetworkResource{
|
|
Mode: "bridge",
|
|
ReservedPorts: []structs.Port{},
|
|
DynamicPorts: []structs.Port{
|
|
{
|
|
Label: "http",
|
|
Value: 8080,
|
|
To: 8080,
|
|
HostNetwork: "${meta.public_network}",
|
|
},
|
|
{
|
|
Label: "metrics",
|
|
Value: 9090,
|
|
To: 9090,
|
|
HostNetwork: "${meta.private_network}",
|
|
},
|
|
},
|
|
},
|
|
results: []bool{true, true, false},
|
|
},
|
|
{
|
|
network: &structs.NetworkResource{
|
|
Mode: "bridge",
|
|
ReservedPorts: []structs.Port{},
|
|
DynamicPorts: []structs.Port{
|
|
{
|
|
Label: "metrics",
|
|
Value: 9090,
|
|
To: 9090,
|
|
HostNetwork: "${meta.wrong_network}",
|
|
},
|
|
},
|
|
},
|
|
results: []bool{false, false, false},
|
|
},
|
|
{
|
|
network: &structs.NetworkResource{
|
|
Mode: "bridge",
|
|
ReservedPorts: []structs.Port{},
|
|
DynamicPorts: []structs.Port{
|
|
{
|
|
Label: "metrics",
|
|
Value: 9090,
|
|
To: 9090,
|
|
HostNetwork: "${meta.nonetwork}",
|
|
},
|
|
},
|
|
},
|
|
results: []bool{false, false, false},
|
|
},
|
|
{
|
|
network: &structs.NetworkResource{
|
|
Mode: "bridge",
|
|
ReservedPorts: []structs.Port{},
|
|
DynamicPorts: []structs.Port{
|
|
{
|
|
Label: "metrics",
|
|
Value: 9090,
|
|
To: 9090,
|
|
HostNetwork: "public",
|
|
},
|
|
},
|
|
},
|
|
results: []bool{true, true, false},
|
|
},
|
|
{
|
|
network: &structs.NetworkResource{
|
|
Mode: "bridge",
|
|
ReservedPorts: []structs.Port{},
|
|
DynamicPorts: []structs.Port{
|
|
{
|
|
Label: "metrics",
|
|
Value: 9090,
|
|
To: 9090,
|
|
HostNetwork: "${meta.private_network}-nonexisting",
|
|
},
|
|
},
|
|
},
|
|
results: []bool{false, false, false},
|
|
},
|
|
{
|
|
network: &structs.NetworkResource{Mode: "cni/mynet"},
|
|
results: []bool{false, false, true},
|
|
},
|
|
{
|
|
network: &structs.NetworkResource{Mode: "cni/nonexistent"},
|
|
results: []bool{false, false, false},
|
|
},
|
|
}
|
|
|
|
for _, c := range cases {
|
|
checker.SetNetwork(c.network)
|
|
for i, node := range nodes {
|
|
require.Equal(t, c.results[i], checker.Feasible(node), "mode=%q, idx=%d", c.network.Mode, i)
|
|
}
|
|
}
|
|
}
|
|
|
|
func TestNetworkChecker_bridge_upgrade_path(t *testing.T) {
|
|
ci.Parallel(t)
|
|
|
|
_, ctx := testContext(t)
|
|
|
|
t.Run("older client", func(t *testing.T) {
|
|
// Create a client that is still on v0.11, which does not have the bridge
|
|
// network finger-printer (and thus no bridge network resource)
|
|
oldClient := mock.Node()
|
|
oldClient.Attributes["nomad.version"] = "0.11.0"
|
|
|
|
checker := NewNetworkChecker(ctx)
|
|
checker.SetNetwork(&structs.NetworkResource{Mode: "bridge"})
|
|
|
|
ok := checker.Feasible(oldClient)
|
|
require.True(t, ok)
|
|
})
|
|
|
|
t.Run("updated client", func(t *testing.T) {
|
|
// Create a client that is updated to 0.12, but did not detect a bridge
|
|
// network resource.
|
|
oldClient := mock.Node()
|
|
oldClient.Attributes["nomad.version"] = "0.12.0"
|
|
|
|
checker := NewNetworkChecker(ctx)
|
|
checker.SetNetwork(&structs.NetworkResource{Mode: "bridge"})
|
|
|
|
ok := checker.Feasible(oldClient)
|
|
require.False(t, ok)
|
|
})
|
|
}
|
|
|
|
func TestDriverChecker_DriverInfo(t *testing.T) {
|
|
ci.Parallel(t)
|
|
|
|
_, ctx := testContext(t)
|
|
nodes := []*structs.Node{
|
|
mock.Node(),
|
|
mock.Node(),
|
|
mock.Node(),
|
|
}
|
|
nodes[0].Drivers["foo"] = &structs.DriverInfo{
|
|
Detected: true,
|
|
Healthy: true,
|
|
}
|
|
nodes[1].Drivers["foo"] = &structs.DriverInfo{
|
|
Detected: true,
|
|
Healthy: false,
|
|
}
|
|
nodes[2].Drivers["foo"] = &structs.DriverInfo{
|
|
Detected: false,
|
|
Healthy: false,
|
|
}
|
|
|
|
drivers := map[string]struct{}{
|
|
"exec": {},
|
|
"foo": {},
|
|
}
|
|
checker := NewDriverChecker(ctx, drivers)
|
|
cases := []struct {
|
|
Node *structs.Node
|
|
Result bool
|
|
}{
|
|
{
|
|
Node: nodes[0],
|
|
Result: true,
|
|
},
|
|
{
|
|
Node: nodes[1],
|
|
Result: false,
|
|
},
|
|
{
|
|
Node: nodes[2],
|
|
Result: false,
|
|
},
|
|
}
|
|
|
|
for i, c := range cases {
|
|
if act := checker.Feasible(c.Node); act != c.Result {
|
|
t.Fatalf("case(%d) failed: got %v; want %v", i, act, c.Result)
|
|
}
|
|
}
|
|
}
|
|
func TestDriverChecker_Compatibility(t *testing.T) {
|
|
ci.Parallel(t)
|
|
|
|
_, ctx := testContext(t)
|
|
nodes := []*structs.Node{
|
|
mock.Node(),
|
|
mock.Node(),
|
|
mock.Node(),
|
|
mock.Node(),
|
|
}
|
|
for _, n := range nodes {
|
|
// force compatibility mode
|
|
n.Drivers = nil
|
|
}
|
|
nodes[0].Attributes["driver.foo"] = "1"
|
|
nodes[1].Attributes["driver.foo"] = "0"
|
|
nodes[2].Attributes["driver.foo"] = "true"
|
|
nodes[3].Attributes["driver.foo"] = "False"
|
|
|
|
drivers := map[string]struct{}{
|
|
"exec": {},
|
|
"foo": {},
|
|
}
|
|
checker := NewDriverChecker(ctx, drivers)
|
|
cases := []struct {
|
|
Node *structs.Node
|
|
Result bool
|
|
}{
|
|
{
|
|
Node: nodes[0],
|
|
Result: true,
|
|
},
|
|
{
|
|
Node: nodes[1],
|
|
Result: false,
|
|
},
|
|
{
|
|
Node: nodes[2],
|
|
Result: true,
|
|
},
|
|
{
|
|
Node: nodes[3],
|
|
Result: false,
|
|
},
|
|
}
|
|
|
|
for i, c := range cases {
|
|
if act := checker.Feasible(c.Node); act != c.Result {
|
|
t.Fatalf("case(%d) failed: got %v; want %v", i, act, c.Result)
|
|
}
|
|
}
|
|
}
|
|
|
|
func Test_HealthChecks(t *testing.T) {
|
|
ci.Parallel(t)
|
|
|
|
require := require.New(t)
|
|
_, ctx := testContext(t)
|
|
|
|
nodes := []*structs.Node{
|
|
mock.Node(),
|
|
mock.Node(),
|
|
mock.Node(),
|
|
}
|
|
for _, e := range nodes {
|
|
e.Drivers = make(map[string]*structs.DriverInfo)
|
|
}
|
|
nodes[0].Attributes["driver.foo"] = "1"
|
|
nodes[0].Drivers["foo"] = &structs.DriverInfo{
|
|
Detected: true,
|
|
Healthy: true,
|
|
HealthDescription: "running",
|
|
UpdateTime: time.Now(),
|
|
}
|
|
nodes[1].Attributes["driver.bar"] = "1"
|
|
nodes[1].Drivers["bar"] = &structs.DriverInfo{
|
|
Detected: true,
|
|
Healthy: false,
|
|
HealthDescription: "not running",
|
|
UpdateTime: time.Now(),
|
|
}
|
|
nodes[2].Attributes["driver.baz"] = "0"
|
|
nodes[2].Drivers["baz"] = &structs.DriverInfo{
|
|
Detected: false,
|
|
Healthy: false,
|
|
HealthDescription: "not running",
|
|
UpdateTime: time.Now(),
|
|
}
|
|
|
|
testDrivers := []string{"foo", "bar", "baz"}
|
|
cases := []struct {
|
|
Node *structs.Node
|
|
Result bool
|
|
}{
|
|
{
|
|
Node: nodes[0],
|
|
Result: true,
|
|
},
|
|
{
|
|
Node: nodes[1],
|
|
Result: false,
|
|
},
|
|
{
|
|
Node: nodes[2],
|
|
Result: false,
|
|
},
|
|
}
|
|
|
|
for i, c := range cases {
|
|
drivers := map[string]struct{}{
|
|
testDrivers[i]: {},
|
|
}
|
|
checker := NewDriverChecker(ctx, drivers)
|
|
act := checker.Feasible(c.Node)
|
|
require.Equal(act, c.Result)
|
|
}
|
|
}
|
|
|
|
func TestConstraintChecker(t *testing.T) {
|
|
ci.Parallel(t)
|
|
|
|
_, ctx := testContext(t)
|
|
nodes := []*structs.Node{
|
|
mock.Node(),
|
|
mock.Node(),
|
|
mock.Node(),
|
|
mock.Node(),
|
|
}
|
|
|
|
nodes[0].Attributes["kernel.name"] = "freebsd"
|
|
nodes[1].Datacenter = "dc2"
|
|
nodes[2].NodeClass = "large"
|
|
nodes[2].Attributes["foo"] = "bar"
|
|
nodes[3].NodePool = "prod"
|
|
|
|
constraints := []*structs.Constraint{
|
|
{
|
|
Operand: "=",
|
|
LTarget: "${node.datacenter}",
|
|
RTarget: "dc1",
|
|
},
|
|
{
|
|
Operand: "is",
|
|
LTarget: "${attr.kernel.name}",
|
|
RTarget: "linux",
|
|
},
|
|
{
|
|
Operand: "!=",
|
|
LTarget: "${node.class}",
|
|
RTarget: "linux-medium-pci",
|
|
},
|
|
{
|
|
Operand: "!=",
|
|
LTarget: "${node.pool}",
|
|
RTarget: "prod",
|
|
},
|
|
{
|
|
Operand: "is_set",
|
|
LTarget: "${attr.foo}",
|
|
},
|
|
}
|
|
checker := NewConstraintChecker(ctx, constraints)
|
|
cases := []struct {
|
|
Node *structs.Node
|
|
Result bool
|
|
}{
|
|
{
|
|
Node: nodes[0],
|
|
Result: false,
|
|
},
|
|
{
|
|
Node: nodes[1],
|
|
Result: false,
|
|
},
|
|
{
|
|
Node: nodes[2],
|
|
Result: true,
|
|
},
|
|
{
|
|
Node: nodes[3],
|
|
Result: false,
|
|
},
|
|
}
|
|
|
|
for i, c := range cases {
|
|
if act := checker.Feasible(c.Node); act != c.Result {
|
|
t.Fatalf("case(%d) failed: got %v; want %v", i, act, c.Result)
|
|
}
|
|
}
|
|
}
|
|
|
|
func TestResolveConstraintTarget(t *testing.T) {
|
|
ci.Parallel(t)
|
|
|
|
type tcase struct {
|
|
target string
|
|
node *structs.Node
|
|
val interface{}
|
|
result bool
|
|
}
|
|
node := mock.Node()
|
|
cases := []tcase{
|
|
{
|
|
target: "${node.unique.id}",
|
|
node: node,
|
|
val: node.ID,
|
|
result: true,
|
|
},
|
|
{
|
|
target: "${node.datacenter}",
|
|
node: node,
|
|
val: node.Datacenter,
|
|
result: true,
|
|
},
|
|
{
|
|
target: "${node.unique.name}",
|
|
node: node,
|
|
val: node.Name,
|
|
result: true,
|
|
},
|
|
{
|
|
target: "${node.class}",
|
|
node: node,
|
|
val: node.NodeClass,
|
|
result: true,
|
|
},
|
|
{
|
|
target: "${node.foo}",
|
|
node: node,
|
|
result: false,
|
|
},
|
|
{
|
|
target: "${attr.kernel.name}",
|
|
node: node,
|
|
val: node.Attributes["kernel.name"],
|
|
result: true,
|
|
},
|
|
{
|
|
target: "${attr.rand}",
|
|
node: node,
|
|
val: "",
|
|
result: false,
|
|
},
|
|
{
|
|
target: "${meta.pci-dss}",
|
|
node: node,
|
|
val: node.Meta["pci-dss"],
|
|
result: true,
|
|
},
|
|
{
|
|
target: "${meta.rand}",
|
|
node: node,
|
|
val: "",
|
|
result: false,
|
|
},
|
|
}
|
|
|
|
for _, tc := range cases {
|
|
res, ok := resolveTarget(tc.target, tc.node)
|
|
if ok != tc.result {
|
|
t.Fatalf("TC: %#v, Result: %v %v", tc, res, ok)
|
|
}
|
|
if ok && !reflect.DeepEqual(res, tc.val) {
|
|
t.Fatalf("TC: %#v, Result: %v %v", tc, res, ok)
|
|
}
|
|
}
|
|
}
|
|
|
|
func TestCheckConstraint(t *testing.T) {
|
|
ci.Parallel(t)
|
|
|
|
type tcase struct {
|
|
op string
|
|
lVal, rVal interface{}
|
|
result bool
|
|
}
|
|
cases := []tcase{
|
|
{
|
|
op: "=",
|
|
lVal: "foo", rVal: "foo",
|
|
result: true,
|
|
},
|
|
{
|
|
op: "is",
|
|
lVal: "foo", rVal: "foo",
|
|
result: true,
|
|
},
|
|
{
|
|
op: "==",
|
|
lVal: "foo", rVal: "foo",
|
|
result: true,
|
|
},
|
|
{
|
|
op: "==",
|
|
lVal: "foo", rVal: nil,
|
|
result: false,
|
|
},
|
|
{
|
|
op: "==",
|
|
lVal: nil, rVal: "foo",
|
|
result: false,
|
|
},
|
|
{
|
|
op: "==",
|
|
lVal: nil, rVal: nil,
|
|
result: false,
|
|
},
|
|
{
|
|
op: "!=",
|
|
lVal: "foo", rVal: "foo",
|
|
result: false,
|
|
},
|
|
{
|
|
op: "!=",
|
|
lVal: "foo", rVal: "bar",
|
|
result: true,
|
|
},
|
|
{
|
|
op: "!=",
|
|
lVal: nil, rVal: "foo",
|
|
result: true,
|
|
},
|
|
{
|
|
op: "!=",
|
|
lVal: "foo", rVal: nil,
|
|
result: true,
|
|
},
|
|
{
|
|
op: "!=",
|
|
lVal: nil, rVal: nil,
|
|
result: false,
|
|
},
|
|
{
|
|
op: "not",
|
|
lVal: "foo", rVal: "bar",
|
|
result: true,
|
|
},
|
|
{
|
|
op: structs.ConstraintVersion,
|
|
lVal: "1.2.3", rVal: "~> 1.0",
|
|
result: true,
|
|
},
|
|
{
|
|
op: structs.ConstraintVersion,
|
|
lVal: nil, rVal: "~> 1.0",
|
|
result: false,
|
|
},
|
|
{
|
|
op: structs.ConstraintRegex,
|
|
lVal: "foobarbaz", rVal: "[\\w]+",
|
|
result: true,
|
|
},
|
|
{
|
|
op: structs.ConstraintRegex,
|
|
lVal: nil, rVal: "[\\w]+",
|
|
result: false,
|
|
},
|
|
{
|
|
op: "<",
|
|
lVal: "foo", rVal: "bar",
|
|
result: false,
|
|
},
|
|
{
|
|
op: "<",
|
|
lVal: nil, rVal: "bar",
|
|
result: false,
|
|
},
|
|
{
|
|
op: structs.ConstraintSetContains,
|
|
lVal: "foo,bar,baz", rVal: "foo, bar ",
|
|
result: true,
|
|
},
|
|
{
|
|
op: structs.ConstraintSetContains,
|
|
lVal: "foo,bar,baz", rVal: "foo,bam",
|
|
result: false,
|
|
},
|
|
{
|
|
op: structs.ConstraintAttributeIsSet,
|
|
lVal: "foo",
|
|
result: true,
|
|
},
|
|
{
|
|
op: structs.ConstraintAttributeIsSet,
|
|
lVal: nil,
|
|
result: false,
|
|
},
|
|
{
|
|
op: structs.ConstraintAttributeIsNotSet,
|
|
lVal: nil,
|
|
result: true,
|
|
},
|
|
{
|
|
op: structs.ConstraintAttributeIsNotSet,
|
|
lVal: "foo",
|
|
result: false,
|
|
},
|
|
}
|
|
|
|
for _, tc := range cases {
|
|
_, ctx := testContext(t)
|
|
if res := checkConstraint(ctx, tc.op, tc.lVal, tc.rVal, tc.lVal != nil, tc.rVal != nil); res != tc.result {
|
|
t.Fatalf("TC: %#v, Result: %v", tc, res)
|
|
}
|
|
}
|
|
}
|
|
|
|
func TestCheckOrder(t *testing.T) {
|
|
ci.Parallel(t)
|
|
|
|
cases := []struct {
|
|
op string
|
|
lVal, rVal any
|
|
exp bool
|
|
}{
|
|
{
|
|
op: "<",
|
|
lVal: "bar", rVal: "foo",
|
|
exp: true,
|
|
},
|
|
{
|
|
op: "<=",
|
|
lVal: "foo", rVal: "foo",
|
|
exp: true,
|
|
},
|
|
{
|
|
op: ">",
|
|
lVal: "bar", rVal: "foo",
|
|
exp: false,
|
|
},
|
|
{
|
|
op: ">=",
|
|
lVal: "bar", rVal: "bar",
|
|
exp: true,
|
|
},
|
|
{
|
|
op: ">",
|
|
lVal: "1", rVal: "foo",
|
|
exp: false,
|
|
},
|
|
{
|
|
op: "<",
|
|
lVal: "10", rVal: "1",
|
|
exp: false,
|
|
},
|
|
{
|
|
op: "<",
|
|
lVal: "1", rVal: "10",
|
|
exp: true,
|
|
},
|
|
{
|
|
op: "<",
|
|
lVal: "10.5", rVal: "1.5",
|
|
exp: false,
|
|
},
|
|
{
|
|
op: "<",
|
|
lVal: "1.5", rVal: "10.5",
|
|
exp: true,
|
|
},
|
|
}
|
|
for _, tc := range cases {
|
|
name := fmt.Sprintf("%v %s %v", tc.lVal, tc.op, tc.rVal)
|
|
t.Run(name, func(t *testing.T) {
|
|
must.Eq(t, tc.exp, checkOrder(tc.op, tc.lVal, tc.rVal))
|
|
})
|
|
}
|
|
}
|
|
|
|
func TestCheckVersionConstraint(t *testing.T) {
|
|
ci.Parallel(t)
|
|
|
|
type tcase struct {
|
|
lVal, rVal interface{}
|
|
result bool
|
|
}
|
|
cases := []tcase{
|
|
{
|
|
lVal: "1.2.3", rVal: "~> 1.0",
|
|
result: true,
|
|
},
|
|
{
|
|
lVal: "1.2.3", rVal: ">= 1.0, < 1.4",
|
|
result: true,
|
|
},
|
|
{
|
|
lVal: "2.0.1", rVal: "~> 1.0",
|
|
result: false,
|
|
},
|
|
{
|
|
lVal: "1.4", rVal: ">= 1.0, < 1.4",
|
|
result: false,
|
|
},
|
|
{
|
|
lVal: 1, rVal: "~> 1.0",
|
|
result: true,
|
|
},
|
|
{
|
|
// Prereleases are never > final releases
|
|
lVal: "1.3.0-beta1", rVal: ">= 0.6.1",
|
|
result: false,
|
|
},
|
|
{
|
|
// Prerelease X.Y.Z must match
|
|
lVal: "1.7.0-alpha1", rVal: ">= 1.6.0-beta1",
|
|
result: false,
|
|
},
|
|
{
|
|
// Meta is ignored
|
|
lVal: "1.3.0-beta1+ent", rVal: "= 1.3.0-beta1",
|
|
result: true,
|
|
},
|
|
}
|
|
for _, tc := range cases {
|
|
_, ctx := testContext(t)
|
|
p := newVersionConstraintParser(ctx)
|
|
if res := checkVersionMatch(ctx, p, tc.lVal, tc.rVal); res != tc.result {
|
|
t.Fatalf("TC: %#v, Result: %v", tc, res)
|
|
}
|
|
}
|
|
}
|
|
|
|
func TestCheckSemverConstraint(t *testing.T) {
|
|
ci.Parallel(t)
|
|
|
|
type tcase struct {
|
|
name string
|
|
lVal, rVal interface{}
|
|
result bool
|
|
}
|
|
cases := []tcase{
|
|
{
|
|
name: "Pessimistic operator always fails 1",
|
|
lVal: "1.2.3", rVal: "~> 1.0",
|
|
result: false,
|
|
},
|
|
{
|
|
name: "1.2.3 does satisfy >= 1.0, < 1.4",
|
|
lVal: "1.2.3", rVal: ">= 1.0, < 1.4",
|
|
result: true,
|
|
},
|
|
{
|
|
name: "Pessimistic operator always fails 2",
|
|
lVal: "2.0.1", rVal: "~> 1.0",
|
|
result: false,
|
|
},
|
|
{
|
|
name: "1.4 does not satisfy >= 1.0, < 1.4",
|
|
lVal: "1.4", rVal: ">= 1.0, < 1.4",
|
|
result: false,
|
|
},
|
|
{
|
|
name: "Pessimistic operator always fails 3",
|
|
lVal: 1, rVal: "~> 1.0",
|
|
result: false,
|
|
},
|
|
{
|
|
name: "Prereleases are handled according to semver 1",
|
|
lVal: "1.3.0-beta1", rVal: ">= 0.6.1",
|
|
result: true,
|
|
},
|
|
{
|
|
name: "Prereleases are handled according to semver 2",
|
|
lVal: "1.7.0-alpha1", rVal: ">= 1.6.0-beta1",
|
|
result: true,
|
|
},
|
|
{
|
|
name: "Meta is ignored according to semver",
|
|
lVal: "1.3.0-beta1+ent", rVal: "= 1.3.0-beta1",
|
|
result: true,
|
|
},
|
|
}
|
|
|
|
for _, tc := range cases {
|
|
tc := tc
|
|
t.Run(tc.name, func(t *testing.T) {
|
|
_, ctx := testContext(t)
|
|
p := newSemverConstraintParser(ctx)
|
|
actual := checkVersionMatch(ctx, p, tc.lVal, tc.rVal)
|
|
require.Equal(t, tc.result, actual)
|
|
})
|
|
}
|
|
}
|
|
|
|
func TestCheckRegexpConstraint(t *testing.T) {
|
|
ci.Parallel(t)
|
|
|
|
type tcase struct {
|
|
lVal, rVal interface{}
|
|
result bool
|
|
}
|
|
cases := []tcase{
|
|
{
|
|
lVal: "foobar", rVal: "bar",
|
|
result: true,
|
|
},
|
|
{
|
|
lVal: "foobar", rVal: "^foo",
|
|
result: true,
|
|
},
|
|
{
|
|
lVal: "foobar", rVal: "^bar",
|
|
result: false,
|
|
},
|
|
{
|
|
lVal: "zipzap", rVal: "foo",
|
|
result: false,
|
|
},
|
|
{
|
|
lVal: 1, rVal: "foo",
|
|
result: false,
|
|
},
|
|
}
|
|
for _, tc := range cases {
|
|
_, ctx := testContext(t)
|
|
if res := checkRegexpMatch(ctx, tc.lVal, tc.rVal); res != tc.result {
|
|
t.Fatalf("TC: %#v, Result: %v", tc, res)
|
|
}
|
|
}
|
|
}
|
|
|
|
// This test puts allocations on the node to test if it detects infeasibility of
|
|
// nodes correctly and picks the only feasible one
|
|
func TestDistinctHostsIterator_JobDistinctHosts(t *testing.T) {
|
|
ci.Parallel(t)
|
|
|
|
_, ctx := testContext(t)
|
|
nodes := []*structs.Node{
|
|
mock.Node(),
|
|
mock.Node(),
|
|
mock.Node(),
|
|
}
|
|
static := NewStaticIterator(ctx, nodes)
|
|
|
|
// Create a job with a distinct_hosts constraint and two task groups.
|
|
tg1 := &structs.TaskGroup{Name: "bar"}
|
|
tg2 := &structs.TaskGroup{Name: "baz"}
|
|
|
|
job := &structs.Job{
|
|
ID: "foo",
|
|
Namespace: structs.DefaultNamespace,
|
|
Constraints: []*structs.Constraint{{Operand: structs.ConstraintDistinctHosts}},
|
|
TaskGroups: []*structs.TaskGroup{tg1, tg2},
|
|
}
|
|
|
|
// Add allocs placing tg1 on node1 and tg2 on node2. This should make the
|
|
// job unsatisfiable on all nodes but node3
|
|
plan := ctx.Plan()
|
|
plan.NodeAllocation[nodes[0].ID] = []*structs.Allocation{
|
|
{
|
|
Namespace: structs.DefaultNamespace,
|
|
TaskGroup: tg1.Name,
|
|
JobID: job.ID,
|
|
Job: job,
|
|
ID: uuid.Generate(),
|
|
},
|
|
|
|
// Should be ignored as it is a different job.
|
|
{
|
|
Namespace: structs.DefaultNamespace,
|
|
TaskGroup: tg2.Name,
|
|
JobID: "ignore 2",
|
|
Job: job,
|
|
ID: uuid.Generate(),
|
|
},
|
|
}
|
|
plan.NodeAllocation[nodes[1].ID] = []*structs.Allocation{
|
|
{
|
|
Namespace: structs.DefaultNamespace,
|
|
TaskGroup: tg2.Name,
|
|
JobID: job.ID,
|
|
Job: job,
|
|
ID: uuid.Generate(),
|
|
},
|
|
|
|
// Should be ignored as it is a different job.
|
|
{
|
|
Namespace: structs.DefaultNamespace,
|
|
TaskGroup: tg1.Name,
|
|
JobID: "ignore 2",
|
|
Job: job,
|
|
ID: uuid.Generate(),
|
|
},
|
|
}
|
|
|
|
proposed := NewDistinctHostsIterator(ctx, static)
|
|
proposed.SetTaskGroup(tg1)
|
|
proposed.SetJob(job)
|
|
|
|
out := collectFeasible(proposed)
|
|
if len(out) != 1 {
|
|
t.Fatalf("Bad: %#v", out)
|
|
}
|
|
|
|
if out[0].ID != nodes[2].ID {
|
|
t.Fatalf("wrong node picked")
|
|
}
|
|
}
|
|
|
|
func TestDistinctHostsIterator_JobDistinctHosts_Table(t *testing.T) {
|
|
ci.Parallel(t)
|
|
|
|
// Create a job with a distinct_hosts constraint and a task group.
|
|
tg1 := &structs.TaskGroup{Name: "bar"}
|
|
job := &structs.Job{
|
|
ID: "foo",
|
|
Namespace: structs.DefaultNamespace,
|
|
Constraints: []*structs.Constraint{{
|
|
Operand: structs.ConstraintDistinctHosts,
|
|
}},
|
|
TaskGroups: []*structs.TaskGroup{tg1},
|
|
}
|
|
makeJobAllocs := func(js []*structs.Job) []*structs.Allocation {
|
|
na := make([]*structs.Allocation, len(js))
|
|
for i, j := range js {
|
|
allocID := uuid.Generate()
|
|
ns := structs.DefaultNamespace
|
|
if j.Namespace != "" {
|
|
ns = j.Namespace
|
|
}
|
|
na[i] = &structs.Allocation{
|
|
Namespace: ns,
|
|
TaskGroup: j.TaskGroups[0].Name,
|
|
JobID: j.ID,
|
|
Job: j,
|
|
ID: allocID,
|
|
}
|
|
}
|
|
return na
|
|
}
|
|
|
|
n1 := mock.Node()
|
|
n2 := mock.Node()
|
|
n3 := mock.Node()
|
|
|
|
testCases := []struct {
|
|
name string
|
|
RTarget string
|
|
expectLen int
|
|
expectNodes []*structs.Node
|
|
}{
|
|
{
|
|
name: "unset",
|
|
RTarget: "",
|
|
expectLen: 1,
|
|
expectNodes: []*structs.Node{n3},
|
|
},
|
|
{
|
|
name: "true",
|
|
RTarget: "true",
|
|
expectLen: 1,
|
|
expectNodes: []*structs.Node{n3},
|
|
},
|
|
{
|
|
name: "false",
|
|
RTarget: "false",
|
|
expectLen: 3,
|
|
expectNodes: []*structs.Node{n1, n2, n3},
|
|
},
|
|
{
|
|
name: "unparsable",
|
|
RTarget: "not_a_bool",
|
|
expectLen: 1,
|
|
expectNodes: []*structs.Node{n3},
|
|
},
|
|
}
|
|
for _, tc := range testCases {
|
|
t.Run(tc.name, func(t *testing.T) {
|
|
tc := tc
|
|
ci.Parallel(t)
|
|
|
|
_, ctx := testContext(t)
|
|
static := NewStaticIterator(ctx, []*structs.Node{n1, n2, n3})
|
|
|
|
j := job.Copy()
|
|
j.Constraints[0].RTarget = tc.RTarget
|
|
|
|
// This job has all the same identifiers as the first; however, it is
|
|
// placed in a different namespace to ensure that it doesn't interact
|
|
// with the feasibility of this placement.
|
|
oj := j.Copy()
|
|
oj.Namespace = "different"
|
|
|
|
plan := ctx.Plan()
|
|
// Add allocations so that some of the nodes will be ineligible
|
|
// to receive the job when the distinct_hosts constraint
|
|
// is active. This will require the job be placed on n3.
|
|
//
|
|
// Another job (oj) is placed on all of the nodes to ensure that
|
|
// there are no unexpected interactions between namespaces.
|
|
|
|
plan.NodeAllocation[n1.ID] = makeJobAllocs([]*structs.Job{j, oj})
|
|
plan.NodeAllocation[n2.ID] = makeJobAllocs([]*structs.Job{j, oj})
|
|
plan.NodeAllocation[n3.ID] = makeJobAllocs([]*structs.Job{oj})
|
|
|
|
proposed := NewDistinctHostsIterator(ctx, static)
|
|
proposed.SetTaskGroup(tg1)
|
|
proposed.SetJob(j)
|
|
|
|
out := collectFeasible(proposed)
|
|
must.Len(t, tc.expectLen, out, must.Sprintf("Bad: %v", out))
|
|
must.SliceContainsAll(t, tc.expectNodes, out)
|
|
})
|
|
}
|
|
}
|
|
|
|
func TestDistinctHostsIterator_JobDistinctHosts_InfeasibleCount(t *testing.T) {
|
|
ci.Parallel(t)
|
|
|
|
_, ctx := testContext(t)
|
|
nodes := []*structs.Node{
|
|
mock.Node(),
|
|
mock.Node(),
|
|
}
|
|
static := NewStaticIterator(ctx, nodes)
|
|
|
|
// Create a job with a distinct_hosts constraint and three task groups.
|
|
tg1 := &structs.TaskGroup{Name: "bar"}
|
|
tg2 := &structs.TaskGroup{Name: "baz"}
|
|
tg3 := &structs.TaskGroup{Name: "bam"}
|
|
|
|
job := &structs.Job{
|
|
ID: "foo",
|
|
Namespace: structs.DefaultNamespace,
|
|
Constraints: []*structs.Constraint{{Operand: structs.ConstraintDistinctHosts}},
|
|
TaskGroups: []*structs.TaskGroup{tg1, tg2, tg3},
|
|
}
|
|
|
|
// Add allocs placing tg1 on node1 and tg2 on node2. This should make the
|
|
// job unsatisfiable for tg3
|
|
plan := ctx.Plan()
|
|
plan.NodeAllocation[nodes[0].ID] = []*structs.Allocation{
|
|
{
|
|
Namespace: structs.DefaultNamespace,
|
|
TaskGroup: tg1.Name,
|
|
JobID: job.ID,
|
|
ID: uuid.Generate(),
|
|
},
|
|
}
|
|
plan.NodeAllocation[nodes[1].ID] = []*structs.Allocation{
|
|
{
|
|
Namespace: structs.DefaultNamespace,
|
|
TaskGroup: tg2.Name,
|
|
JobID: job.ID,
|
|
ID: uuid.Generate(),
|
|
},
|
|
}
|
|
|
|
proposed := NewDistinctHostsIterator(ctx, static)
|
|
proposed.SetTaskGroup(tg3)
|
|
proposed.SetJob(job)
|
|
|
|
// It should not be able to place 3 tasks with only two nodes.
|
|
out := collectFeasible(proposed)
|
|
if len(out) != 0 {
|
|
t.Fatalf("Bad: %#v", out)
|
|
}
|
|
}
|
|
|
|
func TestDistinctHostsIterator_TaskGroupDistinctHosts(t *testing.T) {
|
|
ci.Parallel(t)
|
|
|
|
_, ctx := testContext(t)
|
|
nodes := []*structs.Node{
|
|
mock.Node(),
|
|
mock.Node(),
|
|
}
|
|
static := NewStaticIterator(ctx, nodes)
|
|
|
|
// Create a task group with a distinct_hosts constraint.
|
|
tg1 := &structs.TaskGroup{
|
|
Name: "example",
|
|
Constraints: []*structs.Constraint{
|
|
{Operand: structs.ConstraintDistinctHosts},
|
|
},
|
|
}
|
|
tg2 := &structs.TaskGroup{Name: "baz"}
|
|
|
|
// Add a planned alloc to node1.
|
|
plan := ctx.Plan()
|
|
plan.NodeAllocation[nodes[0].ID] = []*structs.Allocation{
|
|
{
|
|
Namespace: structs.DefaultNamespace,
|
|
TaskGroup: tg1.Name,
|
|
JobID: "foo",
|
|
},
|
|
}
|
|
|
|
// Add a planned alloc to node2 with the same task group name but a
|
|
// different job.
|
|
plan.NodeAllocation[nodes[1].ID] = []*structs.Allocation{
|
|
{
|
|
Namespace: structs.DefaultNamespace,
|
|
TaskGroup: tg1.Name,
|
|
JobID: "bar",
|
|
},
|
|
}
|
|
|
|
proposed := NewDistinctHostsIterator(ctx, static)
|
|
proposed.SetTaskGroup(tg1)
|
|
proposed.SetJob(&structs.Job{
|
|
ID: "foo",
|
|
Namespace: structs.DefaultNamespace,
|
|
})
|
|
|
|
out := collectFeasible(proposed)
|
|
if len(out) != 1 {
|
|
t.Fatalf("Bad: %#v", out)
|
|
}
|
|
|
|
// Expect it to skip the first node as there is a previous alloc on it for
|
|
// the same task group.
|
|
if out[0] != nodes[1] {
|
|
t.Fatalf("Bad: %v", out)
|
|
}
|
|
|
|
// Since the other task group doesn't have the constraint, both nodes should
|
|
// be feasible.
|
|
proposed.Reset()
|
|
proposed.SetTaskGroup(tg2)
|
|
out = collectFeasible(proposed)
|
|
if len(out) != 2 {
|
|
t.Fatalf("Bad: %#v", out)
|
|
}
|
|
}
|
|
|
|
// This test puts creates allocations across task groups that use a property
|
|
// value to detect if the constraint at the job level properly considers all
|
|
// task groups.
|
|
func TestDistinctPropertyIterator_JobDistinctProperty(t *testing.T) {
|
|
ci.Parallel(t)
|
|
|
|
state, ctx := testContext(t)
|
|
nodes := []*structs.Node{
|
|
mock.Node(),
|
|
mock.Node(),
|
|
mock.Node(),
|
|
mock.Node(),
|
|
mock.Node(),
|
|
}
|
|
|
|
for i, n := range nodes {
|
|
n.Meta["rack"] = fmt.Sprintf("%d", i)
|
|
|
|
// Add to state store
|
|
if err := state.UpsertNode(structs.MsgTypeTestSetup, uint64(100+i), n); err != nil {
|
|
t.Fatalf("failed to upsert node: %v", err)
|
|
}
|
|
}
|
|
|
|
static := NewStaticIterator(ctx, nodes)
|
|
|
|
// Create a job with a distinct_property constraint and a task groups.
|
|
tg1 := &structs.TaskGroup{Name: "bar"}
|
|
tg2 := &structs.TaskGroup{Name: "baz"}
|
|
|
|
job := &structs.Job{
|
|
ID: "foo",
|
|
Namespace: structs.DefaultNamespace,
|
|
Constraints: []*structs.Constraint{
|
|
{
|
|
Operand: structs.ConstraintDistinctProperty,
|
|
LTarget: "${meta.rack}",
|
|
},
|
|
},
|
|
TaskGroups: []*structs.TaskGroup{tg1, tg2},
|
|
}
|
|
|
|
// Add allocs placing tg1 on node1 and 2 and tg2 on node3 and 4. This should make the
|
|
// job unsatisfiable on all nodes but node5. Also mix the allocations
|
|
// existing in the plan and the state store.
|
|
plan := ctx.Plan()
|
|
alloc1ID := uuid.Generate()
|
|
plan.NodeAllocation[nodes[0].ID] = []*structs.Allocation{
|
|
{
|
|
Namespace: structs.DefaultNamespace,
|
|
TaskGroup: tg1.Name,
|
|
JobID: job.ID,
|
|
Job: job,
|
|
ID: alloc1ID,
|
|
NodeID: nodes[0].ID,
|
|
},
|
|
|
|
// Should be ignored as it is a different job.
|
|
{
|
|
Namespace: structs.DefaultNamespace,
|
|
TaskGroup: tg2.Name,
|
|
JobID: "ignore 2",
|
|
Job: job,
|
|
ID: uuid.Generate(),
|
|
NodeID: nodes[0].ID,
|
|
},
|
|
}
|
|
plan.NodeAllocation[nodes[2].ID] = []*structs.Allocation{
|
|
{
|
|
Namespace: structs.DefaultNamespace,
|
|
TaskGroup: tg2.Name,
|
|
JobID: job.ID,
|
|
Job: job,
|
|
ID: uuid.Generate(),
|
|
NodeID: nodes[2].ID,
|
|
},
|
|
|
|
// Should be ignored as it is a different job.
|
|
{
|
|
Namespace: structs.DefaultNamespace,
|
|
TaskGroup: tg1.Name,
|
|
JobID: "ignore 2",
|
|
Job: job,
|
|
ID: uuid.Generate(),
|
|
NodeID: nodes[2].ID,
|
|
},
|
|
}
|
|
|
|
// Put an allocation on Node 5 but make it stopped in the plan
|
|
stoppingAllocID := uuid.Generate()
|
|
plan.NodeUpdate[nodes[4].ID] = []*structs.Allocation{
|
|
{
|
|
Namespace: structs.DefaultNamespace,
|
|
TaskGroup: tg2.Name,
|
|
JobID: job.ID,
|
|
Job: job,
|
|
ID: stoppingAllocID,
|
|
NodeID: nodes[4].ID,
|
|
},
|
|
}
|
|
|
|
upserting := []*structs.Allocation{
|
|
// Have one of the allocations exist in both the plan and the state
|
|
// store. This resembles an allocation update
|
|
{
|
|
Namespace: structs.DefaultNamespace,
|
|
TaskGroup: tg1.Name,
|
|
JobID: job.ID,
|
|
Job: job,
|
|
ID: alloc1ID,
|
|
EvalID: uuid.Generate(),
|
|
NodeID: nodes[0].ID,
|
|
},
|
|
|
|
{
|
|
Namespace: structs.DefaultNamespace,
|
|
TaskGroup: tg1.Name,
|
|
JobID: job.ID,
|
|
Job: job,
|
|
ID: uuid.Generate(),
|
|
EvalID: uuid.Generate(),
|
|
NodeID: nodes[1].ID,
|
|
},
|
|
|
|
// Should be ignored as it is a different job.
|
|
{
|
|
Namespace: structs.DefaultNamespace,
|
|
TaskGroup: tg2.Name,
|
|
JobID: "ignore 2",
|
|
Job: job,
|
|
ID: uuid.Generate(),
|
|
EvalID: uuid.Generate(),
|
|
NodeID: nodes[1].ID,
|
|
},
|
|
{
|
|
Namespace: structs.DefaultNamespace,
|
|
TaskGroup: tg2.Name,
|
|
JobID: job.ID,
|
|
Job: job,
|
|
ID: uuid.Generate(),
|
|
EvalID: uuid.Generate(),
|
|
NodeID: nodes[3].ID,
|
|
},
|
|
|
|
// Should be ignored as it is a different job.
|
|
{
|
|
Namespace: structs.DefaultNamespace,
|
|
TaskGroup: tg1.Name,
|
|
JobID: "ignore 2",
|
|
Job: job,
|
|
ID: uuid.Generate(),
|
|
EvalID: uuid.Generate(),
|
|
NodeID: nodes[3].ID,
|
|
},
|
|
{
|
|
Namespace: structs.DefaultNamespace,
|
|
TaskGroup: tg2.Name,
|
|
JobID: job.ID,
|
|
Job: job,
|
|
ID: stoppingAllocID,
|
|
EvalID: uuid.Generate(),
|
|
NodeID: nodes[4].ID,
|
|
},
|
|
}
|
|
if err := state.UpsertAllocs(structs.MsgTypeTestSetup, 1000, upserting); err != nil {
|
|
t.Fatalf("failed to UpsertAllocs: %v", err)
|
|
}
|
|
|
|
proposed := NewDistinctPropertyIterator(ctx, static)
|
|
proposed.SetJob(job)
|
|
proposed.SetTaskGroup(tg2)
|
|
proposed.Reset()
|
|
|
|
out := collectFeasible(proposed)
|
|
if len(out) != 1 {
|
|
t.Fatalf("Bad: %#v", out)
|
|
}
|
|
if out[0].ID != nodes[4].ID {
|
|
t.Fatalf("wrong node picked")
|
|
}
|
|
}
|
|
|
|
// This test creates allocations across task groups that use a property value to
|
|
// detect if the constraint at the job level properly considers all task groups
|
|
// when the constraint allows a count greater than one
|
|
func TestDistinctPropertyIterator_JobDistinctProperty_Count(t *testing.T) {
|
|
ci.Parallel(t)
|
|
|
|
state, ctx := testContext(t)
|
|
nodes := []*structs.Node{
|
|
mock.Node(),
|
|
mock.Node(),
|
|
mock.Node(),
|
|
}
|
|
|
|
for i, n := range nodes {
|
|
n.Meta["rack"] = fmt.Sprintf("%d", i)
|
|
|
|
// Add to state store
|
|
if err := state.UpsertNode(structs.MsgTypeTestSetup, uint64(100+i), n); err != nil {
|
|
t.Fatalf("failed to upsert node: %v", err)
|
|
}
|
|
}
|
|
|
|
static := NewStaticIterator(ctx, nodes)
|
|
|
|
// Create a job with a distinct_property constraint and a task groups.
|
|
tg1 := &structs.TaskGroup{Name: "bar"}
|
|
tg2 := &structs.TaskGroup{Name: "baz"}
|
|
|
|
job := &structs.Job{
|
|
ID: "foo",
|
|
Namespace: structs.DefaultNamespace,
|
|
Constraints: []*structs.Constraint{
|
|
{
|
|
Operand: structs.ConstraintDistinctProperty,
|
|
LTarget: "${meta.rack}",
|
|
RTarget: "2",
|
|
},
|
|
},
|
|
TaskGroups: []*structs.TaskGroup{tg1, tg2},
|
|
}
|
|
|
|
// Add allocs placing two allocations on both node 1 and 2 and only one on
|
|
// node 3. This should make the job unsatisfiable on all nodes but node5.
|
|
// Also mix the allocations existing in the plan and the state store.
|
|
plan := ctx.Plan()
|
|
alloc1ID := uuid.Generate()
|
|
plan.NodeAllocation[nodes[0].ID] = []*structs.Allocation{
|
|
{
|
|
Namespace: structs.DefaultNamespace,
|
|
TaskGroup: tg1.Name,
|
|
JobID: job.ID,
|
|
Job: job,
|
|
ID: alloc1ID,
|
|
NodeID: nodes[0].ID,
|
|
},
|
|
|
|
{
|
|
Namespace: structs.DefaultNamespace,
|
|
TaskGroup: tg2.Name,
|
|
JobID: job.ID,
|
|
Job: job,
|
|
ID: alloc1ID,
|
|
NodeID: nodes[0].ID,
|
|
},
|
|
|
|
// Should be ignored as it is a different job.
|
|
{
|
|
Namespace: structs.DefaultNamespace,
|
|
TaskGroup: tg2.Name,
|
|
JobID: "ignore 2",
|
|
Job: job,
|
|
ID: uuid.Generate(),
|
|
NodeID: nodes[0].ID,
|
|
},
|
|
}
|
|
plan.NodeAllocation[nodes[1].ID] = []*structs.Allocation{
|
|
{
|
|
Namespace: structs.DefaultNamespace,
|
|
TaskGroup: tg1.Name,
|
|
JobID: job.ID,
|
|
Job: job,
|
|
ID: uuid.Generate(),
|
|
NodeID: nodes[1].ID,
|
|
},
|
|
|
|
{
|
|
Namespace: structs.DefaultNamespace,
|
|
TaskGroup: tg2.Name,
|
|
JobID: job.ID,
|
|
Job: job,
|
|
ID: uuid.Generate(),
|
|
NodeID: nodes[1].ID,
|
|
},
|
|
|
|
// Should be ignored as it is a different job.
|
|
{
|
|
Namespace: structs.DefaultNamespace,
|
|
TaskGroup: tg1.Name,
|
|
JobID: "ignore 2",
|
|
Job: job,
|
|
ID: uuid.Generate(),
|
|
NodeID: nodes[1].ID,
|
|
},
|
|
}
|
|
plan.NodeAllocation[nodes[2].ID] = []*structs.Allocation{
|
|
{
|
|
Namespace: structs.DefaultNamespace,
|
|
TaskGroup: tg1.Name,
|
|
JobID: job.ID,
|
|
Job: job,
|
|
ID: uuid.Generate(),
|
|
NodeID: nodes[2].ID,
|
|
},
|
|
|
|
// Should be ignored as it is a different job.
|
|
{
|
|
Namespace: structs.DefaultNamespace,
|
|
TaskGroup: tg1.Name,
|
|
JobID: "ignore 2",
|
|
Job: job,
|
|
ID: uuid.Generate(),
|
|
NodeID: nodes[2].ID,
|
|
},
|
|
}
|
|
|
|
// Put an allocation on Node 3 but make it stopped in the plan
|
|
stoppingAllocID := uuid.Generate()
|
|
plan.NodeUpdate[nodes[2].ID] = []*structs.Allocation{
|
|
{
|
|
Namespace: structs.DefaultNamespace,
|
|
TaskGroup: tg2.Name,
|
|
JobID: job.ID,
|
|
Job: job,
|
|
ID: stoppingAllocID,
|
|
NodeID: nodes[2].ID,
|
|
},
|
|
}
|
|
|
|
upserting := []*structs.Allocation{
|
|
// Have one of the allocations exist in both the plan and the state
|
|
// store. This resembles an allocation update
|
|
{
|
|
Namespace: structs.DefaultNamespace,
|
|
TaskGroup: tg1.Name,
|
|
JobID: job.ID,
|
|
Job: job,
|
|
ID: alloc1ID,
|
|
EvalID: uuid.Generate(),
|
|
NodeID: nodes[0].ID,
|
|
},
|
|
|
|
{
|
|
Namespace: structs.DefaultNamespace,
|
|
TaskGroup: tg1.Name,
|
|
JobID: job.ID,
|
|
Job: job,
|
|
ID: uuid.Generate(),
|
|
EvalID: uuid.Generate(),
|
|
NodeID: nodes[1].ID,
|
|
},
|
|
|
|
{
|
|
Namespace: structs.DefaultNamespace,
|
|
TaskGroup: tg2.Name,
|
|
JobID: job.ID,
|
|
Job: job,
|
|
ID: uuid.Generate(),
|
|
EvalID: uuid.Generate(),
|
|
NodeID: nodes[0].ID,
|
|
},
|
|
|
|
// Should be ignored as it is a different job.
|
|
{
|
|
Namespace: structs.DefaultNamespace,
|
|
TaskGroup: tg1.Name,
|
|
JobID: "ignore 2",
|
|
Job: job,
|
|
ID: uuid.Generate(),
|
|
EvalID: uuid.Generate(),
|
|
NodeID: nodes[1].ID,
|
|
},
|
|
{
|
|
Namespace: structs.DefaultNamespace,
|
|
TaskGroup: tg2.Name,
|
|
JobID: "ignore 2",
|
|
Job: job,
|
|
ID: uuid.Generate(),
|
|
EvalID: uuid.Generate(),
|
|
NodeID: nodes[1].ID,
|
|
},
|
|
}
|
|
if err := state.UpsertAllocs(structs.MsgTypeTestSetup, 1000, upserting); err != nil {
|
|
t.Fatalf("failed to UpsertAllocs: %v", err)
|
|
}
|
|
|
|
proposed := NewDistinctPropertyIterator(ctx, static)
|
|
proposed.SetJob(job)
|
|
proposed.SetTaskGroup(tg2)
|
|
proposed.Reset()
|
|
|
|
out := collectFeasible(proposed)
|
|
if len(out) != 1 {
|
|
t.Fatalf("Bad: %#v", out)
|
|
}
|
|
if out[0].ID != nodes[2].ID {
|
|
t.Fatalf("wrong node picked")
|
|
}
|
|
}
|
|
|
|
// This test checks that if a node has an allocation on it that gets stopped,
|
|
// there is a plan to re-use that for a new allocation, that the next select
|
|
// won't select that node.
|
|
func TestDistinctPropertyIterator_JobDistinctProperty_RemoveAndReplace(t *testing.T) {
|
|
ci.Parallel(t)
|
|
|
|
state, ctx := testContext(t)
|
|
nodes := []*structs.Node{
|
|
mock.Node(),
|
|
}
|
|
|
|
nodes[0].Meta["rack"] = "1"
|
|
|
|
// Add to state store
|
|
if err := state.UpsertNode(structs.MsgTypeTestSetup, uint64(100), nodes[0]); err != nil {
|
|
t.Fatalf("failed to upsert node: %v", err)
|
|
}
|
|
|
|
static := NewStaticIterator(ctx, nodes)
|
|
|
|
// Create a job with a distinct_property constraint and a task groups.
|
|
tg1 := &structs.TaskGroup{Name: "bar"}
|
|
job := &structs.Job{
|
|
Namespace: structs.DefaultNamespace,
|
|
ID: "foo",
|
|
Constraints: []*structs.Constraint{
|
|
{
|
|
Operand: structs.ConstraintDistinctProperty,
|
|
LTarget: "${meta.rack}",
|
|
},
|
|
},
|
|
TaskGroups: []*structs.TaskGroup{tg1},
|
|
}
|
|
|
|
plan := ctx.Plan()
|
|
plan.NodeAllocation[nodes[0].ID] = []*structs.Allocation{
|
|
{
|
|
Namespace: structs.DefaultNamespace,
|
|
TaskGroup: tg1.Name,
|
|
JobID: job.ID,
|
|
Job: job,
|
|
ID: uuid.Generate(),
|
|
NodeID: nodes[0].ID,
|
|
},
|
|
}
|
|
|
|
stoppingAllocID := uuid.Generate()
|
|
plan.NodeUpdate[nodes[0].ID] = []*structs.Allocation{
|
|
{
|
|
Namespace: structs.DefaultNamespace,
|
|
TaskGroup: tg1.Name,
|
|
JobID: job.ID,
|
|
Job: job,
|
|
ID: stoppingAllocID,
|
|
NodeID: nodes[0].ID,
|
|
},
|
|
}
|
|
|
|
upserting := []*structs.Allocation{
|
|
{
|
|
Namespace: structs.DefaultNamespace,
|
|
TaskGroup: tg1.Name,
|
|
JobID: job.ID,
|
|
Job: job,
|
|
ID: stoppingAllocID,
|
|
EvalID: uuid.Generate(),
|
|
NodeID: nodes[0].ID,
|
|
},
|
|
}
|
|
if err := state.UpsertAllocs(structs.MsgTypeTestSetup, 1000, upserting); err != nil {
|
|
t.Fatalf("failed to UpsertAllocs: %v", err)
|
|
}
|
|
|
|
proposed := NewDistinctPropertyIterator(ctx, static)
|
|
proposed.SetJob(job)
|
|
proposed.SetTaskGroup(tg1)
|
|
proposed.Reset()
|
|
|
|
out := collectFeasible(proposed)
|
|
if len(out) != 0 {
|
|
t.Fatalf("Bad: %#v", out)
|
|
}
|
|
}
|
|
|
|
// This test creates previous allocations selecting certain property values to
|
|
// test if it detects infeasibility of property values correctly and picks the
|
|
// only feasible one
|
|
func TestDistinctPropertyIterator_JobDistinctProperty_Infeasible(t *testing.T) {
|
|
ci.Parallel(t)
|
|
|
|
state, ctx := testContext(t)
|
|
nodes := []*structs.Node{
|
|
mock.Node(),
|
|
mock.Node(),
|
|
}
|
|
|
|
for i, n := range nodes {
|
|
n.Meta["rack"] = fmt.Sprintf("%d", i)
|
|
|
|
// Add to state store
|
|
if err := state.UpsertNode(structs.MsgTypeTestSetup, uint64(100+i), n); err != nil {
|
|
t.Fatalf("failed to upsert node: %v", err)
|
|
}
|
|
}
|
|
|
|
static := NewStaticIterator(ctx, nodes)
|
|
|
|
// Create a job with a distinct_property constraint and a task groups.
|
|
tg1 := &structs.TaskGroup{Name: "bar"}
|
|
tg2 := &structs.TaskGroup{Name: "baz"}
|
|
tg3 := &structs.TaskGroup{Name: "bam"}
|
|
|
|
job := &structs.Job{
|
|
Namespace: structs.DefaultNamespace,
|
|
ID: "foo",
|
|
Constraints: []*structs.Constraint{
|
|
{
|
|
Operand: structs.ConstraintDistinctProperty,
|
|
LTarget: "${meta.rack}",
|
|
},
|
|
},
|
|
TaskGroups: []*structs.TaskGroup{tg1, tg2, tg3},
|
|
}
|
|
|
|
// Add allocs placing tg1 on node1 and tg2 on node2. This should make the
|
|
// job unsatisfiable for tg3.
|
|
plan := ctx.Plan()
|
|
plan.NodeAllocation[nodes[0].ID] = []*structs.Allocation{
|
|
{
|
|
Namespace: structs.DefaultNamespace,
|
|
TaskGroup: tg1.Name,
|
|
JobID: job.ID,
|
|
Job: job,
|
|
ID: uuid.Generate(),
|
|
NodeID: nodes[0].ID,
|
|
},
|
|
}
|
|
upserting := []*structs.Allocation{
|
|
{
|
|
Namespace: structs.DefaultNamespace,
|
|
TaskGroup: tg2.Name,
|
|
JobID: job.ID,
|
|
Job: job,
|
|
ID: uuid.Generate(),
|
|
EvalID: uuid.Generate(),
|
|
NodeID: nodes[1].ID,
|
|
},
|
|
}
|
|
if err := state.UpsertAllocs(structs.MsgTypeTestSetup, 1000, upserting); err != nil {
|
|
t.Fatalf("failed to UpsertAllocs: %v", err)
|
|
}
|
|
|
|
proposed := NewDistinctPropertyIterator(ctx, static)
|
|
proposed.SetJob(job)
|
|
proposed.SetTaskGroup(tg3)
|
|
proposed.Reset()
|
|
|
|
out := collectFeasible(proposed)
|
|
if len(out) != 0 {
|
|
t.Fatalf("Bad: %#v", out)
|
|
}
|
|
}
|
|
|
|
// This test creates previous allocations selecting certain property values to
|
|
// test if it detects infeasibility of property values correctly and picks the
|
|
// only feasible one
|
|
func TestDistinctPropertyIterator_JobDistinctProperty_Infeasible_Count(t *testing.T) {
|
|
ci.Parallel(t)
|
|
|
|
state, ctx := testContext(t)
|
|
nodes := []*structs.Node{
|
|
mock.Node(),
|
|
mock.Node(),
|
|
}
|
|
|
|
for i, n := range nodes {
|
|
n.Meta["rack"] = fmt.Sprintf("%d", i)
|
|
|
|
// Add to state store
|
|
if err := state.UpsertNode(structs.MsgTypeTestSetup, uint64(100+i), n); err != nil {
|
|
t.Fatalf("failed to upsert node: %v", err)
|
|
}
|
|
}
|
|
|
|
static := NewStaticIterator(ctx, nodes)
|
|
|
|
// Create a job with a distinct_property constraint and a task groups.
|
|
tg1 := &structs.TaskGroup{Name: "bar"}
|
|
tg2 := &structs.TaskGroup{Name: "baz"}
|
|
tg3 := &structs.TaskGroup{Name: "bam"}
|
|
|
|
job := &structs.Job{
|
|
Namespace: structs.DefaultNamespace,
|
|
ID: "foo",
|
|
Constraints: []*structs.Constraint{
|
|
{
|
|
Operand: structs.ConstraintDistinctProperty,
|
|
LTarget: "${meta.rack}",
|
|
RTarget: "2",
|
|
},
|
|
},
|
|
TaskGroups: []*structs.TaskGroup{tg1, tg2, tg3},
|
|
}
|
|
|
|
// Add allocs placing two tg1's on node1 and two tg2's on node2. This should
|
|
// make the job unsatisfiable for tg3.
|
|
plan := ctx.Plan()
|
|
plan.NodeAllocation[nodes[0].ID] = []*structs.Allocation{
|
|
{
|
|
Namespace: structs.DefaultNamespace,
|
|
TaskGroup: tg1.Name,
|
|
JobID: job.ID,
|
|
Job: job,
|
|
ID: uuid.Generate(),
|
|
NodeID: nodes[0].ID,
|
|
},
|
|
{
|
|
Namespace: structs.DefaultNamespace,
|
|
TaskGroup: tg2.Name,
|
|
JobID: job.ID,
|
|
Job: job,
|
|
ID: uuid.Generate(),
|
|
NodeID: nodes[0].ID,
|
|
},
|
|
}
|
|
upserting := []*structs.Allocation{
|
|
{
|
|
Namespace: structs.DefaultNamespace,
|
|
TaskGroup: tg1.Name,
|
|
JobID: job.ID,
|
|
Job: job,
|
|
ID: uuid.Generate(),
|
|
EvalID: uuid.Generate(),
|
|
NodeID: nodes[1].ID,
|
|
},
|
|
{
|
|
Namespace: structs.DefaultNamespace,
|
|
TaskGroup: tg2.Name,
|
|
JobID: job.ID,
|
|
Job: job,
|
|
ID: uuid.Generate(),
|
|
EvalID: uuid.Generate(),
|
|
NodeID: nodes[1].ID,
|
|
},
|
|
}
|
|
if err := state.UpsertAllocs(structs.MsgTypeTestSetup, 1000, upserting); err != nil {
|
|
t.Fatalf("failed to UpsertAllocs: %v", err)
|
|
}
|
|
|
|
proposed := NewDistinctPropertyIterator(ctx, static)
|
|
proposed.SetJob(job)
|
|
proposed.SetTaskGroup(tg3)
|
|
proposed.Reset()
|
|
|
|
out := collectFeasible(proposed)
|
|
if len(out) != 0 {
|
|
t.Fatalf("Bad: %#v", out)
|
|
}
|
|
}
|
|
|
|
// This test creates previous allocations selecting certain property values to
|
|
// test if it detects infeasibility of property values correctly and picks the
|
|
// only feasible one when the constraint is at the task group.
|
|
func TestDistinctPropertyIterator_TaskGroupDistinctProperty(t *testing.T) {
|
|
ci.Parallel(t)
|
|
|
|
state, ctx := testContext(t)
|
|
nodes := []*structs.Node{
|
|
mock.Node(),
|
|
mock.Node(),
|
|
mock.Node(),
|
|
}
|
|
|
|
for i, n := range nodes {
|
|
n.Meta["rack"] = fmt.Sprintf("%d", i)
|
|
|
|
// Add to state store
|
|
if err := state.UpsertNode(structs.MsgTypeTestSetup, uint64(100+i), n); err != nil {
|
|
t.Fatalf("failed to upsert node: %v", err)
|
|
}
|
|
}
|
|
|
|
static := NewStaticIterator(ctx, nodes)
|
|
|
|
// Create a job with a task group with the distinct_property constraint
|
|
tg1 := &structs.TaskGroup{
|
|
Name: "example",
|
|
Constraints: []*structs.Constraint{
|
|
{
|
|
Operand: structs.ConstraintDistinctProperty,
|
|
LTarget: "${meta.rack}",
|
|
},
|
|
},
|
|
}
|
|
tg2 := &structs.TaskGroup{Name: "baz"}
|
|
|
|
job := &structs.Job{
|
|
Namespace: structs.DefaultNamespace,
|
|
ID: "foo",
|
|
TaskGroups: []*structs.TaskGroup{tg1, tg2},
|
|
}
|
|
|
|
// Add allocs placing tg1 on node1 and 2. This should make the
|
|
// job unsatisfiable on all nodes but node3. Also mix the allocations
|
|
// existing in the plan and the state store.
|
|
plan := ctx.Plan()
|
|
plan.NodeAllocation[nodes[0].ID] = []*structs.Allocation{
|
|
{
|
|
Namespace: structs.DefaultNamespace,
|
|
TaskGroup: tg1.Name,
|
|
JobID: job.ID,
|
|
Job: job,
|
|
ID: uuid.Generate(),
|
|
NodeID: nodes[0].ID,
|
|
},
|
|
}
|
|
|
|
// Put an allocation on Node 3 but make it stopped in the plan
|
|
stoppingAllocID := uuid.Generate()
|
|
plan.NodeUpdate[nodes[2].ID] = []*structs.Allocation{
|
|
{
|
|
Namespace: structs.DefaultNamespace,
|
|
TaskGroup: tg1.Name,
|
|
JobID: job.ID,
|
|
Job: job,
|
|
ID: stoppingAllocID,
|
|
NodeID: nodes[2].ID,
|
|
},
|
|
}
|
|
|
|
upserting := []*structs.Allocation{
|
|
{
|
|
Namespace: structs.DefaultNamespace,
|
|
TaskGroup: tg1.Name,
|
|
JobID: job.ID,
|
|
Job: job,
|
|
ID: uuid.Generate(),
|
|
EvalID: uuid.Generate(),
|
|
NodeID: nodes[1].ID,
|
|
},
|
|
|
|
// Should be ignored as it is a different job.
|
|
{
|
|
Namespace: structs.DefaultNamespace,
|
|
TaskGroup: tg1.Name,
|
|
JobID: "ignore 2",
|
|
Job: job,
|
|
ID: uuid.Generate(),
|
|
EvalID: uuid.Generate(),
|
|
NodeID: nodes[2].ID,
|
|
},
|
|
|
|
{
|
|
Namespace: structs.DefaultNamespace,
|
|
TaskGroup: tg1.Name,
|
|
JobID: job.ID,
|
|
Job: job,
|
|
ID: stoppingAllocID,
|
|
EvalID: uuid.Generate(),
|
|
NodeID: nodes[2].ID,
|
|
},
|
|
}
|
|
if err := state.UpsertAllocs(structs.MsgTypeTestSetup, 1000, upserting); err != nil {
|
|
t.Fatalf("failed to UpsertAllocs: %v", err)
|
|
}
|
|
|
|
proposed := NewDistinctPropertyIterator(ctx, static)
|
|
proposed.SetJob(job)
|
|
proposed.SetTaskGroup(tg1)
|
|
proposed.Reset()
|
|
|
|
out := collectFeasible(proposed)
|
|
if len(out) != 1 {
|
|
t.Fatalf("Bad: %#v", out)
|
|
}
|
|
if out[0].ID != nodes[2].ID {
|
|
t.Fatalf("wrong node picked")
|
|
}
|
|
|
|
// Since the other task group doesn't have the constraint, both nodes should
|
|
// be feasible.
|
|
proposed.SetTaskGroup(tg2)
|
|
proposed.Reset()
|
|
|
|
out = collectFeasible(proposed)
|
|
if len(out) != 3 {
|
|
t.Fatalf("Bad: %#v", out)
|
|
}
|
|
}
|
|
|
|
func collectFeasible(iter FeasibleIterator) (out []*structs.Node) {
|
|
for {
|
|
next := iter.Next()
|
|
if next == nil {
|
|
break
|
|
}
|
|
out = append(out, next)
|
|
}
|
|
return
|
|
}
|
|
|
|
// mockFeasibilityChecker is a FeasibilityChecker that returns predetermined
|
|
// feasibility values.
|
|
type mockFeasibilityChecker struct {
|
|
retVals []bool
|
|
i int
|
|
}
|
|
|
|
func newMockFeasibilityChecker(values ...bool) *mockFeasibilityChecker {
|
|
return &mockFeasibilityChecker{retVals: values}
|
|
}
|
|
|
|
func (c *mockFeasibilityChecker) Feasible(*structs.Node) bool {
|
|
if c.i >= len(c.retVals) {
|
|
c.i++
|
|
return false
|
|
}
|
|
|
|
f := c.retVals[c.i]
|
|
c.i++
|
|
return f
|
|
}
|
|
|
|
// calls returns how many times the checker was called.
|
|
func (c *mockFeasibilityChecker) calls() int { return c.i }
|
|
|
|
func TestFeasibilityWrapper_JobIneligible(t *testing.T) {
|
|
ci.Parallel(t)
|
|
|
|
_, ctx := testContext(t)
|
|
nodes := []*structs.Node{mock.Node()}
|
|
static := NewStaticIterator(ctx, nodes)
|
|
mocked := newMockFeasibilityChecker(false)
|
|
wrapper := NewFeasibilityWrapper(ctx, static, []FeasibilityChecker{mocked}, nil, nil)
|
|
|
|
// Set the job to ineligible
|
|
ctx.Eligibility().SetJobEligibility(false, nodes[0].ComputedClass)
|
|
|
|
// Run the wrapper.
|
|
out := collectFeasible(wrapper)
|
|
|
|
if out != nil || mocked.calls() != 0 {
|
|
t.Fatalf("bad: %#v %d", out, mocked.calls())
|
|
}
|
|
}
|
|
|
|
func TestFeasibilityWrapper_JobEscapes(t *testing.T) {
|
|
ci.Parallel(t)
|
|
|
|
_, ctx := testContext(t)
|
|
nodes := []*structs.Node{mock.Node()}
|
|
static := NewStaticIterator(ctx, nodes)
|
|
mocked := newMockFeasibilityChecker(false)
|
|
wrapper := NewFeasibilityWrapper(ctx, static, []FeasibilityChecker{mocked}, nil, nil)
|
|
|
|
// Set the job to escaped
|
|
cc := nodes[0].ComputedClass
|
|
ctx.Eligibility().job[cc] = EvalComputedClassEscaped
|
|
|
|
// Run the wrapper.
|
|
out := collectFeasible(wrapper)
|
|
|
|
if out != nil || mocked.calls() != 1 {
|
|
t.Fatalf("bad: %#v", out)
|
|
}
|
|
|
|
// Ensure that the job status didn't change from escaped even though the
|
|
// option failed.
|
|
if status := ctx.Eligibility().JobStatus(cc); status != EvalComputedClassEscaped {
|
|
t.Fatalf("job status is %v; want %v", status, EvalComputedClassEscaped)
|
|
}
|
|
}
|
|
|
|
func TestFeasibilityWrapper_JobAndTg_Eligible(t *testing.T) {
|
|
ci.Parallel(t)
|
|
|
|
_, ctx := testContext(t)
|
|
nodes := []*structs.Node{mock.Node()}
|
|
static := NewStaticIterator(ctx, nodes)
|
|
jobMock := newMockFeasibilityChecker(true)
|
|
tgMock := newMockFeasibilityChecker(false)
|
|
wrapper := NewFeasibilityWrapper(ctx, static, []FeasibilityChecker{jobMock}, []FeasibilityChecker{tgMock}, nil)
|
|
|
|
// Set the job to escaped
|
|
cc := nodes[0].ComputedClass
|
|
ctx.Eligibility().job[cc] = EvalComputedClassEligible
|
|
ctx.Eligibility().SetTaskGroupEligibility(true, "foo", cc)
|
|
wrapper.SetTaskGroup("foo")
|
|
|
|
// Run the wrapper.
|
|
out := collectFeasible(wrapper)
|
|
|
|
if out == nil || tgMock.calls() != 0 {
|
|
t.Fatalf("bad: %#v %v", out, tgMock.calls())
|
|
}
|
|
}
|
|
|
|
func TestFeasibilityWrapper_JobEligible_TgIneligible(t *testing.T) {
|
|
ci.Parallel(t)
|
|
|
|
_, ctx := testContext(t)
|
|
nodes := []*structs.Node{mock.Node()}
|
|
static := NewStaticIterator(ctx, nodes)
|
|
jobMock := newMockFeasibilityChecker(true)
|
|
tgMock := newMockFeasibilityChecker(false)
|
|
wrapper := NewFeasibilityWrapper(ctx, static, []FeasibilityChecker{jobMock}, []FeasibilityChecker{tgMock}, nil)
|
|
|
|
// Set the job to escaped
|
|
cc := nodes[0].ComputedClass
|
|
ctx.Eligibility().job[cc] = EvalComputedClassEligible
|
|
ctx.Eligibility().SetTaskGroupEligibility(false, "foo", cc)
|
|
wrapper.SetTaskGroup("foo")
|
|
|
|
// Run the wrapper.
|
|
out := collectFeasible(wrapper)
|
|
|
|
if out != nil || tgMock.calls() != 0 {
|
|
t.Fatalf("bad: %#v %v", out, tgMock.calls())
|
|
}
|
|
}
|
|
|
|
func TestFeasibilityWrapper_JobEligible_TgEscaped(t *testing.T) {
|
|
ci.Parallel(t)
|
|
|
|
_, ctx := testContext(t)
|
|
nodes := []*structs.Node{mock.Node()}
|
|
static := NewStaticIterator(ctx, nodes)
|
|
jobMock := newMockFeasibilityChecker(true)
|
|
tgMock := newMockFeasibilityChecker(true)
|
|
wrapper := NewFeasibilityWrapper(ctx, static, []FeasibilityChecker{jobMock}, []FeasibilityChecker{tgMock}, nil)
|
|
|
|
// Set the job to escaped
|
|
cc := nodes[0].ComputedClass
|
|
ctx.Eligibility().job[cc] = EvalComputedClassEligible
|
|
ctx.Eligibility().taskGroups["foo"] =
|
|
map[string]ComputedClassFeasibility{cc: EvalComputedClassEscaped}
|
|
wrapper.SetTaskGroup("foo")
|
|
|
|
// Run the wrapper.
|
|
out := collectFeasible(wrapper)
|
|
|
|
if out == nil || tgMock.calls() != 1 {
|
|
t.Fatalf("bad: %#v %v", out, tgMock.calls())
|
|
}
|
|
|
|
if e, ok := ctx.Eligibility().taskGroups["foo"][cc]; !ok || e != EvalComputedClassEscaped {
|
|
t.Fatalf("bad: %v %v", e, ok)
|
|
}
|
|
}
|
|
|
|
func TestSetContainsAny(t *testing.T) {
|
|
ci.Parallel(t)
|
|
|
|
require.True(t, checkSetContainsAny("a", "a"))
|
|
require.True(t, checkSetContainsAny("a,b", "a"))
|
|
require.True(t, checkSetContainsAny(" a,b ", "a "))
|
|
require.True(t, checkSetContainsAny("a", "a"))
|
|
require.False(t, checkSetContainsAny("b", "a"))
|
|
}
|
|
|
|
func TestDeviceChecker(t *testing.T) {
|
|
ci.Parallel(t)
|
|
|
|
getTg := func(devices ...*structs.RequestedDevice) *structs.TaskGroup {
|
|
return &structs.TaskGroup{
|
|
Name: "example",
|
|
Tasks: []*structs.Task{
|
|
{
|
|
Resources: &structs.Resources{
|
|
Devices: devices,
|
|
},
|
|
},
|
|
},
|
|
}
|
|
}
|
|
|
|
// Just type
|
|
gpuTypeReq := &structs.RequestedDevice{
|
|
Name: "gpu",
|
|
Count: 1,
|
|
}
|
|
fpgaTypeReq := &structs.RequestedDevice{
|
|
Name: "fpga",
|
|
Count: 1,
|
|
}
|
|
|
|
// vendor/type
|
|
gpuVendorTypeReq := &structs.RequestedDevice{
|
|
Name: "nvidia/gpu",
|
|
Count: 1,
|
|
}
|
|
fpgaVendorTypeReq := &structs.RequestedDevice{
|
|
Name: "nvidia/fpga",
|
|
Count: 1,
|
|
}
|
|
|
|
// vendor/type/model
|
|
gpuFullReq := &structs.RequestedDevice{
|
|
Name: "nvidia/gpu/1080ti",
|
|
Count: 1,
|
|
}
|
|
fpgaFullReq := &structs.RequestedDevice{
|
|
Name: "nvidia/fpga/F100",
|
|
Count: 1,
|
|
}
|
|
|
|
// Just type but high count
|
|
gpuTypeHighCountReq := &structs.RequestedDevice{
|
|
Name: "gpu",
|
|
Count: 3,
|
|
}
|
|
|
|
getNode := func(devices ...*structs.NodeDeviceResource) *structs.Node {
|
|
n := mock.Node()
|
|
n.NodeResources.Devices = devices
|
|
return n
|
|
}
|
|
|
|
nvidia := &structs.NodeDeviceResource{
|
|
Vendor: "nvidia",
|
|
Type: "gpu",
|
|
Name: "1080ti",
|
|
Attributes: map[string]*psstructs.Attribute{
|
|
"memory": psstructs.NewIntAttribute(4, psstructs.UnitGiB),
|
|
"pci_bandwidth": psstructs.NewIntAttribute(995, psstructs.UnitMiBPerS),
|
|
"cores_clock": psstructs.NewIntAttribute(800, psstructs.UnitMHz),
|
|
},
|
|
Instances: []*structs.NodeDevice{
|
|
{
|
|
ID: uuid.Generate(),
|
|
Healthy: true,
|
|
},
|
|
{
|
|
ID: uuid.Generate(),
|
|
Healthy: true,
|
|
},
|
|
},
|
|
}
|
|
|
|
nvidiaUnhealthy := &structs.NodeDeviceResource{
|
|
Vendor: "nvidia",
|
|
Type: "gpu",
|
|
Name: "1080ti",
|
|
Instances: []*structs.NodeDevice{
|
|
{
|
|
ID: uuid.Generate(),
|
|
Healthy: false,
|
|
},
|
|
{
|
|
ID: uuid.Generate(),
|
|
Healthy: false,
|
|
},
|
|
},
|
|
}
|
|
|
|
cases := []struct {
|
|
Name string
|
|
Result bool
|
|
NodeDevices []*structs.NodeDeviceResource
|
|
RequestedDevices []*structs.RequestedDevice
|
|
}{
|
|
{
|
|
Name: "no devices on node",
|
|
Result: false,
|
|
NodeDevices: nil,
|
|
RequestedDevices: []*structs.RequestedDevice{gpuTypeReq},
|
|
},
|
|
{
|
|
Name: "no requested devices on empty node",
|
|
Result: true,
|
|
NodeDevices: nil,
|
|
RequestedDevices: nil,
|
|
},
|
|
{
|
|
Name: "gpu devices by type",
|
|
Result: true,
|
|
NodeDevices: []*structs.NodeDeviceResource{nvidia},
|
|
RequestedDevices: []*structs.RequestedDevice{gpuTypeReq},
|
|
},
|
|
{
|
|
Name: "wrong devices by type",
|
|
Result: false,
|
|
NodeDevices: []*structs.NodeDeviceResource{nvidia},
|
|
RequestedDevices: []*structs.RequestedDevice{fpgaTypeReq},
|
|
},
|
|
{
|
|
Name: "devices by type unhealthy node",
|
|
Result: false,
|
|
NodeDevices: []*structs.NodeDeviceResource{nvidiaUnhealthy},
|
|
RequestedDevices: []*structs.RequestedDevice{gpuTypeReq},
|
|
},
|
|
{
|
|
Name: "gpu devices by vendor/type",
|
|
Result: true,
|
|
NodeDevices: []*structs.NodeDeviceResource{nvidia},
|
|
RequestedDevices: []*structs.RequestedDevice{gpuVendorTypeReq},
|
|
},
|
|
{
|
|
Name: "wrong devices by vendor/type",
|
|
Result: false,
|
|
NodeDevices: []*structs.NodeDeviceResource{nvidia},
|
|
RequestedDevices: []*structs.RequestedDevice{fpgaVendorTypeReq},
|
|
},
|
|
{
|
|
Name: "gpu devices by vendor/type/model",
|
|
Result: true,
|
|
NodeDevices: []*structs.NodeDeviceResource{nvidia},
|
|
RequestedDevices: []*structs.RequestedDevice{gpuFullReq},
|
|
},
|
|
{
|
|
Name: "wrong devices by vendor/type/model",
|
|
Result: false,
|
|
NodeDevices: []*structs.NodeDeviceResource{nvidia},
|
|
RequestedDevices: []*structs.RequestedDevice{fpgaFullReq},
|
|
},
|
|
{
|
|
Name: "too many requested",
|
|
Result: false,
|
|
NodeDevices: []*structs.NodeDeviceResource{nvidia},
|
|
RequestedDevices: []*structs.RequestedDevice{gpuTypeHighCountReq},
|
|
},
|
|
{
|
|
Name: "meets constraints requirement",
|
|
Result: true,
|
|
NodeDevices: []*structs.NodeDeviceResource{nvidia},
|
|
RequestedDevices: []*structs.RequestedDevice{
|
|
{
|
|
Name: "nvidia/gpu",
|
|
Count: 1,
|
|
Constraints: []*structs.Constraint{
|
|
{
|
|
Operand: "=",
|
|
LTarget: "${device.model}",
|
|
RTarget: "1080ti",
|
|
},
|
|
{
|
|
Operand: ">",
|
|
LTarget: "${device.attr.memory}",
|
|
RTarget: "1320.5 MB",
|
|
},
|
|
{
|
|
Operand: "<=",
|
|
LTarget: "${device.attr.pci_bandwidth}",
|
|
RTarget: ".98 GiB/s",
|
|
},
|
|
{
|
|
Operand: "=",
|
|
LTarget: "${device.attr.cores_clock}",
|
|
RTarget: "800MHz",
|
|
},
|
|
{
|
|
Operand: "set_contains",
|
|
LTarget: "${device.ids}",
|
|
RTarget: nvidia.Instances[0].ID,
|
|
},
|
|
},
|
|
},
|
|
},
|
|
},
|
|
{
|
|
Name: "meets constraints requirement multiple count",
|
|
Result: true,
|
|
NodeDevices: []*structs.NodeDeviceResource{nvidia},
|
|
RequestedDevices: []*structs.RequestedDevice{
|
|
{
|
|
Name: "nvidia/gpu",
|
|
Count: 2,
|
|
Constraints: []*structs.Constraint{
|
|
{
|
|
Operand: "=",
|
|
LTarget: "${device.model}",
|
|
RTarget: "1080ti",
|
|
},
|
|
{
|
|
Operand: ">",
|
|
LTarget: "${device.attr.memory}",
|
|
RTarget: "1320.5 MB",
|
|
},
|
|
{
|
|
Operand: "<=",
|
|
LTarget: "${device.attr.pci_bandwidth}",
|
|
RTarget: ".98 GiB/s",
|
|
},
|
|
{
|
|
Operand: "=",
|
|
LTarget: "${device.attr.cores_clock}",
|
|
RTarget: "800MHz",
|
|
},
|
|
{
|
|
Operand: "set_contains",
|
|
LTarget: "${device.ids}",
|
|
RTarget: fmt.Sprintf("%s,%s", nvidia.Instances[1].ID, nvidia.Instances[0].ID),
|
|
},
|
|
},
|
|
},
|
|
},
|
|
},
|
|
{
|
|
Name: "meets constraints requirement over count",
|
|
Result: false,
|
|
NodeDevices: []*structs.NodeDeviceResource{nvidia},
|
|
RequestedDevices: []*structs.RequestedDevice{
|
|
{
|
|
Name: "nvidia/gpu",
|
|
Count: 5,
|
|
Constraints: []*structs.Constraint{
|
|
{
|
|
Operand: "=",
|
|
LTarget: "${device.model}",
|
|
RTarget: "1080ti",
|
|
},
|
|
{
|
|
Operand: ">",
|
|
LTarget: "${device.attr.memory}",
|
|
RTarget: "1320.5 MB",
|
|
},
|
|
{
|
|
Operand: "<=",
|
|
LTarget: "${device.attr.pci_bandwidth}",
|
|
RTarget: ".98 GiB/s",
|
|
},
|
|
{
|
|
Operand: "=",
|
|
LTarget: "${device.attr.cores_clock}",
|
|
RTarget: "800MHz",
|
|
},
|
|
},
|
|
},
|
|
},
|
|
},
|
|
{
|
|
Name: "does not meet first constraint",
|
|
Result: false,
|
|
NodeDevices: []*structs.NodeDeviceResource{nvidia},
|
|
RequestedDevices: []*structs.RequestedDevice{
|
|
{
|
|
Name: "nvidia/gpu",
|
|
Count: 1,
|
|
Constraints: []*structs.Constraint{
|
|
{
|
|
Operand: "=",
|
|
LTarget: "${device.model}",
|
|
RTarget: "2080ti",
|
|
},
|
|
{
|
|
Operand: ">",
|
|
LTarget: "${device.attr.memory}",
|
|
RTarget: "1320.5 MB",
|
|
},
|
|
{
|
|
Operand: "<=",
|
|
LTarget: "${device.attr.pci_bandwidth}",
|
|
RTarget: ".98 GiB/s",
|
|
},
|
|
{
|
|
Operand: "=",
|
|
LTarget: "${device.attr.cores_clock}",
|
|
RTarget: "800MHz",
|
|
},
|
|
},
|
|
},
|
|
},
|
|
},
|
|
{
|
|
Name: "does not meet second constraint",
|
|
Result: false,
|
|
NodeDevices: []*structs.NodeDeviceResource{nvidia},
|
|
RequestedDevices: []*structs.RequestedDevice{
|
|
{
|
|
Name: "nvidia/gpu",
|
|
Count: 1,
|
|
Constraints: []*structs.Constraint{
|
|
{
|
|
Operand: "=",
|
|
LTarget: "${device.model}",
|
|
RTarget: "1080ti",
|
|
},
|
|
{
|
|
Operand: "<",
|
|
LTarget: "${device.attr.memory}",
|
|
RTarget: "1320.5 MB",
|
|
},
|
|
{
|
|
Operand: "<=",
|
|
LTarget: "${device.attr.pci_bandwidth}",
|
|
RTarget: ".98 GiB/s",
|
|
},
|
|
{
|
|
Operand: "=",
|
|
LTarget: "${device.attr.cores_clock}",
|
|
RTarget: "800MHz",
|
|
},
|
|
},
|
|
},
|
|
},
|
|
},
|
|
{
|
|
Name: "does not meet ID constraint",
|
|
Result: false,
|
|
NodeDevices: []*structs.NodeDeviceResource{nvidia},
|
|
RequestedDevices: []*structs.RequestedDevice{
|
|
{
|
|
Name: "nvidia/gpu",
|
|
Count: 1,
|
|
Constraints: []*structs.Constraint{
|
|
{
|
|
Operand: "set_contains",
|
|
LTarget: "${device.ids}",
|
|
RTarget: "not_valid",
|
|
},
|
|
},
|
|
},
|
|
},
|
|
},
|
|
}
|
|
|
|
for _, c := range cases {
|
|
t.Run(c.Name, func(t *testing.T) {
|
|
_, ctx := testContext(t)
|
|
checker := NewDeviceChecker(ctx)
|
|
checker.SetTaskGroup(getTg(c.RequestedDevices...))
|
|
if act := checker.Feasible(getNode(c.NodeDevices...)); act != c.Result {
|
|
t.Fatalf("got %v; want %v", act, c.Result)
|
|
}
|
|
})
|
|
}
|
|
}
|
|
|
|
func TestCheckAttributeConstraint(t *testing.T) {
|
|
ci.Parallel(t)
|
|
|
|
type tcase struct {
|
|
op string
|
|
lVal, rVal *psstructs.Attribute
|
|
result bool
|
|
}
|
|
cases := []tcase{
|
|
{
|
|
op: "=",
|
|
lVal: psstructs.NewStringAttribute("foo"),
|
|
rVal: psstructs.NewStringAttribute("foo"),
|
|
result: true,
|
|
},
|
|
{
|
|
op: "=",
|
|
lVal: nil,
|
|
rVal: nil,
|
|
result: false,
|
|
},
|
|
{
|
|
op: "is",
|
|
lVal: psstructs.NewStringAttribute("foo"),
|
|
rVal: psstructs.NewStringAttribute("foo"),
|
|
result: true,
|
|
},
|
|
{
|
|
op: "==",
|
|
lVal: psstructs.NewStringAttribute("foo"),
|
|
rVal: psstructs.NewStringAttribute("foo"),
|
|
result: true,
|
|
},
|
|
{
|
|
op: "!=",
|
|
lVal: psstructs.NewStringAttribute("foo"),
|
|
rVal: psstructs.NewStringAttribute("foo"),
|
|
result: false,
|
|
},
|
|
{
|
|
op: "!=",
|
|
lVal: nil,
|
|
rVal: psstructs.NewStringAttribute("foo"),
|
|
result: true,
|
|
},
|
|
{
|
|
op: "!=",
|
|
lVal: psstructs.NewStringAttribute("foo"),
|
|
rVal: nil,
|
|
result: true,
|
|
},
|
|
{
|
|
op: "!=",
|
|
lVal: psstructs.NewStringAttribute("foo"),
|
|
rVal: psstructs.NewStringAttribute("bar"),
|
|
result: true,
|
|
},
|
|
{
|
|
op: "not",
|
|
lVal: psstructs.NewStringAttribute("foo"),
|
|
rVal: psstructs.NewStringAttribute("bar"),
|
|
result: true,
|
|
},
|
|
{
|
|
op: structs.ConstraintVersion,
|
|
lVal: psstructs.NewStringAttribute("1.2.3"),
|
|
rVal: psstructs.NewStringAttribute("~> 1.0"),
|
|
result: true,
|
|
},
|
|
{
|
|
op: structs.ConstraintRegex,
|
|
lVal: psstructs.NewStringAttribute("foobarbaz"),
|
|
rVal: psstructs.NewStringAttribute("[\\w]+"),
|
|
result: true,
|
|
},
|
|
{
|
|
op: "<",
|
|
lVal: psstructs.NewStringAttribute("foo"),
|
|
rVal: psstructs.NewStringAttribute("bar"),
|
|
result: false,
|
|
},
|
|
{
|
|
op: structs.ConstraintSetContains,
|
|
lVal: psstructs.NewStringAttribute("foo,bar,baz"),
|
|
rVal: psstructs.NewStringAttribute("foo, bar "),
|
|
result: true,
|
|
},
|
|
{
|
|
op: structs.ConstraintSetContainsAll,
|
|
lVal: psstructs.NewStringAttribute("foo,bar,baz"),
|
|
rVal: psstructs.NewStringAttribute("foo, bar "),
|
|
result: true,
|
|
},
|
|
{
|
|
op: structs.ConstraintSetContains,
|
|
lVal: psstructs.NewStringAttribute("foo,bar,baz"),
|
|
rVal: psstructs.NewStringAttribute("foo,bam"),
|
|
result: false,
|
|
},
|
|
{
|
|
op: structs.ConstraintSetContainsAny,
|
|
lVal: psstructs.NewStringAttribute("foo,bar,baz"),
|
|
rVal: psstructs.NewStringAttribute("foo,bam"),
|
|
result: true,
|
|
},
|
|
{
|
|
op: structs.ConstraintAttributeIsSet,
|
|
lVal: psstructs.NewStringAttribute("foo,bar,baz"),
|
|
result: true,
|
|
},
|
|
{
|
|
op: structs.ConstraintAttributeIsSet,
|
|
lVal: nil,
|
|
result: false,
|
|
},
|
|
{
|
|
op: structs.ConstraintAttributeIsNotSet,
|
|
lVal: psstructs.NewStringAttribute("foo,bar,baz"),
|
|
result: false,
|
|
},
|
|
{
|
|
op: structs.ConstraintAttributeIsNotSet,
|
|
lVal: nil,
|
|
result: true,
|
|
},
|
|
}
|
|
|
|
for _, tc := range cases {
|
|
_, ctx := testContext(t)
|
|
if res := checkAttributeConstraint(ctx, tc.op, tc.lVal, tc.rVal, tc.lVal != nil, tc.rVal != nil); res != tc.result {
|
|
t.Fatalf("TC: %#v, Result: %v", tc, res)
|
|
}
|
|
}
|
|
}
|