3678 lines
88 KiB
Go
3678 lines
88 KiB
Go
package structs
|
|
|
|
import (
|
|
"fmt"
|
|
"os"
|
|
"reflect"
|
|
"strings"
|
|
"testing"
|
|
"time"
|
|
|
|
"github.com/hashicorp/consul/api"
|
|
"github.com/hashicorp/go-multierror"
|
|
"github.com/hashicorp/nomad/helper/uuid"
|
|
"github.com/stretchr/testify/assert"
|
|
"github.com/stretchr/testify/require"
|
|
)
|
|
|
|
func TestJob_Validate(t *testing.T) {
|
|
j := &Job{}
|
|
err := j.Validate()
|
|
mErr := err.(*multierror.Error)
|
|
if !strings.Contains(mErr.Errors[0].Error(), "job region") {
|
|
t.Fatalf("err: %s", err)
|
|
}
|
|
if !strings.Contains(mErr.Errors[1].Error(), "job ID") {
|
|
t.Fatalf("err: %s", err)
|
|
}
|
|
if !strings.Contains(mErr.Errors[2].Error(), "job name") {
|
|
t.Fatalf("err: %s", err)
|
|
}
|
|
if !strings.Contains(mErr.Errors[3].Error(), "namespace") {
|
|
t.Fatalf("err: %s", err)
|
|
}
|
|
if !strings.Contains(mErr.Errors[4].Error(), "job type") {
|
|
t.Fatalf("err: %s", err)
|
|
}
|
|
if !strings.Contains(mErr.Errors[5].Error(), "priority") {
|
|
t.Fatalf("err: %s", err)
|
|
}
|
|
if !strings.Contains(mErr.Errors[6].Error(), "datacenters") {
|
|
t.Fatalf("err: %s", err)
|
|
}
|
|
if !strings.Contains(mErr.Errors[7].Error(), "task groups") {
|
|
t.Fatalf("err: %s", err)
|
|
}
|
|
|
|
j = &Job{
|
|
Type: "invalid-job-type",
|
|
}
|
|
err = j.Validate()
|
|
if expected := `Invalid job type: "invalid-job-type"`; !strings.Contains(err.Error(), expected) {
|
|
t.Errorf("expected %s but found: %v", expected, err)
|
|
}
|
|
|
|
j = &Job{
|
|
Type: JobTypeService,
|
|
Periodic: &PeriodicConfig{
|
|
Enabled: true,
|
|
},
|
|
}
|
|
err = j.Validate()
|
|
mErr = err.(*multierror.Error)
|
|
if !strings.Contains(mErr.Error(), "Periodic") {
|
|
t.Fatalf("err: %s", err)
|
|
}
|
|
|
|
j = &Job{
|
|
Region: "global",
|
|
ID: uuid.Generate(),
|
|
Namespace: "test",
|
|
Name: "my-job",
|
|
Type: JobTypeService,
|
|
Priority: 50,
|
|
Datacenters: []string{"dc1"},
|
|
TaskGroups: []*TaskGroup{
|
|
{
|
|
Name: "web",
|
|
RestartPolicy: &RestartPolicy{
|
|
Interval: 5 * time.Minute,
|
|
Delay: 10 * time.Second,
|
|
Attempts: 10,
|
|
},
|
|
},
|
|
{
|
|
Name: "web",
|
|
RestartPolicy: &RestartPolicy{
|
|
Interval: 5 * time.Minute,
|
|
Delay: 10 * time.Second,
|
|
Attempts: 10,
|
|
},
|
|
},
|
|
{
|
|
RestartPolicy: &RestartPolicy{
|
|
Interval: 5 * time.Minute,
|
|
Delay: 10 * time.Second,
|
|
Attempts: 10,
|
|
},
|
|
},
|
|
},
|
|
}
|
|
err = j.Validate()
|
|
mErr = err.(*multierror.Error)
|
|
if !strings.Contains(mErr.Errors[0].Error(), "2 redefines 'web' from group 1") {
|
|
t.Fatalf("err: %s", err)
|
|
}
|
|
if !strings.Contains(mErr.Errors[1].Error(), "group 3 missing name") {
|
|
t.Fatalf("err: %s", err)
|
|
}
|
|
if !strings.Contains(mErr.Errors[2].Error(), "Task group web validation failed") {
|
|
t.Fatalf("err: %s", err)
|
|
}
|
|
}
|
|
|
|
func TestJob_Warnings(t *testing.T) {
|
|
cases := []struct {
|
|
Name string
|
|
Job *Job
|
|
Expected []string
|
|
}{
|
|
{
|
|
Name: "Higher counts for update stanza",
|
|
Expected: []string{"max parallel count is greater"},
|
|
Job: &Job{
|
|
Type: JobTypeService,
|
|
TaskGroups: []*TaskGroup{
|
|
{
|
|
Name: "foo",
|
|
Count: 2,
|
|
Update: &UpdateStrategy{
|
|
MaxParallel: 10,
|
|
},
|
|
},
|
|
},
|
|
},
|
|
},
|
|
}
|
|
|
|
for _, c := range cases {
|
|
t.Run(c.Name, func(t *testing.T) {
|
|
warnings := c.Job.Warnings()
|
|
if warnings == nil {
|
|
if len(c.Expected) == 0 {
|
|
return
|
|
} else {
|
|
t.Fatal("Got no warnings when they were expected")
|
|
}
|
|
}
|
|
|
|
a := warnings.Error()
|
|
for _, e := range c.Expected {
|
|
if !strings.Contains(a, e) {
|
|
t.Fatalf("Got warnings %q; didn't contain %q", a, e)
|
|
}
|
|
}
|
|
})
|
|
}
|
|
}
|
|
|
|
func TestJob_SpecChanged(t *testing.T) {
|
|
// Get a base test job
|
|
base := testJob()
|
|
|
|
// Only modify the indexes/mutable state of the job
|
|
mutatedBase := base.Copy()
|
|
mutatedBase.Status = "foo"
|
|
mutatedBase.ModifyIndex = base.ModifyIndex + 100
|
|
|
|
// changed contains a spec change that should be detected
|
|
change := base.Copy()
|
|
change.Priority = 99
|
|
|
|
cases := []struct {
|
|
Name string
|
|
Original *Job
|
|
New *Job
|
|
Changed bool
|
|
}{
|
|
{
|
|
Name: "Same job except mutable indexes",
|
|
Changed: false,
|
|
Original: base,
|
|
New: mutatedBase,
|
|
},
|
|
{
|
|
Name: "Different",
|
|
Changed: true,
|
|
Original: base,
|
|
New: change,
|
|
},
|
|
}
|
|
|
|
for _, c := range cases {
|
|
t.Run(c.Name, func(t *testing.T) {
|
|
if actual := c.Original.SpecChanged(c.New); actual != c.Changed {
|
|
t.Fatalf("SpecChanged() returned %v; want %v", actual, c.Changed)
|
|
}
|
|
})
|
|
}
|
|
}
|
|
|
|
func testJob() *Job {
|
|
return &Job{
|
|
Region: "global",
|
|
ID: uuid.Generate(),
|
|
Namespace: "test",
|
|
Name: "my-job",
|
|
Type: JobTypeService,
|
|
Priority: 50,
|
|
AllAtOnce: false,
|
|
Datacenters: []string{"dc1"},
|
|
Constraints: []*Constraint{
|
|
{
|
|
LTarget: "$attr.kernel.name",
|
|
RTarget: "linux",
|
|
Operand: "=",
|
|
},
|
|
},
|
|
Periodic: &PeriodicConfig{
|
|
Enabled: false,
|
|
},
|
|
TaskGroups: []*TaskGroup{
|
|
{
|
|
Name: "web",
|
|
Count: 10,
|
|
EphemeralDisk: DefaultEphemeralDisk(),
|
|
RestartPolicy: &RestartPolicy{
|
|
Mode: RestartPolicyModeFail,
|
|
Attempts: 3,
|
|
Interval: 10 * time.Minute,
|
|
Delay: 1 * time.Minute,
|
|
},
|
|
ReschedulePolicy: &ReschedulePolicy{
|
|
Interval: 5 * time.Minute,
|
|
Attempts: 10,
|
|
Delay: 5 * time.Second,
|
|
DelayFunction: "constant",
|
|
},
|
|
Tasks: []*Task{
|
|
{
|
|
Name: "web",
|
|
Driver: "exec",
|
|
Config: map[string]interface{}{
|
|
"command": "/bin/date",
|
|
},
|
|
Env: map[string]string{
|
|
"FOO": "bar",
|
|
},
|
|
Artifacts: []*TaskArtifact{
|
|
{
|
|
GetterSource: "http://foo.com",
|
|
},
|
|
},
|
|
Services: []*Service{
|
|
{
|
|
Name: "${TASK}-frontend",
|
|
PortLabel: "http",
|
|
},
|
|
},
|
|
Resources: &Resources{
|
|
CPU: 500,
|
|
MemoryMB: 256,
|
|
Networks: []*NetworkResource{
|
|
{
|
|
MBits: 50,
|
|
DynamicPorts: []Port{{Label: "http"}},
|
|
},
|
|
},
|
|
},
|
|
LogConfig: &LogConfig{
|
|
MaxFiles: 10,
|
|
MaxFileSizeMB: 1,
|
|
},
|
|
},
|
|
},
|
|
Meta: map[string]string{
|
|
"elb_check_type": "http",
|
|
"elb_check_interval": "30s",
|
|
"elb_check_min": "3",
|
|
},
|
|
},
|
|
},
|
|
Meta: map[string]string{
|
|
"owner": "armon",
|
|
},
|
|
}
|
|
}
|
|
|
|
func TestJob_Copy(t *testing.T) {
|
|
j := testJob()
|
|
c := j.Copy()
|
|
if !reflect.DeepEqual(j, c) {
|
|
t.Fatalf("Copy() returned an unequal Job; got %#v; want %#v", c, j)
|
|
}
|
|
}
|
|
|
|
func TestJob_IsPeriodic(t *testing.T) {
|
|
j := &Job{
|
|
Type: JobTypeService,
|
|
Periodic: &PeriodicConfig{
|
|
Enabled: true,
|
|
},
|
|
}
|
|
if !j.IsPeriodic() {
|
|
t.Fatalf("IsPeriodic() returned false on periodic job")
|
|
}
|
|
|
|
j = &Job{
|
|
Type: JobTypeService,
|
|
}
|
|
if j.IsPeriodic() {
|
|
t.Fatalf("IsPeriodic() returned true on non-periodic job")
|
|
}
|
|
}
|
|
|
|
func TestJob_IsPeriodicActive(t *testing.T) {
|
|
cases := []struct {
|
|
job *Job
|
|
active bool
|
|
}{
|
|
{
|
|
job: &Job{
|
|
Type: JobTypeService,
|
|
Periodic: &PeriodicConfig{
|
|
Enabled: true,
|
|
},
|
|
},
|
|
active: true,
|
|
},
|
|
{
|
|
job: &Job{
|
|
Type: JobTypeService,
|
|
Periodic: &PeriodicConfig{
|
|
Enabled: false,
|
|
},
|
|
},
|
|
active: false,
|
|
},
|
|
{
|
|
job: &Job{
|
|
Type: JobTypeService,
|
|
Periodic: &PeriodicConfig{
|
|
Enabled: true,
|
|
},
|
|
Stop: true,
|
|
},
|
|
active: false,
|
|
},
|
|
{
|
|
job: &Job{
|
|
Type: JobTypeService,
|
|
Periodic: &PeriodicConfig{
|
|
Enabled: false,
|
|
},
|
|
ParameterizedJob: &ParameterizedJobConfig{},
|
|
},
|
|
active: false,
|
|
},
|
|
}
|
|
|
|
for i, c := range cases {
|
|
if act := c.job.IsPeriodicActive(); act != c.active {
|
|
t.Fatalf("case %d failed: got %v; want %v", i, act, c.active)
|
|
}
|
|
}
|
|
}
|
|
|
|
func TestJob_SystemJob_Validate(t *testing.T) {
|
|
j := testJob()
|
|
j.Type = JobTypeSystem
|
|
j.Canonicalize()
|
|
|
|
err := j.Validate()
|
|
if err == nil || !strings.Contains(err.Error(), "exceed") {
|
|
t.Fatalf("expect error due to count")
|
|
}
|
|
|
|
j.TaskGroups[0].Count = 0
|
|
if err := j.Validate(); err != nil {
|
|
t.Fatalf("unexpected err: %v", err)
|
|
}
|
|
|
|
j.TaskGroups[0].Count = 1
|
|
if err := j.Validate(); err != nil {
|
|
t.Fatalf("unexpected err: %v", err)
|
|
}
|
|
}
|
|
|
|
func TestJob_VaultPolicies(t *testing.T) {
|
|
j0 := &Job{}
|
|
e0 := make(map[string]map[string]*Vault, 0)
|
|
|
|
vj1 := &Vault{
|
|
Policies: []string{
|
|
"p1",
|
|
"p2",
|
|
},
|
|
}
|
|
vj2 := &Vault{
|
|
Policies: []string{
|
|
"p3",
|
|
"p4",
|
|
},
|
|
}
|
|
vj3 := &Vault{
|
|
Policies: []string{
|
|
"p5",
|
|
},
|
|
}
|
|
j1 := &Job{
|
|
TaskGroups: []*TaskGroup{
|
|
{
|
|
Name: "foo",
|
|
Tasks: []*Task{
|
|
{
|
|
Name: "t1",
|
|
},
|
|
{
|
|
Name: "t2",
|
|
Vault: vj1,
|
|
},
|
|
},
|
|
},
|
|
{
|
|
Name: "bar",
|
|
Tasks: []*Task{
|
|
{
|
|
Name: "t3",
|
|
Vault: vj2,
|
|
},
|
|
{
|
|
Name: "t4",
|
|
Vault: vj3,
|
|
},
|
|
},
|
|
},
|
|
},
|
|
}
|
|
|
|
e1 := map[string]map[string]*Vault{
|
|
"foo": {
|
|
"t2": vj1,
|
|
},
|
|
"bar": {
|
|
"t3": vj2,
|
|
"t4": vj3,
|
|
},
|
|
}
|
|
|
|
cases := []struct {
|
|
Job *Job
|
|
Expected map[string]map[string]*Vault
|
|
}{
|
|
{
|
|
Job: j0,
|
|
Expected: e0,
|
|
},
|
|
{
|
|
Job: j1,
|
|
Expected: e1,
|
|
},
|
|
}
|
|
|
|
for i, c := range cases {
|
|
got := c.Job.VaultPolicies()
|
|
if !reflect.DeepEqual(got, c.Expected) {
|
|
t.Fatalf("case %d: got %#v; want %#v", i+1, got, c.Expected)
|
|
}
|
|
}
|
|
}
|
|
|
|
func TestJob_RequiredSignals(t *testing.T) {
|
|
j0 := &Job{}
|
|
e0 := make(map[string]map[string][]string, 0)
|
|
|
|
vj1 := &Vault{
|
|
Policies: []string{"p1"},
|
|
ChangeMode: VaultChangeModeNoop,
|
|
}
|
|
vj2 := &Vault{
|
|
Policies: []string{"p1"},
|
|
ChangeMode: VaultChangeModeSignal,
|
|
ChangeSignal: "SIGUSR1",
|
|
}
|
|
tj1 := &Template{
|
|
SourcePath: "foo",
|
|
DestPath: "bar",
|
|
ChangeMode: TemplateChangeModeNoop,
|
|
}
|
|
tj2 := &Template{
|
|
SourcePath: "foo",
|
|
DestPath: "bar",
|
|
ChangeMode: TemplateChangeModeSignal,
|
|
ChangeSignal: "SIGUSR2",
|
|
}
|
|
j1 := &Job{
|
|
TaskGroups: []*TaskGroup{
|
|
{
|
|
Name: "foo",
|
|
Tasks: []*Task{
|
|
{
|
|
Name: "t1",
|
|
},
|
|
{
|
|
Name: "t2",
|
|
Vault: vj2,
|
|
Templates: []*Template{tj2},
|
|
},
|
|
},
|
|
},
|
|
{
|
|
Name: "bar",
|
|
Tasks: []*Task{
|
|
{
|
|
Name: "t3",
|
|
Vault: vj1,
|
|
Templates: []*Template{tj1},
|
|
},
|
|
{
|
|
Name: "t4",
|
|
Vault: vj2,
|
|
},
|
|
},
|
|
},
|
|
},
|
|
}
|
|
|
|
e1 := map[string]map[string][]string{
|
|
"foo": {
|
|
"t2": {"SIGUSR1", "SIGUSR2"},
|
|
},
|
|
"bar": {
|
|
"t4": {"SIGUSR1"},
|
|
},
|
|
}
|
|
|
|
j2 := &Job{
|
|
TaskGroups: []*TaskGroup{
|
|
{
|
|
Name: "foo",
|
|
Tasks: []*Task{
|
|
{
|
|
Name: "t1",
|
|
KillSignal: "SIGQUIT",
|
|
},
|
|
},
|
|
},
|
|
},
|
|
}
|
|
|
|
e2 := map[string]map[string][]string{
|
|
"foo": {
|
|
"t1": {"SIGQUIT"},
|
|
},
|
|
}
|
|
|
|
cases := []struct {
|
|
Job *Job
|
|
Expected map[string]map[string][]string
|
|
}{
|
|
{
|
|
Job: j0,
|
|
Expected: e0,
|
|
},
|
|
{
|
|
Job: j1,
|
|
Expected: e1,
|
|
},
|
|
{
|
|
Job: j2,
|
|
Expected: e2,
|
|
},
|
|
}
|
|
|
|
for i, c := range cases {
|
|
got := c.Job.RequiredSignals()
|
|
if !reflect.DeepEqual(got, c.Expected) {
|
|
t.Fatalf("case %d: got %#v; want %#v", i+1, got, c.Expected)
|
|
}
|
|
}
|
|
}
|
|
|
|
func TestTaskGroup_Validate(t *testing.T) {
|
|
j := testJob()
|
|
tg := &TaskGroup{
|
|
Count: -1,
|
|
RestartPolicy: &RestartPolicy{
|
|
Interval: 5 * time.Minute,
|
|
Delay: 10 * time.Second,
|
|
Attempts: 10,
|
|
Mode: RestartPolicyModeDelay,
|
|
},
|
|
ReschedulePolicy: &ReschedulePolicy{
|
|
Interval: 5 * time.Minute,
|
|
Attempts: 5,
|
|
Delay: 5 * time.Second,
|
|
},
|
|
}
|
|
err := tg.Validate(j)
|
|
mErr := err.(*multierror.Error)
|
|
if !strings.Contains(mErr.Errors[0].Error(), "group name") {
|
|
t.Fatalf("err: %s", err)
|
|
}
|
|
if !strings.Contains(mErr.Errors[1].Error(), "count can't be negative") {
|
|
t.Fatalf("err: %s", err)
|
|
}
|
|
if !strings.Contains(mErr.Errors[2].Error(), "Missing tasks") {
|
|
t.Fatalf("err: %s", err)
|
|
}
|
|
|
|
tg = &TaskGroup{
|
|
Tasks: []*Task{
|
|
{
|
|
Name: "task-a",
|
|
Resources: &Resources{
|
|
Networks: []*NetworkResource{
|
|
{
|
|
ReservedPorts: []Port{{Label: "foo", Value: 123}},
|
|
},
|
|
},
|
|
},
|
|
},
|
|
{
|
|
Name: "task-b",
|
|
Resources: &Resources{
|
|
Networks: []*NetworkResource{
|
|
{
|
|
ReservedPorts: []Port{{Label: "foo", Value: 123}},
|
|
},
|
|
},
|
|
},
|
|
},
|
|
},
|
|
}
|
|
err = tg.Validate(&Job{})
|
|
expected := `Static port 123 already reserved by task-a:foo`
|
|
if !strings.Contains(err.Error(), expected) {
|
|
t.Errorf("expected %s but found: %v", expected, err)
|
|
}
|
|
|
|
tg = &TaskGroup{
|
|
Tasks: []*Task{
|
|
{
|
|
Name: "task-a",
|
|
Resources: &Resources{
|
|
Networks: []*NetworkResource{
|
|
{
|
|
ReservedPorts: []Port{
|
|
{Label: "foo", Value: 123},
|
|
{Label: "bar", Value: 123},
|
|
},
|
|
},
|
|
},
|
|
},
|
|
},
|
|
},
|
|
}
|
|
err = tg.Validate(&Job{})
|
|
expected = `Static port 123 already reserved by task-a:foo`
|
|
if !strings.Contains(err.Error(), expected) {
|
|
t.Errorf("expected %s but found: %v", expected, err)
|
|
}
|
|
|
|
tg = &TaskGroup{
|
|
Name: "web",
|
|
Count: 1,
|
|
Tasks: []*Task{
|
|
{Name: "web", Leader: true},
|
|
{Name: "web", Leader: true},
|
|
{},
|
|
},
|
|
RestartPolicy: &RestartPolicy{
|
|
Interval: 5 * time.Minute,
|
|
Delay: 10 * time.Second,
|
|
Attempts: 10,
|
|
Mode: RestartPolicyModeDelay,
|
|
},
|
|
ReschedulePolicy: &ReschedulePolicy{
|
|
Interval: 5 * time.Minute,
|
|
Attempts: 10,
|
|
Delay: 5 * time.Second,
|
|
DelayFunction: "constant",
|
|
},
|
|
}
|
|
|
|
err = tg.Validate(j)
|
|
mErr = err.(*multierror.Error)
|
|
if !strings.Contains(mErr.Errors[0].Error(), "should have an ephemeral disk object") {
|
|
t.Fatalf("err: %s", err)
|
|
}
|
|
if !strings.Contains(mErr.Errors[1].Error(), "2 redefines 'web' from task 1") {
|
|
t.Fatalf("err: %s", err)
|
|
}
|
|
if !strings.Contains(mErr.Errors[2].Error(), "Task 3 missing name") {
|
|
t.Fatalf("err: %s", err)
|
|
}
|
|
if !strings.Contains(mErr.Errors[3].Error(), "Only one task may be marked as leader") {
|
|
t.Fatalf("err: %s", err)
|
|
}
|
|
if !strings.Contains(mErr.Errors[4].Error(), "Task web validation failed") {
|
|
t.Fatalf("err: %s", err)
|
|
}
|
|
|
|
tg = &TaskGroup{
|
|
Name: "web",
|
|
Count: 1,
|
|
Tasks: []*Task{
|
|
{Name: "web", Leader: true},
|
|
},
|
|
Update: DefaultUpdateStrategy.Copy(),
|
|
}
|
|
j.Type = JobTypeBatch
|
|
err = tg.Validate(j)
|
|
if !strings.Contains(err.Error(), "does not allow update block") {
|
|
t.Fatalf("err: %s", err)
|
|
}
|
|
|
|
tg = &TaskGroup{
|
|
Count: -1,
|
|
RestartPolicy: &RestartPolicy{
|
|
Interval: 5 * time.Minute,
|
|
Delay: 10 * time.Second,
|
|
Attempts: 10,
|
|
Mode: RestartPolicyModeDelay,
|
|
},
|
|
ReschedulePolicy: &ReschedulePolicy{
|
|
Interval: 5 * time.Minute,
|
|
Attempts: 5,
|
|
Delay: 5 * time.Second,
|
|
},
|
|
}
|
|
j.Type = JobTypeSystem
|
|
err = tg.Validate(j)
|
|
if !strings.Contains(err.Error(), "System jobs should not have a reschedule policy") {
|
|
t.Fatalf("err: %s", err)
|
|
}
|
|
}
|
|
|
|
func TestTask_Validate(t *testing.T) {
|
|
task := &Task{}
|
|
ephemeralDisk := DefaultEphemeralDisk()
|
|
err := task.Validate(ephemeralDisk)
|
|
mErr := err.(*multierror.Error)
|
|
if !strings.Contains(mErr.Errors[0].Error(), "task name") {
|
|
t.Fatalf("err: %s", err)
|
|
}
|
|
if !strings.Contains(mErr.Errors[1].Error(), "task driver") {
|
|
t.Fatalf("err: %s", err)
|
|
}
|
|
if !strings.Contains(mErr.Errors[2].Error(), "task resources") {
|
|
t.Fatalf("err: %s", err)
|
|
}
|
|
|
|
task = &Task{Name: "web/foo"}
|
|
err = task.Validate(ephemeralDisk)
|
|
mErr = err.(*multierror.Error)
|
|
if !strings.Contains(mErr.Errors[0].Error(), "slashes") {
|
|
t.Fatalf("err: %s", err)
|
|
}
|
|
|
|
task = &Task{
|
|
Name: "web",
|
|
Driver: "docker",
|
|
Resources: &Resources{
|
|
CPU: 100,
|
|
MemoryMB: 100,
|
|
IOPS: 10,
|
|
},
|
|
LogConfig: DefaultLogConfig(),
|
|
}
|
|
ephemeralDisk.SizeMB = 200
|
|
err = task.Validate(ephemeralDisk)
|
|
if err != nil {
|
|
t.Fatalf("err: %s", err)
|
|
}
|
|
|
|
task.Constraints = append(task.Constraints,
|
|
&Constraint{
|
|
Operand: ConstraintDistinctHosts,
|
|
},
|
|
&Constraint{
|
|
Operand: ConstraintDistinctProperty,
|
|
LTarget: "${meta.rack}",
|
|
})
|
|
|
|
err = task.Validate(ephemeralDisk)
|
|
mErr = err.(*multierror.Error)
|
|
if !strings.Contains(mErr.Errors[0].Error(), "task level: distinct_hosts") {
|
|
t.Fatalf("err: %s", err)
|
|
}
|
|
if !strings.Contains(mErr.Errors[1].Error(), "task level: distinct_property") {
|
|
t.Fatalf("err: %s", err)
|
|
}
|
|
}
|
|
|
|
func TestTask_Validate_Services(t *testing.T) {
|
|
s1 := &Service{
|
|
Name: "service-name",
|
|
PortLabel: "bar",
|
|
Checks: []*ServiceCheck{
|
|
{
|
|
Name: "check-name",
|
|
Type: ServiceCheckTCP,
|
|
Interval: 0 * time.Second,
|
|
},
|
|
{
|
|
Name: "check-name",
|
|
Type: ServiceCheckTCP,
|
|
Timeout: 2 * time.Second,
|
|
},
|
|
{
|
|
Name: "check-name",
|
|
Type: ServiceCheckTCP,
|
|
Interval: 1 * time.Second,
|
|
},
|
|
},
|
|
}
|
|
|
|
s2 := &Service{
|
|
Name: "service-name",
|
|
PortLabel: "bar",
|
|
}
|
|
|
|
s3 := &Service{
|
|
Name: "service-A",
|
|
PortLabel: "a",
|
|
}
|
|
s4 := &Service{
|
|
Name: "service-A",
|
|
PortLabel: "b",
|
|
}
|
|
|
|
ephemeralDisk := DefaultEphemeralDisk()
|
|
ephemeralDisk.SizeMB = 200
|
|
task := &Task{
|
|
Name: "web",
|
|
Driver: "docker",
|
|
Resources: &Resources{
|
|
CPU: 100,
|
|
MemoryMB: 100,
|
|
IOPS: 10,
|
|
},
|
|
Services: []*Service{s1, s2},
|
|
}
|
|
|
|
task1 := &Task{
|
|
Name: "web",
|
|
Driver: "docker",
|
|
Resources: DefaultResources(),
|
|
Services: []*Service{s3, s4},
|
|
LogConfig: DefaultLogConfig(),
|
|
}
|
|
task1.Resources.Networks = []*NetworkResource{
|
|
{
|
|
MBits: 10,
|
|
DynamicPorts: []Port{
|
|
{
|
|
Label: "a",
|
|
Value: 1000,
|
|
},
|
|
{
|
|
Label: "b",
|
|
Value: 2000,
|
|
},
|
|
},
|
|
},
|
|
}
|
|
|
|
err := task.Validate(ephemeralDisk)
|
|
if err == nil {
|
|
t.Fatal("expected an error")
|
|
}
|
|
|
|
if !strings.Contains(err.Error(), "service \"service-name\" is duplicate") {
|
|
t.Fatalf("err: %v", err)
|
|
}
|
|
|
|
if !strings.Contains(err.Error(), "check \"check-name\" is duplicate") {
|
|
t.Fatalf("err: %v", err)
|
|
}
|
|
|
|
if !strings.Contains(err.Error(), "missing required value interval") {
|
|
t.Fatalf("err: %v", err)
|
|
}
|
|
|
|
if !strings.Contains(err.Error(), "cannot be less than") {
|
|
t.Fatalf("err: %v", err)
|
|
}
|
|
|
|
if err = task1.Validate(ephemeralDisk); err != nil {
|
|
t.Fatalf("err : %v", err)
|
|
}
|
|
}
|
|
|
|
func TestTask_Validate_Service_AddressMode_Ok(t *testing.T) {
|
|
ephemeralDisk := DefaultEphemeralDisk()
|
|
getTask := func(s *Service) *Task {
|
|
task := &Task{
|
|
Name: "web",
|
|
Driver: "docker",
|
|
Resources: DefaultResources(),
|
|
Services: []*Service{s},
|
|
LogConfig: DefaultLogConfig(),
|
|
}
|
|
task.Resources.Networks = []*NetworkResource{
|
|
{
|
|
MBits: 10,
|
|
DynamicPorts: []Port{
|
|
{
|
|
Label: "http",
|
|
Value: 80,
|
|
},
|
|
},
|
|
},
|
|
}
|
|
return task
|
|
}
|
|
|
|
cases := []*Service{
|
|
{
|
|
// https://github.com/hashicorp/nomad/issues/3681#issuecomment-357274177
|
|
Name: "DriverModeWithLabel",
|
|
PortLabel: "http",
|
|
AddressMode: AddressModeDriver,
|
|
},
|
|
{
|
|
Name: "DriverModeWithPort",
|
|
PortLabel: "80",
|
|
AddressMode: AddressModeDriver,
|
|
},
|
|
{
|
|
Name: "HostModeWithLabel",
|
|
PortLabel: "http",
|
|
AddressMode: AddressModeHost,
|
|
},
|
|
{
|
|
Name: "HostModeWithoutLabel",
|
|
AddressMode: AddressModeHost,
|
|
},
|
|
{
|
|
Name: "DriverModeWithoutLabel",
|
|
AddressMode: AddressModeDriver,
|
|
},
|
|
}
|
|
|
|
for _, service := range cases {
|
|
task := getTask(service)
|
|
t.Run(service.Name, func(t *testing.T) {
|
|
if err := task.Validate(ephemeralDisk); err != nil {
|
|
t.Fatalf("unexpected err: %v", err)
|
|
}
|
|
})
|
|
}
|
|
}
|
|
|
|
func TestTask_Validate_Service_AddressMode_Bad(t *testing.T) {
|
|
ephemeralDisk := DefaultEphemeralDisk()
|
|
getTask := func(s *Service) *Task {
|
|
task := &Task{
|
|
Name: "web",
|
|
Driver: "docker",
|
|
Resources: DefaultResources(),
|
|
Services: []*Service{s},
|
|
LogConfig: DefaultLogConfig(),
|
|
}
|
|
task.Resources.Networks = []*NetworkResource{
|
|
{
|
|
MBits: 10,
|
|
DynamicPorts: []Port{
|
|
{
|
|
Label: "http",
|
|
Value: 80,
|
|
},
|
|
},
|
|
},
|
|
}
|
|
return task
|
|
}
|
|
|
|
cases := []*Service{
|
|
{
|
|
// https://github.com/hashicorp/nomad/issues/3681#issuecomment-357274177
|
|
Name: "DriverModeWithLabel",
|
|
PortLabel: "asdf",
|
|
AddressMode: AddressModeDriver,
|
|
},
|
|
{
|
|
Name: "HostModeWithLabel",
|
|
PortLabel: "asdf",
|
|
AddressMode: AddressModeHost,
|
|
},
|
|
{
|
|
Name: "HostModeWithPort",
|
|
PortLabel: "80",
|
|
AddressMode: AddressModeHost,
|
|
},
|
|
}
|
|
|
|
for _, service := range cases {
|
|
task := getTask(service)
|
|
t.Run(service.Name, func(t *testing.T) {
|
|
err := task.Validate(ephemeralDisk)
|
|
if err == nil {
|
|
t.Fatalf("expected an error")
|
|
}
|
|
//t.Logf("err: %v", err)
|
|
})
|
|
}
|
|
}
|
|
|
|
func TestTask_Validate_Service_Check(t *testing.T) {
|
|
|
|
invalidCheck := ServiceCheck{
|
|
Name: "check-name",
|
|
Command: "/bin/true",
|
|
Type: ServiceCheckScript,
|
|
Interval: 10 * time.Second,
|
|
}
|
|
|
|
err := invalidCheck.validate()
|
|
if err == nil || !strings.Contains(err.Error(), "Timeout cannot be less") {
|
|
t.Fatalf("expected a timeout validation error but received: %q", err)
|
|
}
|
|
|
|
check1 := ServiceCheck{
|
|
Name: "check-name",
|
|
Type: ServiceCheckTCP,
|
|
Interval: 10 * time.Second,
|
|
Timeout: 2 * time.Second,
|
|
}
|
|
|
|
if err := check1.validate(); err != nil {
|
|
t.Fatalf("err: %v", err)
|
|
}
|
|
|
|
check1.InitialStatus = "foo"
|
|
err = check1.validate()
|
|
if err == nil {
|
|
t.Fatal("Expected an error")
|
|
}
|
|
|
|
if !strings.Contains(err.Error(), "invalid initial check state (foo)") {
|
|
t.Fatalf("err: %v", err)
|
|
}
|
|
|
|
check1.InitialStatus = api.HealthCritical
|
|
err = check1.validate()
|
|
if err != nil {
|
|
t.Fatalf("err: %v", err)
|
|
}
|
|
|
|
check1.InitialStatus = api.HealthPassing
|
|
err = check1.validate()
|
|
if err != nil {
|
|
t.Fatalf("err: %v", err)
|
|
}
|
|
|
|
check1.InitialStatus = ""
|
|
err = check1.validate()
|
|
if err != nil {
|
|
t.Fatalf("err: %v", err)
|
|
}
|
|
|
|
check2 := ServiceCheck{
|
|
Name: "check-name-2",
|
|
Type: ServiceCheckHTTP,
|
|
Interval: 10 * time.Second,
|
|
Timeout: 2 * time.Second,
|
|
Path: "/foo/bar",
|
|
}
|
|
|
|
err = check2.validate()
|
|
if err != nil {
|
|
t.Fatalf("err: %v", err)
|
|
}
|
|
|
|
check2.Path = ""
|
|
err = check2.validate()
|
|
if err == nil {
|
|
t.Fatal("Expected an error")
|
|
}
|
|
if !strings.Contains(err.Error(), "valid http path") {
|
|
t.Fatalf("err: %v", err)
|
|
}
|
|
|
|
check2.Path = "http://www.example.com"
|
|
err = check2.validate()
|
|
if err == nil {
|
|
t.Fatal("Expected an error")
|
|
}
|
|
if !strings.Contains(err.Error(), "relative http path") {
|
|
t.Fatalf("err: %v", err)
|
|
}
|
|
}
|
|
|
|
// TestTask_Validate_Service_Check_AddressMode asserts that checks do not
|
|
// inherit address mode but do inherit ports.
|
|
func TestTask_Validate_Service_Check_AddressMode(t *testing.T) {
|
|
getTask := func(s *Service) *Task {
|
|
return &Task{
|
|
Resources: &Resources{
|
|
Networks: []*NetworkResource{
|
|
{
|
|
DynamicPorts: []Port{
|
|
{
|
|
Label: "http",
|
|
Value: 9999,
|
|
},
|
|
},
|
|
},
|
|
},
|
|
},
|
|
Services: []*Service{s},
|
|
}
|
|
}
|
|
|
|
cases := []struct {
|
|
Service *Service
|
|
ErrContains string
|
|
}{
|
|
{
|
|
Service: &Service{
|
|
Name: "invalid-driver",
|
|
PortLabel: "80",
|
|
AddressMode: "host",
|
|
},
|
|
ErrContains: `port label "80" referenced`,
|
|
},
|
|
{
|
|
Service: &Service{
|
|
Name: "http-driver-fail-1",
|
|
PortLabel: "80",
|
|
AddressMode: "driver",
|
|
Checks: []*ServiceCheck{
|
|
{
|
|
Name: "invalid-check-1",
|
|
Type: "tcp",
|
|
Interval: time.Second,
|
|
Timeout: time.Second,
|
|
},
|
|
},
|
|
},
|
|
ErrContains: `check "invalid-check-1" cannot use a numeric port`,
|
|
},
|
|
{
|
|
Service: &Service{
|
|
Name: "http-driver-fail-2",
|
|
PortLabel: "80",
|
|
AddressMode: "driver",
|
|
Checks: []*ServiceCheck{
|
|
{
|
|
Name: "invalid-check-2",
|
|
Type: "tcp",
|
|
PortLabel: "80",
|
|
Interval: time.Second,
|
|
Timeout: time.Second,
|
|
},
|
|
},
|
|
},
|
|
ErrContains: `check "invalid-check-2" cannot use a numeric port`,
|
|
},
|
|
{
|
|
Service: &Service{
|
|
Name: "http-driver-fail-3",
|
|
PortLabel: "80",
|
|
AddressMode: "driver",
|
|
Checks: []*ServiceCheck{
|
|
{
|
|
Name: "invalid-check-3",
|
|
Type: "tcp",
|
|
PortLabel: "missing-port-label",
|
|
Interval: time.Second,
|
|
Timeout: time.Second,
|
|
},
|
|
},
|
|
},
|
|
ErrContains: `port label "missing-port-label" referenced`,
|
|
},
|
|
{
|
|
Service: &Service{
|
|
Name: "http-driver-passes",
|
|
PortLabel: "80",
|
|
AddressMode: "driver",
|
|
Checks: []*ServiceCheck{
|
|
{
|
|
Name: "valid-script-check",
|
|
Type: "script",
|
|
Command: "ok",
|
|
Interval: time.Second,
|
|
Timeout: time.Second,
|
|
},
|
|
{
|
|
Name: "valid-host-check",
|
|
Type: "tcp",
|
|
PortLabel: "http",
|
|
Interval: time.Second,
|
|
Timeout: time.Second,
|
|
},
|
|
{
|
|
Name: "valid-driver-check",
|
|
Type: "tcp",
|
|
AddressMode: "driver",
|
|
Interval: time.Second,
|
|
Timeout: time.Second,
|
|
},
|
|
},
|
|
},
|
|
},
|
|
{
|
|
Service: &Service{
|
|
Name: "empty-address-3673-passes-1",
|
|
Checks: []*ServiceCheck{
|
|
{
|
|
Name: "valid-port-label",
|
|
Type: "tcp",
|
|
PortLabel: "http",
|
|
Interval: time.Second,
|
|
Timeout: time.Second,
|
|
},
|
|
{
|
|
Name: "empty-is-ok",
|
|
Type: "script",
|
|
Command: "ok",
|
|
Interval: time.Second,
|
|
Timeout: time.Second,
|
|
},
|
|
},
|
|
},
|
|
},
|
|
{
|
|
Service: &Service{
|
|
Name: "empty-address-3673-passes-2",
|
|
},
|
|
},
|
|
{
|
|
Service: &Service{
|
|
Name: "empty-address-3673-fails",
|
|
Checks: []*ServiceCheck{
|
|
{
|
|
Name: "empty-is-not-ok",
|
|
Type: "tcp",
|
|
Interval: time.Second,
|
|
Timeout: time.Second,
|
|
},
|
|
},
|
|
},
|
|
ErrContains: `invalid: check requires a port but neither check nor service`,
|
|
},
|
|
}
|
|
|
|
for _, tc := range cases {
|
|
tc := tc
|
|
task := getTask(tc.Service)
|
|
t.Run(tc.Service.Name, func(t *testing.T) {
|
|
err := validateServices(task)
|
|
if err == nil && tc.ErrContains == "" {
|
|
// Ok!
|
|
return
|
|
}
|
|
if err == nil {
|
|
t.Fatalf("no error returned. expected: %s", tc.ErrContains)
|
|
}
|
|
if !strings.Contains(err.Error(), tc.ErrContains) {
|
|
t.Fatalf("expected %q but found: %v", tc.ErrContains, err)
|
|
}
|
|
})
|
|
}
|
|
}
|
|
|
|
func TestTask_Validate_Service_Check_CheckRestart(t *testing.T) {
|
|
invalidCheckRestart := &CheckRestart{
|
|
Limit: -1,
|
|
Grace: -1,
|
|
}
|
|
|
|
err := invalidCheckRestart.Validate()
|
|
assert.NotNil(t, err, "invalidateCheckRestart.Validate()")
|
|
assert.Len(t, err.(*multierror.Error).Errors, 2)
|
|
|
|
validCheckRestart := &CheckRestart{}
|
|
assert.Nil(t, validCheckRestart.Validate())
|
|
|
|
validCheckRestart.Limit = 1
|
|
validCheckRestart.Grace = 1
|
|
assert.Nil(t, validCheckRestart.Validate())
|
|
}
|
|
|
|
func TestTask_Validate_LogConfig(t *testing.T) {
|
|
task := &Task{
|
|
LogConfig: DefaultLogConfig(),
|
|
}
|
|
ephemeralDisk := &EphemeralDisk{
|
|
SizeMB: 1,
|
|
}
|
|
|
|
err := task.Validate(ephemeralDisk)
|
|
mErr := err.(*multierror.Error)
|
|
if !strings.Contains(mErr.Errors[3].Error(), "log storage") {
|
|
t.Fatalf("err: %s", err)
|
|
}
|
|
}
|
|
|
|
func TestTask_Validate_Template(t *testing.T) {
|
|
|
|
bad := &Template{}
|
|
task := &Task{
|
|
Templates: []*Template{bad},
|
|
}
|
|
ephemeralDisk := &EphemeralDisk{
|
|
SizeMB: 1,
|
|
}
|
|
|
|
err := task.Validate(ephemeralDisk)
|
|
if !strings.Contains(err.Error(), "Template 1 validation failed") {
|
|
t.Fatalf("err: %s", err)
|
|
}
|
|
|
|
// Have two templates that share the same destination
|
|
good := &Template{
|
|
SourcePath: "foo",
|
|
DestPath: "local/foo",
|
|
ChangeMode: "noop",
|
|
}
|
|
|
|
task.Templates = []*Template{good, good}
|
|
err = task.Validate(ephemeralDisk)
|
|
if !strings.Contains(err.Error(), "same destination as") {
|
|
t.Fatalf("err: %s", err)
|
|
}
|
|
|
|
// Env templates can't use signals
|
|
task.Templates = []*Template{
|
|
{
|
|
Envvars: true,
|
|
ChangeMode: "signal",
|
|
},
|
|
}
|
|
|
|
err = task.Validate(ephemeralDisk)
|
|
if err == nil {
|
|
t.Fatalf("expected error from Template.Validate")
|
|
}
|
|
if expected := "cannot use signals"; !strings.Contains(err.Error(), expected) {
|
|
t.Errorf("expected to find %q but found %v", expected, err)
|
|
}
|
|
}
|
|
|
|
func TestTemplate_Validate(t *testing.T) {
|
|
cases := []struct {
|
|
Tmpl *Template
|
|
Fail bool
|
|
ContainsErrs []string
|
|
}{
|
|
{
|
|
Tmpl: &Template{},
|
|
Fail: true,
|
|
ContainsErrs: []string{
|
|
"specify a source path",
|
|
"specify a destination",
|
|
TemplateChangeModeInvalidError.Error(),
|
|
},
|
|
},
|
|
{
|
|
Tmpl: &Template{
|
|
Splay: -100,
|
|
},
|
|
Fail: true,
|
|
ContainsErrs: []string{
|
|
"positive splay",
|
|
},
|
|
},
|
|
{
|
|
Tmpl: &Template{
|
|
ChangeMode: "foo",
|
|
},
|
|
Fail: true,
|
|
ContainsErrs: []string{
|
|
TemplateChangeModeInvalidError.Error(),
|
|
},
|
|
},
|
|
{
|
|
Tmpl: &Template{
|
|
ChangeMode: "signal",
|
|
},
|
|
Fail: true,
|
|
ContainsErrs: []string{
|
|
"specify signal value",
|
|
},
|
|
},
|
|
{
|
|
Tmpl: &Template{
|
|
SourcePath: "foo",
|
|
DestPath: "../../root",
|
|
ChangeMode: "noop",
|
|
},
|
|
Fail: true,
|
|
ContainsErrs: []string{
|
|
"destination escapes",
|
|
},
|
|
},
|
|
{
|
|
Tmpl: &Template{
|
|
SourcePath: "foo",
|
|
DestPath: "local/foo",
|
|
ChangeMode: "noop",
|
|
},
|
|
Fail: false,
|
|
},
|
|
{
|
|
Tmpl: &Template{
|
|
SourcePath: "foo",
|
|
DestPath: "local/foo",
|
|
ChangeMode: "noop",
|
|
Perms: "0444",
|
|
},
|
|
Fail: false,
|
|
},
|
|
{
|
|
Tmpl: &Template{
|
|
SourcePath: "foo",
|
|
DestPath: "local/foo",
|
|
ChangeMode: "noop",
|
|
Perms: "zza",
|
|
},
|
|
Fail: true,
|
|
ContainsErrs: []string{
|
|
"as octal",
|
|
},
|
|
},
|
|
}
|
|
|
|
for i, c := range cases {
|
|
err := c.Tmpl.Validate()
|
|
if err != nil {
|
|
if !c.Fail {
|
|
t.Fatalf("Case %d: shouldn't have failed: %v", i+1, err)
|
|
}
|
|
|
|
e := err.Error()
|
|
for _, exp := range c.ContainsErrs {
|
|
if !strings.Contains(e, exp) {
|
|
t.Fatalf("Cased %d: should have contained error %q: %q", i+1, exp, e)
|
|
}
|
|
}
|
|
} else if c.Fail {
|
|
t.Fatalf("Case %d: should have failed: %v", i+1, err)
|
|
}
|
|
}
|
|
}
|
|
|
|
func TestConstraint_Validate(t *testing.T) {
|
|
c := &Constraint{}
|
|
err := c.Validate()
|
|
mErr := err.(*multierror.Error)
|
|
if !strings.Contains(mErr.Errors[0].Error(), "Missing constraint operand") {
|
|
t.Fatalf("err: %s", err)
|
|
}
|
|
|
|
c = &Constraint{
|
|
LTarget: "$attr.kernel.name",
|
|
RTarget: "linux",
|
|
Operand: "=",
|
|
}
|
|
err = c.Validate()
|
|
if err != nil {
|
|
t.Fatalf("err: %v", err)
|
|
}
|
|
|
|
// Perform additional regexp validation
|
|
c.Operand = ConstraintRegex
|
|
c.RTarget = "(foo"
|
|
err = c.Validate()
|
|
mErr = err.(*multierror.Error)
|
|
if !strings.Contains(mErr.Errors[0].Error(), "missing closing") {
|
|
t.Fatalf("err: %s", err)
|
|
}
|
|
|
|
// Perform version validation
|
|
c.Operand = ConstraintVersion
|
|
c.RTarget = "~> foo"
|
|
err = c.Validate()
|
|
mErr = err.(*multierror.Error)
|
|
if !strings.Contains(mErr.Errors[0].Error(), "Malformed constraint") {
|
|
t.Fatalf("err: %s", err)
|
|
}
|
|
|
|
// Perform distinct_property validation
|
|
c.Operand = ConstraintDistinctProperty
|
|
c.RTarget = "0"
|
|
err = c.Validate()
|
|
mErr = err.(*multierror.Error)
|
|
if !strings.Contains(mErr.Errors[0].Error(), "count of 1 or greater") {
|
|
t.Fatalf("err: %s", err)
|
|
}
|
|
|
|
c.RTarget = "-1"
|
|
err = c.Validate()
|
|
mErr = err.(*multierror.Error)
|
|
if !strings.Contains(mErr.Errors[0].Error(), "to uint64") {
|
|
t.Fatalf("err: %s", err)
|
|
}
|
|
|
|
// Perform distinct_hosts validation
|
|
c.Operand = ConstraintDistinctHosts
|
|
c.LTarget = ""
|
|
c.RTarget = ""
|
|
if err := c.Validate(); err != nil {
|
|
t.Fatalf("expected valid constraint: %v", err)
|
|
}
|
|
|
|
// Perform set_contains validation
|
|
c.Operand = ConstraintSetContains
|
|
c.RTarget = ""
|
|
err = c.Validate()
|
|
mErr = err.(*multierror.Error)
|
|
if !strings.Contains(mErr.Errors[0].Error(), "requires an RTarget") {
|
|
t.Fatalf("err: %s", err)
|
|
}
|
|
|
|
// Perform LTarget validation
|
|
c.Operand = ConstraintRegex
|
|
c.RTarget = "foo"
|
|
c.LTarget = ""
|
|
err = c.Validate()
|
|
mErr = err.(*multierror.Error)
|
|
if !strings.Contains(mErr.Errors[0].Error(), "No LTarget") {
|
|
t.Fatalf("err: %s", err)
|
|
}
|
|
|
|
// Perform constraint type validation
|
|
c.Operand = "foo"
|
|
err = c.Validate()
|
|
mErr = err.(*multierror.Error)
|
|
if !strings.Contains(mErr.Errors[0].Error(), "Unknown constraint type") {
|
|
t.Fatalf("err: %s", err)
|
|
}
|
|
}
|
|
|
|
func TestUpdateStrategy_Validate(t *testing.T) {
|
|
u := &UpdateStrategy{
|
|
MaxParallel: 0,
|
|
HealthCheck: "foo",
|
|
MinHealthyTime: -10,
|
|
HealthyDeadline: -15,
|
|
AutoRevert: false,
|
|
Canary: -1,
|
|
}
|
|
|
|
err := u.Validate()
|
|
mErr := err.(*multierror.Error)
|
|
if !strings.Contains(mErr.Errors[0].Error(), "Invalid health check given") {
|
|
t.Fatalf("err: %s", err)
|
|
}
|
|
if !strings.Contains(mErr.Errors[1].Error(), "Max parallel can not be less than one") {
|
|
t.Fatalf("err: %s", err)
|
|
}
|
|
if !strings.Contains(mErr.Errors[2].Error(), "Canary count can not be less than zero") {
|
|
t.Fatalf("err: %s", err)
|
|
}
|
|
if !strings.Contains(mErr.Errors[3].Error(), "Minimum healthy time may not be less than zero") {
|
|
t.Fatalf("err: %s", err)
|
|
}
|
|
if !strings.Contains(mErr.Errors[4].Error(), "Healthy deadline must be greater than zero") {
|
|
t.Fatalf("err: %s", err)
|
|
}
|
|
if !strings.Contains(mErr.Errors[5].Error(), "Minimum healthy time must be less than healthy deadline") {
|
|
t.Fatalf("err: %s", err)
|
|
}
|
|
}
|
|
|
|
func TestResource_NetIndex(t *testing.T) {
|
|
r := &Resources{
|
|
Networks: []*NetworkResource{
|
|
{Device: "eth0"},
|
|
{Device: "lo0"},
|
|
{Device: ""},
|
|
},
|
|
}
|
|
if idx := r.NetIndex(&NetworkResource{Device: "eth0"}); idx != 0 {
|
|
t.Fatalf("Bad: %d", idx)
|
|
}
|
|
if idx := r.NetIndex(&NetworkResource{Device: "lo0"}); idx != 1 {
|
|
t.Fatalf("Bad: %d", idx)
|
|
}
|
|
if idx := r.NetIndex(&NetworkResource{Device: "eth1"}); idx != -1 {
|
|
t.Fatalf("Bad: %d", idx)
|
|
}
|
|
}
|
|
|
|
func TestResource_Superset(t *testing.T) {
|
|
r1 := &Resources{
|
|
CPU: 2000,
|
|
MemoryMB: 2048,
|
|
DiskMB: 10000,
|
|
IOPS: 100,
|
|
}
|
|
r2 := &Resources{
|
|
CPU: 2000,
|
|
MemoryMB: 1024,
|
|
DiskMB: 5000,
|
|
IOPS: 50,
|
|
}
|
|
|
|
if s, _ := r1.Superset(r1); !s {
|
|
t.Fatalf("bad")
|
|
}
|
|
if s, _ := r1.Superset(r2); !s {
|
|
t.Fatalf("bad")
|
|
}
|
|
if s, _ := r2.Superset(r1); s {
|
|
t.Fatalf("bad")
|
|
}
|
|
if s, _ := r2.Superset(r2); !s {
|
|
t.Fatalf("bad")
|
|
}
|
|
}
|
|
|
|
func TestResource_Add(t *testing.T) {
|
|
r1 := &Resources{
|
|
CPU: 2000,
|
|
MemoryMB: 2048,
|
|
DiskMB: 10000,
|
|
IOPS: 100,
|
|
Networks: []*NetworkResource{
|
|
{
|
|
CIDR: "10.0.0.0/8",
|
|
MBits: 100,
|
|
ReservedPorts: []Port{{"ssh", 22}},
|
|
},
|
|
},
|
|
}
|
|
r2 := &Resources{
|
|
CPU: 2000,
|
|
MemoryMB: 1024,
|
|
DiskMB: 5000,
|
|
IOPS: 50,
|
|
Networks: []*NetworkResource{
|
|
{
|
|
IP: "10.0.0.1",
|
|
MBits: 50,
|
|
ReservedPorts: []Port{{"web", 80}},
|
|
},
|
|
},
|
|
}
|
|
|
|
err := r1.Add(r2)
|
|
if err != nil {
|
|
t.Fatalf("Err: %v", err)
|
|
}
|
|
|
|
expect := &Resources{
|
|
CPU: 3000,
|
|
MemoryMB: 3072,
|
|
DiskMB: 15000,
|
|
IOPS: 150,
|
|
Networks: []*NetworkResource{
|
|
{
|
|
CIDR: "10.0.0.0/8",
|
|
MBits: 150,
|
|
ReservedPorts: []Port{{"ssh", 22}, {"web", 80}},
|
|
},
|
|
},
|
|
}
|
|
|
|
if !reflect.DeepEqual(expect.Networks, r1.Networks) {
|
|
t.Fatalf("bad: %#v %#v", expect, r1)
|
|
}
|
|
}
|
|
|
|
func TestResource_Add_Network(t *testing.T) {
|
|
r1 := &Resources{}
|
|
r2 := &Resources{
|
|
Networks: []*NetworkResource{
|
|
{
|
|
MBits: 50,
|
|
DynamicPorts: []Port{{"http", 0}, {"https", 0}},
|
|
},
|
|
},
|
|
}
|
|
r3 := &Resources{
|
|
Networks: []*NetworkResource{
|
|
{
|
|
MBits: 25,
|
|
DynamicPorts: []Port{{"admin", 0}},
|
|
},
|
|
},
|
|
}
|
|
|
|
err := r1.Add(r2)
|
|
if err != nil {
|
|
t.Fatalf("Err: %v", err)
|
|
}
|
|
err = r1.Add(r3)
|
|
if err != nil {
|
|
t.Fatalf("Err: %v", err)
|
|
}
|
|
|
|
expect := &Resources{
|
|
Networks: []*NetworkResource{
|
|
{
|
|
MBits: 75,
|
|
DynamicPorts: []Port{{"http", 0}, {"https", 0}, {"admin", 0}},
|
|
},
|
|
},
|
|
}
|
|
|
|
if !reflect.DeepEqual(expect.Networks, r1.Networks) {
|
|
t.Fatalf("bad: %#v %#v", expect.Networks[0], r1.Networks[0])
|
|
}
|
|
}
|
|
|
|
func TestEncodeDecode(t *testing.T) {
|
|
type FooRequest struct {
|
|
Foo string
|
|
Bar int
|
|
Baz bool
|
|
}
|
|
arg := &FooRequest{
|
|
Foo: "test",
|
|
Bar: 42,
|
|
Baz: true,
|
|
}
|
|
buf, err := Encode(1, arg)
|
|
if err != nil {
|
|
t.Fatalf("err: %v", err)
|
|
}
|
|
|
|
var out FooRequest
|
|
err = Decode(buf[1:], &out)
|
|
if err != nil {
|
|
t.Fatalf("err: %v", err)
|
|
}
|
|
|
|
if !reflect.DeepEqual(arg, &out) {
|
|
t.Fatalf("bad: %#v %#v", arg, out)
|
|
}
|
|
}
|
|
|
|
func BenchmarkEncodeDecode(b *testing.B) {
|
|
job := testJob()
|
|
|
|
for i := 0; i < b.N; i++ {
|
|
buf, err := Encode(1, job)
|
|
if err != nil {
|
|
b.Fatalf("err: %v", err)
|
|
}
|
|
|
|
var out Job
|
|
err = Decode(buf[1:], &out)
|
|
if err != nil {
|
|
b.Fatalf("err: %v", err)
|
|
}
|
|
}
|
|
}
|
|
|
|
func TestInvalidServiceCheck(t *testing.T) {
|
|
s := Service{
|
|
Name: "service-name",
|
|
PortLabel: "bar",
|
|
Checks: []*ServiceCheck{
|
|
{
|
|
Name: "check-name",
|
|
Type: "lol",
|
|
},
|
|
},
|
|
}
|
|
if err := s.Validate(); err == nil {
|
|
t.Fatalf("Service should be invalid (invalid type)")
|
|
}
|
|
|
|
s = Service{
|
|
Name: "service.name",
|
|
PortLabel: "bar",
|
|
}
|
|
if err := s.ValidateName(s.Name); err == nil {
|
|
t.Fatalf("Service should be invalid (contains a dot): %v", err)
|
|
}
|
|
|
|
s = Service{
|
|
Name: "-my-service",
|
|
PortLabel: "bar",
|
|
}
|
|
if err := s.Validate(); err == nil {
|
|
t.Fatalf("Service should be invalid (begins with a hyphen): %v", err)
|
|
}
|
|
|
|
s = Service{
|
|
Name: "my-service-${NOMAD_META_FOO}",
|
|
PortLabel: "bar",
|
|
}
|
|
if err := s.Validate(); err != nil {
|
|
t.Fatalf("Service should be valid: %v", err)
|
|
}
|
|
|
|
s = Service{
|
|
Name: "my_service-${NOMAD_META_FOO}",
|
|
PortLabel: "bar",
|
|
}
|
|
if err := s.Validate(); err == nil {
|
|
t.Fatalf("Service should be invalid (contains underscore but not in a variable name): %v", err)
|
|
}
|
|
|
|
s = Service{
|
|
Name: "abcdef0123456789-abcdef0123456789-abcdef0123456789-abcdef0123456",
|
|
PortLabel: "bar",
|
|
}
|
|
if err := s.ValidateName(s.Name); err == nil {
|
|
t.Fatalf("Service should be invalid (too long): %v", err)
|
|
}
|
|
|
|
s = Service{
|
|
Name: "service-name",
|
|
Checks: []*ServiceCheck{
|
|
{
|
|
Name: "check-tcp",
|
|
Type: ServiceCheckTCP,
|
|
Interval: 5 * time.Second,
|
|
Timeout: 2 * time.Second,
|
|
},
|
|
{
|
|
Name: "check-http",
|
|
Type: ServiceCheckHTTP,
|
|
Path: "/foo",
|
|
Interval: 5 * time.Second,
|
|
Timeout: 2 * time.Second,
|
|
},
|
|
},
|
|
}
|
|
if err := s.Validate(); err == nil {
|
|
t.Fatalf("service should be invalid (tcp/http checks with no port): %v", err)
|
|
}
|
|
|
|
s = Service{
|
|
Name: "service-name",
|
|
Checks: []*ServiceCheck{
|
|
{
|
|
Name: "check-script",
|
|
Type: ServiceCheckScript,
|
|
Command: "/bin/date",
|
|
Interval: 5 * time.Second,
|
|
Timeout: 2 * time.Second,
|
|
},
|
|
},
|
|
}
|
|
if err := s.Validate(); err != nil {
|
|
t.Fatalf("un-expected error: %v", err)
|
|
}
|
|
}
|
|
|
|
func TestDistinctCheckID(t *testing.T) {
|
|
c1 := ServiceCheck{
|
|
Name: "web-health",
|
|
Type: "http",
|
|
Path: "/health",
|
|
Interval: 2 * time.Second,
|
|
Timeout: 3 * time.Second,
|
|
}
|
|
c2 := ServiceCheck{
|
|
Name: "web-health",
|
|
Type: "http",
|
|
Path: "/health1",
|
|
Interval: 2 * time.Second,
|
|
Timeout: 3 * time.Second,
|
|
}
|
|
|
|
c3 := ServiceCheck{
|
|
Name: "web-health",
|
|
Type: "http",
|
|
Path: "/health",
|
|
Interval: 4 * time.Second,
|
|
Timeout: 3 * time.Second,
|
|
}
|
|
serviceID := "123"
|
|
c1Hash := c1.Hash(serviceID)
|
|
c2Hash := c2.Hash(serviceID)
|
|
c3Hash := c3.Hash(serviceID)
|
|
|
|
if c1Hash == c2Hash || c1Hash == c3Hash || c3Hash == c2Hash {
|
|
t.Fatalf("Checks need to be uniq c1: %s, c2: %s, c3: %s", c1Hash, c2Hash, c3Hash)
|
|
}
|
|
|
|
}
|
|
|
|
func TestService_Canonicalize(t *testing.T) {
|
|
job := "example"
|
|
taskGroup := "cache"
|
|
task := "redis"
|
|
|
|
s := Service{
|
|
Name: "${TASK}-db",
|
|
}
|
|
|
|
s.Canonicalize(job, taskGroup, task)
|
|
if s.Name != "redis-db" {
|
|
t.Fatalf("Expected name: %v, Actual: %v", "redis-db", s.Name)
|
|
}
|
|
|
|
s.Name = "db"
|
|
s.Canonicalize(job, taskGroup, task)
|
|
if s.Name != "db" {
|
|
t.Fatalf("Expected name: %v, Actual: %v", "redis-db", s.Name)
|
|
}
|
|
|
|
s.Name = "${JOB}-${TASKGROUP}-${TASK}-db"
|
|
s.Canonicalize(job, taskGroup, task)
|
|
if s.Name != "example-cache-redis-db" {
|
|
t.Fatalf("Expected name: %v, Actual: %v", "example-cache-redis-db", s.Name)
|
|
}
|
|
|
|
s.Name = "${BASE}-db"
|
|
s.Canonicalize(job, taskGroup, task)
|
|
if s.Name != "example-cache-redis-db" {
|
|
t.Fatalf("Expected name: %v, Actual: %v", "example-cache-redis-db", s.Name)
|
|
}
|
|
|
|
}
|
|
|
|
func TestJob_ExpandServiceNames(t *testing.T) {
|
|
j := &Job{
|
|
Name: "my-job",
|
|
TaskGroups: []*TaskGroup{
|
|
{
|
|
Name: "web",
|
|
Tasks: []*Task{
|
|
{
|
|
Name: "frontend",
|
|
Services: []*Service{
|
|
{
|
|
Name: "${BASE}-default",
|
|
},
|
|
{
|
|
Name: "jmx",
|
|
},
|
|
},
|
|
},
|
|
},
|
|
},
|
|
{
|
|
Name: "admin",
|
|
Tasks: []*Task{
|
|
{
|
|
Name: "admin-web",
|
|
},
|
|
},
|
|
},
|
|
},
|
|
}
|
|
|
|
j.Canonicalize()
|
|
|
|
service1Name := j.TaskGroups[0].Tasks[0].Services[0].Name
|
|
if service1Name != "my-job-web-frontend-default" {
|
|
t.Fatalf("Expected Service Name: %s, Actual: %s", "my-job-web-frontend-default", service1Name)
|
|
}
|
|
|
|
service2Name := j.TaskGroups[0].Tasks[0].Services[1].Name
|
|
if service2Name != "jmx" {
|
|
t.Fatalf("Expected Service Name: %s, Actual: %s", "jmx", service2Name)
|
|
}
|
|
|
|
}
|
|
|
|
func TestPeriodicConfig_EnabledInvalid(t *testing.T) {
|
|
// Create a config that is enabled but with no interval specified.
|
|
p := &PeriodicConfig{Enabled: true}
|
|
if err := p.Validate(); err == nil {
|
|
t.Fatal("Enabled PeriodicConfig with no spec or type shouldn't be valid")
|
|
}
|
|
|
|
// Create a config that is enabled, with a spec but no type specified.
|
|
p = &PeriodicConfig{Enabled: true, Spec: "foo"}
|
|
if err := p.Validate(); err == nil {
|
|
t.Fatal("Enabled PeriodicConfig with no spec type shouldn't be valid")
|
|
}
|
|
|
|
// Create a config that is enabled, with a spec type but no spec specified.
|
|
p = &PeriodicConfig{Enabled: true, SpecType: PeriodicSpecCron}
|
|
if err := p.Validate(); err == nil {
|
|
t.Fatal("Enabled PeriodicConfig with no spec shouldn't be valid")
|
|
}
|
|
|
|
// Create a config that is enabled, with a bad time zone.
|
|
p = &PeriodicConfig{Enabled: true, TimeZone: "FOO"}
|
|
if err := p.Validate(); err == nil || !strings.Contains(err.Error(), "time zone") {
|
|
t.Fatalf("Enabled PeriodicConfig with bad time zone shouldn't be valid: %v", err)
|
|
}
|
|
}
|
|
|
|
func TestPeriodicConfig_InvalidCron(t *testing.T) {
|
|
specs := []string{"foo", "* *", "@foo"}
|
|
for _, spec := range specs {
|
|
p := &PeriodicConfig{Enabled: true, SpecType: PeriodicSpecCron, Spec: spec}
|
|
p.Canonicalize()
|
|
if err := p.Validate(); err == nil {
|
|
t.Fatal("Invalid cron spec")
|
|
}
|
|
}
|
|
}
|
|
|
|
func TestPeriodicConfig_ValidCron(t *testing.T) {
|
|
specs := []string{"0 0 29 2 *", "@hourly", "0 0-15 * * *"}
|
|
for _, spec := range specs {
|
|
p := &PeriodicConfig{Enabled: true, SpecType: PeriodicSpecCron, Spec: spec}
|
|
p.Canonicalize()
|
|
if err := p.Validate(); err != nil {
|
|
t.Fatal("Passed valid cron")
|
|
}
|
|
}
|
|
}
|
|
|
|
func TestPeriodicConfig_NextCron(t *testing.T) {
|
|
from := time.Date(2009, time.November, 10, 23, 22, 30, 0, time.UTC)
|
|
specs := []string{"0 0 29 2 * 1980", "*/5 * * * *"}
|
|
expected := []time.Time{{}, time.Date(2009, time.November, 10, 23, 25, 0, 0, time.UTC)}
|
|
for i, spec := range specs {
|
|
p := &PeriodicConfig{Enabled: true, SpecType: PeriodicSpecCron, Spec: spec}
|
|
p.Canonicalize()
|
|
n := p.Next(from)
|
|
if expected[i] != n {
|
|
t.Fatalf("Next(%v) returned %v; want %v", from, n, expected[i])
|
|
}
|
|
}
|
|
}
|
|
|
|
func TestPeriodicConfig_ValidTimeZone(t *testing.T) {
|
|
zones := []string{"Africa/Abidjan", "America/Chicago", "Europe/Minsk", "UTC"}
|
|
for _, zone := range zones {
|
|
p := &PeriodicConfig{Enabled: true, SpecType: PeriodicSpecCron, Spec: "0 0 29 2 * 1980", TimeZone: zone}
|
|
p.Canonicalize()
|
|
if err := p.Validate(); err != nil {
|
|
t.Fatalf("Valid tz errored: %v", err)
|
|
}
|
|
}
|
|
}
|
|
|
|
func TestPeriodicConfig_DST(t *testing.T) {
|
|
// On Sun, Mar 12, 2:00 am 2017: +1 hour UTC
|
|
p := &PeriodicConfig{
|
|
Enabled: true,
|
|
SpecType: PeriodicSpecCron,
|
|
Spec: "0 2 11-12 3 * 2017",
|
|
TimeZone: "America/Los_Angeles",
|
|
}
|
|
p.Canonicalize()
|
|
|
|
t1 := time.Date(2017, time.March, 11, 1, 0, 0, 0, p.location)
|
|
t2 := time.Date(2017, time.March, 12, 1, 0, 0, 0, p.location)
|
|
|
|
// E1 is an 8 hour adjustment, E2 is a 7 hour adjustment
|
|
e1 := time.Date(2017, time.March, 11, 10, 0, 0, 0, time.UTC)
|
|
e2 := time.Date(2017, time.March, 12, 9, 0, 0, 0, time.UTC)
|
|
|
|
n1 := p.Next(t1).UTC()
|
|
n2 := p.Next(t2).UTC()
|
|
|
|
if !reflect.DeepEqual(e1, n1) {
|
|
t.Fatalf("Got %v; want %v", n1, e1)
|
|
}
|
|
if !reflect.DeepEqual(e2, n2) {
|
|
t.Fatalf("Got %v; want %v", n1, e1)
|
|
}
|
|
}
|
|
|
|
func TestRestartPolicy_Validate(t *testing.T) {
|
|
// Policy with acceptable restart options passes
|
|
p := &RestartPolicy{
|
|
Mode: RestartPolicyModeFail,
|
|
Attempts: 0,
|
|
Interval: 5 * time.Second,
|
|
}
|
|
if err := p.Validate(); err != nil {
|
|
t.Fatalf("err: %v", err)
|
|
}
|
|
|
|
// Policy with ambiguous restart options fails
|
|
p = &RestartPolicy{
|
|
Mode: RestartPolicyModeDelay,
|
|
Attempts: 0,
|
|
Interval: 5 * time.Second,
|
|
}
|
|
if err := p.Validate(); err == nil || !strings.Contains(err.Error(), "ambiguous") {
|
|
t.Fatalf("expect ambiguity error, got: %v", err)
|
|
}
|
|
|
|
// Bad policy mode fails
|
|
p = &RestartPolicy{
|
|
Mode: "nope",
|
|
Attempts: 1,
|
|
Interval: 5 * time.Second,
|
|
}
|
|
if err := p.Validate(); err == nil || !strings.Contains(err.Error(), "mode") {
|
|
t.Fatalf("expect mode error, got: %v", err)
|
|
}
|
|
|
|
// Fails when attempts*delay does not fit inside interval
|
|
p = &RestartPolicy{
|
|
Mode: RestartPolicyModeDelay,
|
|
Attempts: 3,
|
|
Delay: 5 * time.Second,
|
|
Interval: 5 * time.Second,
|
|
}
|
|
if err := p.Validate(); err == nil || !strings.Contains(err.Error(), "can't restart") {
|
|
t.Fatalf("expect restart interval error, got: %v", err)
|
|
}
|
|
|
|
// Fails when interval is to small
|
|
p = &RestartPolicy{
|
|
Mode: RestartPolicyModeDelay,
|
|
Attempts: 3,
|
|
Delay: 5 * time.Second,
|
|
Interval: 2 * time.Second,
|
|
}
|
|
if err := p.Validate(); err == nil || !strings.Contains(err.Error(), "Interval can not be less than") {
|
|
t.Fatalf("expect interval too small error, got: %v", err)
|
|
}
|
|
}
|
|
|
|
func TestReschedulePolicy_Validate(t *testing.T) {
|
|
type testCase struct {
|
|
desc string
|
|
ReschedulePolicy *ReschedulePolicy
|
|
errors []error
|
|
}
|
|
|
|
testCases := []testCase{
|
|
{
|
|
desc: "Nil",
|
|
},
|
|
{
|
|
desc: "Disabled",
|
|
ReschedulePolicy: &ReschedulePolicy{
|
|
Attempts: 0,
|
|
Interval: 0 * time.Second},
|
|
},
|
|
{
|
|
desc: "Disabled",
|
|
ReschedulePolicy: &ReschedulePolicy{
|
|
Attempts: -1,
|
|
Interval: 5 * time.Minute},
|
|
},
|
|
{
|
|
desc: "Valid Linear Delay",
|
|
ReschedulePolicy: &ReschedulePolicy{
|
|
Attempts: 1,
|
|
Interval: 5 * time.Minute,
|
|
Delay: 10 * time.Second,
|
|
DelayFunction: "constant"},
|
|
},
|
|
{
|
|
desc: "Valid Exponential Delay",
|
|
ReschedulePolicy: &ReschedulePolicy{
|
|
Attempts: 5,
|
|
Interval: 1 * time.Hour,
|
|
Delay: 30 * time.Second,
|
|
MaxDelay: 5 * time.Minute,
|
|
DelayFunction: "exponential"},
|
|
},
|
|
{
|
|
desc: "Valid Fibonacci Delay",
|
|
ReschedulePolicy: &ReschedulePolicy{
|
|
Attempts: 5,
|
|
Interval: 15 * time.Minute,
|
|
Delay: 10 * time.Second,
|
|
MaxDelay: 5 * time.Minute,
|
|
DelayFunction: "fibonacci"},
|
|
},
|
|
{
|
|
desc: "Invalid delay function",
|
|
ReschedulePolicy: &ReschedulePolicy{
|
|
Attempts: 1,
|
|
Interval: 1 * time.Second,
|
|
DelayFunction: "blah"},
|
|
errors: []error{
|
|
fmt.Errorf("Interval cannot be less than %v (got %v)", ReschedulePolicyMinInterval, time.Second),
|
|
fmt.Errorf("Delay cannot be less than %v (got %v)", ReschedulePolicyMinDelay, 0*time.Second),
|
|
fmt.Errorf("Invalid delay function %q, must be one of %q", "blah", RescheduleDelayFunctions),
|
|
},
|
|
},
|
|
{
|
|
desc: "Invalid delay ceiling",
|
|
ReschedulePolicy: &ReschedulePolicy{
|
|
Attempts: 1,
|
|
Interval: 8 * time.Second,
|
|
DelayFunction: "exponential",
|
|
Delay: 15 * time.Second,
|
|
MaxDelay: 5 * time.Second},
|
|
errors: []error{
|
|
fmt.Errorf("Max Delay cannot be less than Delay %v (got %v)",
|
|
15*time.Second, 5*time.Second),
|
|
},
|
|
},
|
|
{
|
|
desc: "Invalid delay and interval",
|
|
ReschedulePolicy: &ReschedulePolicy{
|
|
Attempts: 1,
|
|
Interval: 1 * time.Second,
|
|
DelayFunction: "constant"},
|
|
errors: []error{
|
|
fmt.Errorf("Interval cannot be less than %v (got %v)", ReschedulePolicyMinInterval, time.Second),
|
|
fmt.Errorf("Delay cannot be less than %v (got %v)", ReschedulePolicyMinDelay, 0*time.Second),
|
|
},
|
|
}, {
|
|
// Should suggest 2h40m as the interval
|
|
desc: "Invalid Attempts - linear delay",
|
|
ReschedulePolicy: &ReschedulePolicy{
|
|
Attempts: 10,
|
|
Interval: 1 * time.Hour,
|
|
Delay: 20 * time.Minute,
|
|
DelayFunction: "constant",
|
|
},
|
|
errors: []error{
|
|
fmt.Errorf("Nomad can only make %v attempts in %v with initial delay %v and"+
|
|
" delay function %q", 3, time.Hour, 20*time.Minute, "constant"),
|
|
fmt.Errorf("Set the interval to at least %v to accommodate %v attempts",
|
|
200*time.Minute, 10),
|
|
},
|
|
},
|
|
{
|
|
// Should suggest 4h40m as the interval
|
|
// Delay progression in minutes {5, 10, 20, 40, 40, 40, 40, 40, 40, 40}
|
|
desc: "Invalid Attempts - exponential delay",
|
|
ReschedulePolicy: &ReschedulePolicy{
|
|
Attempts: 10,
|
|
Interval: 30 * time.Minute,
|
|
Delay: 5 * time.Minute,
|
|
MaxDelay: 40 * time.Minute,
|
|
DelayFunction: "exponential",
|
|
},
|
|
errors: []error{
|
|
fmt.Errorf("Nomad can only make %v attempts in %v with initial delay %v, "+
|
|
"delay function %q, and delay ceiling %v", 3, 30*time.Minute, 5*time.Minute,
|
|
"exponential", 40*time.Minute),
|
|
fmt.Errorf("Set the interval to at least %v to accommodate %v attempts",
|
|
280*time.Minute, 10),
|
|
},
|
|
},
|
|
{
|
|
// Should suggest 8h as the interval
|
|
// Delay progression in minutes {20, 20, 40, 60, 80, 80, 80, 80, 80, 80}
|
|
desc: "Invalid Attempts - fibonacci delay",
|
|
ReschedulePolicy: &ReschedulePolicy{
|
|
Attempts: 10,
|
|
Interval: 1 * time.Hour,
|
|
Delay: 20 * time.Minute,
|
|
MaxDelay: 80 * time.Minute,
|
|
DelayFunction: "fibonacci",
|
|
},
|
|
errors: []error{
|
|
fmt.Errorf("Nomad can only make %v attempts in %v with initial delay %v, "+
|
|
"delay function %q, and delay ceiling %v", 4, 1*time.Hour, 20*time.Minute,
|
|
"fibonacci", 80*time.Minute),
|
|
fmt.Errorf("Set the interval to at least %v to accommodate %v attempts",
|
|
480*time.Minute, 10),
|
|
},
|
|
},
|
|
{
|
|
desc: "Ambiguous Unlimited config, has both attempts and unlimited set",
|
|
ReschedulePolicy: &ReschedulePolicy{
|
|
Attempts: 1,
|
|
Unlimited: true,
|
|
DelayFunction: "exponential",
|
|
Delay: 5 * time.Minute,
|
|
MaxDelay: 1 * time.Hour,
|
|
},
|
|
errors: []error{
|
|
fmt.Errorf("Interval must be a non zero value if Attempts > 0"),
|
|
fmt.Errorf("Reschedule Policy with Attempts = %v, Interval = %v, and Unlimited = %v is ambiguous", 1, time.Duration(0), true),
|
|
},
|
|
},
|
|
{
|
|
desc: "Invalid Unlimited config",
|
|
ReschedulePolicy: &ReschedulePolicy{
|
|
Attempts: 1,
|
|
Interval: 1 * time.Second,
|
|
Unlimited: true,
|
|
DelayFunction: "exponential",
|
|
},
|
|
errors: []error{
|
|
fmt.Errorf("Delay cannot be less than %v (got %v)", ReschedulePolicyMinDelay, 0*time.Second),
|
|
fmt.Errorf("Max Delay cannot be less than %v (got %v)", ReschedulePolicyMinDelay, 0*time.Second),
|
|
},
|
|
},
|
|
{
|
|
desc: "Valid Unlimited config",
|
|
ReschedulePolicy: &ReschedulePolicy{
|
|
Unlimited: true,
|
|
DelayFunction: "exponential",
|
|
Delay: 5 * time.Second,
|
|
MaxDelay: 1 * time.Hour,
|
|
},
|
|
},
|
|
}
|
|
|
|
for _, tc := range testCases {
|
|
t.Run(tc.desc, func(t *testing.T) {
|
|
require := require.New(t)
|
|
gotErr := tc.ReschedulePolicy.Validate()
|
|
if tc.errors != nil {
|
|
// Validate all errors
|
|
for _, err := range tc.errors {
|
|
require.Contains(gotErr.Error(), err.Error())
|
|
}
|
|
} else {
|
|
require.Nil(gotErr)
|
|
}
|
|
})
|
|
}
|
|
}
|
|
|
|
func TestAllocation_Index(t *testing.T) {
|
|
a1 := Allocation{
|
|
Name: "example.cache[1]",
|
|
TaskGroup: "cache",
|
|
JobID: "example",
|
|
Job: &Job{
|
|
ID: "example",
|
|
TaskGroups: []*TaskGroup{{Name: "cache"}}},
|
|
}
|
|
e1 := uint(1)
|
|
a2 := a1.Copy()
|
|
a2.Name = "example.cache[713127]"
|
|
e2 := uint(713127)
|
|
|
|
if a1.Index() != e1 || a2.Index() != e2 {
|
|
t.Fatalf("Got %d and %d", a1.Index(), a2.Index())
|
|
}
|
|
}
|
|
|
|
func TestTaskArtifact_Validate_Source(t *testing.T) {
|
|
valid := &TaskArtifact{GetterSource: "google.com"}
|
|
if err := valid.Validate(); err != nil {
|
|
t.Fatalf("unexpected error: %v", err)
|
|
}
|
|
}
|
|
|
|
func TestTaskArtifact_Validate_Dest(t *testing.T) {
|
|
valid := &TaskArtifact{GetterSource: "google.com"}
|
|
if err := valid.Validate(); err != nil {
|
|
t.Fatalf("unexpected error: %v", err)
|
|
}
|
|
|
|
valid.RelativeDest = "local/"
|
|
if err := valid.Validate(); err != nil {
|
|
t.Fatalf("unexpected error: %v", err)
|
|
}
|
|
|
|
valid.RelativeDest = "local/.."
|
|
if err := valid.Validate(); err != nil {
|
|
t.Fatalf("unexpected error: %v", err)
|
|
}
|
|
|
|
valid.RelativeDest = "local/../../.."
|
|
if err := valid.Validate(); err == nil {
|
|
t.Fatalf("expected error: %v", err)
|
|
}
|
|
}
|
|
|
|
func TestAllocation_ShouldMigrate(t *testing.T) {
|
|
alloc := Allocation{
|
|
PreviousAllocation: "123",
|
|
TaskGroup: "foo",
|
|
Job: &Job{
|
|
TaskGroups: []*TaskGroup{
|
|
{
|
|
Name: "foo",
|
|
EphemeralDisk: &EphemeralDisk{
|
|
Migrate: true,
|
|
Sticky: true,
|
|
},
|
|
},
|
|
},
|
|
},
|
|
}
|
|
|
|
if !alloc.ShouldMigrate() {
|
|
t.Fatalf("bad: %v", alloc)
|
|
}
|
|
|
|
alloc1 := Allocation{
|
|
PreviousAllocation: "123",
|
|
TaskGroup: "foo",
|
|
Job: &Job{
|
|
TaskGroups: []*TaskGroup{
|
|
{
|
|
Name: "foo",
|
|
EphemeralDisk: &EphemeralDisk{},
|
|
},
|
|
},
|
|
},
|
|
}
|
|
|
|
if alloc1.ShouldMigrate() {
|
|
t.Fatalf("bad: %v", alloc)
|
|
}
|
|
|
|
alloc2 := Allocation{
|
|
PreviousAllocation: "123",
|
|
TaskGroup: "foo",
|
|
Job: &Job{
|
|
TaskGroups: []*TaskGroup{
|
|
{
|
|
Name: "foo",
|
|
EphemeralDisk: &EphemeralDisk{
|
|
Sticky: false,
|
|
Migrate: true,
|
|
},
|
|
},
|
|
},
|
|
},
|
|
}
|
|
|
|
if alloc2.ShouldMigrate() {
|
|
t.Fatalf("bad: %v", alloc)
|
|
}
|
|
|
|
alloc3 := Allocation{
|
|
PreviousAllocation: "123",
|
|
TaskGroup: "foo",
|
|
Job: &Job{
|
|
TaskGroups: []*TaskGroup{
|
|
{
|
|
Name: "foo",
|
|
},
|
|
},
|
|
},
|
|
}
|
|
|
|
if alloc3.ShouldMigrate() {
|
|
t.Fatalf("bad: %v", alloc)
|
|
}
|
|
|
|
// No previous
|
|
alloc4 := Allocation{
|
|
TaskGroup: "foo",
|
|
Job: &Job{
|
|
TaskGroups: []*TaskGroup{
|
|
{
|
|
Name: "foo",
|
|
EphemeralDisk: &EphemeralDisk{
|
|
Migrate: true,
|
|
Sticky: true,
|
|
},
|
|
},
|
|
},
|
|
},
|
|
}
|
|
|
|
if alloc4.ShouldMigrate() {
|
|
t.Fatalf("bad: %v", alloc4)
|
|
}
|
|
}
|
|
|
|
func TestTaskArtifact_Validate_Checksum(t *testing.T) {
|
|
cases := []struct {
|
|
Input *TaskArtifact
|
|
Err bool
|
|
}{
|
|
{
|
|
&TaskArtifact{
|
|
GetterSource: "foo.com",
|
|
GetterOptions: map[string]string{
|
|
"checksum": "no-type",
|
|
},
|
|
},
|
|
true,
|
|
},
|
|
{
|
|
&TaskArtifact{
|
|
GetterSource: "foo.com",
|
|
GetterOptions: map[string]string{
|
|
"checksum": "md5:toosmall",
|
|
},
|
|
},
|
|
true,
|
|
},
|
|
{
|
|
&TaskArtifact{
|
|
GetterSource: "foo.com",
|
|
GetterOptions: map[string]string{
|
|
"checksum": "invalid:type",
|
|
},
|
|
},
|
|
true,
|
|
},
|
|
}
|
|
|
|
for i, tc := range cases {
|
|
err := tc.Input.Validate()
|
|
if (err != nil) != tc.Err {
|
|
t.Fatalf("case %d: %v", i, err)
|
|
continue
|
|
}
|
|
}
|
|
}
|
|
|
|
func TestAllocation_Terminated(t *testing.T) {
|
|
type desiredState struct {
|
|
ClientStatus string
|
|
DesiredStatus string
|
|
Terminated bool
|
|
}
|
|
|
|
harness := []desiredState{
|
|
{
|
|
ClientStatus: AllocClientStatusPending,
|
|
DesiredStatus: AllocDesiredStatusStop,
|
|
Terminated: false,
|
|
},
|
|
{
|
|
ClientStatus: AllocClientStatusRunning,
|
|
DesiredStatus: AllocDesiredStatusStop,
|
|
Terminated: false,
|
|
},
|
|
{
|
|
ClientStatus: AllocClientStatusFailed,
|
|
DesiredStatus: AllocDesiredStatusStop,
|
|
Terminated: true,
|
|
},
|
|
{
|
|
ClientStatus: AllocClientStatusFailed,
|
|
DesiredStatus: AllocDesiredStatusRun,
|
|
Terminated: true,
|
|
},
|
|
}
|
|
|
|
for _, state := range harness {
|
|
alloc := Allocation{}
|
|
alloc.DesiredStatus = state.DesiredStatus
|
|
alloc.ClientStatus = state.ClientStatus
|
|
if alloc.Terminated() != state.Terminated {
|
|
t.Fatalf("expected: %v, actual: %v", state.Terminated, alloc.Terminated())
|
|
}
|
|
}
|
|
}
|
|
|
|
func TestAllocation_ShouldReschedule(t *testing.T) {
|
|
type testCase struct {
|
|
Desc string
|
|
FailTime time.Time
|
|
ClientStatus string
|
|
DesiredStatus string
|
|
ReschedulePolicy *ReschedulePolicy
|
|
RescheduleTrackers []*RescheduleEvent
|
|
ShouldReschedule bool
|
|
}
|
|
|
|
fail := time.Now()
|
|
|
|
harness := []testCase{
|
|
{
|
|
Desc: "Reschedule when desired state is stop",
|
|
ClientStatus: AllocClientStatusPending,
|
|
DesiredStatus: AllocDesiredStatusStop,
|
|
FailTime: fail,
|
|
ReschedulePolicy: nil,
|
|
ShouldReschedule: false,
|
|
},
|
|
{
|
|
Desc: "Disabled rescheduling",
|
|
ClientStatus: AllocClientStatusFailed,
|
|
DesiredStatus: AllocDesiredStatusRun,
|
|
FailTime: fail,
|
|
ReschedulePolicy: &ReschedulePolicy{Attempts: 0, Interval: 1 * time.Minute},
|
|
ShouldReschedule: false,
|
|
},
|
|
{
|
|
Desc: "Reschedule when client status is complete",
|
|
ClientStatus: AllocClientStatusComplete,
|
|
DesiredStatus: AllocDesiredStatusRun,
|
|
FailTime: fail,
|
|
ReschedulePolicy: nil,
|
|
ShouldReschedule: false,
|
|
},
|
|
{
|
|
Desc: "Reschedule with nil reschedule policy",
|
|
ClientStatus: AllocClientStatusFailed,
|
|
DesiredStatus: AllocDesiredStatusRun,
|
|
FailTime: fail,
|
|
ReschedulePolicy: nil,
|
|
ShouldReschedule: false,
|
|
},
|
|
{
|
|
Desc: "Reschedule with unlimited and attempts >0",
|
|
ClientStatus: AllocClientStatusFailed,
|
|
DesiredStatus: AllocDesiredStatusRun,
|
|
FailTime: fail,
|
|
ReschedulePolicy: &ReschedulePolicy{Attempts: 1, Unlimited: true},
|
|
ShouldReschedule: true,
|
|
},
|
|
{
|
|
Desc: "Reschedule when client status is complete",
|
|
ClientStatus: AllocClientStatusComplete,
|
|
DesiredStatus: AllocDesiredStatusRun,
|
|
FailTime: fail,
|
|
ReschedulePolicy: nil,
|
|
ShouldReschedule: false,
|
|
},
|
|
{
|
|
Desc: "Reschedule with policy when client status complete",
|
|
ClientStatus: AllocClientStatusComplete,
|
|
DesiredStatus: AllocDesiredStatusRun,
|
|
FailTime: fail,
|
|
ReschedulePolicy: &ReschedulePolicy{Attempts: 1, Interval: 1 * time.Minute},
|
|
ShouldReschedule: false,
|
|
},
|
|
{
|
|
Desc: "Reschedule with no previous attempts",
|
|
ClientStatus: AllocClientStatusFailed,
|
|
DesiredStatus: AllocDesiredStatusRun,
|
|
FailTime: fail,
|
|
ReschedulePolicy: &ReschedulePolicy{Attempts: 1, Interval: 1 * time.Minute},
|
|
ShouldReschedule: true,
|
|
},
|
|
{
|
|
Desc: "Reschedule with leftover attempts",
|
|
ClientStatus: AllocClientStatusFailed,
|
|
DesiredStatus: AllocDesiredStatusRun,
|
|
ReschedulePolicy: &ReschedulePolicy{Attempts: 2, Interval: 5 * time.Minute},
|
|
FailTime: fail,
|
|
RescheduleTrackers: []*RescheduleEvent{
|
|
{
|
|
RescheduleTime: fail.Add(-1 * time.Minute).UTC().UnixNano(),
|
|
},
|
|
},
|
|
ShouldReschedule: true,
|
|
},
|
|
{
|
|
Desc: "Reschedule with too old previous attempts",
|
|
ClientStatus: AllocClientStatusFailed,
|
|
DesiredStatus: AllocDesiredStatusRun,
|
|
FailTime: fail,
|
|
ReschedulePolicy: &ReschedulePolicy{Attempts: 1, Interval: 5 * time.Minute},
|
|
RescheduleTrackers: []*RescheduleEvent{
|
|
{
|
|
RescheduleTime: fail.Add(-6 * time.Minute).UTC().UnixNano(),
|
|
},
|
|
},
|
|
ShouldReschedule: true,
|
|
},
|
|
{
|
|
Desc: "Reschedule with no leftover attempts",
|
|
ClientStatus: AllocClientStatusFailed,
|
|
DesiredStatus: AllocDesiredStatusRun,
|
|
FailTime: fail,
|
|
ReschedulePolicy: &ReschedulePolicy{Attempts: 2, Interval: 5 * time.Minute},
|
|
RescheduleTrackers: []*RescheduleEvent{
|
|
{
|
|
RescheduleTime: fail.Add(-3 * time.Minute).UTC().UnixNano(),
|
|
},
|
|
{
|
|
RescheduleTime: fail.Add(-4 * time.Minute).UTC().UnixNano(),
|
|
},
|
|
},
|
|
ShouldReschedule: false,
|
|
},
|
|
}
|
|
|
|
for _, state := range harness {
|
|
alloc := Allocation{}
|
|
alloc.DesiredStatus = state.DesiredStatus
|
|
alloc.ClientStatus = state.ClientStatus
|
|
alloc.RescheduleTracker = &RescheduleTracker{state.RescheduleTrackers}
|
|
|
|
t.Run(state.Desc, func(t *testing.T) {
|
|
if got := alloc.ShouldReschedule(state.ReschedulePolicy, state.FailTime); got != state.ShouldReschedule {
|
|
t.Fatalf("expected %v but got %v", state.ShouldReschedule, got)
|
|
}
|
|
})
|
|
|
|
}
|
|
}
|
|
|
|
func TestAllocation_LastEventTime(t *testing.T) {
|
|
type testCase struct {
|
|
desc string
|
|
taskState map[string]*TaskState
|
|
expectedLastEventTime time.Time
|
|
}
|
|
|
|
t1 := time.Now().UTC()
|
|
|
|
testCases := []testCase{
|
|
{
|
|
desc: "nil task state",
|
|
expectedLastEventTime: t1,
|
|
},
|
|
{
|
|
desc: "empty task state",
|
|
taskState: make(map[string]*TaskState),
|
|
expectedLastEventTime: t1,
|
|
},
|
|
{
|
|
desc: "Finished At not set",
|
|
taskState: map[string]*TaskState{"foo": {State: "start",
|
|
StartedAt: t1.Add(-2 * time.Hour)}},
|
|
expectedLastEventTime: t1,
|
|
},
|
|
{
|
|
desc: "One finished ",
|
|
taskState: map[string]*TaskState{"foo": {State: "start",
|
|
StartedAt: t1.Add(-2 * time.Hour),
|
|
FinishedAt: t1.Add(-1 * time.Hour)}},
|
|
expectedLastEventTime: t1.Add(-1 * time.Hour),
|
|
},
|
|
{
|
|
desc: "Multiple task groups",
|
|
taskState: map[string]*TaskState{"foo": {State: "start",
|
|
StartedAt: t1.Add(-2 * time.Hour),
|
|
FinishedAt: t1.Add(-1 * time.Hour)},
|
|
"bar": {State: "start",
|
|
StartedAt: t1.Add(-2 * time.Hour),
|
|
FinishedAt: t1.Add(-40 * time.Minute)}},
|
|
expectedLastEventTime: t1.Add(-40 * time.Minute),
|
|
},
|
|
{
|
|
desc: "No finishedAt set, one task event, should use modify time",
|
|
taskState: map[string]*TaskState{"foo": {
|
|
State: "run",
|
|
StartedAt: t1.Add(-2 * time.Hour),
|
|
Events: []*TaskEvent{
|
|
{Type: "start", Time: t1.Add(-20 * time.Minute).UnixNano()},
|
|
}},
|
|
},
|
|
expectedLastEventTime: t1,
|
|
},
|
|
}
|
|
for _, tc := range testCases {
|
|
t.Run(tc.desc, func(t *testing.T) {
|
|
alloc := &Allocation{CreateTime: t1.UnixNano(), ModifyTime: t1.UnixNano()}
|
|
alloc.TaskStates = tc.taskState
|
|
require.Equal(t, tc.expectedLastEventTime, alloc.LastEventTime())
|
|
})
|
|
}
|
|
}
|
|
|
|
func TestAllocation_NextDelay(t *testing.T) {
|
|
type testCase struct {
|
|
desc string
|
|
reschedulePolicy *ReschedulePolicy
|
|
alloc *Allocation
|
|
expectedRescheduleTime time.Time
|
|
expectedRescheduleEligible bool
|
|
}
|
|
now := time.Now()
|
|
testCases := []testCase{
|
|
{
|
|
desc: "Allocation hasn't failed yet",
|
|
reschedulePolicy: &ReschedulePolicy{
|
|
DelayFunction: "constant",
|
|
Delay: 5 * time.Second,
|
|
},
|
|
alloc: &Allocation{},
|
|
expectedRescheduleTime: time.Time{},
|
|
expectedRescheduleEligible: false,
|
|
},
|
|
{
|
|
desc: "Allocation lacks task state",
|
|
reschedulePolicy: &ReschedulePolicy{
|
|
DelayFunction: "constant",
|
|
Delay: 5 * time.Second,
|
|
Unlimited: true,
|
|
},
|
|
alloc: &Allocation{ClientStatus: AllocClientStatusFailed, ModifyTime: now.UnixNano()},
|
|
expectedRescheduleTime: now.UTC().Add(5 * time.Second),
|
|
expectedRescheduleEligible: true,
|
|
},
|
|
{
|
|
desc: "linear delay, unlimited restarts, no reschedule tracker",
|
|
reschedulePolicy: &ReschedulePolicy{
|
|
DelayFunction: "constant",
|
|
Delay: 5 * time.Second,
|
|
Unlimited: true,
|
|
},
|
|
alloc: &Allocation{
|
|
ClientStatus: AllocClientStatusFailed,
|
|
TaskStates: map[string]*TaskState{"foo": {State: "dead",
|
|
StartedAt: now.Add(-1 * time.Hour),
|
|
FinishedAt: now.Add(-2 * time.Second)}},
|
|
},
|
|
expectedRescheduleTime: now.Add(-2 * time.Second).Add(5 * time.Second),
|
|
expectedRescheduleEligible: true,
|
|
},
|
|
{
|
|
desc: "linear delay with reschedule tracker",
|
|
reschedulePolicy: &ReschedulePolicy{
|
|
DelayFunction: "constant",
|
|
Delay: 5 * time.Second,
|
|
Interval: 10 * time.Minute,
|
|
Attempts: 2,
|
|
},
|
|
alloc: &Allocation{
|
|
ClientStatus: AllocClientStatusFailed,
|
|
TaskStates: map[string]*TaskState{"foo": {State: "start",
|
|
StartedAt: now.Add(-1 * time.Hour),
|
|
FinishedAt: now.Add(-2 * time.Second)}},
|
|
RescheduleTracker: &RescheduleTracker{
|
|
Events: []*RescheduleEvent{{
|
|
RescheduleTime: now.Add(-2 * time.Minute).UTC().UnixNano(),
|
|
Delay: 5 * time.Second,
|
|
}},
|
|
}},
|
|
expectedRescheduleTime: now.Add(-2 * time.Second).Add(5 * time.Second),
|
|
expectedRescheduleEligible: true,
|
|
},
|
|
{
|
|
desc: "linear delay with reschedule tracker, attempts exhausted",
|
|
reschedulePolicy: &ReschedulePolicy{
|
|
DelayFunction: "constant",
|
|
Delay: 5 * time.Second,
|
|
Interval: 10 * time.Minute,
|
|
Attempts: 2,
|
|
},
|
|
alloc: &Allocation{
|
|
ClientStatus: AllocClientStatusFailed,
|
|
TaskStates: map[string]*TaskState{"foo": {State: "start",
|
|
StartedAt: now.Add(-1 * time.Hour),
|
|
FinishedAt: now.Add(-2 * time.Second)}},
|
|
RescheduleTracker: &RescheduleTracker{
|
|
Events: []*RescheduleEvent{
|
|
{
|
|
RescheduleTime: now.Add(-3 * time.Minute).UTC().UnixNano(),
|
|
Delay: 5 * time.Second,
|
|
},
|
|
{
|
|
RescheduleTime: now.Add(-2 * time.Minute).UTC().UnixNano(),
|
|
Delay: 5 * time.Second,
|
|
},
|
|
},
|
|
}},
|
|
expectedRescheduleTime: now.Add(-2 * time.Second).Add(5 * time.Second),
|
|
expectedRescheduleEligible: false,
|
|
},
|
|
{
|
|
desc: "exponential delay - no reschedule tracker",
|
|
reschedulePolicy: &ReschedulePolicy{
|
|
DelayFunction: "exponential",
|
|
Delay: 5 * time.Second,
|
|
MaxDelay: 90 * time.Second,
|
|
Unlimited: true,
|
|
},
|
|
alloc: &Allocation{
|
|
ClientStatus: AllocClientStatusFailed,
|
|
TaskStates: map[string]*TaskState{"foo": {State: "start",
|
|
StartedAt: now.Add(-1 * time.Hour),
|
|
FinishedAt: now.Add(-2 * time.Second)}},
|
|
},
|
|
expectedRescheduleTime: now.Add(-2 * time.Second).Add(5 * time.Second),
|
|
expectedRescheduleEligible: true,
|
|
},
|
|
{
|
|
desc: "exponential delay with reschedule tracker",
|
|
reschedulePolicy: &ReschedulePolicy{
|
|
DelayFunction: "exponential",
|
|
Delay: 5 * time.Second,
|
|
MaxDelay: 90 * time.Second,
|
|
Unlimited: true,
|
|
},
|
|
alloc: &Allocation{
|
|
ClientStatus: AllocClientStatusFailed,
|
|
TaskStates: map[string]*TaskState{"foo": {State: "start",
|
|
StartedAt: now.Add(-1 * time.Hour),
|
|
FinishedAt: now.Add(-2 * time.Second)}},
|
|
RescheduleTracker: &RescheduleTracker{
|
|
Events: []*RescheduleEvent{
|
|
{
|
|
RescheduleTime: now.Add(-2 * time.Hour).UTC().UnixNano(),
|
|
Delay: 5 * time.Second,
|
|
},
|
|
{
|
|
RescheduleTime: now.Add(-1 * time.Hour).UTC().UnixNano(),
|
|
Delay: 10 * time.Second,
|
|
},
|
|
{
|
|
RescheduleTime: now.Add(-1 * time.Hour).UTC().UnixNano(),
|
|
Delay: 20 * time.Second,
|
|
},
|
|
},
|
|
}},
|
|
expectedRescheduleTime: now.Add(-2 * time.Second).Add(40 * time.Second),
|
|
expectedRescheduleEligible: true,
|
|
},
|
|
{
|
|
desc: "exponential delay with delay ceiling reached",
|
|
reschedulePolicy: &ReschedulePolicy{
|
|
DelayFunction: "exponential",
|
|
Delay: 5 * time.Second,
|
|
MaxDelay: 90 * time.Second,
|
|
Unlimited: true,
|
|
},
|
|
alloc: &Allocation{
|
|
ClientStatus: AllocClientStatusFailed,
|
|
TaskStates: map[string]*TaskState{"foo": {State: "start",
|
|
StartedAt: now.Add(-1 * time.Hour),
|
|
FinishedAt: now.Add(-15 * time.Second)}},
|
|
RescheduleTracker: &RescheduleTracker{
|
|
Events: []*RescheduleEvent{
|
|
{
|
|
RescheduleTime: now.Add(-2 * time.Hour).UTC().UnixNano(),
|
|
Delay: 5 * time.Second,
|
|
},
|
|
{
|
|
RescheduleTime: now.Add(-1 * time.Hour).UTC().UnixNano(),
|
|
Delay: 10 * time.Second,
|
|
},
|
|
{
|
|
RescheduleTime: now.Add(-1 * time.Hour).UTC().UnixNano(),
|
|
Delay: 20 * time.Second,
|
|
},
|
|
{
|
|
RescheduleTime: now.Add(-1 * time.Hour).UTC().UnixNano(),
|
|
Delay: 40 * time.Second,
|
|
},
|
|
{
|
|
RescheduleTime: now.Add(-40 * time.Second).UTC().UnixNano(),
|
|
Delay: 80 * time.Second,
|
|
},
|
|
},
|
|
}},
|
|
expectedRescheduleTime: now.Add(-15 * time.Second).Add(90 * time.Second),
|
|
expectedRescheduleEligible: true,
|
|
},
|
|
{
|
|
// Test case where most recent reschedule ran longer than delay ceiling
|
|
desc: "exponential delay, delay ceiling reset condition met",
|
|
reschedulePolicy: &ReschedulePolicy{
|
|
DelayFunction: "exponential",
|
|
Delay: 5 * time.Second,
|
|
MaxDelay: 90 * time.Second,
|
|
Unlimited: true,
|
|
},
|
|
alloc: &Allocation{
|
|
ClientStatus: AllocClientStatusFailed,
|
|
TaskStates: map[string]*TaskState{"foo": {State: "start",
|
|
StartedAt: now.Add(-1 * time.Hour),
|
|
FinishedAt: now.Add(-15 * time.Minute)}},
|
|
RescheduleTracker: &RescheduleTracker{
|
|
Events: []*RescheduleEvent{
|
|
{
|
|
RescheduleTime: now.Add(-2 * time.Hour).UTC().UnixNano(),
|
|
Delay: 5 * time.Second,
|
|
},
|
|
{
|
|
RescheduleTime: now.Add(-1 * time.Hour).UTC().UnixNano(),
|
|
Delay: 10 * time.Second,
|
|
},
|
|
{
|
|
RescheduleTime: now.Add(-1 * time.Hour).UTC().UnixNano(),
|
|
Delay: 20 * time.Second,
|
|
},
|
|
{
|
|
RescheduleTime: now.Add(-1 * time.Hour).UTC().UnixNano(),
|
|
Delay: 40 * time.Second,
|
|
},
|
|
{
|
|
RescheduleTime: now.Add(-1 * time.Hour).UTC().UnixNano(),
|
|
Delay: 80 * time.Second,
|
|
},
|
|
{
|
|
RescheduleTime: now.Add(-1 * time.Hour).UTC().UnixNano(),
|
|
Delay: 90 * time.Second,
|
|
},
|
|
{
|
|
RescheduleTime: now.Add(-1 * time.Hour).UTC().UnixNano(),
|
|
Delay: 90 * time.Second,
|
|
},
|
|
},
|
|
}},
|
|
expectedRescheduleTime: now.Add(-15 * time.Minute).Add(5 * time.Second),
|
|
expectedRescheduleEligible: true,
|
|
},
|
|
{
|
|
desc: "fibonacci delay - no reschedule tracker",
|
|
reschedulePolicy: &ReschedulePolicy{
|
|
DelayFunction: "fibonacci",
|
|
Delay: 5 * time.Second,
|
|
MaxDelay: 90 * time.Second,
|
|
Unlimited: true,
|
|
},
|
|
alloc: &Allocation{
|
|
ClientStatus: AllocClientStatusFailed,
|
|
TaskStates: map[string]*TaskState{"foo": {State: "start",
|
|
StartedAt: now.Add(-1 * time.Hour),
|
|
FinishedAt: now.Add(-2 * time.Second)}}},
|
|
expectedRescheduleTime: now.Add(-2 * time.Second).Add(5 * time.Second),
|
|
expectedRescheduleEligible: true,
|
|
},
|
|
{
|
|
desc: "fibonacci delay with reschedule tracker",
|
|
reschedulePolicy: &ReschedulePolicy{
|
|
DelayFunction: "fibonacci",
|
|
Delay: 5 * time.Second,
|
|
MaxDelay: 90 * time.Second,
|
|
Unlimited: true,
|
|
},
|
|
alloc: &Allocation{
|
|
ClientStatus: AllocClientStatusFailed,
|
|
TaskStates: map[string]*TaskState{"foo": {State: "start",
|
|
StartedAt: now.Add(-1 * time.Hour),
|
|
FinishedAt: now.Add(-2 * time.Second)}},
|
|
RescheduleTracker: &RescheduleTracker{
|
|
Events: []*RescheduleEvent{
|
|
{
|
|
RescheduleTime: now.Add(-2 * time.Hour).UTC().UnixNano(),
|
|
Delay: 5 * time.Second,
|
|
},
|
|
{
|
|
RescheduleTime: now.Add(-5 * time.Second).UTC().UnixNano(),
|
|
Delay: 5 * time.Second,
|
|
},
|
|
},
|
|
}},
|
|
expectedRescheduleTime: now.Add(-2 * time.Second).Add(10 * time.Second),
|
|
expectedRescheduleEligible: true,
|
|
},
|
|
{
|
|
desc: "fibonacci delay with more events",
|
|
reschedulePolicy: &ReschedulePolicy{
|
|
DelayFunction: "fibonacci",
|
|
Delay: 5 * time.Second,
|
|
MaxDelay: 90 * time.Second,
|
|
Unlimited: true,
|
|
},
|
|
alloc: &Allocation{
|
|
ClientStatus: AllocClientStatusFailed,
|
|
TaskStates: map[string]*TaskState{"foo": {State: "start",
|
|
StartedAt: now.Add(-1 * time.Hour),
|
|
FinishedAt: now.Add(-2 * time.Second)}},
|
|
RescheduleTracker: &RescheduleTracker{
|
|
Events: []*RescheduleEvent{
|
|
{
|
|
RescheduleTime: now.Add(-2 * time.Hour).UTC().UnixNano(),
|
|
Delay: 5 * time.Second,
|
|
},
|
|
{
|
|
RescheduleTime: now.Add(-1 * time.Hour).UTC().UnixNano(),
|
|
Delay: 5 * time.Second,
|
|
},
|
|
{
|
|
RescheduleTime: now.Add(-1 * time.Hour).UTC().UnixNano(),
|
|
Delay: 10 * time.Second,
|
|
},
|
|
{
|
|
RescheduleTime: now.Add(-1 * time.Hour).UTC().UnixNano(),
|
|
Delay: 15 * time.Second,
|
|
},
|
|
{
|
|
RescheduleTime: now.Add(-1 * time.Hour).UTC().UnixNano(),
|
|
Delay: 25 * time.Second,
|
|
},
|
|
},
|
|
}},
|
|
expectedRescheduleTime: now.Add(-2 * time.Second).Add(40 * time.Second),
|
|
expectedRescheduleEligible: true,
|
|
},
|
|
{
|
|
desc: "fibonacci delay with delay ceiling reached",
|
|
reschedulePolicy: &ReschedulePolicy{
|
|
DelayFunction: "fibonacci",
|
|
Delay: 5 * time.Second,
|
|
MaxDelay: 50 * time.Second,
|
|
Unlimited: true,
|
|
},
|
|
alloc: &Allocation{
|
|
ClientStatus: AllocClientStatusFailed,
|
|
TaskStates: map[string]*TaskState{"foo": {State: "start",
|
|
StartedAt: now.Add(-1 * time.Hour),
|
|
FinishedAt: now.Add(-15 * time.Second)}},
|
|
RescheduleTracker: &RescheduleTracker{
|
|
Events: []*RescheduleEvent{
|
|
{
|
|
RescheduleTime: now.Add(-2 * time.Hour).UTC().UnixNano(),
|
|
Delay: 5 * time.Second,
|
|
},
|
|
{
|
|
RescheduleTime: now.Add(-1 * time.Hour).UTC().UnixNano(),
|
|
Delay: 5 * time.Second,
|
|
},
|
|
{
|
|
RescheduleTime: now.Add(-1 * time.Hour).UTC().UnixNano(),
|
|
Delay: 10 * time.Second,
|
|
},
|
|
{
|
|
RescheduleTime: now.Add(-1 * time.Hour).UTC().UnixNano(),
|
|
Delay: 15 * time.Second,
|
|
},
|
|
{
|
|
RescheduleTime: now.Add(-1 * time.Hour).UTC().UnixNano(),
|
|
Delay: 25 * time.Second,
|
|
},
|
|
{
|
|
RescheduleTime: now.Add(-40 * time.Second).UTC().UnixNano(),
|
|
Delay: 40 * time.Second,
|
|
},
|
|
},
|
|
}},
|
|
expectedRescheduleTime: now.Add(-15 * time.Second).Add(50 * time.Second),
|
|
expectedRescheduleEligible: true,
|
|
},
|
|
{
|
|
desc: "fibonacci delay with delay reset condition met",
|
|
reschedulePolicy: &ReschedulePolicy{
|
|
DelayFunction: "fibonacci",
|
|
Delay: 5 * time.Second,
|
|
MaxDelay: 50 * time.Second,
|
|
Unlimited: true,
|
|
},
|
|
alloc: &Allocation{
|
|
ClientStatus: AllocClientStatusFailed,
|
|
TaskStates: map[string]*TaskState{"foo": {State: "start",
|
|
StartedAt: now.Add(-1 * time.Hour),
|
|
FinishedAt: now.Add(-5 * time.Minute)}},
|
|
RescheduleTracker: &RescheduleTracker{
|
|
Events: []*RescheduleEvent{
|
|
{
|
|
RescheduleTime: now.Add(-2 * time.Hour).UTC().UnixNano(),
|
|
Delay: 5 * time.Second,
|
|
},
|
|
{
|
|
RescheduleTime: now.Add(-1 * time.Hour).UTC().UnixNano(),
|
|
Delay: 5 * time.Second,
|
|
},
|
|
{
|
|
RescheduleTime: now.Add(-1 * time.Hour).UTC().UnixNano(),
|
|
Delay: 10 * time.Second,
|
|
},
|
|
{
|
|
RescheduleTime: now.Add(-1 * time.Hour).UTC().UnixNano(),
|
|
Delay: 15 * time.Second,
|
|
},
|
|
{
|
|
RescheduleTime: now.Add(-1 * time.Hour).UTC().UnixNano(),
|
|
Delay: 25 * time.Second,
|
|
},
|
|
{
|
|
RescheduleTime: now.Add(-1 * time.Hour).UTC().UnixNano(),
|
|
Delay: 40 * time.Second,
|
|
},
|
|
},
|
|
}},
|
|
expectedRescheduleTime: now.Add(-5 * time.Minute).Add(5 * time.Second),
|
|
expectedRescheduleEligible: true,
|
|
},
|
|
{
|
|
desc: "fibonacci delay with the most recent event that reset delay value",
|
|
reschedulePolicy: &ReschedulePolicy{
|
|
DelayFunction: "fibonacci",
|
|
Delay: 5 * time.Second,
|
|
MaxDelay: 50 * time.Second,
|
|
Unlimited: true,
|
|
},
|
|
alloc: &Allocation{
|
|
ClientStatus: AllocClientStatusFailed,
|
|
TaskStates: map[string]*TaskState{"foo": {State: "start",
|
|
StartedAt: now.Add(-1 * time.Hour),
|
|
FinishedAt: now.Add(-5 * time.Second)}},
|
|
RescheduleTracker: &RescheduleTracker{
|
|
Events: []*RescheduleEvent{
|
|
{
|
|
RescheduleTime: now.Add(-2 * time.Hour).UTC().UnixNano(),
|
|
Delay: 5 * time.Second,
|
|
},
|
|
{
|
|
RescheduleTime: now.Add(-1 * time.Hour).UTC().UnixNano(),
|
|
Delay: 5 * time.Second,
|
|
},
|
|
{
|
|
RescheduleTime: now.Add(-1 * time.Hour).UTC().UnixNano(),
|
|
Delay: 10 * time.Second,
|
|
},
|
|
{
|
|
RescheduleTime: now.Add(-1 * time.Hour).UTC().UnixNano(),
|
|
Delay: 15 * time.Second,
|
|
},
|
|
{
|
|
RescheduleTime: now.Add(-1 * time.Hour).UTC().UnixNano(),
|
|
Delay: 25 * time.Second,
|
|
},
|
|
{
|
|
RescheduleTime: now.Add(-1 * time.Hour).UTC().UnixNano(),
|
|
Delay: 40 * time.Second,
|
|
},
|
|
{
|
|
RescheduleTime: now.Add(-1 * time.Hour).UTC().UnixNano(),
|
|
Delay: 50 * time.Second,
|
|
},
|
|
{
|
|
RescheduleTime: now.Add(-1 * time.Minute).UTC().UnixNano(),
|
|
Delay: 5 * time.Second,
|
|
},
|
|
},
|
|
}},
|
|
expectedRescheduleTime: now.Add(-5 * time.Second).Add(5 * time.Second),
|
|
expectedRescheduleEligible: true,
|
|
},
|
|
}
|
|
for _, tc := range testCases {
|
|
t.Run(tc.desc, func(t *testing.T) {
|
|
require := require.New(t)
|
|
j := testJob()
|
|
j.TaskGroups[0].ReschedulePolicy = tc.reschedulePolicy
|
|
tc.alloc.Job = j
|
|
tc.alloc.TaskGroup = j.TaskGroups[0].Name
|
|
reschedTime, allowed := tc.alloc.NextRescheduleTime()
|
|
require.Equal(tc.expectedRescheduleEligible, allowed)
|
|
require.Equal(tc.expectedRescheduleTime, reschedTime)
|
|
})
|
|
}
|
|
|
|
}
|
|
|
|
func TestRescheduleTracker_Copy(t *testing.T) {
|
|
type testCase struct {
|
|
original *RescheduleTracker
|
|
expected *RescheduleTracker
|
|
}
|
|
|
|
cases := []testCase{
|
|
{nil, nil},
|
|
{&RescheduleTracker{Events: []*RescheduleEvent{
|
|
{RescheduleTime: 2,
|
|
PrevAllocID: "12",
|
|
PrevNodeID: "12",
|
|
Delay: 30 * time.Second},
|
|
}}, &RescheduleTracker{Events: []*RescheduleEvent{
|
|
{RescheduleTime: 2,
|
|
PrevAllocID: "12",
|
|
PrevNodeID: "12",
|
|
Delay: 30 * time.Second},
|
|
}}},
|
|
}
|
|
|
|
for _, tc := range cases {
|
|
if got := tc.original.Copy(); !reflect.DeepEqual(got, tc.expected) {
|
|
t.Fatalf("expected %v but got %v", *tc.expected, *got)
|
|
}
|
|
}
|
|
}
|
|
|
|
func TestVault_Validate(t *testing.T) {
|
|
v := &Vault{
|
|
Env: true,
|
|
ChangeMode: VaultChangeModeNoop,
|
|
}
|
|
|
|
if err := v.Validate(); err == nil || !strings.Contains(err.Error(), "Policy list") {
|
|
t.Fatalf("Expected policy list empty error")
|
|
}
|
|
|
|
v.Policies = []string{"foo", "root"}
|
|
v.ChangeMode = VaultChangeModeSignal
|
|
|
|
err := v.Validate()
|
|
if err == nil {
|
|
t.Fatalf("Expected validation errors")
|
|
}
|
|
|
|
if !strings.Contains(err.Error(), "Signal must") {
|
|
t.Fatalf("Expected signal empty error")
|
|
}
|
|
if !strings.Contains(err.Error(), "root") {
|
|
t.Fatalf("Expected root error")
|
|
}
|
|
}
|
|
|
|
func TestParameterizedJobConfig_Validate(t *testing.T) {
|
|
d := &ParameterizedJobConfig{
|
|
Payload: "foo",
|
|
}
|
|
|
|
if err := d.Validate(); err == nil || !strings.Contains(err.Error(), "payload") {
|
|
t.Fatalf("Expected unknown payload requirement: %v", err)
|
|
}
|
|
|
|
d.Payload = DispatchPayloadOptional
|
|
d.MetaOptional = []string{"foo", "bar"}
|
|
d.MetaRequired = []string{"bar", "baz"}
|
|
|
|
if err := d.Validate(); err == nil || !strings.Contains(err.Error(), "disjoint") {
|
|
t.Fatalf("Expected meta not being disjoint error: %v", err)
|
|
}
|
|
}
|
|
|
|
func TestParameterizedJobConfig_Validate_NonBatch(t *testing.T) {
|
|
job := testJob()
|
|
job.ParameterizedJob = &ParameterizedJobConfig{
|
|
Payload: DispatchPayloadOptional,
|
|
}
|
|
job.Type = JobTypeSystem
|
|
|
|
if err := job.Validate(); err == nil || !strings.Contains(err.Error(), "only be used with") {
|
|
t.Fatalf("Expected bad scheduler tpye: %v", err)
|
|
}
|
|
}
|
|
|
|
func TestParameterizedJobConfig_Canonicalize(t *testing.T) {
|
|
d := &ParameterizedJobConfig{}
|
|
d.Canonicalize()
|
|
if d.Payload != DispatchPayloadOptional {
|
|
t.Fatalf("Canonicalize failed")
|
|
}
|
|
}
|
|
|
|
func TestDispatchPayloadConfig_Validate(t *testing.T) {
|
|
d := &DispatchPayloadConfig{
|
|
File: "foo",
|
|
}
|
|
|
|
// task/local/haha
|
|
if err := d.Validate(); err != nil {
|
|
t.Fatalf("bad: %v", err)
|
|
}
|
|
|
|
// task/haha
|
|
d.File = "../haha"
|
|
if err := d.Validate(); err != nil {
|
|
t.Fatalf("bad: %v", err)
|
|
}
|
|
|
|
// ../haha
|
|
d.File = "../../../haha"
|
|
if err := d.Validate(); err == nil {
|
|
t.Fatalf("bad: %v", err)
|
|
}
|
|
}
|
|
|
|
func TestIsRecoverable(t *testing.T) {
|
|
if IsRecoverable(nil) {
|
|
t.Errorf("nil should not be recoverable")
|
|
}
|
|
if IsRecoverable(NewRecoverableError(nil, true)) {
|
|
t.Errorf("NewRecoverableError(nil, true) should not be recoverable")
|
|
}
|
|
if IsRecoverable(fmt.Errorf("i promise im recoverable")) {
|
|
t.Errorf("Custom errors should not be recoverable")
|
|
}
|
|
if IsRecoverable(NewRecoverableError(fmt.Errorf(""), false)) {
|
|
t.Errorf("Explicitly unrecoverable errors should not be recoverable")
|
|
}
|
|
if !IsRecoverable(NewRecoverableError(fmt.Errorf(""), true)) {
|
|
t.Errorf("Explicitly recoverable errors *should* be recoverable")
|
|
}
|
|
}
|
|
|
|
func TestACLTokenValidate(t *testing.T) {
|
|
tk := &ACLToken{}
|
|
|
|
// Missing a type
|
|
err := tk.Validate()
|
|
assert.NotNil(t, err)
|
|
if !strings.Contains(err.Error(), "client or management") {
|
|
t.Fatalf("bad: %v", err)
|
|
}
|
|
|
|
// Missing policies
|
|
tk.Type = ACLClientToken
|
|
err = tk.Validate()
|
|
assert.NotNil(t, err)
|
|
if !strings.Contains(err.Error(), "missing policies") {
|
|
t.Fatalf("bad: %v", err)
|
|
}
|
|
|
|
// Invalid policies
|
|
tk.Type = ACLManagementToken
|
|
tk.Policies = []string{"foo"}
|
|
err = tk.Validate()
|
|
assert.NotNil(t, err)
|
|
if !strings.Contains(err.Error(), "associated with policies") {
|
|
t.Fatalf("bad: %v", err)
|
|
}
|
|
|
|
// Name too long policies
|
|
tk.Name = ""
|
|
for i := 0; i < 8; i++ {
|
|
tk.Name += uuid.Generate()
|
|
}
|
|
tk.Policies = nil
|
|
err = tk.Validate()
|
|
assert.NotNil(t, err)
|
|
if !strings.Contains(err.Error(), "too long") {
|
|
t.Fatalf("bad: %v", err)
|
|
}
|
|
|
|
// Make it valid
|
|
tk.Name = "foo"
|
|
err = tk.Validate()
|
|
assert.Nil(t, err)
|
|
}
|
|
|
|
func TestACLTokenPolicySubset(t *testing.T) {
|
|
tk := &ACLToken{
|
|
Type: ACLClientToken,
|
|
Policies: []string{"foo", "bar", "baz"},
|
|
}
|
|
|
|
assert.Equal(t, true, tk.PolicySubset([]string{"foo", "bar", "baz"}))
|
|
assert.Equal(t, true, tk.PolicySubset([]string{"foo", "bar"}))
|
|
assert.Equal(t, true, tk.PolicySubset([]string{"foo"}))
|
|
assert.Equal(t, true, tk.PolicySubset([]string{}))
|
|
assert.Equal(t, false, tk.PolicySubset([]string{"foo", "bar", "new"}))
|
|
assert.Equal(t, false, tk.PolicySubset([]string{"new"}))
|
|
|
|
tk = &ACLToken{
|
|
Type: ACLManagementToken,
|
|
}
|
|
|
|
assert.Equal(t, true, tk.PolicySubset([]string{"foo", "bar", "baz"}))
|
|
assert.Equal(t, true, tk.PolicySubset([]string{"foo", "bar"}))
|
|
assert.Equal(t, true, tk.PolicySubset([]string{"foo"}))
|
|
assert.Equal(t, true, tk.PolicySubset([]string{}))
|
|
assert.Equal(t, true, tk.PolicySubset([]string{"foo", "bar", "new"}))
|
|
assert.Equal(t, true, tk.PolicySubset([]string{"new"}))
|
|
}
|
|
|
|
func TestACLTokenSetHash(t *testing.T) {
|
|
tk := &ACLToken{
|
|
Name: "foo",
|
|
Type: ACLClientToken,
|
|
Policies: []string{"foo", "bar"},
|
|
Global: false,
|
|
}
|
|
out1 := tk.SetHash()
|
|
assert.NotNil(t, out1)
|
|
assert.NotNil(t, tk.Hash)
|
|
assert.Equal(t, out1, tk.Hash)
|
|
|
|
tk.Policies = []string{"foo"}
|
|
out2 := tk.SetHash()
|
|
assert.NotNil(t, out2)
|
|
assert.NotNil(t, tk.Hash)
|
|
assert.Equal(t, out2, tk.Hash)
|
|
assert.NotEqual(t, out1, out2)
|
|
}
|
|
|
|
func TestACLPolicySetHash(t *testing.T) {
|
|
ap := &ACLPolicy{
|
|
Name: "foo",
|
|
Description: "great policy",
|
|
Rules: "node { policy = \"read\" }",
|
|
}
|
|
out1 := ap.SetHash()
|
|
assert.NotNil(t, out1)
|
|
assert.NotNil(t, ap.Hash)
|
|
assert.Equal(t, out1, ap.Hash)
|
|
|
|
ap.Rules = "node { policy = \"write\" }"
|
|
out2 := ap.SetHash()
|
|
assert.NotNil(t, out2)
|
|
assert.NotNil(t, ap.Hash)
|
|
assert.Equal(t, out2, ap.Hash)
|
|
assert.NotEqual(t, out1, out2)
|
|
}
|
|
|
|
func TestTaskEventPopulate(t *testing.T) {
|
|
prepopulatedEvent := NewTaskEvent(TaskSetup)
|
|
prepopulatedEvent.DisplayMessage = "Hola"
|
|
testcases := []struct {
|
|
event *TaskEvent
|
|
expectedMsg string
|
|
}{
|
|
{nil, ""},
|
|
{prepopulatedEvent, "Hola"},
|
|
{NewTaskEvent(TaskSetup).SetMessage("Setup"), "Setup"},
|
|
{NewTaskEvent(TaskStarted), "Task started by client"},
|
|
{NewTaskEvent(TaskReceived), "Task received by client"},
|
|
{NewTaskEvent(TaskFailedValidation), "Validation of task failed"},
|
|
{NewTaskEvent(TaskFailedValidation).SetValidationError(fmt.Errorf("task failed validation")), "task failed validation"},
|
|
{NewTaskEvent(TaskSetupFailure), "Task setup failed"},
|
|
{NewTaskEvent(TaskSetupFailure).SetSetupError(fmt.Errorf("task failed setup")), "task failed setup"},
|
|
{NewTaskEvent(TaskDriverFailure), "Failed to start task"},
|
|
{NewTaskEvent(TaskDownloadingArtifacts), "Client is downloading artifacts"},
|
|
{NewTaskEvent(TaskArtifactDownloadFailed), "Failed to download artifacts"},
|
|
{NewTaskEvent(TaskArtifactDownloadFailed).SetDownloadError(fmt.Errorf("connection reset by peer")), "connection reset by peer"},
|
|
{NewTaskEvent(TaskRestarting).SetRestartDelay(2 * time.Second).SetRestartReason(ReasonWithinPolicy), "Task restarting in 2s"},
|
|
{NewTaskEvent(TaskRestarting).SetRestartReason("Chaos Monkey did it"), "Chaos Monkey did it - Task restarting in 0s"},
|
|
{NewTaskEvent(TaskKilling), "Sent interrupt"},
|
|
{NewTaskEvent(TaskKilling).SetKillReason("Its time for you to die"), "Its time for you to die"},
|
|
{NewTaskEvent(TaskKilling).SetKillTimeout(1 * time.Second), "Sent interrupt. Waiting 1s before force killing"},
|
|
{NewTaskEvent(TaskTerminated).SetExitCode(-1).SetSignal(3), "Exit Code: -1, Signal: 3"},
|
|
{NewTaskEvent(TaskTerminated).SetMessage("Goodbye"), "Exit Code: 0, Exit Message: \"Goodbye\""},
|
|
{NewTaskEvent(TaskKilled), "Task successfully killed"},
|
|
{NewTaskEvent(TaskKilled).SetKillError(fmt.Errorf("undead creatures can't be killed")), "undead creatures can't be killed"},
|
|
{NewTaskEvent(TaskNotRestarting).SetRestartReason("Chaos Monkey did it"), "Chaos Monkey did it"},
|
|
{NewTaskEvent(TaskNotRestarting), "Task exceeded restart policy"},
|
|
{NewTaskEvent(TaskLeaderDead), "Leader Task in Group dead"},
|
|
{NewTaskEvent(TaskSiblingFailed), "Task's sibling failed"},
|
|
{NewTaskEvent(TaskSiblingFailed).SetFailedSibling("patient zero"), "Task's sibling \"patient zero\" failed"},
|
|
{NewTaskEvent(TaskSignaling), "Task being sent a signal"},
|
|
{NewTaskEvent(TaskSignaling).SetTaskSignal(os.Interrupt), "Task being sent signal interrupt"},
|
|
{NewTaskEvent(TaskSignaling).SetTaskSignal(os.Interrupt).SetTaskSignalReason("process interrupted"), "Task being sent signal interrupt: process interrupted"},
|
|
{NewTaskEvent(TaskRestartSignal), "Task signaled to restart"},
|
|
{NewTaskEvent(TaskRestartSignal).SetRestartReason("Chaos Monkey restarted it"), "Chaos Monkey restarted it"},
|
|
{NewTaskEvent(TaskDriverMessage).SetDriverMessage("YOLO"), "YOLO"},
|
|
{NewTaskEvent("Unknown Type, No message"), ""},
|
|
{NewTaskEvent("Unknown Type").SetMessage("Hello world"), "Hello world"},
|
|
}
|
|
|
|
for _, tc := range testcases {
|
|
tc.event.PopulateEventDisplayMessage()
|
|
if tc.event != nil && tc.event.DisplayMessage != tc.expectedMsg {
|
|
t.Fatalf("Expected %v but got %v", tc.expectedMsg, tc.event.DisplayMessage)
|
|
}
|
|
}
|
|
}
|
|
|
|
func TestNetworkResourcesEquals(t *testing.T) {
|
|
require := require.New(t)
|
|
var networkResourcesTest = []struct {
|
|
input []*NetworkResource
|
|
expected bool
|
|
errorMsg string
|
|
}{
|
|
{
|
|
[]*NetworkResource{
|
|
{
|
|
IP: "10.0.0.1",
|
|
MBits: 50,
|
|
ReservedPorts: []Port{{"web", 80}},
|
|
},
|
|
{
|
|
IP: "10.0.0.1",
|
|
MBits: 50,
|
|
ReservedPorts: []Port{{"web", 80}},
|
|
},
|
|
},
|
|
true,
|
|
"Equal network resources should return true",
|
|
},
|
|
{
|
|
[]*NetworkResource{
|
|
{
|
|
IP: "10.0.0.0",
|
|
MBits: 50,
|
|
ReservedPorts: []Port{{"web", 80}},
|
|
},
|
|
{
|
|
IP: "10.0.0.1",
|
|
MBits: 50,
|
|
ReservedPorts: []Port{{"web", 80}},
|
|
},
|
|
},
|
|
false,
|
|
"Different IP addresses should return false",
|
|
},
|
|
{
|
|
[]*NetworkResource{
|
|
{
|
|
IP: "10.0.0.1",
|
|
MBits: 40,
|
|
ReservedPorts: []Port{{"web", 80}},
|
|
},
|
|
{
|
|
IP: "10.0.0.1",
|
|
MBits: 50,
|
|
ReservedPorts: []Port{{"web", 80}},
|
|
},
|
|
},
|
|
false,
|
|
"Different MBits values should return false",
|
|
},
|
|
{
|
|
[]*NetworkResource{
|
|
{
|
|
IP: "10.0.0.1",
|
|
MBits: 50,
|
|
ReservedPorts: []Port{{"web", 80}},
|
|
},
|
|
{
|
|
IP: "10.0.0.1",
|
|
MBits: 50,
|
|
ReservedPorts: []Port{{"web", 80}, {"web", 80}},
|
|
},
|
|
},
|
|
false,
|
|
"Different ReservedPorts lengths should return false",
|
|
},
|
|
{
|
|
[]*NetworkResource{
|
|
{
|
|
IP: "10.0.0.1",
|
|
MBits: 50,
|
|
ReservedPorts: []Port{{"web", 80}},
|
|
},
|
|
{
|
|
IP: "10.0.0.1",
|
|
MBits: 50,
|
|
ReservedPorts: []Port{},
|
|
},
|
|
},
|
|
false,
|
|
"Empty and non empty ReservedPorts values should return false",
|
|
},
|
|
{
|
|
[]*NetworkResource{
|
|
{
|
|
IP: "10.0.0.1",
|
|
MBits: 50,
|
|
ReservedPorts: []Port{{"web", 80}},
|
|
},
|
|
{
|
|
IP: "10.0.0.1",
|
|
MBits: 50,
|
|
ReservedPorts: []Port{{"notweb", 80}},
|
|
},
|
|
},
|
|
false,
|
|
"Different valued ReservedPorts values should return false",
|
|
},
|
|
{
|
|
[]*NetworkResource{
|
|
{
|
|
IP: "10.0.0.1",
|
|
MBits: 50,
|
|
DynamicPorts: []Port{{"web", 80}},
|
|
},
|
|
{
|
|
IP: "10.0.0.1",
|
|
MBits: 50,
|
|
DynamicPorts: []Port{{"web", 80}, {"web", 80}},
|
|
},
|
|
},
|
|
false,
|
|
"Different DynamicPorts lengths should return false",
|
|
},
|
|
{
|
|
[]*NetworkResource{
|
|
{
|
|
IP: "10.0.0.1",
|
|
MBits: 50,
|
|
DynamicPorts: []Port{{"web", 80}},
|
|
},
|
|
{
|
|
IP: "10.0.0.1",
|
|
MBits: 50,
|
|
DynamicPorts: []Port{},
|
|
},
|
|
},
|
|
false,
|
|
"Empty and non empty DynamicPorts values should return false",
|
|
},
|
|
{
|
|
[]*NetworkResource{
|
|
{
|
|
IP: "10.0.0.1",
|
|
MBits: 50,
|
|
DynamicPorts: []Port{{"web", 80}},
|
|
},
|
|
{
|
|
IP: "10.0.0.1",
|
|
MBits: 50,
|
|
DynamicPorts: []Port{{"notweb", 80}},
|
|
},
|
|
},
|
|
false,
|
|
"Different valued DynamicPorts values should return false",
|
|
},
|
|
}
|
|
for _, testCase := range networkResourcesTest {
|
|
first := testCase.input[0]
|
|
second := testCase.input[1]
|
|
require.Equal(testCase.expected, first.Equals(second), testCase.errorMsg)
|
|
}
|
|
}
|
|
|
|
func TestNode_Canonicalize(t *testing.T) {
|
|
t.Parallel()
|
|
require := require.New(t)
|
|
|
|
// Make sure the eligiblity is set properly
|
|
node := &Node{}
|
|
node.Canonicalize()
|
|
require.Equal(NodeSchedulingEligible, node.SchedulingEligibility)
|
|
|
|
node = &Node{
|
|
Drain: true,
|
|
}
|
|
node.Canonicalize()
|
|
require.Equal(NodeSchedulingIneligible, node.SchedulingEligibility)
|
|
}
|