2017-02-01 00:43:57 +00:00
|
|
|
package consul
|
|
|
|
|
|
|
|
import (
|
|
|
|
"context"
|
|
|
|
"fmt"
|
|
|
|
"io/ioutil"
|
|
|
|
"log"
|
|
|
|
"os"
|
|
|
|
"reflect"
|
2017-08-16 18:27:30 +00:00
|
|
|
"strings"
|
2017-09-14 22:17:21 +00:00
|
|
|
"sync/atomic"
|
2017-02-01 00:43:57 +00:00
|
|
|
"testing"
|
|
|
|
"time"
|
|
|
|
|
|
|
|
"github.com/hashicorp/consul/api"
|
2017-06-13 21:02:11 +00:00
|
|
|
cstructs "github.com/hashicorp/nomad/client/structs"
|
2017-02-01 00:43:57 +00:00
|
|
|
"github.com/hashicorp/nomad/nomad/structs"
|
2017-08-16 18:27:30 +00:00
|
|
|
"github.com/kr/pretty"
|
2017-12-05 19:39:42 +00:00
|
|
|
"github.com/stretchr/testify/assert"
|
2017-02-01 00:43:57 +00:00
|
|
|
)
|
|
|
|
|
2017-04-18 04:15:13 +00:00
|
|
|
const (
|
|
|
|
// Ports used in testTask
|
|
|
|
xPort = 1234
|
|
|
|
yPort = 1235
|
|
|
|
)
|
|
|
|
|
2017-02-01 00:43:57 +00:00
|
|
|
func testLogger() *log.Logger {
|
|
|
|
if testing.Verbose() {
|
|
|
|
return log.New(os.Stderr, "", log.LstdFlags)
|
|
|
|
}
|
|
|
|
return log.New(ioutil.Discard, "", 0)
|
|
|
|
}
|
|
|
|
|
|
|
|
func testTask() *structs.Task {
|
|
|
|
return &structs.Task{
|
|
|
|
Name: "taskname",
|
|
|
|
Resources: &structs.Resources{
|
|
|
|
Networks: []*structs.NetworkResource{
|
|
|
|
{
|
2017-04-18 04:15:13 +00:00
|
|
|
DynamicPorts: []structs.Port{
|
|
|
|
{Label: "x", Value: xPort},
|
|
|
|
{Label: "y", Value: yPort},
|
|
|
|
},
|
2017-02-01 00:43:57 +00:00
|
|
|
},
|
|
|
|
},
|
|
|
|
},
|
|
|
|
Services: []*structs.Service{
|
|
|
|
{
|
|
|
|
Name: "taskname-service",
|
|
|
|
PortLabel: "x",
|
|
|
|
Tags: []string{"tag1", "tag2"},
|
|
|
|
},
|
|
|
|
},
|
|
|
|
}
|
|
|
|
}
|
|
|
|
|
2017-09-14 22:17:21 +00:00
|
|
|
// restartRecorder is a minimal TaskRestarter implementation that simply
|
|
|
|
// counts how many restarts were triggered.
|
|
|
|
type restartRecorder struct {
|
|
|
|
restarts int64
|
2017-09-13 06:15:46 +00:00
|
|
|
}
|
|
|
|
|
2017-09-14 22:17:21 +00:00
|
|
|
func (r *restartRecorder) Restart(source, reason string, failure bool) {
|
|
|
|
atomic.AddInt64(&r.restarts, 1)
|
2017-09-13 06:15:46 +00:00
|
|
|
}
|
|
|
|
|
2017-02-01 00:43:57 +00:00
|
|
|
// testFakeCtx contains a fake Consul AgentAPI and implements the Exec
|
|
|
|
// interface to allow testing without running Consul.
|
|
|
|
type testFakeCtx struct {
|
|
|
|
ServiceClient *ServiceClient
|
2017-09-13 06:15:46 +00:00
|
|
|
FakeConsul *MockAgent
|
2017-02-01 00:43:57 +00:00
|
|
|
Task *structs.Task
|
2017-09-14 22:17:21 +00:00
|
|
|
Restarter *restartRecorder
|
2017-02-01 00:43:57 +00:00
|
|
|
|
2017-04-18 04:15:13 +00:00
|
|
|
// Ticked whenever a script is called
|
|
|
|
execs chan int
|
|
|
|
|
|
|
|
// If non-nil will be called by script checks
|
2017-02-01 00:43:57 +00:00
|
|
|
ExecFunc func(ctx context.Context, cmd string, args []string) ([]byte, int, error)
|
|
|
|
}
|
|
|
|
|
|
|
|
// Exec implements the ScriptExecutor interface and will use an alternate
|
|
|
|
// implementation t.ExecFunc if non-nil.
|
|
|
|
func (t *testFakeCtx) Exec(ctx context.Context, cmd string, args []string) ([]byte, int, error) {
|
2017-04-18 04:15:13 +00:00
|
|
|
select {
|
|
|
|
case t.execs <- 1:
|
|
|
|
default:
|
|
|
|
}
|
2017-02-01 00:43:57 +00:00
|
|
|
if t.ExecFunc == nil {
|
|
|
|
// Default impl is just "ok"
|
|
|
|
return []byte("ok"), 0, nil
|
|
|
|
}
|
|
|
|
return t.ExecFunc(ctx, cmd, args)
|
|
|
|
}
|
|
|
|
|
2017-04-08 00:10:26 +00:00
|
|
|
var errNoOps = fmt.Errorf("testing error: no pending operations")
|
|
|
|
|
|
|
|
// syncOps simulates one iteration of the ServiceClient.Run loop and returns
|
|
|
|
// any errors returned by sync() or errNoOps if no pending operations.
|
|
|
|
func (t *testFakeCtx) syncOnce() error {
|
|
|
|
select {
|
|
|
|
case ops := <-t.ServiceClient.opCh:
|
|
|
|
t.ServiceClient.merge(ops)
|
|
|
|
return t.ServiceClient.sync()
|
|
|
|
default:
|
|
|
|
return errNoOps
|
|
|
|
}
|
|
|
|
}
|
|
|
|
|
2017-02-01 00:43:57 +00:00
|
|
|
// setupFake creates a testFakeCtx with a ServiceClient backed by a fakeConsul.
|
|
|
|
// A test Task is also provided.
|
|
|
|
func setupFake() *testFakeCtx {
|
2017-09-13 06:15:46 +00:00
|
|
|
fc := NewMockAgent()
|
2017-02-01 00:43:57 +00:00
|
|
|
return &testFakeCtx{
|
2017-04-19 19:18:06 +00:00
|
|
|
ServiceClient: NewServiceClient(fc, true, testLogger()),
|
2017-02-01 00:43:57 +00:00
|
|
|
FakeConsul: fc,
|
|
|
|
Task: testTask(),
|
2017-09-14 22:17:21 +00:00
|
|
|
Restarter: &restartRecorder{},
|
2017-04-18 04:15:13 +00:00
|
|
|
execs: make(chan int, 100),
|
2017-02-01 00:43:57 +00:00
|
|
|
}
|
|
|
|
}
|
|
|
|
|
|
|
|
func TestConsul_ChangeTags(t *testing.T) {
|
|
|
|
ctx := setupFake()
|
|
|
|
|
2017-08-07 22:54:05 +00:00
|
|
|
allocID := "allocid"
|
2017-09-13 06:15:46 +00:00
|
|
|
if err := ctx.ServiceClient.RegisterTask(allocID, ctx.Task, ctx.Restarter, nil, nil); err != nil {
|
2017-02-01 00:43:57 +00:00
|
|
|
t.Fatalf("unexpected error registering task: %v", err)
|
|
|
|
}
|
|
|
|
|
2017-04-08 00:10:26 +00:00
|
|
|
if err := ctx.syncOnce(); err != nil {
|
2017-02-01 00:43:57 +00:00
|
|
|
t.Fatalf("unexpected error syncing task: %v", err)
|
|
|
|
}
|
|
|
|
|
|
|
|
if n := len(ctx.FakeConsul.services); n != 1 {
|
|
|
|
t.Fatalf("expected 1 service but found %d:\n%#v", n, ctx.FakeConsul.services)
|
|
|
|
}
|
|
|
|
|
2017-08-07 22:54:05 +00:00
|
|
|
// Query the allocs registrations and then again when we update. The IDs
|
|
|
|
// should change
|
|
|
|
reg1, err := ctx.ServiceClient.AllocRegistrations(allocID)
|
|
|
|
if err != nil {
|
|
|
|
t.Fatalf("Looking up alloc registration failed: %v", err)
|
|
|
|
}
|
|
|
|
if reg1 == nil {
|
|
|
|
t.Fatalf("Nil alloc registrations: %v", err)
|
|
|
|
}
|
|
|
|
if num := reg1.NumServices(); num != 1 {
|
|
|
|
t.Fatalf("Wrong number of servies: got %d; want 1", num)
|
|
|
|
}
|
|
|
|
if num := reg1.NumChecks(); num != 0 {
|
|
|
|
t.Fatalf("Wrong number of checks: got %d; want 0", num)
|
|
|
|
}
|
|
|
|
|
2017-02-01 00:43:57 +00:00
|
|
|
origKey := ""
|
|
|
|
for k, v := range ctx.FakeConsul.services {
|
|
|
|
origKey = k
|
|
|
|
if v.Name != ctx.Task.Services[0].Name {
|
|
|
|
t.Errorf("expected Name=%q != %q", ctx.Task.Services[0].Name, v.Name)
|
|
|
|
}
|
|
|
|
if !reflect.DeepEqual(v.Tags, ctx.Task.Services[0].Tags) {
|
|
|
|
t.Errorf("expected Tags=%v != %v", ctx.Task.Services[0].Tags, v.Tags)
|
|
|
|
}
|
|
|
|
}
|
|
|
|
|
2017-04-08 00:10:26 +00:00
|
|
|
origTask := ctx.Task
|
|
|
|
ctx.Task = testTask()
|
2017-02-01 00:43:57 +00:00
|
|
|
ctx.Task.Services[0].Tags[0] = "newtag"
|
2017-09-13 06:15:46 +00:00
|
|
|
if err := ctx.ServiceClient.UpdateTask("allocid", origTask, ctx.Task, nil, nil, nil); err != nil {
|
2017-02-01 00:43:57 +00:00
|
|
|
t.Fatalf("unexpected error registering task: %v", err)
|
|
|
|
}
|
2017-04-08 00:10:26 +00:00
|
|
|
if err := ctx.syncOnce(); err != nil {
|
2017-02-01 00:43:57 +00:00
|
|
|
t.Fatalf("unexpected error syncing task: %v", err)
|
|
|
|
}
|
|
|
|
|
|
|
|
if n := len(ctx.FakeConsul.services); n != 1 {
|
|
|
|
t.Fatalf("expected 1 service but found %d:\n%#v", n, ctx.FakeConsul.services)
|
|
|
|
}
|
|
|
|
|
|
|
|
for k, v := range ctx.FakeConsul.services {
|
|
|
|
if k == origKey {
|
|
|
|
t.Errorf("expected key to change but found %q", k)
|
|
|
|
}
|
|
|
|
if v.Name != ctx.Task.Services[0].Name {
|
|
|
|
t.Errorf("expected Name=%q != %q", ctx.Task.Services[0].Name, v.Name)
|
|
|
|
}
|
|
|
|
if !reflect.DeepEqual(v.Tags, ctx.Task.Services[0].Tags) {
|
|
|
|
t.Errorf("expected Tags=%v != %v", ctx.Task.Services[0].Tags, v.Tags)
|
|
|
|
}
|
|
|
|
}
|
2017-08-07 22:54:05 +00:00
|
|
|
|
|
|
|
// Check again and ensure the IDs changed
|
|
|
|
reg2, err := ctx.ServiceClient.AllocRegistrations(allocID)
|
|
|
|
if err != nil {
|
|
|
|
t.Fatalf("Looking up alloc registration failed: %v", err)
|
|
|
|
}
|
|
|
|
if reg2 == nil {
|
|
|
|
t.Fatalf("Nil alloc registrations: %v", err)
|
|
|
|
}
|
|
|
|
if num := reg2.NumServices(); num != 1 {
|
|
|
|
t.Fatalf("Wrong number of servies: got %d; want 1", num)
|
|
|
|
}
|
|
|
|
if num := reg2.NumChecks(); num != 0 {
|
|
|
|
t.Fatalf("Wrong number of checks: got %d; want 0", num)
|
|
|
|
}
|
|
|
|
|
|
|
|
for task, treg := range reg1.Tasks {
|
|
|
|
otherTaskReg, ok := reg2.Tasks[task]
|
|
|
|
if !ok {
|
|
|
|
t.Fatalf("Task %q not in second reg", task)
|
|
|
|
}
|
|
|
|
|
|
|
|
for sID := range treg.Services {
|
|
|
|
if _, ok := otherTaskReg.Services[sID]; ok {
|
|
|
|
t.Fatalf("service ID didn't change")
|
|
|
|
}
|
|
|
|
}
|
|
|
|
}
|
2017-02-01 00:43:57 +00:00
|
|
|
}
|
|
|
|
|
2017-04-18 04:15:13 +00:00
|
|
|
// TestConsul_ChangePorts asserts that changing the ports on a service updates
|
2017-12-08 01:08:25 +00:00
|
|
|
// it in Consul. Pre-0.7.1 ports were not part of the service ID and this was a
|
|
|
|
// slightly different code path than changing tags.
|
2017-04-18 04:15:13 +00:00
|
|
|
func TestConsul_ChangePorts(t *testing.T) {
|
|
|
|
ctx := setupFake()
|
|
|
|
ctx.Task.Services[0].Checks = []*structs.ServiceCheck{
|
|
|
|
{
|
|
|
|
Name: "c1",
|
|
|
|
Type: "tcp",
|
|
|
|
Interval: time.Second,
|
|
|
|
Timeout: time.Second,
|
|
|
|
PortLabel: "x",
|
|
|
|
},
|
|
|
|
{
|
|
|
|
Name: "c2",
|
|
|
|
Type: "script",
|
|
|
|
Interval: 9000 * time.Hour,
|
|
|
|
Timeout: time.Second,
|
|
|
|
},
|
|
|
|
{
|
|
|
|
Name: "c3",
|
|
|
|
Type: "http",
|
|
|
|
Protocol: "http",
|
|
|
|
Path: "/",
|
|
|
|
Interval: time.Second,
|
|
|
|
Timeout: time.Second,
|
|
|
|
PortLabel: "y",
|
|
|
|
},
|
|
|
|
}
|
|
|
|
|
2017-09-13 06:15:46 +00:00
|
|
|
if err := ctx.ServiceClient.RegisterTask("allocid", ctx.Task, ctx.Restarter, ctx, nil); err != nil {
|
2017-04-18 04:15:13 +00:00
|
|
|
t.Fatalf("unexpected error registering task: %v", err)
|
|
|
|
}
|
|
|
|
|
|
|
|
if err := ctx.syncOnce(); err != nil {
|
|
|
|
t.Fatalf("unexpected error syncing task: %v", err)
|
|
|
|
}
|
|
|
|
|
|
|
|
if n := len(ctx.FakeConsul.services); n != 1 {
|
|
|
|
t.Fatalf("expected 1 service but found %d:\n%#v", n, ctx.FakeConsul.services)
|
|
|
|
}
|
|
|
|
|
|
|
|
origServiceKey := ""
|
|
|
|
for k, v := range ctx.FakeConsul.services {
|
|
|
|
origServiceKey = k
|
|
|
|
if v.Name != ctx.Task.Services[0].Name {
|
|
|
|
t.Errorf("expected Name=%q != %q", ctx.Task.Services[0].Name, v.Name)
|
|
|
|
}
|
|
|
|
if !reflect.DeepEqual(v.Tags, ctx.Task.Services[0].Tags) {
|
|
|
|
t.Errorf("expected Tags=%v != %v", ctx.Task.Services[0].Tags, v.Tags)
|
|
|
|
}
|
|
|
|
if v.Port != xPort {
|
|
|
|
t.Errorf("expected Port x=%v but found: %v", xPort, v.Port)
|
|
|
|
}
|
|
|
|
}
|
|
|
|
|
|
|
|
if n := len(ctx.FakeConsul.checks); n != 3 {
|
|
|
|
t.Fatalf("expected 3 checks but found %d:\n%#v", n, ctx.FakeConsul.checks)
|
|
|
|
}
|
|
|
|
|
|
|
|
origTCPKey := ""
|
|
|
|
origScriptKey := ""
|
|
|
|
origHTTPKey := ""
|
|
|
|
for k, v := range ctx.FakeConsul.checks {
|
|
|
|
switch v.Name {
|
|
|
|
case "c1":
|
|
|
|
origTCPKey = k
|
|
|
|
if expected := fmt.Sprintf(":%d", xPort); v.TCP != expected {
|
|
|
|
t.Errorf("expected Port x=%v but found: %v", expected, v.TCP)
|
|
|
|
}
|
|
|
|
case "c2":
|
|
|
|
origScriptKey = k
|
|
|
|
select {
|
|
|
|
case <-ctx.execs:
|
|
|
|
if n := len(ctx.execs); n > 0 {
|
|
|
|
t.Errorf("expected 1 exec but found: %d", n+1)
|
|
|
|
}
|
|
|
|
case <-time.After(3 * time.Second):
|
|
|
|
t.Errorf("script not called in time")
|
|
|
|
}
|
|
|
|
case "c3":
|
|
|
|
origHTTPKey = k
|
|
|
|
if expected := fmt.Sprintf("http://:%d/", yPort); v.HTTP != expected {
|
|
|
|
t.Errorf("expected Port y=%v but found: %v", expected, v.HTTP)
|
|
|
|
}
|
|
|
|
default:
|
|
|
|
t.Fatalf("unexpected check: %q", v.Name)
|
|
|
|
}
|
|
|
|
}
|
|
|
|
|
|
|
|
// Now update the PortLabel on the Service and Check c3
|
|
|
|
origTask := ctx.Task
|
|
|
|
ctx.Task = testTask()
|
|
|
|
ctx.Task.Services[0].PortLabel = "y"
|
|
|
|
ctx.Task.Services[0].Checks = []*structs.ServiceCheck{
|
|
|
|
{
|
|
|
|
Name: "c1",
|
|
|
|
Type: "tcp",
|
|
|
|
Interval: time.Second,
|
|
|
|
Timeout: time.Second,
|
|
|
|
PortLabel: "x",
|
|
|
|
},
|
|
|
|
{
|
|
|
|
Name: "c2",
|
|
|
|
Type: "script",
|
|
|
|
Interval: 9000 * time.Hour,
|
|
|
|
Timeout: time.Second,
|
|
|
|
},
|
|
|
|
{
|
|
|
|
Name: "c3",
|
|
|
|
Type: "http",
|
|
|
|
Protocol: "http",
|
|
|
|
Path: "/",
|
|
|
|
Interval: time.Second,
|
|
|
|
Timeout: time.Second,
|
2017-06-13 21:02:11 +00:00
|
|
|
// Removed PortLabel; should default to service's (y)
|
2017-04-18 04:15:13 +00:00
|
|
|
},
|
|
|
|
}
|
2017-09-13 06:15:46 +00:00
|
|
|
if err := ctx.ServiceClient.UpdateTask("allocid", origTask, ctx.Task, nil, ctx, nil); err != nil {
|
2017-04-18 04:15:13 +00:00
|
|
|
t.Fatalf("unexpected error registering task: %v", err)
|
|
|
|
}
|
|
|
|
if err := ctx.syncOnce(); err != nil {
|
|
|
|
t.Fatalf("unexpected error syncing task: %v", err)
|
|
|
|
}
|
|
|
|
|
|
|
|
if n := len(ctx.FakeConsul.services); n != 1 {
|
|
|
|
t.Fatalf("expected 1 service but found %d:\n%#v", n, ctx.FakeConsul.services)
|
|
|
|
}
|
|
|
|
|
|
|
|
for k, v := range ctx.FakeConsul.services {
|
2017-12-08 01:08:25 +00:00
|
|
|
if k == origServiceKey {
|
|
|
|
t.Errorf("expected key change; still: %q", k)
|
2017-04-18 04:15:13 +00:00
|
|
|
}
|
|
|
|
if v.Name != ctx.Task.Services[0].Name {
|
|
|
|
t.Errorf("expected Name=%q != %q", ctx.Task.Services[0].Name, v.Name)
|
|
|
|
}
|
|
|
|
if !reflect.DeepEqual(v.Tags, ctx.Task.Services[0].Tags) {
|
|
|
|
t.Errorf("expected Tags=%v != %v", ctx.Task.Services[0].Tags, v.Tags)
|
|
|
|
}
|
|
|
|
if v.Port != yPort {
|
|
|
|
t.Errorf("expected Port y=%v but found: %v", yPort, v.Port)
|
|
|
|
}
|
|
|
|
}
|
|
|
|
|
|
|
|
if n := len(ctx.FakeConsul.checks); n != 3 {
|
|
|
|
t.Fatalf("expected 3 check but found %d:\n%#v", n, ctx.FakeConsul.checks)
|
|
|
|
}
|
|
|
|
|
|
|
|
for k, v := range ctx.FakeConsul.checks {
|
|
|
|
switch v.Name {
|
|
|
|
case "c1":
|
2017-12-08 01:08:25 +00:00
|
|
|
if k == origTCPKey {
|
|
|
|
t.Errorf("expected key change for %s from %q", v.Name, origTCPKey)
|
2017-04-18 04:15:13 +00:00
|
|
|
}
|
|
|
|
if expected := fmt.Sprintf(":%d", xPort); v.TCP != expected {
|
|
|
|
t.Errorf("expected Port x=%v but found: %v", expected, v.TCP)
|
|
|
|
}
|
|
|
|
case "c2":
|
2017-12-08 01:08:25 +00:00
|
|
|
if k == origScriptKey {
|
|
|
|
t.Errorf("expected key change for %s from %q", v.Name, origScriptKey)
|
2017-04-18 04:15:13 +00:00
|
|
|
}
|
|
|
|
select {
|
|
|
|
case <-ctx.execs:
|
|
|
|
if n := len(ctx.execs); n > 0 {
|
|
|
|
t.Errorf("expected 1 exec but found: %d", n+1)
|
|
|
|
}
|
|
|
|
case <-time.After(3 * time.Second):
|
|
|
|
t.Errorf("script not called in time")
|
|
|
|
}
|
|
|
|
case "c3":
|
|
|
|
if k == origHTTPKey {
|
|
|
|
t.Errorf("expected %s key to change from %q", v.Name, k)
|
|
|
|
}
|
|
|
|
if expected := fmt.Sprintf("http://:%d/", yPort); v.HTTP != expected {
|
|
|
|
t.Errorf("expected Port y=%v but found: %v", expected, v.HTTP)
|
|
|
|
}
|
|
|
|
default:
|
2017-08-07 21:13:05 +00:00
|
|
|
t.Errorf("Unknown check: %q", k)
|
2017-04-18 04:15:13 +00:00
|
|
|
}
|
|
|
|
}
|
|
|
|
}
|
|
|
|
|
2017-06-16 23:35:16 +00:00
|
|
|
// TestConsul_ChangeChecks asserts that updating only the checks on a service
|
|
|
|
// properly syncs with Consul.
|
|
|
|
func TestConsul_ChangeChecks(t *testing.T) {
|
|
|
|
ctx := setupFake()
|
|
|
|
ctx.Task.Services[0].Checks = []*structs.ServiceCheck{
|
|
|
|
{
|
|
|
|
Name: "c1",
|
|
|
|
Type: "tcp",
|
|
|
|
Interval: time.Second,
|
|
|
|
Timeout: time.Second,
|
|
|
|
PortLabel: "x",
|
2017-09-14 22:17:21 +00:00
|
|
|
CheckRestart: &structs.CheckRestart{
|
|
|
|
Limit: 3,
|
|
|
|
},
|
2017-06-16 23:35:16 +00:00
|
|
|
},
|
|
|
|
}
|
|
|
|
|
2017-08-07 22:54:05 +00:00
|
|
|
allocID := "allocid"
|
2017-09-13 06:15:46 +00:00
|
|
|
if err := ctx.ServiceClient.RegisterTask(allocID, ctx.Task, ctx.Restarter, ctx, nil); err != nil {
|
2017-06-16 23:35:16 +00:00
|
|
|
t.Fatalf("unexpected error registering task: %v", err)
|
|
|
|
}
|
|
|
|
|
|
|
|
if err := ctx.syncOnce(); err != nil {
|
|
|
|
t.Fatalf("unexpected error syncing task: %v", err)
|
|
|
|
}
|
|
|
|
|
|
|
|
if n := len(ctx.FakeConsul.services); n != 1 {
|
|
|
|
t.Fatalf("expected 1 service but found %d:\n%#v", n, ctx.FakeConsul.services)
|
|
|
|
}
|
|
|
|
|
2017-09-14 22:17:21 +00:00
|
|
|
// Assert a check restart watch update was enqueued and clear it
|
|
|
|
if n := len(ctx.ServiceClient.checkWatcher.checkUpdateCh); n != 1 {
|
|
|
|
t.Fatalf("expected 1 check restart update but found %d", n)
|
|
|
|
}
|
|
|
|
upd := <-ctx.ServiceClient.checkWatcher.checkUpdateCh
|
|
|
|
c1ID := upd.checkID
|
|
|
|
|
2017-08-07 22:54:05 +00:00
|
|
|
// Query the allocs registrations and then again when we update. The IDs
|
|
|
|
// should change
|
|
|
|
reg1, err := ctx.ServiceClient.AllocRegistrations(allocID)
|
|
|
|
if err != nil {
|
|
|
|
t.Fatalf("Looking up alloc registration failed: %v", err)
|
|
|
|
}
|
|
|
|
if reg1 == nil {
|
|
|
|
t.Fatalf("Nil alloc registrations: %v", err)
|
|
|
|
}
|
|
|
|
if num := reg1.NumServices(); num != 1 {
|
|
|
|
t.Fatalf("Wrong number of servies: got %d; want 1", num)
|
|
|
|
}
|
|
|
|
if num := reg1.NumChecks(); num != 1 {
|
|
|
|
t.Fatalf("Wrong number of checks: got %d; want 1", num)
|
|
|
|
}
|
|
|
|
|
2017-06-16 23:35:16 +00:00
|
|
|
origServiceKey := ""
|
|
|
|
for k, v := range ctx.FakeConsul.services {
|
|
|
|
origServiceKey = k
|
|
|
|
if v.Name != ctx.Task.Services[0].Name {
|
|
|
|
t.Errorf("expected Name=%q != %q", ctx.Task.Services[0].Name, v.Name)
|
|
|
|
}
|
|
|
|
if v.Port != xPort {
|
|
|
|
t.Errorf("expected Port x=%v but found: %v", xPort, v.Port)
|
|
|
|
}
|
|
|
|
}
|
|
|
|
|
|
|
|
if n := len(ctx.FakeConsul.checks); n != 1 {
|
|
|
|
t.Fatalf("expected 1 check but found %d:\n%#v", n, ctx.FakeConsul.checks)
|
|
|
|
}
|
|
|
|
for _, v := range ctx.FakeConsul.checks {
|
|
|
|
if v.Name != "c1" {
|
|
|
|
t.Fatalf("expected check c1 but found %q", v.Name)
|
|
|
|
}
|
|
|
|
}
|
|
|
|
|
2017-08-07 22:54:05 +00:00
|
|
|
// Now add a check and modify the original
|
2017-06-16 23:35:16 +00:00
|
|
|
origTask := ctx.Task.Copy()
|
|
|
|
ctx.Task.Services[0].Checks = []*structs.ServiceCheck{
|
|
|
|
{
|
|
|
|
Name: "c1",
|
|
|
|
Type: "tcp",
|
2017-08-07 22:54:05 +00:00
|
|
|
Interval: 2 * time.Second,
|
2017-06-16 23:35:16 +00:00
|
|
|
Timeout: time.Second,
|
|
|
|
PortLabel: "x",
|
2017-09-14 22:17:21 +00:00
|
|
|
CheckRestart: &structs.CheckRestart{
|
|
|
|
Limit: 3,
|
|
|
|
},
|
2017-06-16 23:35:16 +00:00
|
|
|
},
|
|
|
|
{
|
|
|
|
Name: "c2",
|
|
|
|
Type: "http",
|
|
|
|
Path: "/",
|
|
|
|
Interval: time.Second,
|
|
|
|
Timeout: time.Second,
|
|
|
|
PortLabel: "x",
|
|
|
|
},
|
|
|
|
}
|
2017-09-13 06:15:46 +00:00
|
|
|
if err := ctx.ServiceClient.UpdateTask("allocid", origTask, ctx.Task, nil, ctx, nil); err != nil {
|
2017-06-16 23:35:16 +00:00
|
|
|
t.Fatalf("unexpected error registering task: %v", err)
|
|
|
|
}
|
2017-09-14 22:17:21 +00:00
|
|
|
|
|
|
|
// Assert 2 check restart watch updates was enqueued
|
|
|
|
if n := len(ctx.ServiceClient.checkWatcher.checkUpdateCh); n != 2 {
|
|
|
|
t.Fatalf("expected 2 check restart updates but found %d", n)
|
|
|
|
}
|
|
|
|
|
|
|
|
// First the new watch
|
|
|
|
upd = <-ctx.ServiceClient.checkWatcher.checkUpdateCh
|
|
|
|
if upd.checkID == c1ID || upd.remove {
|
|
|
|
t.Fatalf("expected check watch update to be an add of checkID=%q but found remove=%t checkID=%q",
|
|
|
|
c1ID, upd.remove, upd.checkID)
|
|
|
|
}
|
|
|
|
|
|
|
|
// Then remove the old watch
|
|
|
|
upd = <-ctx.ServiceClient.checkWatcher.checkUpdateCh
|
|
|
|
if upd.checkID != c1ID || !upd.remove {
|
|
|
|
t.Fatalf("expected check watch update to be a removal of checkID=%q but found remove=%t checkID=%q",
|
|
|
|
c1ID, upd.remove, upd.checkID)
|
|
|
|
}
|
|
|
|
|
2017-06-16 23:35:16 +00:00
|
|
|
if err := ctx.syncOnce(); err != nil {
|
|
|
|
t.Fatalf("unexpected error syncing task: %v", err)
|
|
|
|
}
|
|
|
|
|
|
|
|
if n := len(ctx.FakeConsul.services); n != 1 {
|
|
|
|
t.Fatalf("expected 1 service but found %d:\n%#v", n, ctx.FakeConsul.services)
|
|
|
|
}
|
|
|
|
|
|
|
|
if _, ok := ctx.FakeConsul.services[origServiceKey]; !ok {
|
|
|
|
t.Errorf("unexpected key change; was: %q -- but found %#v", origServiceKey, ctx.FakeConsul.services)
|
|
|
|
}
|
|
|
|
|
|
|
|
if n := len(ctx.FakeConsul.checks); n != 2 {
|
|
|
|
t.Fatalf("expected 2 check but found %d:\n%#v", n, ctx.FakeConsul.checks)
|
|
|
|
}
|
|
|
|
|
|
|
|
for k, v := range ctx.FakeConsul.checks {
|
|
|
|
switch v.Name {
|
|
|
|
case "c1":
|
|
|
|
if expected := fmt.Sprintf(":%d", xPort); v.TCP != expected {
|
|
|
|
t.Errorf("expected Port x=%v but found: %v", expected, v.TCP)
|
|
|
|
}
|
2017-09-14 22:17:21 +00:00
|
|
|
|
|
|
|
// update id
|
|
|
|
c1ID = k
|
2017-06-16 23:35:16 +00:00
|
|
|
case "c2":
|
|
|
|
if expected := fmt.Sprintf("http://:%d/", xPort); v.HTTP != expected {
|
|
|
|
t.Errorf("expected Port x=%v but found: %v", expected, v.HTTP)
|
|
|
|
}
|
|
|
|
default:
|
2017-08-07 21:13:05 +00:00
|
|
|
t.Errorf("Unknown check: %q", k)
|
2017-06-16 23:35:16 +00:00
|
|
|
}
|
|
|
|
}
|
2017-08-07 22:54:05 +00:00
|
|
|
|
|
|
|
// Check again and ensure the IDs changed
|
|
|
|
reg2, err := ctx.ServiceClient.AllocRegistrations(allocID)
|
|
|
|
if err != nil {
|
|
|
|
t.Fatalf("Looking up alloc registration failed: %v", err)
|
|
|
|
}
|
|
|
|
if reg2 == nil {
|
|
|
|
t.Fatalf("Nil alloc registrations: %v", err)
|
|
|
|
}
|
|
|
|
if num := reg2.NumServices(); num != 1 {
|
|
|
|
t.Fatalf("Wrong number of servies: got %d; want 1", num)
|
|
|
|
}
|
|
|
|
if num := reg2.NumChecks(); num != 2 {
|
|
|
|
t.Fatalf("Wrong number of checks: got %d; want 2", num)
|
|
|
|
}
|
|
|
|
|
|
|
|
for task, treg := range reg1.Tasks {
|
|
|
|
otherTaskReg, ok := reg2.Tasks[task]
|
|
|
|
if !ok {
|
|
|
|
t.Fatalf("Task %q not in second reg", task)
|
|
|
|
}
|
|
|
|
|
|
|
|
for sID, sreg := range treg.Services {
|
|
|
|
otherServiceReg, ok := otherTaskReg.Services[sID]
|
|
|
|
if !ok {
|
|
|
|
t.Fatalf("service ID changed")
|
|
|
|
}
|
|
|
|
|
|
|
|
for newID := range sreg.checkIDs {
|
|
|
|
if _, ok := otherServiceReg.checkIDs[newID]; ok {
|
|
|
|
t.Fatalf("check IDs should change")
|
|
|
|
}
|
|
|
|
}
|
|
|
|
}
|
|
|
|
}
|
2017-09-14 22:17:21 +00:00
|
|
|
|
|
|
|
// Alter a CheckRestart and make sure the watcher is updated but nothing else
|
|
|
|
origTask = ctx.Task.Copy()
|
|
|
|
ctx.Task.Services[0].Checks = []*structs.ServiceCheck{
|
|
|
|
{
|
|
|
|
Name: "c1",
|
|
|
|
Type: "tcp",
|
|
|
|
Interval: 2 * time.Second,
|
|
|
|
Timeout: time.Second,
|
|
|
|
PortLabel: "x",
|
|
|
|
CheckRestart: &structs.CheckRestart{
|
|
|
|
Limit: 11,
|
|
|
|
},
|
|
|
|
},
|
|
|
|
{
|
|
|
|
Name: "c2",
|
|
|
|
Type: "http",
|
|
|
|
Path: "/",
|
|
|
|
Interval: time.Second,
|
|
|
|
Timeout: time.Second,
|
|
|
|
PortLabel: "x",
|
|
|
|
},
|
|
|
|
}
|
|
|
|
if err := ctx.ServiceClient.UpdateTask("allocid", origTask, ctx.Task, nil, ctx, nil); err != nil {
|
|
|
|
t.Fatalf("unexpected error registering task: %v", err)
|
|
|
|
}
|
|
|
|
if err := ctx.syncOnce(); err != nil {
|
|
|
|
t.Fatalf("unexpected error syncing task: %v", err)
|
|
|
|
}
|
|
|
|
|
|
|
|
if n := len(ctx.FakeConsul.checks); n != 2 {
|
|
|
|
t.Fatalf("expected 2 check but found %d:\n%#v", n, ctx.FakeConsul.checks)
|
|
|
|
}
|
|
|
|
|
|
|
|
for k, v := range ctx.FakeConsul.checks {
|
|
|
|
if v.Name == "c1" {
|
|
|
|
if k != c1ID {
|
|
|
|
t.Errorf("expected c1 to still have id %q but found %q", c1ID, k)
|
|
|
|
}
|
|
|
|
break
|
|
|
|
}
|
|
|
|
}
|
|
|
|
|
|
|
|
// Assert a check restart watch update was enqueued for a removal and an add
|
|
|
|
if n := len(ctx.ServiceClient.checkWatcher.checkUpdateCh); n != 1 {
|
|
|
|
t.Fatalf("expected 1 check restart update but found %d", n)
|
|
|
|
}
|
|
|
|
<-ctx.ServiceClient.checkWatcher.checkUpdateCh
|
2017-06-16 23:35:16 +00:00
|
|
|
}
|
|
|
|
|
2017-02-01 00:43:57 +00:00
|
|
|
// TestConsul_RegServices tests basic service registration.
|
|
|
|
func TestConsul_RegServices(t *testing.T) {
|
|
|
|
ctx := setupFake()
|
|
|
|
|
2017-09-14 22:17:21 +00:00
|
|
|
// Add a check w/restarting
|
|
|
|
ctx.Task.Services[0].Checks = []*structs.ServiceCheck{
|
|
|
|
{
|
|
|
|
Name: "testcheck",
|
|
|
|
Type: "tcp",
|
|
|
|
Interval: 100 * time.Millisecond,
|
|
|
|
CheckRestart: &structs.CheckRestart{
|
|
|
|
Limit: 3,
|
|
|
|
},
|
|
|
|
},
|
|
|
|
}
|
|
|
|
|
2017-09-13 06:15:46 +00:00
|
|
|
if err := ctx.ServiceClient.RegisterTask("allocid", ctx.Task, ctx.Restarter, nil, nil); err != nil {
|
2017-02-01 00:43:57 +00:00
|
|
|
t.Fatalf("unexpected error registering task: %v", err)
|
|
|
|
}
|
|
|
|
|
2017-04-08 00:10:26 +00:00
|
|
|
if err := ctx.syncOnce(); err != nil {
|
2017-02-01 00:43:57 +00:00
|
|
|
t.Fatalf("unexpected error syncing task: %v", err)
|
|
|
|
}
|
|
|
|
|
|
|
|
if n := len(ctx.FakeConsul.services); n != 1 {
|
|
|
|
t.Fatalf("expected 1 service but found %d:\n%#v", n, ctx.FakeConsul.services)
|
|
|
|
}
|
2017-09-14 22:17:21 +00:00
|
|
|
|
2017-02-01 00:43:57 +00:00
|
|
|
for _, v := range ctx.FakeConsul.services {
|
|
|
|
if v.Name != ctx.Task.Services[0].Name {
|
|
|
|
t.Errorf("expected Name=%q != %q", ctx.Task.Services[0].Name, v.Name)
|
|
|
|
}
|
|
|
|
if !reflect.DeepEqual(v.Tags, ctx.Task.Services[0].Tags) {
|
|
|
|
t.Errorf("expected Tags=%v != %v", ctx.Task.Services[0].Tags, v.Tags)
|
|
|
|
}
|
2017-04-19 19:18:06 +00:00
|
|
|
if v.Port != xPort {
|
|
|
|
t.Errorf("expected Port=%d != %d", xPort, v.Port)
|
2017-02-01 00:43:57 +00:00
|
|
|
}
|
|
|
|
}
|
|
|
|
|
2017-09-14 22:17:21 +00:00
|
|
|
// Assert the check update is pending
|
|
|
|
if n := len(ctx.ServiceClient.checkWatcher.checkUpdateCh); n != 1 {
|
|
|
|
t.Fatalf("expected 1 check restart update but found %d", n)
|
|
|
|
}
|
|
|
|
|
|
|
|
// Assert the check update is properly formed
|
|
|
|
checkUpd := <-ctx.ServiceClient.checkWatcher.checkUpdateCh
|
|
|
|
if checkUpd.checkRestart.allocID != "allocid" {
|
|
|
|
t.Fatalf("expected check's allocid to be %q but found %q", "allocid", checkUpd.checkRestart.allocID)
|
|
|
|
}
|
|
|
|
if expected := 200 * time.Millisecond; checkUpd.checkRestart.timeLimit != expected {
|
|
|
|
t.Fatalf("expected check's time limit to be %v but found %v", expected, checkUpd.checkRestart.timeLimit)
|
|
|
|
}
|
|
|
|
|
2017-02-01 00:43:57 +00:00
|
|
|
// Make a change which will register a new service
|
|
|
|
ctx.Task.Services[0].Name = "taskname-service2"
|
|
|
|
ctx.Task.Services[0].Tags[0] = "tag3"
|
2017-09-13 06:15:46 +00:00
|
|
|
if err := ctx.ServiceClient.RegisterTask("allocid", ctx.Task, ctx.Restarter, nil, nil); err != nil {
|
2017-02-01 00:43:57 +00:00
|
|
|
t.Fatalf("unpexpected error registering task: %v", err)
|
|
|
|
}
|
|
|
|
|
2017-09-14 22:17:21 +00:00
|
|
|
// Assert check update is pending
|
|
|
|
if n := len(ctx.ServiceClient.checkWatcher.checkUpdateCh); n != 1 {
|
|
|
|
t.Fatalf("expected 1 check restart update but found %d", n)
|
|
|
|
}
|
|
|
|
|
|
|
|
// Assert the check update's id has changed
|
|
|
|
checkUpd2 := <-ctx.ServiceClient.checkWatcher.checkUpdateCh
|
|
|
|
if checkUpd.checkID == checkUpd2.checkID {
|
|
|
|
t.Fatalf("expected new check update to have a new ID both both have: %q", checkUpd.checkID)
|
|
|
|
}
|
|
|
|
|
2017-02-01 00:43:57 +00:00
|
|
|
// Make sure changes don't take affect until sync() is called (since
|
|
|
|
// Run() isn't running)
|
|
|
|
if n := len(ctx.FakeConsul.services); n != 1 {
|
|
|
|
t.Fatalf("expected 1 service but found %d:\n%#v", n, ctx.FakeConsul.services)
|
|
|
|
}
|
|
|
|
for _, v := range ctx.FakeConsul.services {
|
|
|
|
if reflect.DeepEqual(v.Tags, ctx.Task.Services[0].Tags) {
|
|
|
|
t.Errorf("expected Tags to differ, changes applied before sync()")
|
|
|
|
}
|
|
|
|
}
|
|
|
|
|
|
|
|
// Now sync() and re-check for the applied updates
|
2017-04-08 00:10:26 +00:00
|
|
|
if err := ctx.syncOnce(); err != nil {
|
2017-02-01 00:43:57 +00:00
|
|
|
t.Fatalf("unexpected error syncing task: %v", err)
|
|
|
|
}
|
|
|
|
if n := len(ctx.FakeConsul.services); n != 2 {
|
|
|
|
t.Fatalf("expected 2 services but found %d:\n%#v", n, ctx.FakeConsul.services)
|
|
|
|
}
|
|
|
|
found := false
|
|
|
|
for _, v := range ctx.FakeConsul.services {
|
|
|
|
if v.Name == ctx.Task.Services[0].Name {
|
|
|
|
if found {
|
|
|
|
t.Fatalf("found new service name %q twice", v.Name)
|
|
|
|
}
|
|
|
|
found = true
|
|
|
|
if !reflect.DeepEqual(v.Tags, ctx.Task.Services[0].Tags) {
|
|
|
|
t.Errorf("expected Tags=%v != %v", ctx.Task.Services[0].Tags, v.Tags)
|
|
|
|
}
|
|
|
|
}
|
|
|
|
}
|
|
|
|
if !found {
|
|
|
|
t.Fatalf("did not find new service %q", ctx.Task.Services[0].Name)
|
|
|
|
}
|
|
|
|
|
|
|
|
// Remove the new task
|
|
|
|
ctx.ServiceClient.RemoveTask("allocid", ctx.Task)
|
2017-04-08 00:10:26 +00:00
|
|
|
if err := ctx.syncOnce(); err != nil {
|
2017-02-01 00:43:57 +00:00
|
|
|
t.Fatalf("unexpected error syncing task: %v", err)
|
|
|
|
}
|
|
|
|
if n := len(ctx.FakeConsul.services); n != 1 {
|
|
|
|
t.Fatalf("expected 1 service but found %d:\n%#v", n, ctx.FakeConsul.services)
|
|
|
|
}
|
|
|
|
for _, v := range ctx.FakeConsul.services {
|
|
|
|
if v.Name != "taskname-service" {
|
|
|
|
t.Errorf("expected original task to survive not %q", v.Name)
|
|
|
|
}
|
|
|
|
}
|
2017-09-14 22:17:21 +00:00
|
|
|
|
|
|
|
// Assert check update is pending
|
|
|
|
if n := len(ctx.ServiceClient.checkWatcher.checkUpdateCh); n != 1 {
|
|
|
|
t.Fatalf("expected 1 check restart update but found %d", n)
|
|
|
|
}
|
|
|
|
|
|
|
|
// Assert the check update's id is correct and that it's a removal
|
|
|
|
checkUpd3 := <-ctx.ServiceClient.checkWatcher.checkUpdateCh
|
|
|
|
if checkUpd2.checkID != checkUpd3.checkID {
|
|
|
|
t.Fatalf("expected checkid %q but found %q", checkUpd2.checkID, checkUpd3.checkID)
|
|
|
|
}
|
|
|
|
if !checkUpd3.remove {
|
|
|
|
t.Fatalf("expected check watch removal update but found: %#v", checkUpd3)
|
|
|
|
}
|
2017-02-01 00:43:57 +00:00
|
|
|
}
|
|
|
|
|
|
|
|
// TestConsul_ShutdownOK tests the ok path for the shutdown logic in
|
|
|
|
// ServiceClient.
|
|
|
|
func TestConsul_ShutdownOK(t *testing.T) {
|
|
|
|
ctx := setupFake()
|
|
|
|
|
|
|
|
// Add a script check to make sure its TTL gets updated
|
|
|
|
ctx.Task.Services[0].Checks = []*structs.ServiceCheck{
|
|
|
|
{
|
|
|
|
Name: "scriptcheck",
|
|
|
|
Type: "script",
|
|
|
|
Command: "true",
|
|
|
|
// Make check block until shutdown
|
|
|
|
Interval: 9000 * time.Hour,
|
|
|
|
Timeout: 10 * time.Second,
|
|
|
|
InitialStatus: "warning",
|
|
|
|
},
|
|
|
|
}
|
|
|
|
|
2017-04-08 00:10:26 +00:00
|
|
|
go ctx.ServiceClient.Run()
|
2017-02-01 00:43:57 +00:00
|
|
|
|
|
|
|
// Register a task and agent
|
2017-09-13 06:15:46 +00:00
|
|
|
if err := ctx.ServiceClient.RegisterTask("allocid", ctx.Task, ctx.Restarter, ctx, nil); err != nil {
|
2017-02-01 00:43:57 +00:00
|
|
|
t.Fatalf("unexpected error registering task: %v", err)
|
|
|
|
}
|
|
|
|
|
|
|
|
agentServices := []*structs.Service{
|
|
|
|
{
|
|
|
|
Name: "http",
|
|
|
|
Tags: []string{"nomad"},
|
|
|
|
PortLabel: "localhost:2345",
|
|
|
|
},
|
|
|
|
}
|
|
|
|
if err := ctx.ServiceClient.RegisterAgent("client", agentServices); err != nil {
|
|
|
|
t.Fatalf("unexpected error registering agent: %v", err)
|
|
|
|
}
|
|
|
|
|
2017-04-08 00:10:26 +00:00
|
|
|
// Shutdown should block until scripts finish
|
2017-02-01 00:43:57 +00:00
|
|
|
if err := ctx.ServiceClient.Shutdown(); err != nil {
|
|
|
|
t.Errorf("unexpected error shutting down client: %v", err)
|
|
|
|
}
|
|
|
|
|
|
|
|
// UpdateTTL should have been called once for the script check
|
|
|
|
if n := len(ctx.FakeConsul.checkTTLs); n != 1 {
|
|
|
|
t.Fatalf("expected 1 checkTTL entry but found: %d", n)
|
|
|
|
}
|
|
|
|
for _, v := range ctx.FakeConsul.checkTTLs {
|
|
|
|
if v != 1 {
|
|
|
|
t.Fatalf("expected script check to be updated once but found %d", v)
|
|
|
|
}
|
|
|
|
}
|
|
|
|
for _, v := range ctx.FakeConsul.checks {
|
|
|
|
if v.Status != "passing" {
|
|
|
|
t.Fatalf("expected check to be passing but found %q", v.Status)
|
|
|
|
}
|
|
|
|
}
|
|
|
|
}
|
|
|
|
|
|
|
|
// TestConsul_ShutdownSlow tests the slow but ok path for the shutdown logic in
|
|
|
|
// ServiceClient.
|
|
|
|
func TestConsul_ShutdownSlow(t *testing.T) {
|
|
|
|
t.Parallel() // run the slow tests in parallel
|
|
|
|
ctx := setupFake()
|
|
|
|
|
|
|
|
// Add a script check to make sure its TTL gets updated
|
|
|
|
ctx.Task.Services[0].Checks = []*structs.ServiceCheck{
|
|
|
|
{
|
|
|
|
Name: "scriptcheck",
|
|
|
|
Type: "script",
|
|
|
|
Command: "true",
|
|
|
|
// Make check block until shutdown
|
|
|
|
Interval: 9000 * time.Hour,
|
|
|
|
Timeout: 5 * time.Second,
|
|
|
|
InitialStatus: "warning",
|
|
|
|
},
|
|
|
|
}
|
|
|
|
|
|
|
|
// Make Exec slow, but not too slow
|
2017-04-08 00:10:26 +00:00
|
|
|
waiter := make(chan struct{})
|
2017-02-01 00:43:57 +00:00
|
|
|
ctx.ExecFunc = func(ctx context.Context, cmd string, args []string) ([]byte, int, error) {
|
2017-04-08 00:10:26 +00:00
|
|
|
select {
|
|
|
|
case <-waiter:
|
|
|
|
default:
|
|
|
|
close(waiter)
|
|
|
|
}
|
2017-02-01 00:43:57 +00:00
|
|
|
time.Sleep(time.Second)
|
|
|
|
return []byte{}, 0, nil
|
|
|
|
}
|
|
|
|
|
|
|
|
// Make shutdown wait time just a bit longer than ctx.Exec takes
|
|
|
|
ctx.ServiceClient.shutdownWait = 3 * time.Second
|
|
|
|
|
2017-04-08 00:10:26 +00:00
|
|
|
go ctx.ServiceClient.Run()
|
2017-02-01 00:43:57 +00:00
|
|
|
|
|
|
|
// Register a task and agent
|
2017-09-13 06:15:46 +00:00
|
|
|
if err := ctx.ServiceClient.RegisterTask("allocid", ctx.Task, ctx.Restarter, ctx, nil); err != nil {
|
2017-02-01 00:43:57 +00:00
|
|
|
t.Fatalf("unexpected error registering task: %v", err)
|
|
|
|
}
|
|
|
|
|
2017-04-08 00:10:26 +00:00
|
|
|
// wait for Exec to get called before shutting down
|
|
|
|
<-waiter
|
|
|
|
|
2017-02-01 00:43:57 +00:00
|
|
|
// Shutdown should block until all enqueued operations finish.
|
|
|
|
preShutdown := time.Now()
|
|
|
|
if err := ctx.ServiceClient.Shutdown(); err != nil {
|
|
|
|
t.Errorf("unexpected error shutting down client: %v", err)
|
|
|
|
}
|
|
|
|
|
|
|
|
// Shutdown time should have taken: 1s <= shutdown <= 3s
|
|
|
|
shutdownTime := time.Now().Sub(preShutdown)
|
|
|
|
if shutdownTime < time.Second || shutdownTime > ctx.ServiceClient.shutdownWait {
|
|
|
|
t.Errorf("expected shutdown to take >1s and <%s but took: %s", ctx.ServiceClient.shutdownWait, shutdownTime)
|
|
|
|
}
|
|
|
|
|
|
|
|
// UpdateTTL should have been called once for the script check
|
|
|
|
if n := len(ctx.FakeConsul.checkTTLs); n != 1 {
|
|
|
|
t.Fatalf("expected 1 checkTTL entry but found: %d", n)
|
|
|
|
}
|
|
|
|
for _, v := range ctx.FakeConsul.checkTTLs {
|
|
|
|
if v != 1 {
|
|
|
|
t.Fatalf("expected script check to be updated once but found %d", v)
|
|
|
|
}
|
|
|
|
}
|
|
|
|
for _, v := range ctx.FakeConsul.checks {
|
|
|
|
if v.Status != "passing" {
|
|
|
|
t.Fatalf("expected check to be passing but found %q", v.Status)
|
|
|
|
}
|
|
|
|
}
|
|
|
|
}
|
|
|
|
|
|
|
|
// TestConsul_ShutdownBlocked tests the blocked past deadline path for the
|
|
|
|
// shutdown logic in ServiceClient.
|
|
|
|
func TestConsul_ShutdownBlocked(t *testing.T) {
|
|
|
|
t.Parallel() // run the slow tests in parallel
|
|
|
|
ctx := setupFake()
|
|
|
|
|
|
|
|
// Add a script check to make sure its TTL gets updated
|
|
|
|
ctx.Task.Services[0].Checks = []*structs.ServiceCheck{
|
|
|
|
{
|
|
|
|
Name: "scriptcheck",
|
|
|
|
Type: "script",
|
|
|
|
Command: "true",
|
|
|
|
// Make check block until shutdown
|
|
|
|
Interval: 9000 * time.Hour,
|
|
|
|
Timeout: 9000 * time.Hour,
|
|
|
|
InitialStatus: "warning",
|
|
|
|
},
|
|
|
|
}
|
|
|
|
|
|
|
|
block := make(chan struct{})
|
|
|
|
defer close(block) // cleanup after test
|
|
|
|
|
2017-04-08 00:10:26 +00:00
|
|
|
// Make Exec block forever
|
|
|
|
waiter := make(chan struct{})
|
2017-02-01 00:43:57 +00:00
|
|
|
ctx.ExecFunc = func(ctx context.Context, cmd string, args []string) ([]byte, int, error) {
|
2017-04-08 00:10:26 +00:00
|
|
|
close(waiter)
|
2017-02-01 00:43:57 +00:00
|
|
|
<-block
|
|
|
|
return []byte{}, 0, nil
|
|
|
|
}
|
|
|
|
|
|
|
|
// Use a short shutdown deadline since we're intentionally blocking forever
|
|
|
|
ctx.ServiceClient.shutdownWait = time.Second
|
|
|
|
|
2017-04-08 00:10:26 +00:00
|
|
|
go ctx.ServiceClient.Run()
|
2017-02-01 00:43:57 +00:00
|
|
|
|
|
|
|
// Register a task and agent
|
2017-09-13 06:15:46 +00:00
|
|
|
if err := ctx.ServiceClient.RegisterTask("allocid", ctx.Task, ctx.Restarter, ctx, nil); err != nil {
|
2017-02-01 00:43:57 +00:00
|
|
|
t.Fatalf("unexpected error registering task: %v", err)
|
|
|
|
}
|
|
|
|
|
2017-04-08 00:10:26 +00:00
|
|
|
// Wait for exec to be called
|
|
|
|
<-waiter
|
|
|
|
|
2017-02-01 00:43:57 +00:00
|
|
|
// Shutdown should block until all enqueued operations finish.
|
|
|
|
preShutdown := time.Now()
|
|
|
|
err := ctx.ServiceClient.Shutdown()
|
|
|
|
if err == nil {
|
|
|
|
t.Errorf("expected a timed out error from shutdown")
|
|
|
|
}
|
|
|
|
|
2017-04-08 00:10:26 +00:00
|
|
|
// Shutdown time should have taken shutdownWait; to avoid timing
|
|
|
|
// related errors simply test for wait <= shutdown <= wait+3s
|
2017-02-01 00:43:57 +00:00
|
|
|
shutdownTime := time.Now().Sub(preShutdown)
|
2017-04-08 00:10:26 +00:00
|
|
|
maxWait := ctx.ServiceClient.shutdownWait + (3 * time.Second)
|
2017-02-01 00:43:57 +00:00
|
|
|
if shutdownTime < ctx.ServiceClient.shutdownWait || shutdownTime > maxWait {
|
|
|
|
t.Errorf("expected shutdown to take >%s and <%s but took: %s", ctx.ServiceClient.shutdownWait, maxWait, shutdownTime)
|
|
|
|
}
|
|
|
|
|
|
|
|
// UpdateTTL should not have been called for the script check
|
|
|
|
if n := len(ctx.FakeConsul.checkTTLs); n != 0 {
|
|
|
|
t.Fatalf("expected 0 checkTTL entry but found: %d", n)
|
|
|
|
}
|
|
|
|
for _, v := range ctx.FakeConsul.checks {
|
|
|
|
if expected := "warning"; v.Status != expected {
|
|
|
|
t.Fatalf("expected check to be %q but found %q", expected, v.Status)
|
|
|
|
}
|
|
|
|
}
|
|
|
|
}
|
2017-04-19 19:18:06 +00:00
|
|
|
|
|
|
|
// TestConsul_NoTLSSkipVerifySupport asserts that checks with
|
|
|
|
// TLSSkipVerify=true are skipped when Consul doesn't support TLSSkipVerify.
|
|
|
|
func TestConsul_NoTLSSkipVerifySupport(t *testing.T) {
|
|
|
|
ctx := setupFake()
|
|
|
|
ctx.ServiceClient = NewServiceClient(ctx.FakeConsul, false, testLogger())
|
|
|
|
ctx.Task.Services[0].Checks = []*structs.ServiceCheck{
|
|
|
|
// This check sets TLSSkipVerify so it should get dropped
|
|
|
|
{
|
|
|
|
Name: "tls-check-skip",
|
|
|
|
Type: "http",
|
|
|
|
Protocol: "https",
|
|
|
|
Path: "/",
|
|
|
|
TLSSkipVerify: true,
|
|
|
|
},
|
|
|
|
// This check doesn't set TLSSkipVerify so it should work fine
|
|
|
|
{
|
|
|
|
Name: "tls-check-noskip",
|
|
|
|
Type: "http",
|
|
|
|
Protocol: "https",
|
|
|
|
Path: "/",
|
|
|
|
TLSSkipVerify: false,
|
|
|
|
},
|
|
|
|
}
|
|
|
|
|
2017-09-13 06:15:46 +00:00
|
|
|
if err := ctx.ServiceClient.RegisterTask("allocid", ctx.Task, ctx.Restarter, nil, nil); err != nil {
|
2017-04-19 19:18:06 +00:00
|
|
|
t.Fatalf("unexpected error registering task: %v", err)
|
|
|
|
}
|
|
|
|
|
|
|
|
if err := ctx.syncOnce(); err != nil {
|
|
|
|
t.Fatalf("unexpected error syncing task: %v", err)
|
|
|
|
}
|
|
|
|
|
|
|
|
if len(ctx.FakeConsul.checks) != 1 {
|
|
|
|
t.Errorf("expected 1 check but found %d", len(ctx.FakeConsul.checks))
|
|
|
|
}
|
|
|
|
for _, v := range ctx.FakeConsul.checks {
|
|
|
|
if expected := "tls-check-noskip"; v.Name != expected {
|
|
|
|
t.Errorf("only expected %q but found: %q", expected, v.Name)
|
|
|
|
}
|
|
|
|
if v.TLSSkipVerify {
|
|
|
|
t.Errorf("TLSSkipVerify=true when TLSSkipVerify not supported!")
|
|
|
|
}
|
|
|
|
}
|
|
|
|
}
|
2017-04-26 18:22:01 +00:00
|
|
|
|
|
|
|
// TestConsul_RemoveScript assert removing a script check removes all objects
|
|
|
|
// related to that check.
|
|
|
|
func TestConsul_CancelScript(t *testing.T) {
|
|
|
|
ctx := setupFake()
|
|
|
|
ctx.Task.Services[0].Checks = []*structs.ServiceCheck{
|
|
|
|
{
|
|
|
|
Name: "scriptcheckDel",
|
|
|
|
Type: "script",
|
|
|
|
Interval: 9000 * time.Hour,
|
|
|
|
Timeout: 9000 * time.Hour,
|
|
|
|
},
|
|
|
|
{
|
|
|
|
Name: "scriptcheckKeep",
|
|
|
|
Type: "script",
|
|
|
|
Interval: 9000 * time.Hour,
|
|
|
|
Timeout: 9000 * time.Hour,
|
|
|
|
},
|
|
|
|
}
|
|
|
|
|
2017-09-13 06:15:46 +00:00
|
|
|
if err := ctx.ServiceClient.RegisterTask("allocid", ctx.Task, ctx.Restarter, ctx, nil); err != nil {
|
2017-04-26 18:22:01 +00:00
|
|
|
t.Fatalf("unexpected error registering task: %v", err)
|
|
|
|
}
|
|
|
|
|
|
|
|
if err := ctx.syncOnce(); err != nil {
|
|
|
|
t.Fatalf("unexpected error syncing task: %v", err)
|
|
|
|
}
|
|
|
|
|
|
|
|
if len(ctx.FakeConsul.checks) != 2 {
|
|
|
|
t.Errorf("expected 2 checks but found %d", len(ctx.FakeConsul.checks))
|
|
|
|
}
|
|
|
|
|
|
|
|
if len(ctx.ServiceClient.scripts) != 2 && len(ctx.ServiceClient.runningScripts) != 2 {
|
|
|
|
t.Errorf("expected 2 running script but found scripts=%d runningScripts=%d",
|
|
|
|
len(ctx.ServiceClient.scripts), len(ctx.ServiceClient.runningScripts))
|
|
|
|
}
|
|
|
|
|
|
|
|
for i := 0; i < 2; i++ {
|
|
|
|
select {
|
|
|
|
case <-ctx.execs:
|
|
|
|
// Script ran as expected!
|
|
|
|
case <-time.After(3 * time.Second):
|
|
|
|
t.Fatalf("timed out waiting for script check to run")
|
|
|
|
}
|
|
|
|
}
|
|
|
|
|
|
|
|
// Remove a check and update the task
|
|
|
|
origTask := ctx.Task.Copy()
|
|
|
|
ctx.Task.Services[0].Checks = []*structs.ServiceCheck{
|
|
|
|
{
|
|
|
|
Name: "scriptcheckKeep",
|
|
|
|
Type: "script",
|
|
|
|
Interval: 9000 * time.Hour,
|
|
|
|
Timeout: 9000 * time.Hour,
|
|
|
|
},
|
|
|
|
}
|
|
|
|
|
2017-09-13 06:15:46 +00:00
|
|
|
if err := ctx.ServiceClient.UpdateTask("allocid", origTask, ctx.Task, ctx.Restarter, ctx, nil); err != nil {
|
2017-04-26 18:22:01 +00:00
|
|
|
t.Fatalf("unexpected error registering task: %v", err)
|
|
|
|
}
|
|
|
|
|
|
|
|
if err := ctx.syncOnce(); err != nil {
|
|
|
|
t.Fatalf("unexpected error syncing task: %v", err)
|
|
|
|
}
|
|
|
|
|
|
|
|
if len(ctx.FakeConsul.checks) != 1 {
|
|
|
|
t.Errorf("expected 1 check but found %d", len(ctx.FakeConsul.checks))
|
|
|
|
}
|
|
|
|
|
|
|
|
if len(ctx.ServiceClient.scripts) != 1 && len(ctx.ServiceClient.runningScripts) != 1 {
|
|
|
|
t.Errorf("expected 1 running script but found scripts=%d runningScripts=%d",
|
|
|
|
len(ctx.ServiceClient.scripts), len(ctx.ServiceClient.runningScripts))
|
|
|
|
}
|
|
|
|
|
|
|
|
// Make sure exec wasn't called again
|
|
|
|
select {
|
|
|
|
case <-ctx.execs:
|
|
|
|
t.Errorf("unexpected execution of script; was goroutine not cancelled?")
|
|
|
|
case <-time.After(100 * time.Millisecond):
|
|
|
|
// No unexpected script execs
|
|
|
|
}
|
|
|
|
|
|
|
|
// Don't leak goroutines
|
|
|
|
for _, scriptHandle := range ctx.ServiceClient.runningScripts {
|
|
|
|
scriptHandle.cancel()
|
|
|
|
}
|
|
|
|
}
|
2017-06-13 21:02:11 +00:00
|
|
|
|
|
|
|
// TestConsul_DriverNetwork_AutoUse asserts that if a driver network has
|
|
|
|
// auto-use set then services should advertise it unless explicitly set to
|
|
|
|
// host. Checks should always use host.
|
|
|
|
func TestConsul_DriverNetwork_AutoUse(t *testing.T) {
|
|
|
|
ctx := setupFake()
|
|
|
|
|
|
|
|
ctx.Task.Services = []*structs.Service{
|
|
|
|
{
|
|
|
|
Name: "auto-advertise-x",
|
|
|
|
PortLabel: "x",
|
|
|
|
AddressMode: structs.AddressModeAuto,
|
|
|
|
Checks: []*structs.ServiceCheck{
|
|
|
|
{
|
|
|
|
Name: "default-check-x",
|
|
|
|
Type: "tcp",
|
|
|
|
Interval: time.Second,
|
|
|
|
Timeout: time.Second,
|
|
|
|
},
|
|
|
|
{
|
|
|
|
Name: "weird-y-check",
|
2017-06-16 23:35:16 +00:00
|
|
|
Type: "http",
|
2017-06-13 21:02:11 +00:00
|
|
|
Interval: time.Second,
|
|
|
|
Timeout: time.Second,
|
|
|
|
PortLabel: "y",
|
|
|
|
},
|
|
|
|
},
|
|
|
|
},
|
|
|
|
{
|
|
|
|
Name: "driver-advertise-y",
|
|
|
|
PortLabel: "y",
|
|
|
|
AddressMode: structs.AddressModeDriver,
|
|
|
|
Checks: []*structs.ServiceCheck{
|
|
|
|
{
|
|
|
|
Name: "default-check-y",
|
|
|
|
Type: "tcp",
|
|
|
|
Interval: time.Second,
|
|
|
|
Timeout: time.Second,
|
|
|
|
},
|
|
|
|
},
|
|
|
|
},
|
|
|
|
{
|
|
|
|
Name: "host-advertise-y",
|
|
|
|
PortLabel: "y",
|
|
|
|
AddressMode: structs.AddressModeHost,
|
|
|
|
},
|
|
|
|
}
|
|
|
|
|
|
|
|
net := &cstructs.DriverNetwork{
|
|
|
|
PortMap: map[string]int{
|
|
|
|
"x": 8888,
|
|
|
|
"y": 9999,
|
|
|
|
},
|
2017-06-16 23:35:16 +00:00
|
|
|
IP: "172.18.0.2",
|
|
|
|
AutoAdvertise: true,
|
2017-06-13 21:02:11 +00:00
|
|
|
}
|
|
|
|
|
2017-09-13 06:15:46 +00:00
|
|
|
if err := ctx.ServiceClient.RegisterTask("allocid", ctx.Task, ctx.Restarter, ctx, net); err != nil {
|
2017-06-13 21:02:11 +00:00
|
|
|
t.Fatalf("unexpected error registering task: %v", err)
|
|
|
|
}
|
|
|
|
|
|
|
|
if err := ctx.syncOnce(); err != nil {
|
|
|
|
t.Fatalf("unexpected error syncing task: %v", err)
|
|
|
|
}
|
|
|
|
|
|
|
|
if n := len(ctx.FakeConsul.services); n != 3 {
|
|
|
|
t.Fatalf("expected 2 services but found: %d", n)
|
|
|
|
}
|
|
|
|
|
|
|
|
for _, v := range ctx.FakeConsul.services {
|
|
|
|
switch v.Name {
|
|
|
|
case ctx.Task.Services[0].Name: // x
|
2017-06-16 23:35:16 +00:00
|
|
|
// Since DriverNetwork.AutoAdvertise=true, driver ports should be used
|
2017-06-13 21:02:11 +00:00
|
|
|
if v.Port != net.PortMap["x"] {
|
|
|
|
t.Errorf("expected service %s's port to be %d but found %d",
|
|
|
|
v.Name, net.PortMap["x"], v.Port)
|
|
|
|
}
|
2017-06-16 23:35:16 +00:00
|
|
|
// The order of checks in Consul is not guaranteed to
|
|
|
|
// be the same as their order in the Task definition,
|
|
|
|
// so check in a loop
|
|
|
|
if expected := 2; len(v.Checks) != expected {
|
2017-06-26 19:53:43 +00:00
|
|
|
t.Errorf("expected %d checks but found %d", expected, len(v.Checks))
|
2017-06-13 21:02:11 +00:00
|
|
|
}
|
2017-06-16 23:35:16 +00:00
|
|
|
for _, c := range v.Checks {
|
|
|
|
// No name on AgentServiceChecks, use type
|
|
|
|
switch {
|
|
|
|
case c.TCP != "":
|
|
|
|
// Checks should always use host port though
|
|
|
|
if c.TCP != ":1234" { // xPort
|
|
|
|
t.Errorf("exepcted service %s check 1's port to be %d but found %q",
|
|
|
|
v.Name, xPort, c.TCP)
|
|
|
|
}
|
|
|
|
case c.HTTP != "":
|
|
|
|
if c.HTTP != "http://:1235" { // yPort
|
|
|
|
t.Errorf("exepcted service %s check 2's port to be %d but found %q",
|
|
|
|
v.Name, yPort, c.HTTP)
|
|
|
|
}
|
|
|
|
default:
|
|
|
|
t.Errorf("unexpected check %#v on service %q", c, v.Name)
|
|
|
|
}
|
2017-06-13 21:02:11 +00:00
|
|
|
}
|
|
|
|
case ctx.Task.Services[1].Name: // y
|
|
|
|
// Service should be container ip:port
|
|
|
|
if v.Address != net.IP {
|
|
|
|
t.Errorf("expected service %s's address to be %s but found %s",
|
|
|
|
v.Name, net.IP, v.Address)
|
|
|
|
}
|
|
|
|
if v.Port != net.PortMap["y"] {
|
|
|
|
t.Errorf("expected service %s's port to be %d but found %d",
|
|
|
|
v.Name, net.PortMap["x"], v.Port)
|
|
|
|
}
|
|
|
|
// Check should be host ip:port
|
|
|
|
if v.Checks[0].TCP != ":1235" { // yPort
|
|
|
|
t.Errorf("expected service %s check's port to be %d but found %s",
|
|
|
|
v.Name, yPort, v.Checks[0].TCP)
|
|
|
|
}
|
|
|
|
case ctx.Task.Services[2].Name: // y + host mode
|
|
|
|
if v.Port != yPort {
|
2017-06-26 19:53:43 +00:00
|
|
|
t.Errorf("expected service %s's port to be %d but found %d",
|
2017-06-13 21:02:11 +00:00
|
|
|
v.Name, yPort, v.Port)
|
|
|
|
}
|
|
|
|
default:
|
|
|
|
t.Errorf("unexpected service name: %q", v.Name)
|
|
|
|
}
|
|
|
|
}
|
|
|
|
}
|
|
|
|
|
|
|
|
// TestConsul_DriverNetwork_NoAutoUse asserts that if a driver network doesn't
|
|
|
|
// set auto-use only services which request the driver's network should
|
|
|
|
// advertise it.
|
|
|
|
func TestConsul_DriverNetwork_NoAutoUse(t *testing.T) {
|
|
|
|
ctx := setupFake()
|
|
|
|
|
|
|
|
ctx.Task.Services = []*structs.Service{
|
|
|
|
{
|
|
|
|
Name: "auto-advertise-x",
|
|
|
|
PortLabel: "x",
|
|
|
|
AddressMode: structs.AddressModeAuto,
|
|
|
|
},
|
|
|
|
{
|
|
|
|
Name: "driver-advertise-y",
|
|
|
|
PortLabel: "y",
|
|
|
|
AddressMode: structs.AddressModeDriver,
|
|
|
|
},
|
|
|
|
{
|
|
|
|
Name: "host-advertise-y",
|
|
|
|
PortLabel: "y",
|
|
|
|
AddressMode: structs.AddressModeHost,
|
|
|
|
},
|
|
|
|
}
|
|
|
|
|
|
|
|
net := &cstructs.DriverNetwork{
|
|
|
|
PortMap: map[string]int{
|
|
|
|
"x": 8888,
|
|
|
|
"y": 9999,
|
|
|
|
},
|
2017-06-16 23:35:16 +00:00
|
|
|
IP: "172.18.0.2",
|
|
|
|
AutoAdvertise: false,
|
2017-06-13 21:02:11 +00:00
|
|
|
}
|
|
|
|
|
2017-09-13 06:15:46 +00:00
|
|
|
if err := ctx.ServiceClient.RegisterTask("allocid", ctx.Task, ctx.Restarter, ctx, net); err != nil {
|
2017-06-13 21:02:11 +00:00
|
|
|
t.Fatalf("unexpected error registering task: %v", err)
|
|
|
|
}
|
|
|
|
|
|
|
|
if err := ctx.syncOnce(); err != nil {
|
|
|
|
t.Fatalf("unexpected error syncing task: %v", err)
|
|
|
|
}
|
|
|
|
|
|
|
|
if n := len(ctx.FakeConsul.services); n != 3 {
|
2017-06-16 23:35:16 +00:00
|
|
|
t.Fatalf("expected 3 services but found: %d", n)
|
2017-06-13 21:02:11 +00:00
|
|
|
}
|
|
|
|
|
|
|
|
for _, v := range ctx.FakeConsul.services {
|
|
|
|
switch v.Name {
|
|
|
|
case ctx.Task.Services[0].Name: // x + auto
|
2017-06-16 23:35:16 +00:00
|
|
|
// Since DriverNetwork.AutoAdvertise=false, host ports should be used
|
2017-06-13 21:02:11 +00:00
|
|
|
if v.Port != xPort {
|
|
|
|
t.Errorf("expected service %s's port to be %d but found %d",
|
|
|
|
v.Name, xPort, v.Port)
|
|
|
|
}
|
|
|
|
case ctx.Task.Services[1].Name: // y + driver mode
|
|
|
|
// Service should be container ip:port
|
|
|
|
if v.Address != net.IP {
|
|
|
|
t.Errorf("expected service %s's address to be %s but found %s",
|
|
|
|
v.Name, net.IP, v.Address)
|
|
|
|
}
|
|
|
|
if v.Port != net.PortMap["y"] {
|
|
|
|
t.Errorf("expected service %s's port to be %d but found %d",
|
|
|
|
v.Name, net.PortMap["x"], v.Port)
|
|
|
|
}
|
|
|
|
case ctx.Task.Services[2].Name: // y + host mode
|
|
|
|
if v.Port != yPort {
|
2017-06-26 19:53:43 +00:00
|
|
|
t.Errorf("expected service %s's port to be %d but found %d",
|
2017-06-13 21:02:11 +00:00
|
|
|
v.Name, yPort, v.Port)
|
|
|
|
}
|
|
|
|
default:
|
|
|
|
t.Errorf("unexpected service name: %q", v.Name)
|
|
|
|
}
|
|
|
|
}
|
|
|
|
}
|
2017-06-16 23:35:16 +00:00
|
|
|
|
|
|
|
// TestConsul_DriverNetwork_Change asserts that if a driver network is
|
|
|
|
// specified and a service updates its use its properly updated in Consul.
|
|
|
|
func TestConsul_DriverNetwork_Change(t *testing.T) {
|
|
|
|
ctx := setupFake()
|
|
|
|
|
|
|
|
ctx.Task.Services = []*structs.Service{
|
|
|
|
{
|
|
|
|
Name: "service-foo",
|
|
|
|
PortLabel: "x",
|
|
|
|
AddressMode: structs.AddressModeAuto,
|
|
|
|
},
|
|
|
|
}
|
|
|
|
|
|
|
|
net := &cstructs.DriverNetwork{
|
|
|
|
PortMap: map[string]int{
|
|
|
|
"x": 8888,
|
|
|
|
"y": 9999,
|
|
|
|
},
|
|
|
|
IP: "172.18.0.2",
|
|
|
|
AutoAdvertise: false,
|
|
|
|
}
|
|
|
|
|
|
|
|
syncAndAssertPort := func(port int) {
|
|
|
|
if err := ctx.syncOnce(); err != nil {
|
|
|
|
t.Fatalf("unexpected error syncing task: %v", err)
|
|
|
|
}
|
|
|
|
|
|
|
|
if n := len(ctx.FakeConsul.services); n != 1 {
|
|
|
|
t.Fatalf("expected 1 service but found: %d", n)
|
|
|
|
}
|
|
|
|
|
|
|
|
for _, v := range ctx.FakeConsul.services {
|
|
|
|
switch v.Name {
|
|
|
|
case ctx.Task.Services[0].Name:
|
|
|
|
if v.Port != port {
|
|
|
|
t.Errorf("expected service %s's port to be %d but found %d",
|
|
|
|
v.Name, port, v.Port)
|
|
|
|
}
|
|
|
|
default:
|
|
|
|
t.Errorf("unexpected service name: %q", v.Name)
|
|
|
|
}
|
|
|
|
}
|
|
|
|
}
|
|
|
|
|
|
|
|
// Initial service should advertise host port x
|
2017-09-13 06:15:46 +00:00
|
|
|
if err := ctx.ServiceClient.RegisterTask("allocid", ctx.Task, ctx.Restarter, ctx, net); err != nil {
|
2017-06-16 23:35:16 +00:00
|
|
|
t.Fatalf("unexpected error registering task: %v", err)
|
|
|
|
}
|
|
|
|
|
|
|
|
syncAndAssertPort(xPort)
|
|
|
|
|
|
|
|
// UpdateTask to use Host (shouldn't change anything)
|
|
|
|
orig := ctx.Task.Copy()
|
|
|
|
ctx.Task.Services[0].AddressMode = structs.AddressModeHost
|
|
|
|
|
2017-09-13 06:15:46 +00:00
|
|
|
if err := ctx.ServiceClient.UpdateTask("allocid", orig, ctx.Task, ctx.Restarter, ctx, net); err != nil {
|
2017-06-16 23:35:16 +00:00
|
|
|
t.Fatalf("unexpected error updating task: %v", err)
|
|
|
|
}
|
|
|
|
|
|
|
|
syncAndAssertPort(xPort)
|
|
|
|
|
|
|
|
// UpdateTask to use Driver (*should* change IP and port)
|
|
|
|
orig = ctx.Task.Copy()
|
|
|
|
ctx.Task.Services[0].AddressMode = structs.AddressModeDriver
|
|
|
|
|
2017-09-13 06:15:46 +00:00
|
|
|
if err := ctx.ServiceClient.UpdateTask("allocid", orig, ctx.Task, ctx.Restarter, ctx, net); err != nil {
|
2017-06-16 23:35:16 +00:00
|
|
|
t.Fatalf("unexpected error updating task: %v", err)
|
|
|
|
}
|
|
|
|
|
|
|
|
syncAndAssertPort(net.PortMap["x"])
|
|
|
|
}
|
2017-07-18 20:23:01 +00:00
|
|
|
|
|
|
|
// TestIsNomadService asserts the isNomadService helper returns true for Nomad
|
|
|
|
// task IDs and false for unknown IDs and Nomad agent IDs (see #2827).
|
|
|
|
func TestIsNomadService(t *testing.T) {
|
|
|
|
tests := []struct {
|
|
|
|
id string
|
|
|
|
result bool
|
|
|
|
}{
|
|
|
|
{"_nomad-client-nomad-client-http", false},
|
|
|
|
{"_nomad-server-nomad-serf", false},
|
2017-12-08 01:08:25 +00:00
|
|
|
|
|
|
|
// Pre-0.7.1 style IDs still match
|
2017-07-18 20:23:01 +00:00
|
|
|
{"_nomad-executor-abc", true},
|
|
|
|
{"_nomad-executor", true},
|
2017-12-08 01:08:25 +00:00
|
|
|
|
|
|
|
// Post-0.7.1 style IDs match
|
|
|
|
{"_nomad-task-FBBK265QN4TMT25ND4EP42TJVMYJ3HR4", true},
|
|
|
|
|
2017-07-18 20:23:01 +00:00
|
|
|
{"not-nomad", false},
|
2017-12-08 01:08:25 +00:00
|
|
|
{"_nomad", false},
|
2017-07-18 20:23:01 +00:00
|
|
|
}
|
|
|
|
|
|
|
|
for _, test := range tests {
|
|
|
|
t.Run(test.id, func(t *testing.T) {
|
|
|
|
actual := isNomadService(test.id)
|
|
|
|
if actual != test.result {
|
2017-07-19 17:00:57 +00:00
|
|
|
t.Errorf("%q should be %t but found %t", test.id, test.result, actual)
|
2017-07-18 20:23:01 +00:00
|
|
|
}
|
|
|
|
})
|
|
|
|
}
|
|
|
|
}
|
2017-08-16 18:27:30 +00:00
|
|
|
|
|
|
|
// TestCreateCheckReg asserts Nomad ServiceCheck structs are properly converted
|
|
|
|
// to Consul API AgentCheckRegistrations.
|
|
|
|
func TestCreateCheckReg(t *testing.T) {
|
|
|
|
check := &structs.ServiceCheck{
|
|
|
|
Name: "name",
|
|
|
|
Type: "http",
|
|
|
|
Path: "/path",
|
|
|
|
PortLabel: "label",
|
|
|
|
Method: "POST",
|
|
|
|
Header: map[string][]string{
|
|
|
|
"Foo": {"bar"},
|
|
|
|
},
|
|
|
|
}
|
|
|
|
|
|
|
|
serviceID := "testService"
|
|
|
|
checkID := check.Hash(serviceID)
|
|
|
|
host := "localhost"
|
|
|
|
port := 41111
|
|
|
|
|
|
|
|
expected := &api.AgentCheckRegistration{
|
|
|
|
ID: checkID,
|
|
|
|
Name: "name",
|
|
|
|
ServiceID: serviceID,
|
2017-08-17 22:18:55 +00:00
|
|
|
AgentServiceCheck: api.AgentServiceCheck{
|
|
|
|
Timeout: "0s",
|
|
|
|
Interval: "0s",
|
|
|
|
HTTP: fmt.Sprintf("http://%s:%d/path", host, port),
|
|
|
|
Method: "POST",
|
|
|
|
Header: map[string][]string{
|
|
|
|
"Foo": {"bar"},
|
|
|
|
},
|
|
|
|
},
|
2017-08-16 18:27:30 +00:00
|
|
|
}
|
|
|
|
|
|
|
|
actual, err := createCheckReg(serviceID, checkID, check, host, port)
|
|
|
|
if err != nil {
|
|
|
|
t.Fatalf("err: %v", err)
|
|
|
|
}
|
|
|
|
|
|
|
|
if diff := pretty.Diff(actual, expected); len(diff) > 0 {
|
|
|
|
t.Fatalf("diff:\n%s\n", strings.Join(diff, "\n"))
|
|
|
|
}
|
|
|
|
}
|
2017-12-05 19:39:42 +00:00
|
|
|
|
|
|
|
// TestGetAddress asserts Nomad uses the correct ip and port for services and
|
|
|
|
// checks depending on port labels, driver networks, and address mode.
|
|
|
|
func TestGetAddress(t *testing.T) {
|
|
|
|
const HostIP = "127.0.0.1"
|
|
|
|
|
|
|
|
cases := []struct {
|
|
|
|
Name string
|
|
|
|
|
|
|
|
// Parameters
|
|
|
|
Mode string
|
|
|
|
PortLabel string
|
|
|
|
Host map[string]int // will be converted to structs.Networks
|
|
|
|
Driver *cstructs.DriverNetwork
|
|
|
|
|
|
|
|
// Results
|
2017-12-20 23:02:34 +00:00
|
|
|
ExpectedIP string
|
|
|
|
ExpectedPort int
|
|
|
|
ExpectedErr string
|
2017-12-05 19:39:42 +00:00
|
|
|
}{
|
|
|
|
{
|
|
|
|
Name: "ExampleService",
|
|
|
|
Mode: structs.AddressModeAuto,
|
|
|
|
PortLabel: "db",
|
|
|
|
Host: map[string]int{"db": 12435},
|
|
|
|
Driver: &cstructs.DriverNetwork{
|
|
|
|
PortMap: map[string]int{"db": 6379},
|
|
|
|
IP: "10.1.2.3",
|
|
|
|
},
|
2017-12-20 23:02:34 +00:00
|
|
|
ExpectedIP: HostIP,
|
|
|
|
ExpectedPort: 12435,
|
2017-12-05 19:39:42 +00:00
|
|
|
},
|
|
|
|
{
|
|
|
|
Name: "Host",
|
|
|
|
Mode: structs.AddressModeHost,
|
|
|
|
PortLabel: "db",
|
|
|
|
Host: map[string]int{"db": 12345},
|
|
|
|
Driver: &cstructs.DriverNetwork{
|
|
|
|
PortMap: map[string]int{"db": 6379},
|
|
|
|
IP: "10.1.2.3",
|
|
|
|
},
|
2017-12-20 23:02:34 +00:00
|
|
|
ExpectedIP: HostIP,
|
|
|
|
ExpectedPort: 12345,
|
2017-12-05 19:39:42 +00:00
|
|
|
},
|
|
|
|
{
|
|
|
|
Name: "Driver",
|
|
|
|
Mode: structs.AddressModeDriver,
|
|
|
|
PortLabel: "db",
|
|
|
|
Host: map[string]int{"db": 12345},
|
|
|
|
Driver: &cstructs.DriverNetwork{
|
|
|
|
PortMap: map[string]int{"db": 6379},
|
|
|
|
IP: "10.1.2.3",
|
|
|
|
},
|
2017-12-20 23:02:34 +00:00
|
|
|
ExpectedIP: "10.1.2.3",
|
|
|
|
ExpectedPort: 6379,
|
2017-12-05 19:39:42 +00:00
|
|
|
},
|
|
|
|
{
|
|
|
|
Name: "AutoDriver",
|
|
|
|
Mode: structs.AddressModeAuto,
|
|
|
|
PortLabel: "db",
|
|
|
|
Host: map[string]int{"db": 12345},
|
|
|
|
Driver: &cstructs.DriverNetwork{
|
|
|
|
PortMap: map[string]int{"db": 6379},
|
|
|
|
IP: "10.1.2.3",
|
|
|
|
AutoAdvertise: true,
|
|
|
|
},
|
2017-12-20 23:02:34 +00:00
|
|
|
ExpectedIP: "10.1.2.3",
|
|
|
|
ExpectedPort: 6379,
|
2017-12-05 19:39:42 +00:00
|
|
|
},
|
|
|
|
{
|
|
|
|
Name: "DriverCustomPort",
|
|
|
|
Mode: structs.AddressModeDriver,
|
|
|
|
PortLabel: "7890",
|
|
|
|
Host: map[string]int{"db": 12345},
|
|
|
|
Driver: &cstructs.DriverNetwork{
|
|
|
|
PortMap: map[string]int{"db": 6379},
|
|
|
|
IP: "10.1.2.3",
|
|
|
|
},
|
2017-12-20 23:02:34 +00:00
|
|
|
ExpectedIP: "10.1.2.3",
|
|
|
|
ExpectedPort: 7890,
|
2017-12-05 19:39:42 +00:00
|
|
|
},
|
|
|
|
{
|
|
|
|
Name: "DriverWithoutNetwork",
|
|
|
|
Mode: structs.AddressModeDriver,
|
|
|
|
PortLabel: "db",
|
|
|
|
Host: map[string]int{"db": 12345},
|
|
|
|
Driver: nil,
|
2017-12-20 23:02:34 +00:00
|
|
|
ExpectedErr: "no driver network exists",
|
2017-12-05 19:39:42 +00:00
|
|
|
},
|
|
|
|
{
|
|
|
|
Name: "DriverBadPort",
|
|
|
|
Mode: structs.AddressModeDriver,
|
|
|
|
PortLabel: "bad-port-label",
|
|
|
|
Host: map[string]int{"db": 12345},
|
|
|
|
Driver: &cstructs.DriverNetwork{
|
|
|
|
PortMap: map[string]int{"db": 6379},
|
|
|
|
IP: "10.1.2.3",
|
|
|
|
},
|
2017-12-20 23:02:34 +00:00
|
|
|
ExpectedErr: "invalid port",
|
2017-12-05 19:39:42 +00:00
|
|
|
},
|
2017-12-08 06:04:22 +00:00
|
|
|
{
|
|
|
|
Name: "DriverZeroPort",
|
|
|
|
Mode: structs.AddressModeDriver,
|
|
|
|
PortLabel: "0",
|
|
|
|
Driver: &cstructs.DriverNetwork{
|
|
|
|
IP: "10.1.2.3",
|
|
|
|
},
|
2017-12-20 23:02:34 +00:00
|
|
|
ExpectedErr: "invalid port",
|
2017-12-08 06:04:22 +00:00
|
|
|
},
|
2017-12-08 05:58:15 +00:00
|
|
|
{
|
|
|
|
Name: "HostBadPort",
|
|
|
|
Mode: structs.AddressModeHost,
|
|
|
|
PortLabel: "bad-port-label",
|
2017-12-20 23:02:34 +00:00
|
|
|
ExpectedErr: "invalid port",
|
2017-12-08 05:58:15 +00:00
|
|
|
},
|
2017-12-05 19:39:42 +00:00
|
|
|
{
|
|
|
|
Name: "InvalidMode",
|
|
|
|
Mode: "invalid-mode",
|
2017-12-19 00:18:42 +00:00
|
|
|
PortLabel: "80",
|
2017-12-20 23:02:34 +00:00
|
|
|
ExpectedErr: "invalid address mode",
|
2017-12-05 19:39:42 +00:00
|
|
|
},
|
2017-12-19 00:18:42 +00:00
|
|
|
{
|
2017-12-20 23:02:34 +00:00
|
|
|
Name: "NoPort_AutoMode",
|
2017-12-20 23:59:47 +00:00
|
|
|
Mode: structs.AddressModeAuto,
|
2017-12-20 23:02:34 +00:00
|
|
|
ExpectedIP: HostIP,
|
|
|
|
},
|
|
|
|
{
|
|
|
|
Name: "NoPort_HostMode",
|
|
|
|
Mode: structs.AddressModeHost,
|
|
|
|
ExpectedIP: HostIP,
|
|
|
|
},
|
|
|
|
{
|
|
|
|
Name: "NoPort_DriverMode",
|
|
|
|
Mode: structs.AddressModeDriver,
|
|
|
|
Driver: &cstructs.DriverNetwork{
|
|
|
|
IP: "10.1.2.3",
|
|
|
|
},
|
|
|
|
ExpectedIP: "10.1.2.3",
|
2017-12-19 00:18:42 +00:00
|
|
|
},
|
2017-12-05 19:39:42 +00:00
|
|
|
}
|
|
|
|
|
|
|
|
for _, tc := range cases {
|
|
|
|
t.Run(tc.Name, func(t *testing.T) {
|
|
|
|
// convert host port map into a structs.Networks
|
|
|
|
networks := []*structs.NetworkResource{
|
|
|
|
{
|
|
|
|
IP: HostIP,
|
|
|
|
ReservedPorts: make([]structs.Port, len(tc.Host)),
|
|
|
|
},
|
|
|
|
}
|
|
|
|
|
|
|
|
i := 0
|
|
|
|
for label, port := range tc.Host {
|
|
|
|
networks[0].ReservedPorts[i].Label = label
|
|
|
|
networks[0].ReservedPorts[i].Value = port
|
|
|
|
i++
|
|
|
|
}
|
|
|
|
|
|
|
|
// Run getAddress
|
|
|
|
ip, port, err := getAddress(tc.Mode, tc.PortLabel, networks, tc.Driver)
|
|
|
|
|
|
|
|
// Assert the results
|
2017-12-20 23:02:34 +00:00
|
|
|
assert.Equal(t, tc.ExpectedIP, ip, "IP mismatch")
|
|
|
|
assert.Equal(t, tc.ExpectedPort, port, "Port mismatch")
|
|
|
|
if tc.ExpectedErr == "" {
|
2017-12-05 19:39:42 +00:00
|
|
|
assert.Nil(t, err)
|
|
|
|
} else {
|
2017-12-08 05:58:15 +00:00
|
|
|
if err == nil {
|
2017-12-20 23:02:34 +00:00
|
|
|
t.Fatalf("expected error containing %q but err=nil", tc.ExpectedErr)
|
2017-12-08 05:58:15 +00:00
|
|
|
} else {
|
2017-12-20 23:02:34 +00:00
|
|
|
assert.Contains(t, err.Error(), tc.ExpectedErr)
|
2017-12-08 05:58:15 +00:00
|
|
|
}
|
2017-12-05 19:39:42 +00:00
|
|
|
}
|
|
|
|
})
|
|
|
|
}
|
|
|
|
}
|