package consul import ( "context" "errors" "fmt" "reflect" "strings" "sync/atomic" "testing" "time" "github.com/hashicorp/consul/api" "github.com/hashicorp/nomad/helper/testlog" "github.com/hashicorp/nomad/helper/uuid" "github.com/hashicorp/nomad/nomad/structs" "github.com/hashicorp/nomad/plugins/drivers" "github.com/kr/pretty" "github.com/stretchr/testify/assert" "github.com/stretchr/testify/require" ) const ( // Ports used in testWorkload xPort = 1234 yPort = 1235 ) func testWorkload() *WorkloadServices { return &WorkloadServices{ AllocID: uuid.Generate(), Task: "taskname", Restarter: &restartRecorder{}, Services: []*structs.Service{ { Name: "taskname-service", PortLabel: "x", Tags: []string{"tag1", "tag2"}, Meta: map[string]string{"meta1": "foo"}, }, }, Networks: []*structs.NetworkResource{ { DynamicPorts: []structs.Port{ {Label: "x", Value: xPort}, {Label: "y", Value: yPort}, }, }, }, } } // restartRecorder is a minimal WorkloadRestarter implementation that simply // counts how many restarts were triggered. type restartRecorder struct { restarts int64 } func (r *restartRecorder) Restart(ctx context.Context, event *structs.TaskEvent, failure bool) error { atomic.AddInt64(&r.restarts, 1) return nil } // testFakeCtx contains a fake Consul AgentAPI type testFakeCtx struct { ServiceClient *ServiceClient FakeConsul *MockAgent Workload *WorkloadServices } var errNoOps = fmt.Errorf("testing error: no pending operations") // syncOps simulates one iteration of the ServiceClient.Run loop and returns // any errors returned by sync() or errNoOps if no pending operations. func (t *testFakeCtx) syncOnce(reason syncReason) error { switch reason { case syncPeriodic: err := t.ServiceClient.sync(syncPeriodic) if err == nil { t.ServiceClient.clearExplicitlyDeregistered() } return err case syncNewOps: select { case ops := <-t.ServiceClient.opCh: t.ServiceClient.merge(ops) err := t.ServiceClient.sync(syncNewOps) if err == nil { t.ServiceClient.clearExplicitlyDeregistered() } return err default: return errNoOps } case syncShutdown: return errors.New("no test for sync due to shutdown") } return errors.New("bad sync reason") } // setupFake creates a testFakeCtx with a ServiceClient backed by a fakeConsul. // A test Workload is also provided. func setupFake(t *testing.T) *testFakeCtx { fc := NewMockAgent() tw := testWorkload() // by default start fake client being out of probation sc := NewServiceClient(fc, testlog.HCLogger(t), true) sc.deregisterProbationExpiry = time.Now().Add(-1 * time.Minute) return &testFakeCtx{ ServiceClient: sc, FakeConsul: fc, Workload: tw, } } func TestConsul_ChangeTags(t *testing.T) { t.Parallel() ctx := setupFake(t) r := require.New(t) r.NoError(ctx.ServiceClient.RegisterWorkload(ctx.Workload)) r.NoError(ctx.syncOnce(syncNewOps)) r.Equal(1, len(ctx.FakeConsul.services), "Expected 1 service to be registered with Consul") // Validate the alloc registration reg1, err := ctx.ServiceClient.AllocRegistrations(ctx.Workload.AllocID) r.NoError(err) r.NotNil(reg1, "Unexpected nil alloc registration") r.Equal(1, reg1.NumServices()) r.Equal(0, reg1.NumChecks()) serviceBefore := ctx.FakeConsul.lookupService("taskname-service")[0] r.Equal(serviceBefore.Name, ctx.Workload.Services[0].Name) r.Equal(serviceBefore.Tags, ctx.Workload.Services[0].Tags) // Update the task definition origWorkload := ctx.Workload.Copy() ctx.Workload.Services[0].Tags[0] = "new-tag" // Register and sync the update r.NoError(ctx.ServiceClient.UpdateWorkload(origWorkload, ctx.Workload)) r.NoError(ctx.syncOnce(syncNewOps)) r.Equal(1, len(ctx.FakeConsul.services), "Expected 1 service to be registered with Consul") // Validate the consul service definition changed serviceAfter := ctx.FakeConsul.lookupService("taskname-service")[0] r.Equal(serviceAfter.Name, ctx.Workload.Services[0].Name) r.Equal(serviceAfter.Tags, ctx.Workload.Services[0].Tags) r.Equal("new-tag", serviceAfter.Tags[0]) } func TestConsul_EnableTagOverride_Syncs(t *testing.T) { t.Parallel() ctx := setupFake(t) r := require.New(t) // Configure our test service to set EnableTagOverride = true ctx.Workload.Services[0].EnableTagOverride = true r.NoError(ctx.ServiceClient.RegisterWorkload(ctx.Workload)) r.NoError(ctx.syncOnce(syncNewOps)) r.Equal(1, len(ctx.FakeConsul.services)) // Validate the alloc registration reg1, err := ctx.ServiceClient.AllocRegistrations(ctx.Workload.AllocID) r.NoError(err) r.NotNil(reg1) r.Equal(1, reg1.NumServices()) r.Equal(0, reg1.NumChecks()) const service = "taskname-service" // sanity check things are what we expect consulServiceDefBefore := ctx.FakeConsul.lookupService(service)[0] r.Equal(ctx.Workload.Services[0].Name, consulServiceDefBefore.Name) r.Equal([]string{"tag1", "tag2"}, consulServiceDefBefore.Tags) r.True(consulServiceDefBefore.EnableTagOverride) // manually set the tags in consul ctx.FakeConsul.lookupService(service)[0].Tags = []string{"new", "tags"} // do a periodic sync (which will respect EnableTagOverride) r.NoError(ctx.syncOnce(syncPeriodic)) r.Equal(1, len(ctx.FakeConsul.services)) consulServiceDefAfter := ctx.FakeConsul.lookupService(service)[0] r.Equal([]string{"new", "tags"}, consulServiceDefAfter.Tags) // manually set tags should still be there // now do a new-ops sync (which will override EnableTagOverride) r.NoError(ctx.ServiceClient.RegisterWorkload(ctx.Workload)) r.NoError(ctx.syncOnce(syncNewOps)) r.Equal(1, len(ctx.FakeConsul.services)) consulServiceDefUpdated := ctx.FakeConsul.lookupService(service)[0] r.Equal([]string{"tag1", "tag2"}, consulServiceDefUpdated.Tags) // jobspec tags should be set now } // TestConsul_ChangePorts asserts that changing the ports on a service updates // it in Consul. Pre-0.7.1 ports were not part of the service ID and this was a // slightly different code path than changing tags. func TestConsul_ChangePorts(t *testing.T) { ctx := setupFake(t) require := require.New(t) ctx.Workload.Services[0].Checks = []*structs.ServiceCheck{ { Name: "c1", Type: "tcp", Interval: time.Second, Timeout: time.Second, PortLabel: "x", }, { Name: "c2", Type: "script", Interval: 9000 * time.Hour, Timeout: time.Second, }, { Name: "c3", Type: "http", Protocol: "http", Path: "/", Interval: time.Second, Timeout: time.Second, PortLabel: "y", }, } require.NoError(ctx.ServiceClient.RegisterWorkload(ctx.Workload)) require.NoError(ctx.syncOnce(syncNewOps)) require.Equal(1, len(ctx.FakeConsul.services), "Expected 1 service to be registered with Consul") for _, v := range ctx.FakeConsul.services { require.Equal(ctx.Workload.Services[0].Name, v.Name) require.Equal(ctx.Workload.Services[0].Tags, v.Tags) require.Equal(xPort, v.Port) } require.Len(ctx.FakeConsul.checks, 3) origTCPKey := "" origScriptKey := "" origHTTPKey := "" for k, v := range ctx.FakeConsul.checks { switch v.Name { case "c1": origTCPKey = k require.Equal(fmt.Sprintf(":%d", xPort), v.TCP) case "c2": origScriptKey = k case "c3": origHTTPKey = k require.Equal(fmt.Sprintf("http://:%d/", yPort), v.HTTP) default: t.Fatalf("unexpected check: %q", v.Name) } } require.NotEmpty(origTCPKey) require.NotEmpty(origScriptKey) require.NotEmpty(origHTTPKey) // Now update the PortLabel on the Service and Check c3 origWorkload := ctx.Workload.Copy() ctx.Workload.Services[0].PortLabel = "y" ctx.Workload.Services[0].Checks = []*structs.ServiceCheck{ { Name: "c1", Type: "tcp", Interval: time.Second, Timeout: time.Second, PortLabel: "x", }, { Name: "c2", Type: "script", Interval: 9000 * time.Hour, Timeout: time.Second, }, { Name: "c3", Type: "http", Protocol: "http", Path: "/", Interval: time.Second, Timeout: time.Second, // Removed PortLabel; should default to service's (y) }, } require.NoError(ctx.ServiceClient.UpdateWorkload(origWorkload, ctx.Workload)) require.NoError(ctx.syncOnce(syncNewOps)) require.Equal(1, len(ctx.FakeConsul.services), "Expected 1 service to be registered with Consul") for _, v := range ctx.FakeConsul.services { require.Equal(ctx.Workload.Services[0].Name, v.Name) require.Equal(ctx.Workload.Services[0].Tags, v.Tags) require.Equal(yPort, v.Port) } require.Equal(3, len(ctx.FakeConsul.checks)) for k, v := range ctx.FakeConsul.checks { switch v.Name { case "c1": // C1 is changed because the service was re-registered require.NotEqual(origTCPKey, k) require.Equal(fmt.Sprintf(":%d", xPort), v.TCP) case "c2": // C2 is changed because the service was re-registered require.NotEqual(origScriptKey, k) case "c3": require.NotEqual(origHTTPKey, k) require.Equal(fmt.Sprintf("http://:%d/", yPort), v.HTTP) default: t.Errorf("Unknown check: %q", k) } } } // TestConsul_ChangeChecks asserts that updating only the checks on a service // properly syncs with Consul. func TestConsul_ChangeChecks(t *testing.T) { ctx := setupFake(t) ctx.Workload.Services[0].Checks = []*structs.ServiceCheck{ { Name: "c1", Type: "tcp", Interval: time.Second, Timeout: time.Second, PortLabel: "x", CheckRestart: &structs.CheckRestart{ Limit: 3, }, }, } if err := ctx.ServiceClient.RegisterWorkload(ctx.Workload); err != nil { t.Fatalf("unexpected error registering task: %v", err) } if err := ctx.syncOnce(syncNewOps); err != nil { t.Fatalf("unexpected error syncing task: %v", err) } if n := len(ctx.FakeConsul.services); n != 1 { t.Fatalf("expected 1 service but found %d:\n%#v", n, ctx.FakeConsul.services) } // Assert a check restart watch update was enqueued and clear it if n := len(ctx.ServiceClient.checkWatcher.checkUpdateCh); n != 1 { t.Fatalf("expected 1 check restart update but found %d", n) } upd := <-ctx.ServiceClient.checkWatcher.checkUpdateCh c1ID := upd.checkID // Query the allocs registrations and then again when we update. The IDs // should change reg1, err := ctx.ServiceClient.AllocRegistrations(ctx.Workload.AllocID) if err != nil { t.Fatalf("Looking up alloc registration failed: %v", err) } if reg1 == nil { t.Fatalf("Nil alloc registrations: %v", err) } if num := reg1.NumServices(); num != 1 { t.Fatalf("Wrong number of services: got %d; want 1", num) } if num := reg1.NumChecks(); num != 1 { t.Fatalf("Wrong number of checks: got %d; want 1", num) } origServiceKey := "" for k, v := range ctx.FakeConsul.services { origServiceKey = k if v.Name != ctx.Workload.Services[0].Name { t.Errorf("expected Name=%q != %q", ctx.Workload.Services[0].Name, v.Name) } if v.Port != xPort { t.Errorf("expected Port x=%v but found: %v", xPort, v.Port) } } if n := len(ctx.FakeConsul.checks); n != 1 { t.Fatalf("expected 1 check but found %d:\n%#v", n, ctx.FakeConsul.checks) } for _, v := range ctx.FakeConsul.checks { if v.Name != "c1" { t.Fatalf("expected check c1 but found %q", v.Name) } } // Now add a check and modify the original origWorkload := ctx.Workload.Copy() ctx.Workload.Services[0].Checks = []*structs.ServiceCheck{ { Name: "c1", Type: "tcp", Interval: 2 * time.Second, Timeout: time.Second, PortLabel: "x", CheckRestart: &structs.CheckRestart{ Limit: 3, }, }, { Name: "c2", Type: "http", Path: "/", Interval: time.Second, Timeout: time.Second, PortLabel: "x", }, } if err := ctx.ServiceClient.UpdateWorkload(origWorkload, ctx.Workload); err != nil { t.Fatalf("unexpected error registering task: %v", err) } // Assert 2 check restart watch updates was enqueued if n := len(ctx.ServiceClient.checkWatcher.checkUpdateCh); n != 2 { t.Fatalf("expected 2 check restart updates but found %d", n) } // First the new watch upd = <-ctx.ServiceClient.checkWatcher.checkUpdateCh if upd.checkID == c1ID || upd.remove { t.Fatalf("expected check watch update to be an add of checkID=%q but found remove=%t checkID=%q", c1ID, upd.remove, upd.checkID) } // Then remove the old watch upd = <-ctx.ServiceClient.checkWatcher.checkUpdateCh if upd.checkID != c1ID || !upd.remove { t.Fatalf("expected check watch update to be a removal of checkID=%q but found remove=%t checkID=%q", c1ID, upd.remove, upd.checkID) } if err := ctx.syncOnce(syncNewOps); err != nil { t.Fatalf("unexpected error syncing task: %v", err) } if n := len(ctx.FakeConsul.services); n != 1 { t.Fatalf("expected 1 service but found %d:\n%#v", n, ctx.FakeConsul.services) } if _, ok := ctx.FakeConsul.services[origServiceKey]; !ok { t.Errorf("unexpected key change; was: %q -- but found %#v", origServiceKey, ctx.FakeConsul.services) } if n := len(ctx.FakeConsul.checks); n != 2 { t.Fatalf("expected 2 check but found %d:\n%#v", n, ctx.FakeConsul.checks) } for k, v := range ctx.FakeConsul.checks { switch v.Name { case "c1": if expected := fmt.Sprintf(":%d", xPort); v.TCP != expected { t.Errorf("expected Port x=%v but found: %v", expected, v.TCP) } // update id c1ID = k case "c2": if expected := fmt.Sprintf("http://:%d/", xPort); v.HTTP != expected { t.Errorf("expected Port x=%v but found: %v", expected, v.HTTP) } default: t.Errorf("Unknown check: %q", k) } } // Check again and ensure the IDs changed reg2, err := ctx.ServiceClient.AllocRegistrations(ctx.Workload.AllocID) if err != nil { t.Fatalf("Looking up alloc registration failed: %v", err) } if reg2 == nil { t.Fatalf("Nil alloc registrations: %v", err) } if num := reg2.NumServices(); num != 1 { t.Fatalf("Wrong number of services: got %d; want 1", num) } if num := reg2.NumChecks(); num != 2 { t.Fatalf("Wrong number of checks: got %d; want 2", num) } for task, treg := range reg1.Tasks { otherTaskReg, ok := reg2.Tasks[task] if !ok { t.Fatalf("Task %q not in second reg", task) } for sID, sreg := range treg.Services { otherServiceReg, ok := otherTaskReg.Services[sID] if !ok { t.Fatalf("service ID changed") } for newID := range sreg.checkIDs { if _, ok := otherServiceReg.checkIDs[newID]; ok { t.Fatalf("check IDs should change") } } } } // Alter a CheckRestart and make sure the watcher is updated but nothing else origWorkload = ctx.Workload.Copy() ctx.Workload.Services[0].Checks = []*structs.ServiceCheck{ { Name: "c1", Type: "tcp", Interval: 2 * time.Second, Timeout: time.Second, PortLabel: "x", CheckRestart: &structs.CheckRestart{ Limit: 11, }, }, { Name: "c2", Type: "http", Path: "/", Interval: time.Second, Timeout: time.Second, PortLabel: "x", }, } if err := ctx.ServiceClient.UpdateWorkload(origWorkload, ctx.Workload); err != nil { t.Fatalf("unexpected error registering task: %v", err) } if err := ctx.syncOnce(syncNewOps); err != nil { t.Fatalf("unexpected error syncing task: %v", err) } if n := len(ctx.FakeConsul.checks); n != 2 { t.Fatalf("expected 2 check but found %d:\n%#v", n, ctx.FakeConsul.checks) } for k, v := range ctx.FakeConsul.checks { if v.Name == "c1" { if k != c1ID { t.Errorf("expected c1 to still have id %q but found %q", c1ID, k) } break } } // Assert a check restart watch update was enqueued for a removal and an add if n := len(ctx.ServiceClient.checkWatcher.checkUpdateCh); n != 1 { t.Fatalf("expected 1 check restart update but found %d", n) } <-ctx.ServiceClient.checkWatcher.checkUpdateCh } // TestConsul_RegServices tests basic service registration. func TestConsul_RegServices(t *testing.T) { ctx := setupFake(t) // Add a check w/restarting ctx.Workload.Services[0].Checks = []*structs.ServiceCheck{ { Name: "testcheck", Type: "tcp", Interval: 100 * time.Millisecond, CheckRestart: &structs.CheckRestart{ Limit: 3, }, }, } if err := ctx.ServiceClient.RegisterWorkload(ctx.Workload); err != nil { t.Fatalf("unexpected error registering task: %v", err) } if err := ctx.syncOnce(syncNewOps); err != nil { t.Fatalf("unexpected error syncing task: %v", err) } if n := len(ctx.FakeConsul.services); n != 1 { t.Fatalf("expected 1 service but found %d:\n%#v", n, ctx.FakeConsul.services) } for _, v := range ctx.FakeConsul.services { if v.Name != ctx.Workload.Services[0].Name { t.Errorf("expected Name=%q != %q", ctx.Workload.Services[0].Name, v.Name) } if !reflect.DeepEqual(v.Tags, ctx.Workload.Services[0].Tags) { t.Errorf("expected Tags=%v != %v", ctx.Workload.Services[0].Tags, v.Tags) } if v.Port != xPort { t.Errorf("expected Port=%d != %d", xPort, v.Port) } } // Assert the check update is pending if n := len(ctx.ServiceClient.checkWatcher.checkUpdateCh); n != 1 { t.Fatalf("expected 1 check restart update but found %d", n) } // Assert the check update is properly formed checkUpd := <-ctx.ServiceClient.checkWatcher.checkUpdateCh if checkUpd.checkRestart.allocID != ctx.Workload.AllocID { t.Fatalf("expected check's allocid to be %q but found %q", "allocid", checkUpd.checkRestart.allocID) } if expected := 200 * time.Millisecond; checkUpd.checkRestart.timeLimit != expected { t.Fatalf("expected check's time limit to be %v but found %v", expected, checkUpd.checkRestart.timeLimit) } // Make a change which will register a new service ctx.Workload.Services[0].Name = "taskname-service2" ctx.Workload.Services[0].Tags[0] = "tag3" if err := ctx.ServiceClient.RegisterWorkload(ctx.Workload); err != nil { t.Fatalf("unexpected error registering task: %v", err) } // Assert check update is pending if n := len(ctx.ServiceClient.checkWatcher.checkUpdateCh); n != 1 { t.Fatalf("expected 1 check restart update but found %d", n) } // Assert the check update's id has changed checkUpd2 := <-ctx.ServiceClient.checkWatcher.checkUpdateCh if checkUpd.checkID == checkUpd2.checkID { t.Fatalf("expected new check update to have a new ID both both have: %q", checkUpd.checkID) } // Make sure changes don't take affect until sync() is called (since // Run() isn't running) if n := len(ctx.FakeConsul.services); n != 1 { t.Fatalf("expected 1 service but found %d:\n%#v", n, ctx.FakeConsul.services) } for _, v := range ctx.FakeConsul.services { if reflect.DeepEqual(v.Tags, ctx.Workload.Services[0].Tags) { t.Errorf("expected Tags to differ, changes applied before sync()") } } // Now sync() and re-check for the applied updates if err := ctx.syncOnce(syncNewOps); err != nil { t.Fatalf("unexpected error syncing task: %v", err) } if n := len(ctx.FakeConsul.services); n != 2 { t.Fatalf("expected 2 services but found %d:\n%#v", n, ctx.FakeConsul.services) } found := false for _, v := range ctx.FakeConsul.services { if v.Name == ctx.Workload.Services[0].Name { if found { t.Fatalf("found new service name %q twice", v.Name) } found = true if !reflect.DeepEqual(v.Tags, ctx.Workload.Services[0].Tags) { t.Errorf("expected Tags=%v != %v", ctx.Workload.Services[0].Tags, v.Tags) } } } if !found { t.Fatalf("did not find new service %q", ctx.Workload.Services[0].Name) } // Remove the new task ctx.ServiceClient.RemoveWorkload(ctx.Workload) if err := ctx.syncOnce(syncNewOps); err != nil { t.Fatalf("unexpected error syncing task: %v", err) } if n := len(ctx.FakeConsul.services); n != 1 { t.Fatalf("expected 1 service but found %d:\n%#v", n, ctx.FakeConsul.services) } for _, v := range ctx.FakeConsul.services { if v.Name != "taskname-service" { t.Errorf("expected original task to survive not %q", v.Name) } } // Assert check update is pending if n := len(ctx.ServiceClient.checkWatcher.checkUpdateCh); n != 1 { t.Fatalf("expected 1 check restart update but found %d", n) } // Assert the check update's id is correct and that it's a removal checkUpd3 := <-ctx.ServiceClient.checkWatcher.checkUpdateCh if checkUpd2.checkID != checkUpd3.checkID { t.Fatalf("expected checkid %q but found %q", checkUpd2.checkID, checkUpd3.checkID) } if !checkUpd3.remove { t.Fatalf("expected check watch removal update but found: %#v", checkUpd3) } } // TestConsul_ShutdownOK tests the ok path for the shutdown logic in // ServiceClient. func TestConsul_ShutdownOK(t *testing.T) { require := require.New(t) ctx := setupFake(t) go ctx.ServiceClient.Run() // register the Nomad agent service and check agentServices := []*structs.Service{ { Name: "http", Tags: []string{"nomad"}, PortLabel: "localhost:2345", Checks: []*structs.ServiceCheck{ { Name: "nomad-tcp", Type: "tcp", Interval: 9000 * time.Hour, // make check block Timeout: 10 * time.Second, InitialStatus: "warning", }, }, }, } require.NoError(ctx.ServiceClient.RegisterAgent("client", agentServices)) require.Eventually(ctx.ServiceClient.hasSeen, time.Second, 10*time.Millisecond) // assert successful registration require.Len(ctx.FakeConsul.services, 1, "expected agent service to be registered") require.Len(ctx.FakeConsul.checks, 1, "expected agent check to be registered") require.Contains(ctx.FakeConsul.services, makeAgentServiceID("client", agentServices[0])) // Shutdown() should block until Nomad agent service/check is deregistered require.NoError(ctx.ServiceClient.Shutdown()) require.Len(ctx.FakeConsul.services, 0, "expected agent service to be deregistered") require.Len(ctx.FakeConsul.checks, 0, "expected agent check to be deregistered") } // TestConsul_ShutdownBlocked tests the blocked past deadline path for the // shutdown logic in ServiceClient. func TestConsul_ShutdownBlocked(t *testing.T) { require := require.New(t) t.Parallel() ctx := setupFake(t) // can be short because we're intentionally blocking, but needs to // be longer than the time we'll block Consul so we can be sure // we're not delayed either. ctx.ServiceClient.shutdownWait = time.Second go ctx.ServiceClient.Run() // register the Nomad agent service and check agentServices := []*structs.Service{ { Name: "http", Tags: []string{"nomad"}, PortLabel: "localhost:2345", Checks: []*structs.ServiceCheck{ { Name: "nomad-tcp", Type: "tcp", Interval: 9000 * time.Hour, // make check block Timeout: 10 * time.Second, InitialStatus: "warning", }, }, }, } require.NoError(ctx.ServiceClient.RegisterAgent("client", agentServices)) require.Eventually(ctx.ServiceClient.hasSeen, time.Second, 10*time.Millisecond) require.Len(ctx.FakeConsul.services, 1, "expected agent service to be registered") require.Len(ctx.FakeConsul.checks, 1, "expected agent check to be registered") // prevent normal shutdown by blocking Consul. the shutdown should wait // until agent deregistration has finished waiter := make(chan struct{}) result := make(chan error) go func() { ctx.FakeConsul.mu.Lock() close(waiter) result <- ctx.ServiceClient.Shutdown() }() <-waiter // wait for lock to be hit // Shutdown should block until all enqueued operations finish. preShutdown := time.Now() select { case <-time.After(200 * time.Millisecond): ctx.FakeConsul.mu.Unlock() require.NoError(<-result) case <-result: t.Fatal("should not have received result until Consul unblocked") } shutdownTime := time.Now().Sub(preShutdown).Seconds() require.Less(shutdownTime, time.Second.Seconds(), "expected shutdown to take >200ms and <1s") require.Greater(shutdownTime, 200*time.Millisecond.Seconds(), "expected shutdown to take >200ms and <1s") require.Len(ctx.FakeConsul.services, 0, "expected agent service to be deregistered") require.Len(ctx.FakeConsul.checks, 0, "expected agent check to be deregistered") } // TestConsul_DriverNetwork_AutoUse asserts that if a driver network has // auto-use set then services should advertise it unless explicitly set to // host. Checks should always use host. func TestConsul_DriverNetwork_AutoUse(t *testing.T) { t.Parallel() ctx := setupFake(t) ctx.Workload.Services = []*structs.Service{ { Name: "auto-advertise-x", PortLabel: "x", AddressMode: structs.AddressModeAuto, Checks: []*structs.ServiceCheck{ { Name: "default-check-x", Type: "tcp", Interval: time.Second, Timeout: time.Second, }, { Name: "weird-y-check", Type: "http", Interval: time.Second, Timeout: time.Second, PortLabel: "y", }, }, }, { Name: "driver-advertise-y", PortLabel: "y", AddressMode: structs.AddressModeDriver, Checks: []*structs.ServiceCheck{ { Name: "default-check-y", Type: "tcp", Interval: time.Second, Timeout: time.Second, }, }, }, { Name: "host-advertise-y", PortLabel: "y", AddressMode: structs.AddressModeHost, }, } ctx.Workload.DriverNetwork = &drivers.DriverNetwork{ PortMap: map[string]int{ "x": 8888, "y": 9999, }, IP: "172.18.0.2", AutoAdvertise: true, } if err := ctx.ServiceClient.RegisterWorkload(ctx.Workload); err != nil { t.Fatalf("unexpected error registering task: %v", err) } if err := ctx.syncOnce(syncNewOps); err != nil { t.Fatalf("unexpected error syncing task: %v", err) } if n := len(ctx.FakeConsul.services); n != 3 { t.Fatalf("expected 2 services but found: %d", n) } for _, v := range ctx.FakeConsul.services { switch v.Name { case ctx.Workload.Services[0].Name: // x // Since DriverNetwork.AutoAdvertise=true, driver ports should be used if v.Port != ctx.Workload.DriverNetwork.PortMap["x"] { t.Errorf("expected service %s's port to be %d but found %d", v.Name, ctx.Workload.DriverNetwork.PortMap["x"], v.Port) } // The order of checks in Consul is not guaranteed to // be the same as their order in the Workload definition, // so check in a loop if expected := 2; len(v.Checks) != expected { t.Errorf("expected %d checks but found %d", expected, len(v.Checks)) } for _, c := range v.Checks { // No name on AgentServiceChecks, use type switch { case c.TCP != "": // Checks should always use host port though if c.TCP != ":1234" { // xPort t.Errorf("expected service %s check 1's port to be %d but found %q", v.Name, xPort, c.TCP) } case c.HTTP != "": if c.HTTP != "http://:1235" { // yPort t.Errorf("expected service %s check 2's port to be %d but found %q", v.Name, yPort, c.HTTP) } default: t.Errorf("unexpected check %#v on service %q", c, v.Name) } } case ctx.Workload.Services[1].Name: // y // Service should be container ip:port if v.Address != ctx.Workload.DriverNetwork.IP { t.Errorf("expected service %s's address to be %s but found %s", v.Name, ctx.Workload.DriverNetwork.IP, v.Address) } if v.Port != ctx.Workload.DriverNetwork.PortMap["y"] { t.Errorf("expected service %s's port to be %d but found %d", v.Name, ctx.Workload.DriverNetwork.PortMap["x"], v.Port) } // Check should be host ip:port if v.Checks[0].TCP != ":1235" { // yPort t.Errorf("expected service %s check's port to be %d but found %s", v.Name, yPort, v.Checks[0].TCP) } case ctx.Workload.Services[2].Name: // y + host mode if v.Port != yPort { t.Errorf("expected service %s's port to be %d but found %d", v.Name, yPort, v.Port) } default: t.Errorf("unexpected service name: %q", v.Name) } } } // TestConsul_DriverNetwork_NoAutoUse asserts that if a driver network doesn't // set auto-use only services which request the driver's network should // advertise it. func TestConsul_DriverNetwork_NoAutoUse(t *testing.T) { t.Parallel() ctx := setupFake(t) ctx.Workload.Services = []*structs.Service{ { Name: "auto-advertise-x", PortLabel: "x", AddressMode: structs.AddressModeAuto, }, { Name: "driver-advertise-y", PortLabel: "y", AddressMode: structs.AddressModeDriver, }, { Name: "host-advertise-y", PortLabel: "y", AddressMode: structs.AddressModeHost, }, } ctx.Workload.DriverNetwork = &drivers.DriverNetwork{ PortMap: map[string]int{ "x": 8888, "y": 9999, }, IP: "172.18.0.2", AutoAdvertise: false, } if err := ctx.ServiceClient.RegisterWorkload(ctx.Workload); err != nil { t.Fatalf("unexpected error registering task: %v", err) } if err := ctx.syncOnce(syncNewOps); err != nil { t.Fatalf("unexpected error syncing task: %v", err) } if n := len(ctx.FakeConsul.services); n != 3 { t.Fatalf("expected 3 services but found: %d", n) } for _, v := range ctx.FakeConsul.services { switch v.Name { case ctx.Workload.Services[0].Name: // x + auto // Since DriverNetwork.AutoAdvertise=false, host ports should be used if v.Port != xPort { t.Errorf("expected service %s's port to be %d but found %d", v.Name, xPort, v.Port) } case ctx.Workload.Services[1].Name: // y + driver mode // Service should be container ip:port if v.Address != ctx.Workload.DriverNetwork.IP { t.Errorf("expected service %s's address to be %s but found %s", v.Name, ctx.Workload.DriverNetwork.IP, v.Address) } if v.Port != ctx.Workload.DriverNetwork.PortMap["y"] { t.Errorf("expected service %s's port to be %d but found %d", v.Name, ctx.Workload.DriverNetwork.PortMap["x"], v.Port) } case ctx.Workload.Services[2].Name: // y + host mode if v.Port != yPort { t.Errorf("expected service %s's port to be %d but found %d", v.Name, yPort, v.Port) } default: t.Errorf("unexpected service name: %q", v.Name) } } } // TestConsul_DriverNetwork_Change asserts that if a driver network is // specified and a service updates its use its properly updated in Consul. func TestConsul_DriverNetwork_Change(t *testing.T) { t.Parallel() ctx := setupFake(t) ctx.Workload.Services = []*structs.Service{ { Name: "service-foo", PortLabel: "x", AddressMode: structs.AddressModeAuto, }, } ctx.Workload.DriverNetwork = &drivers.DriverNetwork{ PortMap: map[string]int{ "x": 8888, "y": 9999, }, IP: "172.18.0.2", AutoAdvertise: false, } syncAndAssertPort := func(port int) { if err := ctx.syncOnce(syncNewOps); err != nil { t.Fatalf("unexpected error syncing task: %v", err) } if n := len(ctx.FakeConsul.services); n != 1 { t.Fatalf("expected 1 service but found: %d", n) } for _, v := range ctx.FakeConsul.services { switch v.Name { case ctx.Workload.Services[0].Name: if v.Port != port { t.Errorf("expected service %s's port to be %d but found %d", v.Name, port, v.Port) } default: t.Errorf("unexpected service name: %q", v.Name) } } } // Initial service should advertise host port x if err := ctx.ServiceClient.RegisterWorkload(ctx.Workload); err != nil { t.Fatalf("unexpected error registering task: %v", err) } syncAndAssertPort(xPort) // UpdateWorkload to use Host (shouldn't change anything) origWorkload := ctx.Workload.Copy() ctx.Workload.Services[0].AddressMode = structs.AddressModeHost if err := ctx.ServiceClient.UpdateWorkload(origWorkload, ctx.Workload); err != nil { t.Fatalf("unexpected error updating task: %v", err) } syncAndAssertPort(xPort) // UpdateWorkload to use Driver (*should* change IP and port) origWorkload = ctx.Workload.Copy() ctx.Workload.Services[0].AddressMode = structs.AddressModeDriver if err := ctx.ServiceClient.UpdateWorkload(origWorkload, ctx.Workload); err != nil { t.Fatalf("unexpected error updating task: %v", err) } syncAndAssertPort(ctx.Workload.DriverNetwork.PortMap["x"]) } // TestConsul_CanaryTags asserts CanaryTags are used when Canary=true func TestConsul_CanaryTags(t *testing.T) { t.Parallel() require := require.New(t) ctx := setupFake(t) canaryTags := []string{"tag1", "canary"} ctx.Workload.Canary = true ctx.Workload.Services[0].CanaryTags = canaryTags require.NoError(ctx.ServiceClient.RegisterWorkload(ctx.Workload)) require.NoError(ctx.syncOnce(syncNewOps)) require.Len(ctx.FakeConsul.services, 1) for _, service := range ctx.FakeConsul.services { require.Equal(canaryTags, service.Tags) } // Disable canary and assert tags are not the canary tags origWorkload := ctx.Workload.Copy() ctx.Workload.Canary = false require.NoError(ctx.ServiceClient.UpdateWorkload(origWorkload, ctx.Workload)) require.NoError(ctx.syncOnce(syncNewOps)) require.Len(ctx.FakeConsul.services, 1) for _, service := range ctx.FakeConsul.services { require.NotEqual(canaryTags, service.Tags) } ctx.ServiceClient.RemoveWorkload(ctx.Workload) require.NoError(ctx.syncOnce(syncNewOps)) require.Len(ctx.FakeConsul.services, 0) } // TestConsul_CanaryTags_NoTags asserts Tags are used when Canary=true and there // are no specified canary tags func TestConsul_CanaryTags_NoTags(t *testing.T) { t.Parallel() require := require.New(t) ctx := setupFake(t) tags := []string{"tag1", "foo"} ctx.Workload.Canary = true ctx.Workload.Services[0].Tags = tags require.NoError(ctx.ServiceClient.RegisterWorkload(ctx.Workload)) require.NoError(ctx.syncOnce(syncNewOps)) require.Len(ctx.FakeConsul.services, 1) for _, service := range ctx.FakeConsul.services { require.Equal(tags, service.Tags) } // Disable canary and assert tags dont change origWorkload := ctx.Workload.Copy() ctx.Workload.Canary = false require.NoError(ctx.ServiceClient.UpdateWorkload(origWorkload, ctx.Workload)) require.NoError(ctx.syncOnce(syncNewOps)) require.Len(ctx.FakeConsul.services, 1) for _, service := range ctx.FakeConsul.services { require.Equal(tags, service.Tags) } ctx.ServiceClient.RemoveWorkload(ctx.Workload) require.NoError(ctx.syncOnce(syncNewOps)) require.Len(ctx.FakeConsul.services, 0) } // TestConsul_CanaryMeta asserts CanaryMeta are used when Canary=true func TestConsul_CanaryMeta(t *testing.T) { t.Parallel() require := require.New(t) ctx := setupFake(t) canaryMeta := map[string]string{"meta1": "canary"} canaryMeta["external-source"] = "nomad" ctx.Workload.Canary = true ctx.Workload.Services[0].CanaryMeta = canaryMeta require.NoError(ctx.ServiceClient.RegisterWorkload(ctx.Workload)) require.NoError(ctx.syncOnce(syncNewOps)) require.Len(ctx.FakeConsul.services, 1) for _, service := range ctx.FakeConsul.services { require.Equal(canaryMeta, service.Meta) } // Disable canary and assert meta are not the canary meta origWorkload := ctx.Workload.Copy() ctx.Workload.Canary = false require.NoError(ctx.ServiceClient.UpdateWorkload(origWorkload, ctx.Workload)) require.NoError(ctx.syncOnce(syncNewOps)) require.Len(ctx.FakeConsul.services, 1) for _, service := range ctx.FakeConsul.services { require.NotEqual(canaryMeta, service.Meta) } ctx.ServiceClient.RemoveWorkload(ctx.Workload) require.NoError(ctx.syncOnce(syncNewOps)) require.Len(ctx.FakeConsul.services, 0) } // TestConsul_CanaryMeta_NoMeta asserts Meta are used when Canary=true and there // are no specified canary meta func TestConsul_CanaryMeta_NoMeta(t *testing.T) { t.Parallel() require := require.New(t) ctx := setupFake(t) meta := map[string]string{"meta1": "foo"} meta["external-source"] = "nomad" ctx.Workload.Canary = true ctx.Workload.Services[0].Meta = meta require.NoError(ctx.ServiceClient.RegisterWorkload(ctx.Workload)) require.NoError(ctx.syncOnce(syncNewOps)) require.Len(ctx.FakeConsul.services, 1) for _, service := range ctx.FakeConsul.services { require.Equal(meta, service.Meta) } // Disable canary and assert meta dont change origWorkload := ctx.Workload.Copy() ctx.Workload.Canary = false require.NoError(ctx.ServiceClient.UpdateWorkload(origWorkload, ctx.Workload)) require.NoError(ctx.syncOnce(syncNewOps)) require.Len(ctx.FakeConsul.services, 1) for _, service := range ctx.FakeConsul.services { require.Equal(meta, service.Meta) } ctx.ServiceClient.RemoveWorkload(ctx.Workload) require.NoError(ctx.syncOnce(syncNewOps)) require.Len(ctx.FakeConsul.services, 0) } // TestConsul_PeriodicSync asserts that Nomad periodically reconciles with // Consul. func TestConsul_PeriodicSync(t *testing.T) { t.Parallel() ctx := setupFake(t) defer ctx.ServiceClient.Shutdown() // Lower periodic sync interval to speed up test ctx.ServiceClient.periodicInterval = 1 * time.Millisecond // Run for 20ms and assert hits >= 5 because each sync() calls multiple // Consul APIs go ctx.ServiceClient.Run() select { case <-ctx.ServiceClient.exitCh: t.Fatalf("exited unexpectedly") case <-time.After(20 * time.Millisecond): } minHits := 5 if hits := ctx.FakeConsul.getHits(); hits < minHits { t.Fatalf("expected at least %d hits but found %d", minHits, hits) } } // TestIsNomadService asserts the isNomadService helper returns true for Nomad // task IDs and false for unknown IDs and Nomad agent IDs (see #2827). func TestIsNomadService(t *testing.T) { t.Parallel() tests := []struct { id string result bool }{ {"_nomad-client-nomad-client-http", false}, {"_nomad-server-nomad-serf", false}, // Pre-0.7.1 style IDs still match {"_nomad-executor-abc", true}, {"_nomad-executor", true}, // Post-0.7.1 style IDs match {"_nomad-task-FBBK265QN4TMT25ND4EP42TJVMYJ3HR4", true}, {"not-nomad", false}, {"_nomad", false}, } for _, test := range tests { t.Run(test.id, func(t *testing.T) { actual := isNomadService(test.id) if actual != test.result { t.Errorf("%q should be %t but found %t", test.id, test.result, actual) } }) } } // TestCreateCheckReg_HTTP asserts Nomad ServiceCheck structs are properly // converted to Consul API AgentCheckRegistrations for HTTP checks. func TestCreateCheckReg_HTTP(t *testing.T) { t.Parallel() check := &structs.ServiceCheck{ Name: "name", Type: "http", Path: "/path", PortLabel: "label", Method: "POST", Header: map[string][]string{ "Foo": {"bar"}, }, } serviceID := "testService" checkID := check.Hash(serviceID) host := "localhost" port := 41111 expected := &api.AgentCheckRegistration{ ID: checkID, Name: "name", ServiceID: serviceID, AgentServiceCheck: api.AgentServiceCheck{ Timeout: "0s", Interval: "0s", HTTP: fmt.Sprintf("http://%s:%d/path", host, port), Method: "POST", Header: map[string][]string{ "Foo": {"bar"}, }, }, } actual, err := createCheckReg(serviceID, checkID, check, host, port) if err != nil { t.Fatalf("err: %v", err) } if diff := pretty.Diff(actual, expected); len(diff) > 0 { t.Fatalf("diff:\n%s\n", strings.Join(diff, "\n")) } } // TestCreateCheckReg_GRPC asserts Nomad ServiceCheck structs are properly // converted to Consul API AgentCheckRegistrations for GRPC checks. func TestCreateCheckReg_GRPC(t *testing.T) { t.Parallel() check := &structs.ServiceCheck{ Name: "name", Type: "grpc", PortLabel: "label", GRPCService: "foo.Bar", GRPCUseTLS: true, TLSSkipVerify: true, Timeout: time.Second, Interval: time.Minute, } serviceID := "testService" checkID := check.Hash(serviceID) expected := &api.AgentCheckRegistration{ ID: checkID, Name: "name", ServiceID: serviceID, AgentServiceCheck: api.AgentServiceCheck{ Timeout: "1s", Interval: "1m0s", GRPC: "localhost:8080/foo.Bar", GRPCUseTLS: true, TLSSkipVerify: true, }, } actual, err := createCheckReg(serviceID, checkID, check, "localhost", 8080) require.NoError(t, err) require.Equal(t, expected, actual) } // TestGetAddress asserts Nomad uses the correct ip and port for services and // checks depending on port labels, driver networks, and address mode. func TestGetAddress(t *testing.T) { const HostIP = "127.0.0.1" cases := []struct { Name string // Parameters Mode string PortLabel string Host map[string]int // will be converted to structs.Networks Driver *drivers.DriverNetwork Ports structs.AllocatedPorts Status *structs.AllocNetworkStatus // Results ExpectedIP string ExpectedPort int ExpectedErr string }{ // Valid Configurations { Name: "ExampleService", Mode: structs.AddressModeAuto, PortLabel: "db", Host: map[string]int{"db": 12435}, Driver: &drivers.DriverNetwork{ PortMap: map[string]int{"db": 6379}, IP: "10.1.2.3", }, ExpectedIP: HostIP, ExpectedPort: 12435, }, { Name: "Host", Mode: structs.AddressModeHost, PortLabel: "db", Host: map[string]int{"db": 12345}, Driver: &drivers.DriverNetwork{ PortMap: map[string]int{"db": 6379}, IP: "10.1.2.3", }, ExpectedIP: HostIP, ExpectedPort: 12345, }, { Name: "Driver", Mode: structs.AddressModeDriver, PortLabel: "db", Host: map[string]int{"db": 12345}, Driver: &drivers.DriverNetwork{ PortMap: map[string]int{"db": 6379}, IP: "10.1.2.3", }, ExpectedIP: "10.1.2.3", ExpectedPort: 6379, }, { Name: "AutoDriver", Mode: structs.AddressModeAuto, PortLabel: "db", Host: map[string]int{"db": 12345}, Driver: &drivers.DriverNetwork{ PortMap: map[string]int{"db": 6379}, IP: "10.1.2.3", AutoAdvertise: true, }, ExpectedIP: "10.1.2.3", ExpectedPort: 6379, }, { Name: "DriverCustomPort", Mode: structs.AddressModeDriver, PortLabel: "7890", Host: map[string]int{"db": 12345}, Driver: &drivers.DriverNetwork{ PortMap: map[string]int{"db": 6379}, IP: "10.1.2.3", }, ExpectedIP: "10.1.2.3", ExpectedPort: 7890, }, // Invalid Configurations { Name: "DriverWithoutNetwork", Mode: structs.AddressModeDriver, PortLabel: "db", Host: map[string]int{"db": 12345}, Driver: nil, ExpectedErr: "no driver network exists", }, { Name: "DriverBadPort", Mode: structs.AddressModeDriver, PortLabel: "bad-port-label", Host: map[string]int{"db": 12345}, Driver: &drivers.DriverNetwork{ PortMap: map[string]int{"db": 6379}, IP: "10.1.2.3", }, ExpectedErr: "invalid port", }, { Name: "DriverZeroPort", Mode: structs.AddressModeDriver, PortLabel: "0", Driver: &drivers.DriverNetwork{ IP: "10.1.2.3", }, ExpectedErr: "invalid port", }, { Name: "HostBadPort", Mode: structs.AddressModeHost, PortLabel: "bad-port-label", ExpectedErr: "invalid port", }, { Name: "InvalidMode", Mode: "invalid-mode", PortLabel: "80", ExpectedErr: "invalid address mode", }, { Name: "NoPort_AutoMode", Mode: structs.AddressModeAuto, ExpectedIP: HostIP, }, { Name: "NoPort_HostMode", Mode: structs.AddressModeHost, ExpectedIP: HostIP, }, { Name: "NoPort_DriverMode", Mode: structs.AddressModeDriver, Driver: &drivers.DriverNetwork{ IP: "10.1.2.3", }, ExpectedIP: "10.1.2.3", }, // Scenarios using port 0.12 networking fields (NetworkStatus, AllocatedPortMapping) { Name: "ExampleServer_withAllocatedPorts", Mode: structs.AddressModeAuto, PortLabel: "db", Ports: []structs.AllocatedPortMapping{ { Label: "db", Value: 12435, To: 6379, HostIP: HostIP, }, }, Status: &structs.AllocNetworkStatus{ InterfaceName: "eth0", Address: "172.26.0.1", }, ExpectedIP: HostIP, ExpectedPort: 12435, }, { Name: "Host_withAllocatedPorts", Mode: structs.AddressModeHost, PortLabel: "db", Ports: []structs.AllocatedPortMapping{ { Label: "db", Value: 12345, To: 6379, HostIP: HostIP, }, }, Status: &structs.AllocNetworkStatus{ InterfaceName: "eth0", Address: "172.26.0.1", }, ExpectedIP: HostIP, ExpectedPort: 12345, }, { Name: "Driver_withAllocatedPorts", Mode: structs.AddressModeDriver, PortLabel: "db", Ports: []structs.AllocatedPortMapping{ { Label: "db", Value: 12345, To: 6379, HostIP: HostIP, }, }, Driver: &drivers.DriverNetwork{ IP: "10.1.2.3", }, Status: &structs.AllocNetworkStatus{ InterfaceName: "eth0", Address: "172.26.0.1", }, ExpectedIP: "10.1.2.3", ExpectedPort: 6379, }, { Name: "AutoDriver_withAllocatedPorts", Mode: structs.AddressModeAuto, PortLabel: "db", Ports: []structs.AllocatedPortMapping{ { Label: "db", Value: 12345, To: 6379, HostIP: HostIP, }, }, Driver: &drivers.DriverNetwork{ IP: "10.1.2.3", AutoAdvertise: true, }, Status: &structs.AllocNetworkStatus{ InterfaceName: "eth0", Address: "172.26.0.1", }, ExpectedIP: "10.1.2.3", ExpectedPort: 6379, }, { Name: "DriverCustomPort_withAllocatedPorts", Mode: structs.AddressModeDriver, PortLabel: "7890", Ports: []structs.AllocatedPortMapping{ { Label: "db", Value: 12345, To: 6379, HostIP: HostIP, }, }, Driver: &drivers.DriverNetwork{ IP: "10.1.2.3", }, Status: &structs.AllocNetworkStatus{ InterfaceName: "eth0", Address: "172.26.0.1", }, ExpectedIP: "10.1.2.3", ExpectedPort: 7890, }, { Name: "Host_MultiHostInterface", Mode: structs.AddressModeAuto, PortLabel: "db", Ports: []structs.AllocatedPortMapping{ { Label: "db", Value: 12345, To: 6379, HostIP: "127.0.0.100", }, }, Status: &structs.AllocNetworkStatus{ InterfaceName: "eth0", Address: "172.26.0.1", }, ExpectedIP: "127.0.0.100", ExpectedPort: 12345, }, { Name: "Alloc", Mode: structs.AddressModeAlloc, PortLabel: "db", Ports: []structs.AllocatedPortMapping{ { Label: "db", Value: 12345, To: 6379, HostIP: HostIP, }, }, Status: &structs.AllocNetworkStatus{ InterfaceName: "eth0", Address: "172.26.0.1", }, ExpectedIP: "172.26.0.1", ExpectedPort: 6379, }, { Name: "Alloc no to value", Mode: structs.AddressModeAlloc, PortLabel: "db", Ports: []structs.AllocatedPortMapping{ { Label: "db", Value: 12345, HostIP: HostIP, }, }, Status: &structs.AllocNetworkStatus{ InterfaceName: "eth0", Address: "172.26.0.1", }, ExpectedIP: "172.26.0.1", ExpectedPort: 12345, }, { Name: "AllocCustomPort", Mode: structs.AddressModeAlloc, PortLabel: "6379", Status: &structs.AllocNetworkStatus{ InterfaceName: "eth0", Address: "172.26.0.1", }, ExpectedIP: "172.26.0.1", ExpectedPort: 6379, }, } for _, tc := range cases { t.Run(tc.Name, func(t *testing.T) { // convert host port map into a structs.Networks networks := []*structs.NetworkResource{ { IP: HostIP, ReservedPorts: make([]structs.Port, len(tc.Host)), }, } i := 0 for label, port := range tc.Host { networks[0].ReservedPorts[i].Label = label networks[0].ReservedPorts[i].Value = port i++ } // Run getAddress ip, port, err := getAddress(tc.Mode, tc.PortLabel, networks, tc.Driver, tc.Ports, tc.Status) // Assert the results assert.Equal(t, tc.ExpectedIP, ip, "IP mismatch") assert.Equal(t, tc.ExpectedPort, port, "Port mismatch") if tc.ExpectedErr == "" { assert.Nil(t, err) } else { if err == nil { t.Fatalf("expected error containing %q but err=nil", tc.ExpectedErr) } else { assert.Contains(t, err.Error(), tc.ExpectedErr) } } }) } } func TestConsul_ServiceName_Duplicates(t *testing.T) { t.Parallel() ctx := setupFake(t) require := require.New(t) ctx.Workload.Services = []*structs.Service{ { Name: "best-service", PortLabel: "x", Tags: []string{ "foo", }, Checks: []*structs.ServiceCheck{ { Name: "check-a", Type: "tcp", Interval: time.Second, Timeout: time.Second, }, }, }, { Name: "best-service", PortLabel: "y", Tags: []string{ "bar", }, Checks: []*structs.ServiceCheck{ { Name: "checky-mccheckface", Type: "tcp", Interval: time.Second, Timeout: time.Second, }, }, }, { Name: "worst-service", PortLabel: "y", }, } require.NoError(ctx.ServiceClient.RegisterWorkload(ctx.Workload)) require.NoError(ctx.syncOnce(syncNewOps)) require.Len(ctx.FakeConsul.services, 3) for _, v := range ctx.FakeConsul.services { if v.Name == ctx.Workload.Services[0].Name && v.Port == xPort { require.ElementsMatch(v.Tags, ctx.Workload.Services[0].Tags) require.Len(v.Checks, 1) } else if v.Name == ctx.Workload.Services[1].Name && v.Port == yPort { require.ElementsMatch(v.Tags, ctx.Workload.Services[1].Tags) require.Len(v.Checks, 1) } else if v.Name == ctx.Workload.Services[2].Name { require.Len(v.Checks, 0) } } } // TestConsul_ServiceDeregistration_OutOfProbation asserts that during in steady // state we remove any services we don't reconize locally func TestConsul_ServiceDeregistration_OutProbation(t *testing.T) { t.Parallel() ctx := setupFake(t) require := require.New(t) ctx.ServiceClient.deregisterProbationExpiry = time.Now().Add(-1 * time.Hour) remainingWorkload := testWorkload() remainingWorkload.Services = []*structs.Service{ { Name: "remaining-service", PortLabel: "x", Checks: []*structs.ServiceCheck{ { Name: "check", Type: "tcp", Interval: time.Second, Timeout: time.Second, }, }, }, } remainingWorkloadServiceID := MakeAllocServiceID(remainingWorkload.AllocID, remainingWorkload.Name(), remainingWorkload.Services[0]) require.NoError(ctx.ServiceClient.RegisterWorkload(remainingWorkload)) require.NoError(ctx.syncOnce(syncNewOps)) require.Len(ctx.FakeConsul.services, 1) require.Len(ctx.FakeConsul.checks, 1) explicitlyRemovedWorkload := testWorkload() explicitlyRemovedWorkload.Services = []*structs.Service{ { Name: "explicitly-removed-service", PortLabel: "y", Checks: []*structs.ServiceCheck{ { Name: "check", Type: "tcp", Interval: time.Second, Timeout: time.Second, }, }, }, } explicitlyRemovedWorkloadServiceID := MakeAllocServiceID(explicitlyRemovedWorkload.AllocID, explicitlyRemovedWorkload.Name(), explicitlyRemovedWorkload.Services[0]) require.NoError(ctx.ServiceClient.RegisterWorkload(explicitlyRemovedWorkload)) require.NoError(ctx.syncOnce(syncNewOps)) require.Len(ctx.FakeConsul.services, 2) require.Len(ctx.FakeConsul.checks, 2) // we register a task through nomad API then remove it out of band outofbandWorkload := testWorkload() outofbandWorkload.Services = []*structs.Service{ { Name: "unknown-service", PortLabel: "x", Checks: []*structs.ServiceCheck{ { Name: "check", Type: "tcp", Interval: time.Second, Timeout: time.Second, }, }, }, } outofbandWorkloadServiceID := MakeAllocServiceID(outofbandWorkload.AllocID, outofbandWorkload.Name(), outofbandWorkload.Services[0]) require.NoError(ctx.ServiceClient.RegisterWorkload(outofbandWorkload)) require.NoError(ctx.syncOnce(syncNewOps)) require.Len(ctx.FakeConsul.services, 3) // remove outofbandWorkload from local services so it appears unknown to client require.Len(ctx.ServiceClient.services, 3) require.Len(ctx.ServiceClient.checks, 3) delete(ctx.ServiceClient.services, outofbandWorkloadServiceID) delete(ctx.ServiceClient.checks, MakeCheckID(outofbandWorkloadServiceID, outofbandWorkload.Services[0].Checks[0])) require.Len(ctx.ServiceClient.services, 2) require.Len(ctx.ServiceClient.checks, 2) // Sync and ensure that explicitly removed service as well as outofbandWorkload were removed ctx.ServiceClient.RemoveWorkload(explicitlyRemovedWorkload) require.NoError(ctx.syncOnce(syncNewOps)) require.NoError(ctx.ServiceClient.sync(syncNewOps)) require.Len(ctx.FakeConsul.services, 1) require.Len(ctx.FakeConsul.checks, 1) require.Contains(ctx.FakeConsul.services, remainingWorkloadServiceID) require.NotContains(ctx.FakeConsul.services, outofbandWorkloadServiceID) require.NotContains(ctx.FakeConsul.services, explicitlyRemovedWorkloadServiceID) require.Contains(ctx.FakeConsul.checks, MakeCheckID(remainingWorkloadServiceID, remainingWorkload.Services[0].Checks[0])) require.NotContains(ctx.FakeConsul.checks, MakeCheckID(outofbandWorkloadServiceID, outofbandWorkload.Services[0].Checks[0])) require.NotContains(ctx.FakeConsul.checks, MakeCheckID(explicitlyRemovedWorkloadServiceID, explicitlyRemovedWorkload.Services[0].Checks[0])) } // TestConsul_ServiceDeregistration_InProbation asserts that during initialization // we only deregister services that were explicitly removed and leave unknown // services untouched. This adds a grace period for restoring recovered tasks // before deregistering them func TestConsul_ServiceDeregistration_InProbation(t *testing.T) { t.Parallel() ctx := setupFake(t) require := require.New(t) ctx.ServiceClient.deregisterProbationExpiry = time.Now().Add(1 * time.Hour) remainingWorkload := testWorkload() remainingWorkload.Services = []*structs.Service{ { Name: "remaining-service", PortLabel: "x", Checks: []*structs.ServiceCheck{ { Name: "check", Type: "tcp", Interval: time.Second, Timeout: time.Second, }, }, }, } remainingWorkloadServiceID := MakeAllocServiceID(remainingWorkload.AllocID, remainingWorkload.Name(), remainingWorkload.Services[0]) require.NoError(ctx.ServiceClient.RegisterWorkload(remainingWorkload)) require.NoError(ctx.syncOnce(syncNewOps)) require.Len(ctx.FakeConsul.services, 1) require.Len(ctx.FakeConsul.checks, 1) explicitlyRemovedWorkload := testWorkload() explicitlyRemovedWorkload.Services = []*structs.Service{ { Name: "explicitly-removed-service", PortLabel: "y", Checks: []*structs.ServiceCheck{ { Name: "check", Type: "tcp", Interval: time.Second, Timeout: time.Second, }, }, }, } explicitlyRemovedWorkloadServiceID := MakeAllocServiceID(explicitlyRemovedWorkload.AllocID, explicitlyRemovedWorkload.Name(), explicitlyRemovedWorkload.Services[0]) require.NoError(ctx.ServiceClient.RegisterWorkload(explicitlyRemovedWorkload)) require.NoError(ctx.syncOnce(syncNewOps)) require.Len(ctx.FakeConsul.services, 2) require.Len(ctx.FakeConsul.checks, 2) // we register a task through nomad API then remove it out of band outofbandWorkload := testWorkload() outofbandWorkload.Services = []*structs.Service{ { Name: "unknown-service", PortLabel: "x", Checks: []*structs.ServiceCheck{ { Name: "check", Type: "tcp", Interval: time.Second, Timeout: time.Second, }, }, }, } outofbandWorkloadServiceID := MakeAllocServiceID(outofbandWorkload.AllocID, outofbandWorkload.Name(), outofbandWorkload.Services[0]) require.NoError(ctx.ServiceClient.RegisterWorkload(outofbandWorkload)) require.NoError(ctx.syncOnce(syncNewOps)) require.Len(ctx.FakeConsul.services, 3) // remove outofbandWorkload from local services so it appears unknown to client require.Len(ctx.ServiceClient.services, 3) require.Len(ctx.ServiceClient.checks, 3) delete(ctx.ServiceClient.services, outofbandWorkloadServiceID) delete(ctx.ServiceClient.checks, MakeCheckID(outofbandWorkloadServiceID, outofbandWorkload.Services[0].Checks[0])) require.Len(ctx.ServiceClient.services, 2) require.Len(ctx.ServiceClient.checks, 2) // Sync and ensure that explicitly removed service was removed, but outofbandWorkload remains ctx.ServiceClient.RemoveWorkload(explicitlyRemovedWorkload) require.NoError(ctx.syncOnce(syncNewOps)) require.NoError(ctx.ServiceClient.sync(syncNewOps)) require.Len(ctx.FakeConsul.services, 2) require.Len(ctx.FakeConsul.checks, 2) require.Contains(ctx.FakeConsul.services, remainingWorkloadServiceID) require.Contains(ctx.FakeConsul.services, outofbandWorkloadServiceID) require.NotContains(ctx.FakeConsul.services, explicitlyRemovedWorkloadServiceID) require.Contains(ctx.FakeConsul.checks, MakeCheckID(remainingWorkloadServiceID, remainingWorkload.Services[0].Checks[0])) require.Contains(ctx.FakeConsul.checks, MakeCheckID(outofbandWorkloadServiceID, outofbandWorkload.Services[0].Checks[0])) require.NotContains(ctx.FakeConsul.checks, MakeCheckID(explicitlyRemovedWorkloadServiceID, explicitlyRemovedWorkload.Services[0].Checks[0])) // after probation, outofband services and checks are removed ctx.ServiceClient.deregisterProbationExpiry = time.Now().Add(-1 * time.Hour) require.NoError(ctx.ServiceClient.sync(syncNewOps)) require.Len(ctx.FakeConsul.services, 1) require.Len(ctx.FakeConsul.checks, 1) require.Contains(ctx.FakeConsul.services, remainingWorkloadServiceID) require.NotContains(ctx.FakeConsul.services, outofbandWorkloadServiceID) require.NotContains(ctx.FakeConsul.services, explicitlyRemovedWorkloadServiceID) require.Contains(ctx.FakeConsul.checks, MakeCheckID(remainingWorkloadServiceID, remainingWorkload.Services[0].Checks[0])) require.NotContains(ctx.FakeConsul.checks, MakeCheckID(outofbandWorkloadServiceID, outofbandWorkload.Services[0].Checks[0])) require.NotContains(ctx.FakeConsul.checks, MakeCheckID(explicitlyRemovedWorkloadServiceID, explicitlyRemovedWorkload.Services[0].Checks[0])) }