From 082ba48809ea144ceeb03e4b92882adc921040b9 Mon Sep 17 00:00:00 2001 From: cskh Date: Tue, 7 Mar 2023 13:27:47 -0500 Subject: [PATCH] upgrade test: use retry with ModifyIndex and remove ent test file (#16553) --- .../test/upgrade/fullstopupgrade_test.go | 38 ++- .../test/upgrade/tenancy_ent_test.go | 271 ------------------ 2 files changed, 25 insertions(+), 284 deletions(-) delete mode 100644 test/integration/consul-container/test/upgrade/tenancy_ent_test.go diff --git a/test/integration/consul-container/test/upgrade/fullstopupgrade_test.go b/test/integration/consul-container/test/upgrade/fullstopupgrade_test.go index c90692b5d..59820c4b0 100644 --- a/test/integration/consul-container/test/upgrade/fullstopupgrade_test.go +++ b/test/integration/consul-container/test/upgrade/fullstopupgrade_test.go @@ -6,6 +6,7 @@ import ( "testing" "time" + goretry "github.com/avast/retry-go" "github.com/stretchr/testify/require" "github.com/hashicorp/consul/api" @@ -75,22 +76,33 @@ func TestStandardUpgradeToTarget_fromLatest(t *testing.T) { const serviceName = "api" index := libservice.ServiceCreate(t, client, serviceName) - ch, errCh := libservice.ServiceHealthBlockingQuery(client, serviceName, index) require.NoError(t, client.Agent().ServiceRegister( &api.AgentServiceRegistration{Name: serviceName, Port: 9998}, )) - - timer := time.NewTimer(3 * time.Second) - select { - case err := <-errCh: - require.NoError(t, err) - case service := <-ch: - require.Len(t, service, 1) - require.Equal(t, serviceName, service[0].Service.Service) - require.Equal(t, 9998, service[0].Service.Port) - case <-timer.C: - t.Fatalf("test timeout") - } + err = goretry.Do( + func() error { + ch, errCh := libservice.ServiceHealthBlockingQuery(client, serviceName, index) + select { + case err := <-errCh: + require.NoError(t, err) + case service := <-ch: + index = service[0].Service.ModifyIndex + if len(service) != 1 { + return fmt.Errorf("service is %d, want 1", len(service)) + } + if serviceName != service[0].Service.Service { + return fmt.Errorf("service name is %s, want %s", service[0].Service.Service, serviceName) + } + if service[0].Service.Port != 9998 { + return fmt.Errorf("service is %d, want 9998", service[0].Service.Port) + } + } + return nil + }, + goretry.Attempts(5), + goretry.Delay(time.Second), + ) + require.NoError(t, err) // upgrade the cluster to the Target version t.Logf("initiating standard upgrade to version=%q", tc.targetVersion) diff --git a/test/integration/consul-container/test/upgrade/tenancy_ent_test.go b/test/integration/consul-container/test/upgrade/tenancy_ent_test.go deleted file mode 100644 index e28664e0b..000000000 --- a/test/integration/consul-container/test/upgrade/tenancy_ent_test.go +++ /dev/null @@ -1,271 +0,0 @@ -//go:build consulent -// +build consulent - -package upgrade - -import ( - "context" - "encoding/json" - "fmt" - "io" - "testing" - - "github.com/hashicorp/consul/api" - "github.com/hashicorp/consul/sdk/testutil" - "github.com/hashicorp/consul/sdk/testutil/retry" - "github.com/stretchr/testify/require" - - libcluster "github.com/hashicorp/consul/test/integration/consul-container/libs/cluster" - libservice "github.com/hashicorp/consul/test/integration/consul-container/libs/service" - "github.com/hashicorp/consul/test/integration/consul-container/libs/utils" -) - -// Test partition crud using Current Clients and Latest GA Servers -func TestLatestGAServersWithCurrentClients_PartitionCRUD(t *testing.T) { - testLatestGAServersWithCurrentClients_TenancyCRUD(t, "Partitions", - func(t *testing.T, client *api.Client) { - ctx, cancel := context.WithCancel(context.Background()) - defer cancel() - - // CRUD partitions - partition, _, err := client.Partitions().Read(ctx, "default", nil) - require.NoError(t, err) - fmt.Printf("%+v\n", partition) - require.NotNil(t, partition) - require.Equal(t, "default", partition.Name) - - fooPartReq := api.Partition{Name: "foo-part"} - fooPart, _, err := client.Partitions().Create(ctx, &api.Partition{Name: "foo-part"}, nil) - require.NoError(t, err) - require.NotNil(t, fooPart) - require.Equal(t, "foo-part", fooPart.Name) - - partition, _, err = client.Partitions().Read(ctx, "foo-part", nil) - require.NoError(t, err) - require.NotNil(t, partition) - require.Equal(t, "foo-part", partition.Name) - - fooPartReq.Description = "foo-part part" - partition, _, err = client.Partitions().Update(ctx, &fooPartReq, nil) - require.NoError(t, err) - require.NotNil(t, partition) - require.Equal(t, "foo-part", partition.Name) - require.Equal(t, "foo-part part", partition.Description) - }, - func(t *testing.T, client *api.Client) { - ctx, cancel := context.WithCancel(context.Background()) - defer cancel() - - //Read partition again - retry.RunWith(libcluster.LongFailer(), t, func(r *retry.R) { - partition, _, err := client.Partitions().Read(ctx, "default", nil) - require.NoError(r, err) - require.NotNil(r, partition) - require.Equal(r, "default", partition.Name) - }) - - retry.RunWith(libcluster.LongFailer(), t, func(r *retry.R) { - partition, _, err := client.Partitions().Read(ctx, "foo-part", nil) - require.NoError(r, err) - require.NotNil(r, partition) - require.Equal(r, "foo-part", partition.Name) - require.Equal(r, "foo-part part", partition.Description) - }) - }, - ) -} - -// Test namespace crud using Current Clients and Latest GA Servers -func TestLatestGAServersWithCurrentClients_NamespaceCRUD(t *testing.T) { - testLatestGAServersWithCurrentClients_TenancyCRUD(t, "Namespaces", - func(t *testing.T, client *api.Client) { - // CRUD namespaces - namespace, _, err := client.Namespaces().Read("default", nil) - require.NoError(t, err) - require.NotNil(t, namespace, "default namespace does not exist yet") - require.Equal(t, "default", namespace.Name) - - fooNsReq := api.Namespace{Name: "foo-ns"} - fooNs, _, err := client.Namespaces().Create(&api.Namespace{Name: "foo-ns"}, nil) - require.NoError(t, err) - require.NotNil(t, fooNs) - require.Equal(t, "foo-ns", fooNs.Name) - - namespace, _, err = client.Namespaces().Read("foo-ns", nil) - require.NoError(t, err) - require.NotNil(t, namespace) - require.Equal(t, "foo-ns", namespace.Name) - - fooNsReq.Description = "foo-ns ns" - namespace, _, err = client.Namespaces().Update(&fooNsReq, nil) - require.NoError(t, err) - require.NotNil(t, namespace) - require.Equal(t, "foo-ns", namespace.Name) - require.Equal(t, "foo-ns ns", namespace.Description) - }, - func(t *testing.T, client *api.Client) { - retry.RunWith(libcluster.LongFailer(), t, func(r *retry.R) { - namespace, _, err := client.Namespaces().Read("default", nil) - require.NoError(r, err) - require.NotNil(r, namespace) - require.Equal(r, "default", namespace.Name) - }) - retry.RunWith(libcluster.LongFailer(), t, func(r *retry.R) { - namespace, _, err := client.Namespaces().Read("foo-ns", nil) - require.NoError(r, err) - require.NotNil(r, namespace) - require.Equal(r, "foo-ns", namespace.Name) - require.Equal(r, "foo-ns ns", namespace.Description) - }) - }, - ) -} - -func testLatestGAServersWithCurrentClients_TenancyCRUD( - t *testing.T, - tenancyName string, - createFn func(t *testing.T, client *api.Client), - readFn func(t *testing.T, client *api.Client), -) { - const ( - numServers = 3 - numClients = 2 - ) - - // Create initial cluster - cluster, _, _ := libtopology.NewCluster(t, &libtopology.ClusterConfig{ - NumServers: numServers, - NumClients: numClients, - BuildOpts: &libcluster.BuildOptions{ - Datacenter: "dc1", - ConsulImageName: utils.LatestImageName, - ConsulVersion: utils.LatestVersion, - }, - ApplyDefaultProxySettings: true, - }) - - client := cluster.APIClient(0) - libcluster.WaitForLeader(t, cluster, client) - libcluster.WaitForMembers(t, client, 5) - - testutil.RunStep(t, "Create "+tenancyName, func(t *testing.T) { - createFn(t, client) - }) - - ctx := context.Background() - - var snapshot io.ReadCloser - testutil.RunStep(t, "Save snapshot", func(t *testing.T) { - var err error - snapshot, _, err = client.Snapshot().Save(nil) - require.NoError(t, err) - }) - - testutil.RunStep(t, "Check "+tenancyName+" after upgrade", func(t *testing.T) { - // Upgrade nodes - leader, err := cluster.Leader() - require.NoError(t, err) - - // upgrade things in the following order: - // - // 1. follower servers - // 2. leader server - // 3. clients - var upgradeOrder []libcluster.Agent - - followers, err := cluster.Followers() - require.NoError(t, err) - upgradeOrder = append(upgradeOrder, followers...) - upgradeOrder = append(upgradeOrder, leader) - upgradeOrder = append(upgradeOrder, cluster.Clients()...) - - for _, n := range upgradeOrder { - conf := n.GetConfig() - - // TODO: ensure this makes sense again, it was doing an apples/orange version!=image comparison - if conf.Version == utils.TargetVersion { - return - } - - conf.Version = utils.TargetVersion - - if n.IsServer() { - // You only ever need bootstrap settings the FIRST time, so we do not need - // them again. - conf.ConfigBuilder.Unset("bootstrap") - } else { - // If we upgrade the clients fast enough - // membership might not be gossiped to all of - // the clients to persist into their serf - // snapshot, so force them to rejoin the - // normal way on restart. - conf.ConfigBuilder.Set("retry_join", []string{"agent-0"}) - } - - newJSON, err := json.MarshalIndent(conf.ConfigBuilder, "", " ") - require.NoError(t, err) - conf.JSON = string(newJSON) - t.Logf("Upgraded cluster config for %q:\n%s", n.GetName(), conf.JSON) - - selfBefore, err := n.GetClient().Agent().Self() - require.NoError(t, err) - - require.NoError(t, n.Upgrade(ctx, conf)) - - selfAfter, err := n.GetClient().Agent().Self() - require.NoError(t, err) - require.Truef(t, - (selfBefore["Config"]["Version"] != selfAfter["Config"]["Version"]) || (selfBefore["Config"]["Revision"] != selfAfter["Config"]["Revision"]), - fmt.Sprintf("upgraded version must be different (%s, %s), (%s, %s)", selfBefore["Config"]["Version"], selfBefore["Config"]["Revision"], selfAfter["Config"]["Version"], selfAfter["Config"]["Revision"]), - ) - - client := n.GetClient() - - libcluster.WaitForLeader(t, cluster, nil) - libcluster.WaitForMembers(t, client, 5) - } - - //get the client again as it changed after upgrade. - client := cluster.APIClient(0) - libcluster.WaitForLeader(t, cluster, client) - - // Read data again - readFn(t, client) - }) - - // Terminate the cluster for the snapshot test - testutil.RunStep(t, "Terminate the cluster", func(t *testing.T) { - require.NoError(t, cluster.Terminate()) - }) - - { // Clear these so they super break if you tried to use them. - cluster = nil - client = nil - } - - // Create a fresh cluster from scratch - cluster2, _, _ := libtopology.NewCluster(t, &libtopology.ClusterConfig{ - NumServers: numServers, - NumClients: numClients, - BuildOpts: &libcluster.BuildOptions{ - Datacenter: "dc1", - ConsulImageName: utils.LatestImageName, - ConsulVersion: utils.LatestVersion, - }, - ApplyDefaultProxySettings: true, - }) - client2 := cluster2.APIClient(0) - - testutil.RunStep(t, "Restore saved snapshot", func(t *testing.T) { - libcluster.WaitForLeader(t, cluster2, client2) - libcluster.WaitForMembers(t, client2, 5) - - // Restore the saved snapshot - require.NoError(t, client2.Snapshot().Restore(nil, snapshot)) - - libcluster.WaitForLeader(t, cluster2, client2) - - // make sure we still have the right data - readFn(t, client2) - }) -}