2017-03-01 22:04:40 +00:00
|
|
|
package consul
|
|
|
|
|
|
|
|
import (
|
|
|
|
"os"
|
|
|
|
"testing"
|
|
|
|
"time"
|
|
|
|
|
2019-03-27 12:54:56 +00:00
|
|
|
"github.com/hashicorp/consul/sdk/testutil/retry"
|
2019-04-25 16:26:33 +00:00
|
|
|
"github.com/hashicorp/consul/testrpc"
|
2017-03-01 22:04:40 +00:00
|
|
|
"github.com/hashicorp/raft"
|
|
|
|
"github.com/hashicorp/serf/serf"
|
|
|
|
)
|
|
|
|
|
2018-02-21 18:19:30 +00:00
|
|
|
func TestAutopilot_IdempotentShutdown(t *testing.T) {
|
|
|
|
dir1, s1 := testServerWithConfig(t, nil)
|
|
|
|
defer os.RemoveAll(dir1)
|
|
|
|
defer s1.Shutdown()
|
|
|
|
retry.Run(t, func(r *retry.R) { r.Check(waitForLeader(s1)) })
|
|
|
|
|
|
|
|
s1.autopilot.Start()
|
|
|
|
s1.autopilot.Start()
|
|
|
|
s1.autopilot.Start()
|
|
|
|
s1.autopilot.Stop()
|
|
|
|
s1.autopilot.Stop()
|
|
|
|
s1.autopilot.Stop()
|
|
|
|
}
|
|
|
|
|
2017-03-01 22:04:40 +00:00
|
|
|
func TestAutopilot_CleanupDeadServer(t *testing.T) {
|
2017-06-27 13:22:18 +00:00
|
|
|
t.Parallel()
|
2017-03-15 23:09:55 +00:00
|
|
|
for i := 1; i <= 3; i++ {
|
|
|
|
testCleanupDeadServer(t, i)
|
|
|
|
}
|
|
|
|
}
|
|
|
|
|
|
|
|
func testCleanupDeadServer(t *testing.T, raftVersion int) {
|
2018-08-06 23:46:09 +00:00
|
|
|
dc := "dc1"
|
2017-03-15 23:09:55 +00:00
|
|
|
conf := func(c *Config) {
|
2018-08-06 23:46:09 +00:00
|
|
|
c.Datacenter = dc
|
2017-03-15 23:09:55 +00:00
|
|
|
c.Bootstrap = false
|
|
|
|
c.BootstrapExpect = 3
|
|
|
|
c.RaftConfig.ProtocolVersion = raft.ProtocolVersion(raftVersion)
|
|
|
|
}
|
|
|
|
dir1, s1 := testServerWithConfig(t, conf)
|
2017-03-01 22:04:40 +00:00
|
|
|
defer os.RemoveAll(dir1)
|
|
|
|
defer s1.Shutdown()
|
|
|
|
|
2017-03-15 23:09:55 +00:00
|
|
|
dir2, s2 := testServerWithConfig(t, conf)
|
2017-03-01 22:04:40 +00:00
|
|
|
defer os.RemoveAll(dir2)
|
|
|
|
defer s2.Shutdown()
|
|
|
|
|
2017-03-15 23:09:55 +00:00
|
|
|
dir3, s3 := testServerWithConfig(t, conf)
|
2017-03-01 22:04:40 +00:00
|
|
|
defer os.RemoveAll(dir3)
|
|
|
|
defer s3.Shutdown()
|
|
|
|
|
|
|
|
servers := []*Server{s1, s2, s3}
|
|
|
|
|
|
|
|
// Try to join
|
2017-05-05 10:29:49 +00:00
|
|
|
joinLAN(t, s2, s1)
|
|
|
|
joinLAN(t, s3, s1)
|
2017-03-01 22:04:40 +00:00
|
|
|
|
|
|
|
for _, s := range servers {
|
2018-08-06 23:46:09 +00:00
|
|
|
testrpc.WaitForLeader(t, s.RPC, dc)
|
2017-05-05 07:23:28 +00:00
|
|
|
retry.Run(t, func(r *retry.R) { r.Check(wantPeers(s, 3)) })
|
2017-03-01 22:04:40 +00:00
|
|
|
}
|
|
|
|
|
2017-03-15 23:09:55 +00:00
|
|
|
// Bring up a new server
|
|
|
|
dir4, s4 := testServerWithConfig(t, conf)
|
|
|
|
defer os.RemoveAll(dir4)
|
|
|
|
defer s4.Shutdown()
|
|
|
|
|
2017-03-01 22:04:40 +00:00
|
|
|
// Kill a non-leader server
|
2017-03-15 23:09:55 +00:00
|
|
|
s3.Shutdown()
|
2017-05-04 22:52:53 +00:00
|
|
|
retry.Run(t, func(r *retry.R) {
|
2017-03-01 22:04:40 +00:00
|
|
|
alive := 0
|
|
|
|
for _, m := range s1.LANMembers() {
|
|
|
|
if m.Status == serf.StatusAlive {
|
|
|
|
alive++
|
|
|
|
}
|
|
|
|
}
|
2017-04-29 16:34:02 +00:00
|
|
|
if alive != 2 {
|
|
|
|
r.Fatal(nil)
|
|
|
|
}
|
|
|
|
})
|
2017-03-01 22:04:40 +00:00
|
|
|
|
2017-03-15 23:09:55 +00:00
|
|
|
// Join the new server
|
2017-05-05 10:29:49 +00:00
|
|
|
joinLAN(t, s4, s1)
|
2017-03-15 23:09:55 +00:00
|
|
|
servers[2] = s4
|
2017-03-01 22:04:40 +00:00
|
|
|
|
|
|
|
// Make sure the dead server is removed and we're back to 3 total peers
|
|
|
|
for _, s := range servers {
|
2017-05-05 07:23:28 +00:00
|
|
|
retry.Run(t, func(r *retry.R) { r.Check(wantPeers(s, 3)) })
|
2017-03-01 22:04:40 +00:00
|
|
|
}
|
|
|
|
}
|
|
|
|
|
2018-08-14 21:24:51 +00:00
|
|
|
func TestAutopilot_CleanupDeadNonvoter(t *testing.T) {
|
|
|
|
dir1, s1 := testServer(t)
|
|
|
|
defer os.RemoveAll(dir1)
|
|
|
|
defer s1.Shutdown()
|
|
|
|
|
|
|
|
dir2, s2 := testServerDCBootstrap(t, "dc1", false)
|
|
|
|
defer os.RemoveAll(dir2)
|
|
|
|
defer s2.Shutdown()
|
|
|
|
|
|
|
|
testrpc.WaitForLeader(t, s1.RPC, "dc1")
|
|
|
|
|
|
|
|
// Have s2 join and then shut it down immediately before it gets a chance to
|
|
|
|
// be promoted to a voter.
|
|
|
|
joinLAN(t, s2, s1)
|
|
|
|
retry.Run(t, func(r *retry.R) {
|
|
|
|
r.Check(wantRaft([]*Server{s1, s2}))
|
|
|
|
})
|
|
|
|
s2.Shutdown()
|
|
|
|
|
|
|
|
retry.Run(t, func(r *retry.R) {
|
|
|
|
r.Check(wantRaft([]*Server{s1}))
|
|
|
|
})
|
|
|
|
}
|
|
|
|
|
2017-03-01 22:04:40 +00:00
|
|
|
func TestAutopilot_CleanupDeadServerPeriodic(t *testing.T) {
|
2017-06-27 13:22:18 +00:00
|
|
|
t.Parallel()
|
2017-03-01 22:04:40 +00:00
|
|
|
dir1, s1 := testServerWithConfig(t, func(c *Config) {
|
|
|
|
c.Datacenter = "dc1"
|
|
|
|
c.Bootstrap = true
|
|
|
|
})
|
|
|
|
defer os.RemoveAll(dir1)
|
|
|
|
defer s1.Shutdown()
|
|
|
|
|
|
|
|
conf := func(c *Config) {
|
|
|
|
c.Datacenter = "dc1"
|
|
|
|
c.Bootstrap = false
|
|
|
|
}
|
2017-09-25 22:27:04 +00:00
|
|
|
|
2017-03-01 22:04:40 +00:00
|
|
|
dir2, s2 := testServerWithConfig(t, conf)
|
|
|
|
defer os.RemoveAll(dir2)
|
|
|
|
defer s2.Shutdown()
|
|
|
|
|
|
|
|
dir3, s3 := testServerWithConfig(t, conf)
|
|
|
|
defer os.RemoveAll(dir3)
|
|
|
|
defer s3.Shutdown()
|
|
|
|
|
2017-03-10 19:41:17 +00:00
|
|
|
dir4, s4 := testServerWithConfig(t, conf)
|
|
|
|
defer os.RemoveAll(dir4)
|
|
|
|
defer s4.Shutdown()
|
|
|
|
|
2017-09-25 22:27:04 +00:00
|
|
|
dir5, s5 := testServerWithConfig(t, conf)
|
|
|
|
defer os.RemoveAll(dir5)
|
|
|
|
defer s5.Shutdown()
|
|
|
|
|
|
|
|
servers := []*Server{s1, s2, s3, s4, s5}
|
2017-03-01 22:04:40 +00:00
|
|
|
|
2017-09-25 22:27:04 +00:00
|
|
|
// Join the servers to s1, and wait until they are all promoted to
|
|
|
|
// voters.
|
2017-03-10 19:41:17 +00:00
|
|
|
for _, s := range servers[1:] {
|
2017-05-05 10:29:49 +00:00
|
|
|
joinLAN(t, s, s1)
|
2017-03-01 22:04:40 +00:00
|
|
|
}
|
2017-09-25 22:27:04 +00:00
|
|
|
retry.Run(t, func(r *retry.R) {
|
|
|
|
r.Check(wantRaft(servers))
|
|
|
|
for _, s := range servers {
|
|
|
|
r.Check(wantPeers(s, 5))
|
|
|
|
}
|
|
|
|
})
|
2017-03-01 22:04:40 +00:00
|
|
|
|
|
|
|
// Kill a non-leader server
|
2017-03-10 19:41:17 +00:00
|
|
|
s4.Shutdown()
|
2017-03-01 22:04:40 +00:00
|
|
|
|
|
|
|
// Should be removed from the peers automatically
|
2017-09-25 22:27:04 +00:00
|
|
|
servers = []*Server{s1, s2, s3, s5}
|
|
|
|
retry.Run(t, func(r *retry.R) {
|
|
|
|
r.Check(wantRaft(servers))
|
|
|
|
for _, s := range servers {
|
|
|
|
r.Check(wantPeers(s, 4))
|
|
|
|
}
|
|
|
|
})
|
2017-03-01 22:04:40 +00:00
|
|
|
}
|
|
|
|
|
2017-10-31 20:16:56 +00:00
|
|
|
func TestAutopilot_RollingUpdate(t *testing.T) {
|
|
|
|
t.Parallel()
|
|
|
|
dir1, s1 := testServerWithConfig(t, func(c *Config) {
|
|
|
|
c.Datacenter = "dc1"
|
|
|
|
c.Bootstrap = true
|
|
|
|
})
|
|
|
|
defer os.RemoveAll(dir1)
|
|
|
|
defer s1.Shutdown()
|
|
|
|
|
|
|
|
conf := func(c *Config) {
|
|
|
|
c.Datacenter = "dc1"
|
|
|
|
c.Bootstrap = false
|
|
|
|
}
|
|
|
|
|
|
|
|
dir2, s2 := testServerWithConfig(t, conf)
|
|
|
|
defer os.RemoveAll(dir2)
|
|
|
|
defer s2.Shutdown()
|
|
|
|
|
|
|
|
dir3, s3 := testServerWithConfig(t, conf)
|
|
|
|
defer os.RemoveAll(dir3)
|
|
|
|
defer s3.Shutdown()
|
|
|
|
|
|
|
|
// Join the servers to s1, and wait until they are all promoted to
|
|
|
|
// voters.
|
|
|
|
servers := []*Server{s1, s2, s3}
|
|
|
|
for _, s := range servers[1:] {
|
|
|
|
joinLAN(t, s, s1)
|
|
|
|
}
|
|
|
|
retry.Run(t, func(r *retry.R) {
|
|
|
|
r.Check(wantRaft(servers))
|
|
|
|
for _, s := range servers {
|
|
|
|
r.Check(wantPeers(s, 3))
|
|
|
|
}
|
|
|
|
})
|
|
|
|
|
|
|
|
// Add one more server like we are doing a rolling update.
|
|
|
|
dir4, s4 := testServerWithConfig(t, conf)
|
|
|
|
defer os.RemoveAll(dir4)
|
|
|
|
defer s4.Shutdown()
|
|
|
|
joinLAN(t, s1, s4)
|
|
|
|
servers = append(servers, s4)
|
|
|
|
retry.Run(t, func(r *retry.R) {
|
|
|
|
r.Check(wantRaft(servers))
|
|
|
|
for _, s := range servers {
|
|
|
|
r.Check(wantPeers(s, 3))
|
|
|
|
}
|
|
|
|
})
|
|
|
|
|
|
|
|
// Now kill one of the "old" nodes like we are doing a rolling update.
|
|
|
|
s3.Shutdown()
|
|
|
|
|
|
|
|
isVoter := func() bool {
|
|
|
|
future := s1.raft.GetConfiguration()
|
|
|
|
if err := future.Error(); err != nil {
|
|
|
|
t.Fatalf("err: %v", err)
|
|
|
|
}
|
|
|
|
for _, s := range future.Configuration().Servers {
|
|
|
|
if string(s.ID) == string(s4.config.NodeID) {
|
|
|
|
return s.Suffrage == raft.Voter
|
|
|
|
}
|
|
|
|
}
|
|
|
|
t.Fatalf("didn't find s4")
|
|
|
|
return false
|
|
|
|
}
|
|
|
|
|
|
|
|
// Wait for s4 to stabilize, get promoted to a voter, and for s3 to be
|
|
|
|
// removed.
|
|
|
|
servers = []*Server{s1, s2, s4}
|
|
|
|
retry.Run(t, func(r *retry.R) {
|
|
|
|
r.Check(wantRaft(servers))
|
|
|
|
for _, s := range servers {
|
|
|
|
r.Check(wantPeers(s, 3))
|
|
|
|
}
|
|
|
|
if !isVoter() {
|
|
|
|
r.Fatalf("should be a voter")
|
|
|
|
}
|
|
|
|
})
|
|
|
|
}
|
|
|
|
|
2017-03-29 19:52:00 +00:00
|
|
|
func TestAutopilot_CleanupStaleRaftServer(t *testing.T) {
|
2017-06-27 13:22:18 +00:00
|
|
|
t.Parallel()
|
2017-03-29 19:52:00 +00:00
|
|
|
dir1, s1 := testServerDCBootstrap(t, "dc1", true)
|
|
|
|
defer os.RemoveAll(dir1)
|
|
|
|
defer s1.Shutdown()
|
|
|
|
|
|
|
|
dir2, s2 := testServerDCBootstrap(t, "dc1", false)
|
|
|
|
defer os.RemoveAll(dir2)
|
|
|
|
defer s2.Shutdown()
|
|
|
|
|
|
|
|
dir3, s3 := testServerDCBootstrap(t, "dc1", false)
|
|
|
|
defer os.RemoveAll(dir3)
|
|
|
|
defer s3.Shutdown()
|
|
|
|
|
|
|
|
dir4, s4 := testServerDCBootstrap(t, "dc1", false)
|
|
|
|
defer os.RemoveAll(dir4)
|
|
|
|
defer s4.Shutdown()
|
|
|
|
|
|
|
|
servers := []*Server{s1, s2, s3}
|
|
|
|
|
|
|
|
// Join the servers to s1
|
|
|
|
for _, s := range servers[1:] {
|
2017-05-05 10:29:49 +00:00
|
|
|
joinLAN(t, s, s1)
|
2017-03-29 19:52:00 +00:00
|
|
|
}
|
|
|
|
|
|
|
|
for _, s := range servers {
|
2017-05-05 07:23:28 +00:00
|
|
|
retry.Run(t, func(r *retry.R) { r.Check(wantPeers(s, 3)) })
|
2017-03-29 19:52:00 +00:00
|
|
|
}
|
|
|
|
|
2017-04-19 23:00:11 +00:00
|
|
|
testrpc.WaitForLeader(t, s1.RPC, "dc1")
|
2017-04-12 22:28:18 +00:00
|
|
|
|
2017-03-29 19:52:00 +00:00
|
|
|
// Add s4 to peers directly
|
2017-05-05 10:29:49 +00:00
|
|
|
s1.raft.AddVoter(raft.ServerID(s4.config.NodeID), raft.ServerAddress(joinAddrLAN(s4)), 0, 0)
|
2017-03-29 19:52:00 +00:00
|
|
|
|
|
|
|
// Verify we have 4 peers
|
|
|
|
peers, err := s1.numPeers()
|
|
|
|
if err != nil {
|
|
|
|
t.Fatal(err)
|
|
|
|
}
|
|
|
|
if peers != 4 {
|
|
|
|
t.Fatalf("bad: %v", peers)
|
|
|
|
}
|
|
|
|
|
|
|
|
// Wait for s4 to be removed
|
|
|
|
for _, s := range []*Server{s1, s2, s3} {
|
2017-05-05 07:23:28 +00:00
|
|
|
retry.Run(t, func(r *retry.R) { r.Check(wantPeers(s, 3)) })
|
2017-03-29 19:52:00 +00:00
|
|
|
}
|
|
|
|
}
|
|
|
|
|
2017-03-01 22:04:40 +00:00
|
|
|
func TestAutopilot_PromoteNonVoter(t *testing.T) {
|
2017-06-27 13:22:18 +00:00
|
|
|
t.Parallel()
|
2017-03-01 22:04:40 +00:00
|
|
|
dir1, s1 := testServerWithConfig(t, func(c *Config) {
|
|
|
|
c.Datacenter = "dc1"
|
|
|
|
c.Bootstrap = true
|
|
|
|
c.RaftConfig.ProtocolVersion = 3
|
|
|
|
c.AutopilotConfig.ServerStabilizationTime = 200 * time.Millisecond
|
|
|
|
c.ServerHealthInterval = 100 * time.Millisecond
|
2017-03-10 00:43:07 +00:00
|
|
|
c.AutopilotInterval = 100 * time.Millisecond
|
2017-03-01 22:04:40 +00:00
|
|
|
})
|
|
|
|
defer os.RemoveAll(dir1)
|
|
|
|
defer s1.Shutdown()
|
|
|
|
codec := rpcClient(t, s1)
|
|
|
|
defer codec.Close()
|
2017-10-31 20:16:56 +00:00
|
|
|
testrpc.WaitForLeader(t, s1.RPC, "dc1")
|
2017-03-01 22:04:40 +00:00
|
|
|
|
|
|
|
dir2, s2 := testServerWithConfig(t, func(c *Config) {
|
|
|
|
c.Datacenter = "dc1"
|
|
|
|
c.Bootstrap = false
|
|
|
|
c.RaftConfig.ProtocolVersion = 3
|
|
|
|
})
|
|
|
|
defer os.RemoveAll(dir2)
|
|
|
|
defer s2.Shutdown()
|
2017-05-05 10:29:49 +00:00
|
|
|
joinLAN(t, s2, s1)
|
2017-03-01 22:04:40 +00:00
|
|
|
|
2017-10-31 20:16:56 +00:00
|
|
|
// Make sure we see it as a nonvoter initially. We wait until half
|
|
|
|
// the stabilization period has passed.
|
2017-05-05 10:14:43 +00:00
|
|
|
retry.Run(t, func(r *retry.R) {
|
|
|
|
future := s1.raft.GetConfiguration()
|
|
|
|
if err := future.Error(); err != nil {
|
|
|
|
r.Fatal(err)
|
|
|
|
}
|
2017-03-01 22:04:40 +00:00
|
|
|
|
2017-05-05 10:14:43 +00:00
|
|
|
servers := future.Configuration().Servers
|
|
|
|
if len(servers) != 2 {
|
|
|
|
r.Fatalf("bad: %v", servers)
|
|
|
|
}
|
|
|
|
if servers[1].Suffrage != raft.Nonvoter {
|
|
|
|
r.Fatalf("bad: %v", servers)
|
|
|
|
}
|
2017-12-12 00:38:52 +00:00
|
|
|
health := s1.autopilot.GetServerHealth(string(servers[1].ID))
|
2017-05-05 10:14:43 +00:00
|
|
|
if health == nil {
|
|
|
|
r.Fatal("nil health")
|
|
|
|
}
|
|
|
|
if !health.Healthy {
|
|
|
|
r.Fatalf("bad: %v", health)
|
|
|
|
}
|
2017-10-31 20:16:56 +00:00
|
|
|
if time.Since(health.StableSince) < s1.config.AutopilotConfig.ServerStabilizationTime/2 {
|
2017-05-05 10:14:43 +00:00
|
|
|
r.Fatal("stable period not elapsed")
|
|
|
|
}
|
|
|
|
})
|
2017-03-01 22:04:40 +00:00
|
|
|
|
2017-10-31 20:16:56 +00:00
|
|
|
// Make sure it ends up as a voter.
|
2017-05-04 22:52:53 +00:00
|
|
|
retry.Run(t, func(r *retry.R) {
|
2017-03-01 22:04:40 +00:00
|
|
|
future := s1.raft.GetConfiguration()
|
|
|
|
if err := future.Error(); err != nil {
|
2017-04-29 16:34:02 +00:00
|
|
|
r.Fatal(err)
|
2017-03-01 22:04:40 +00:00
|
|
|
}
|
|
|
|
|
|
|
|
servers := future.Configuration().Servers
|
2017-10-31 20:16:56 +00:00
|
|
|
if len(servers) != 2 {
|
2017-04-29 16:34:02 +00:00
|
|
|
r.Fatalf("bad: %v", servers)
|
2017-03-01 22:04:40 +00:00
|
|
|
}
|
|
|
|
if servers[1].Suffrage != raft.Voter {
|
2017-04-29 16:34:02 +00:00
|
|
|
r.Fatalf("bad: %v", servers)
|
2017-03-01 22:04:40 +00:00
|
|
|
}
|
2017-04-29 16:34:02 +00:00
|
|
|
})
|
2017-03-01 22:04:40 +00:00
|
|
|
}
|