nomad: testing remove peer
This commit is contained in:
parent
b456588d1c
commit
c18c068bf2
|
@ -123,41 +123,34 @@ func (s *Server) maybeBootstrap() {
|
||||||
|
|
||||||
// nodeFailed is used to handle fail events on the serf cluster
|
// nodeFailed is used to handle fail events on the serf cluster
|
||||||
func (s *Server) nodeFailed(me serf.MemberEvent) {
|
func (s *Server) nodeFailed(me serf.MemberEvent) {
|
||||||
//for _, m := range me.Members {
|
for _, m := range me.Members {
|
||||||
// ok, parts := isConsulServer(m)
|
ok, parts := isNomadServer(m)
|
||||||
// if !ok {
|
if !ok {
|
||||||
// continue
|
continue
|
||||||
// }
|
}
|
||||||
// s.logger.Printf("[INFO] consul: removing server %s", parts)
|
s.logger.Printf("[INFO] nomad: removing server %s", parts)
|
||||||
|
|
||||||
// // Remove the server if known
|
// Remove the server if known
|
||||||
// s.remoteLock.Lock()
|
s.peerLock.Lock()
|
||||||
// existing := s.remoteConsuls[parts.Datacenter]
|
existing := s.peers[parts.Region]
|
||||||
// n := len(existing)
|
n := len(existing)
|
||||||
// for i := 0; i < n; i++ {
|
for i := 0; i < n; i++ {
|
||||||
// if existing[i].Name == parts.Name {
|
if existing[i].Name == parts.Name {
|
||||||
// existing[i], existing[n-1] = existing[n-1], nil
|
existing[i], existing[n-1] = existing[n-1], nil
|
||||||
// existing = existing[:n-1]
|
existing = existing[:n-1]
|
||||||
// n--
|
n--
|
||||||
// break
|
break
|
||||||
// }
|
}
|
||||||
// }
|
}
|
||||||
|
|
||||||
// // Trim the list if all known consuls are dead
|
// Trim the list there are no known servers in a region
|
||||||
// if n == 0 {
|
if n == 0 {
|
||||||
// delete(s.remoteConsuls, parts.Datacenter)
|
delete(s.peers, parts.Region)
|
||||||
// } else {
|
} else {
|
||||||
// s.remoteConsuls[parts.Datacenter] = existing
|
s.peers[parts.Region] = existing
|
||||||
// }
|
}
|
||||||
// s.remoteLock.Unlock()
|
s.peerLock.Unlock()
|
||||||
|
}
|
||||||
// // Remove from the local list as well
|
|
||||||
// if !wan {
|
|
||||||
// s.localLock.Lock()
|
|
||||||
// delete(s.localConsuls, parts.Addr.String())
|
|
||||||
// s.localLock.Unlock()
|
|
||||||
// }
|
|
||||||
//}
|
|
||||||
}
|
}
|
||||||
|
|
||||||
// localMemberEvent is used to reconcile Serf events with the
|
// localMemberEvent is used to reconcile Serf events with the
|
||||||
|
|
|
@ -3,11 +3,17 @@ package nomad
|
||||||
import (
|
import (
|
||||||
"fmt"
|
"fmt"
|
||||||
"testing"
|
"testing"
|
||||||
|
|
||||||
|
"github.com/hashicorp/nomad/testutil"
|
||||||
)
|
)
|
||||||
|
|
||||||
func TestNomad_JoinPeer(t *testing.T) {
|
func TestNomad_JoinPeer(t *testing.T) {
|
||||||
s1 := testServer(t, nil)
|
s1 := testServer(t, nil)
|
||||||
s2 := testServer(t, nil)
|
defer s1.Shutdown()
|
||||||
|
s2 := testServer(t, func(c *Config) {
|
||||||
|
c.Region = "region2"
|
||||||
|
})
|
||||||
|
defer s2.Shutdown()
|
||||||
s2Addr := fmt.Sprintf("127.0.0.1:%d", s2.config.SerfConfig.MemberlistConfig.BindPort)
|
s2Addr := fmt.Sprintf("127.0.0.1:%d", s2.config.SerfConfig.MemberlistConfig.BindPort)
|
||||||
|
|
||||||
num, err := s1.Join([]string{s2Addr})
|
num, err := s1.Join([]string{s2Addr})
|
||||||
|
@ -18,10 +24,73 @@ func TestNomad_JoinPeer(t *testing.T) {
|
||||||
t.Fatalf("bad: %d", num)
|
t.Fatalf("bad: %d", num)
|
||||||
}
|
}
|
||||||
|
|
||||||
if members := s1.Members(); len(members) != 2 {
|
testutil.WaitForResult(func() (bool, error) {
|
||||||
t.Fatalf("bad: %#v", members)
|
if members := s1.Members(); len(members) != 2 {
|
||||||
}
|
return false, fmt.Errorf("bad: %#v", members)
|
||||||
if members := s2.Members(); len(members) != 2 {
|
}
|
||||||
t.Fatalf("bad: %#v", members)
|
if members := s2.Members(); len(members) != 2 {
|
||||||
}
|
return false, fmt.Errorf("bad: %#v", members)
|
||||||
|
}
|
||||||
|
return true, nil
|
||||||
|
}, func(err error) {
|
||||||
|
t.Fatalf("err: %v", err)
|
||||||
|
})
|
||||||
|
|
||||||
|
testutil.WaitForResult(func() (bool, error) {
|
||||||
|
if len(s1.peers) != 2 {
|
||||||
|
return false, fmt.Errorf("bad: %#v", s1.peers)
|
||||||
|
}
|
||||||
|
if len(s2.peers) != 2 {
|
||||||
|
return false, fmt.Errorf("bad: %#v", s2.peers)
|
||||||
|
}
|
||||||
|
return true, nil
|
||||||
|
}, func(err error) {
|
||||||
|
t.Fatalf("err: %v", err)
|
||||||
|
})
|
||||||
|
}
|
||||||
|
|
||||||
|
func TestNomad_RemovePeer(t *testing.T) {
|
||||||
|
s1 := testServer(t, nil)
|
||||||
|
defer s1.Shutdown()
|
||||||
|
s2 := testServer(t, func(c *Config) {
|
||||||
|
c.Region = "region2"
|
||||||
|
})
|
||||||
|
defer s2.Shutdown()
|
||||||
|
s2Addr := fmt.Sprintf("127.0.0.1:%d", s2.config.SerfConfig.MemberlistConfig.BindPort)
|
||||||
|
|
||||||
|
num, err := s1.Join([]string{s2Addr})
|
||||||
|
if err != nil {
|
||||||
|
t.Fatalf("err: %v", err)
|
||||||
|
}
|
||||||
|
if num != 1 {
|
||||||
|
t.Fatalf("bad: %d", num)
|
||||||
|
}
|
||||||
|
|
||||||
|
testutil.WaitForResult(func() (bool, error) {
|
||||||
|
if members := s1.Members(); len(members) != 2 {
|
||||||
|
return false, fmt.Errorf("bad: %#v", members)
|
||||||
|
}
|
||||||
|
if members := s2.Members(); len(members) != 2 {
|
||||||
|
return false, fmt.Errorf("bad: %#v", members)
|
||||||
|
}
|
||||||
|
return true, nil
|
||||||
|
}, func(err error) {
|
||||||
|
t.Fatalf("err: %v", err)
|
||||||
|
})
|
||||||
|
|
||||||
|
// Leave immediately
|
||||||
|
s2.Leave()
|
||||||
|
s2.Shutdown()
|
||||||
|
|
||||||
|
testutil.WaitForResult(func() (bool, error) {
|
||||||
|
if len(s1.peers) != 1 {
|
||||||
|
return false, fmt.Errorf("bad: %#v", s1.peers)
|
||||||
|
}
|
||||||
|
if len(s2.peers) != 1 {
|
||||||
|
return false, fmt.Errorf("bad: %#v", s2.peers)
|
||||||
|
}
|
||||||
|
return true, nil
|
||||||
|
}, func(err error) {
|
||||||
|
t.Fatalf("err: %v", err)
|
||||||
|
})
|
||||||
}
|
}
|
||||||
|
|
|
@ -314,6 +314,7 @@ func (s *Server) setupRaft() error {
|
||||||
log = store
|
log = store
|
||||||
snap = raft.NewDiscardSnapshotStore()
|
snap = raft.NewDiscardSnapshotStore()
|
||||||
peers = &raft.StaticPeers{}
|
peers = &raft.StaticPeers{}
|
||||||
|
s.raftPeers = peers
|
||||||
|
|
||||||
} else {
|
} else {
|
||||||
// Create the base raft path
|
// Create the base raft path
|
||||||
|
|
|
@ -28,6 +28,7 @@ func testServer(t *testing.T, cb func(*Config)) *Server {
|
||||||
config.SerfConfig.MemberlistConfig.BindAddr = "127.0.0.1"
|
config.SerfConfig.MemberlistConfig.BindAddr = "127.0.0.1"
|
||||||
config.SerfConfig.MemberlistConfig.BindPort = getPort()
|
config.SerfConfig.MemberlistConfig.BindPort = getPort()
|
||||||
config.SerfConfig.MemberlistConfig.SuspicionMult = 2
|
config.SerfConfig.MemberlistConfig.SuspicionMult = 2
|
||||||
|
config.SerfConfig.MemberlistConfig.RetransmitMult = 2
|
||||||
config.SerfConfig.MemberlistConfig.ProbeTimeout = 50 * time.Millisecond
|
config.SerfConfig.MemberlistConfig.ProbeTimeout = 50 * time.Millisecond
|
||||||
config.SerfConfig.MemberlistConfig.ProbeInterval = 100 * time.Millisecond
|
config.SerfConfig.MemberlistConfig.ProbeInterval = 100 * time.Millisecond
|
||||||
config.SerfConfig.MemberlistConfig.GossipInterval = 100 * time.Millisecond
|
config.SerfConfig.MemberlistConfig.GossipInterval = 100 * time.Millisecond
|
||||||
|
|
|
@ -0,0 +1,24 @@
|
||||||
|
package testutil
|
||||||
|
|
||||||
|
import "time"
|
||||||
|
|
||||||
|
type testFn func() (bool, error)
|
||||||
|
type errorFn func(error)
|
||||||
|
|
||||||
|
func WaitForResult(test testFn, error errorFn) {
|
||||||
|
retries := 1000
|
||||||
|
|
||||||
|
for retries > 0 {
|
||||||
|
time.Sleep(10 * time.Millisecond)
|
||||||
|
retries--
|
||||||
|
|
||||||
|
success, err := test()
|
||||||
|
if success {
|
||||||
|
return
|
||||||
|
}
|
||||||
|
|
||||||
|
if retries == 0 {
|
||||||
|
error(err)
|
||||||
|
}
|
||||||
|
}
|
||||||
|
}
|
|
@ -0,0 +1 @@
|
||||||
|
package testutil
|
Loading…
Reference in New Issue