diff --git a/consul/server_details/server_details.go b/consul/agent/server.go similarity index 67% rename from consul/server_details/server_details.go rename to consul/agent/server.go index 4f3dc794a..8ba1f0c5d 100644 --- a/consul/server_details/server_details.go +++ b/consul/agent/server.go @@ -1,4 +1,9 @@ -package server_details +// Package agent provides a logical endpoint for Consul agents in the +// network. agent data originates from Serf gossip and is primarily used to +// communicate Consul server information. Gossiped information that ends up +// in Server contains the necessary metadata required for servers.Manager to +// select which server an RPC request should be routed to. +package agent import ( "fmt" @@ -18,8 +23,8 @@ func (k *Key) Equal(x *Key) bool { return k.name == x.name } -// ServerDetails is used to return details of a consul server -type ServerDetails struct { +// Server is used to return details of a consul server +type Server struct { Name string Datacenter string Port int @@ -30,14 +35,14 @@ type ServerDetails struct { } // Key returns the corresponding Key -func (s *ServerDetails) Key() *Key { +func (s *Server) Key() *Key { return &Key{ name: s.Name, } } -// String returns a string representation of ServerDetails -func (s *ServerDetails) String() string { +// String returns a string representation of Server +func (s *Server) String() string { var addrStr, networkStr string if s.Addr != nil { addrStr = s.Addr.String() @@ -47,9 +52,9 @@ func (s *ServerDetails) String() string { return fmt.Sprintf("%s (Addr: %s/%s) (DC: %s)", s.Name, networkStr, addrStr, s.Datacenter) } -// IsConsulServer returns true if a serf member is a consul server. Returns a -// bool and a pointer to the ServerDetails. -func IsConsulServer(m serf.Member) (bool, *ServerDetails) { +// IsConsulServer returns true if a serf member is a consul server +// agent. Returns a bool and a pointer to the Server. +func IsConsulServer(m serf.Member) (bool, *Server) { if m.Tags["role"] != "consul" { return false, nil } @@ -81,7 +86,7 @@ func IsConsulServer(m serf.Member) (bool, *ServerDetails) { addr := &net.TCPAddr{IP: m.Addr, Port: port} - parts := &ServerDetails{ + parts := &Server{ Name: m.Name, Datacenter: datacenter, Port: port, diff --git a/consul/server_details/server_details_internal_test.go b/consul/agent/server_internal_test.go similarity index 86% rename from consul/server_details/server_details_internal_test.go rename to consul/agent/server_internal_test.go index ee09d8f57..3b4f3d7e1 100644 --- a/consul/server_details/server_details_internal_test.go +++ b/consul/agent/server_internal_test.go @@ -1,10 +1,10 @@ -package server_details +package agent import ( "testing" ) -func TestServerDetails_Key_Equal(t *testing.T) { +func TestServer_Key_Equal(t *testing.T) { tests := []struct { name string k1 *Key @@ -47,16 +47,16 @@ func TestServerDetails_Key_Equal(t *testing.T) { } } -func TestServerDetails_Key(t *testing.T) { +func TestServer_Key(t *testing.T) { tests := []struct { name string - sd *ServerDetails + sd *Server k *Key equal bool }{ { name: "Key equality", - sd: &ServerDetails{ + sd: &Server{ Name: "s1", }, k: &Key{ @@ -66,7 +66,7 @@ func TestServerDetails_Key(t *testing.T) { }, { name: "Key inequality", - sd: &ServerDetails{ + sd: &Server{ Name: "s1", }, k: &Key{ diff --git a/consul/server_details/server_details_test.go b/consul/agent/server_test.go similarity index 79% rename from consul/server_details/server_details_test.go rename to consul/agent/server_test.go index b16cbf5a0..0dda84f85 100644 --- a/consul/server_details/server_details_test.go +++ b/consul/agent/server_test.go @@ -1,32 +1,32 @@ -package server_details_test +package agent_test import ( "net" "testing" - "github.com/hashicorp/consul/consul/server_details" + "github.com/hashicorp/consul/consul/agent" "github.com/hashicorp/serf/serf" ) -func TestServerDetails_Key_params(t *testing.T) { +func TestServer_Key_params(t *testing.T) { ipv4a := net.ParseIP("127.0.0.1") ipv4b := net.ParseIP("1.2.3.4") tests := []struct { name string - sd1 *server_details.ServerDetails - sd2 *server_details.ServerDetails + sd1 *agent.Server + sd2 *agent.Server equal bool }{ { name: "Addr inequality", - sd1: &server_details.ServerDetails{ + sd1: &agent.Server{ Name: "s1", Datacenter: "dc1", Port: 8300, Addr: &net.IPAddr{IP: ipv4a}, }, - sd2: &server_details.ServerDetails{ + sd2: &agent.Server{ Name: "s1", Datacenter: "dc1", Port: 8300, @@ -42,7 +42,7 @@ func TestServerDetails_Key_params(t *testing.T) { } // Test Key to make sure it actually works as a key - m := make(map[server_details.Key]bool) + m := make(map[agent.Key]bool) m[*test.sd1.Key()] = true if _, found := m[*test.sd2.Key()]; found != test.equal { t.Errorf("Expected a %v result from map test %s", test.equal, test.name) @@ -61,7 +61,7 @@ func TestIsConsulServer(t *testing.T) { "vsn": "1", }, } - ok, parts := server_details.IsConsulServer(m) + ok, parts := agent.IsConsulServer(m) if !ok || parts.Datacenter != "east-aws" || parts.Port != 10000 { t.Fatalf("bad: %v %v", ok, parts) } @@ -76,7 +76,7 @@ func TestIsConsulServer(t *testing.T) { } m.Tags["bootstrap"] = "1" m.Tags["disabled"] = "1" - ok, parts = server_details.IsConsulServer(m) + ok, parts = agent.IsConsulServer(m) if !ok { t.Fatalf("expected a valid consul server") } @@ -92,7 +92,7 @@ func TestIsConsulServer(t *testing.T) { m.Tags["expect"] = "3" delete(m.Tags, "bootstrap") delete(m.Tags, "disabled") - ok, parts = server_details.IsConsulServer(m) + ok, parts = agent.IsConsulServer(m) if !ok || parts.Expect != 3 { t.Fatalf("bad: %v", parts.Expect) } @@ -101,7 +101,7 @@ func TestIsConsulServer(t *testing.T) { } delete(m.Tags, "role") - ok, parts = server_details.IsConsulServer(m) + ok, parts = agent.IsConsulServer(m) if ok { t.Fatalf("unexpected ok server") } diff --git a/consul/client.go b/consul/client.go index da1241f28..bd5132c66 100644 --- a/consul/client.go +++ b/consul/client.go @@ -10,7 +10,7 @@ import ( "sync" "time" - "github.com/hashicorp/consul/consul/server_details" + "github.com/hashicorp/consul/consul/agent" "github.com/hashicorp/consul/consul/servers" "github.com/hashicorp/consul/consul/structs" "github.com/hashicorp/serf/coordinate" @@ -261,7 +261,7 @@ func (c *Client) lanEventHandler() { // nodeJoin is used to handle join events on the serf cluster func (c *Client) nodeJoin(me serf.MemberEvent) { for _, m := range me.Members { - ok, parts := server_details.IsConsulServer(m) + ok, parts := agent.IsConsulServer(m) if !ok { continue } @@ -283,7 +283,7 @@ func (c *Client) nodeJoin(me serf.MemberEvent) { // nodeFail is used to handle fail events on the serf cluster func (c *Client) nodeFail(me serf.MemberEvent) { for _, m := range me.Members { - ok, parts := server_details.IsConsulServer(m) + ok, parts := agent.IsConsulServer(m) if !ok { continue } diff --git a/consul/leader.go b/consul/leader.go index 3f4bc1484..8aa1f1994 100644 --- a/consul/leader.go +++ b/consul/leader.go @@ -8,7 +8,7 @@ import ( "time" "github.com/armon/go-metrics" - "github.com/hashicorp/consul/consul/server_details" + "github.com/hashicorp/consul/consul/agent" "github.com/hashicorp/consul/consul/structs" "github.com/hashicorp/raft" "github.com/hashicorp/serf/serf" @@ -350,7 +350,7 @@ func (s *Server) shouldHandleMember(member serf.Member) bool { if valid, dc := isConsulNode(member); valid && dc == s.config.Datacenter { return true } - if valid, parts := server_details.IsConsulServer(member); valid && parts.Datacenter == s.config.Datacenter { + if valid, parts := agent.IsConsulServer(member); valid && parts.Datacenter == s.config.Datacenter { return true } return false @@ -361,7 +361,7 @@ func (s *Server) shouldHandleMember(member serf.Member) bool { func (s *Server) handleAliveMember(member serf.Member) error { // Register consul service if a server var service *structs.NodeService - if valid, parts := server_details.IsConsulServer(member); valid { + if valid, parts := agent.IsConsulServer(member); valid { service = &structs.NodeService{ ID: ConsulServiceID, Service: ConsulServiceName, @@ -497,7 +497,7 @@ func (s *Server) handleDeregisterMember(reason string, member serf.Member) error } // Remove from Raft peers if this was a server - if valid, parts := server_details.IsConsulServer(member); valid { + if valid, parts := agent.IsConsulServer(member); valid { if err := s.removeConsulServer(member, parts.Port); err != nil { return err } @@ -524,7 +524,7 @@ func (s *Server) handleDeregisterMember(reason string, member serf.Member) error } // joinConsulServer is used to try to join another consul server -func (s *Server) joinConsulServer(m serf.Member, parts *server_details.ServerDetails) error { +func (s *Server) joinConsulServer(m serf.Member, parts *agent.Server) error { // Do not join ourself if m.Name == s.config.NodeName { return nil @@ -534,7 +534,7 @@ func (s *Server) joinConsulServer(m serf.Member, parts *server_details.ServerDet if parts.Bootstrap { members := s.serfLAN.Members() for _, member := range members { - valid, p := server_details.IsConsulServer(member) + valid, p := agent.IsConsulServer(member) if valid && member.Name != m.Name && p.Bootstrap { s.logger.Printf("[ERR] consul: '%v' and '%v' are both in bootstrap mode. Only one node should be in bootstrap mode, not adding Raft peer.", m.Name, member.Name) return nil diff --git a/consul/merge.go b/consul/merge.go index ed86c6984..defa7ef10 100644 --- a/consul/merge.go +++ b/consul/merge.go @@ -3,7 +3,7 @@ package consul import ( "fmt" - "github.com/hashicorp/consul/consul/server_details" + "github.com/hashicorp/consul/consul/agent" "github.com/hashicorp/serf/serf" ) @@ -25,7 +25,7 @@ func (md *lanMergeDelegate) NotifyMerge(members []*serf.Member) error { continue } - ok, parts := server_details.IsConsulServer(*m) + ok, parts := agent.IsConsulServer(*m) if ok && parts.Datacenter != md.dc { return fmt.Errorf("Member '%s' part of wrong datacenter '%s'", m.Name, parts.Datacenter) @@ -42,7 +42,7 @@ type wanMergeDelegate struct { func (md *wanMergeDelegate) NotifyMerge(members []*serf.Member) error { for _, m := range members { - ok, _ := server_details.IsConsulServer(*m) + ok, _ := agent.IsConsulServer(*m) if !ok { return fmt.Errorf("Member '%s' is not a server", m.Name) } diff --git a/consul/pool.go b/consul/pool.go index cdfd2129d..fd42bf21f 100644 --- a/consul/pool.go +++ b/consul/pool.go @@ -10,7 +10,7 @@ import ( "sync/atomic" "time" - "github.com/hashicorp/consul/consul/server_details" + "github.com/hashicorp/consul/consul/agent" "github.com/hashicorp/consul/tlsutil" "github.com/hashicorp/net-rpc-msgpackrpc" "github.com/hashicorp/yamux" @@ -408,7 +408,7 @@ func (p *ConnPool) RPC(dc string, addr net.Addr, version int, method string, arg // PingConsulServer sends a Status.Ping message to the specified server and // returns true if healthy, false if an error occurred -func (p *ConnPool) PingConsulServer(s *server_details.ServerDetails) (bool, error) { +func (p *ConnPool) PingConsulServer(s *agent.Server) (bool, error) { // Get a usable client conn, sc, err := p.getClient(s.Datacenter, s.Addr, s.Version) if err != nil { diff --git a/consul/serf.go b/consul/serf.go index d9f27309c..1fc489400 100644 --- a/consul/serf.go +++ b/consul/serf.go @@ -4,7 +4,7 @@ import ( "net" "strings" - "github.com/hashicorp/consul/consul/server_details" + "github.com/hashicorp/consul/consul/agent" "github.com/hashicorp/serf/serf" ) @@ -141,7 +141,7 @@ func (s *Server) localEvent(event serf.UserEvent) { // lanNodeJoin is used to handle join events on the LAN pool. func (s *Server) lanNodeJoin(me serf.MemberEvent) { for _, m := range me.Members { - ok, parts := server_details.IsConsulServer(m) + ok, parts := agent.IsConsulServer(m) if !ok { continue } @@ -164,7 +164,7 @@ func (s *Server) lanNodeJoin(me serf.MemberEvent) { // wanNodeJoin is used to handle join events on the WAN pool. func (s *Server) wanNodeJoin(me serf.MemberEvent) { for _, m := range me.Members { - ok, parts := server_details.IsConsulServer(m) + ok, parts := agent.IsConsulServer(m) if !ok { s.logger.Printf("[WARN] consul: non-server in WAN pool: %s", m.Name) continue @@ -210,7 +210,7 @@ func (s *Server) maybeBootstrap() { members := s.serfLAN.Members() addrs := make([]string, 0) for _, member := range members { - valid, p := server_details.IsConsulServer(member) + valid, p := agent.IsConsulServer(member) if !valid { continue } @@ -248,7 +248,7 @@ func (s *Server) maybeBootstrap() { // lanNodeFailed is used to handle fail events on the LAN pool. func (s *Server) lanNodeFailed(me serf.MemberEvent) { for _, m := range me.Members { - ok, parts := server_details.IsConsulServer(m) + ok, parts := agent.IsConsulServer(m) if !ok { continue } @@ -263,7 +263,7 @@ func (s *Server) lanNodeFailed(me serf.MemberEvent) { // wanNodeFailed is used to handle fail events on the WAN pool. func (s *Server) wanNodeFailed(me serf.MemberEvent) { for _, m := range me.Members { - ok, parts := server_details.IsConsulServer(m) + ok, parts := agent.IsConsulServer(m) if !ok { continue } diff --git a/consul/server.go b/consul/server.go index 347cc6f26..4d0005f08 100644 --- a/consul/server.go +++ b/consul/server.go @@ -15,7 +15,7 @@ import ( "time" "github.com/hashicorp/consul/acl" - "github.com/hashicorp/consul/consul/server_details" + "github.com/hashicorp/consul/consul/agent" "github.com/hashicorp/consul/consul/state" "github.com/hashicorp/consul/tlsutil" "github.com/hashicorp/raft" @@ -98,7 +98,7 @@ type Server struct { // localConsuls is used to track the known consuls // in the local datacenter. Used to do leader forwarding. - localConsuls map[string]*server_details.ServerDetails + localConsuls map[string]*agent.Server localLock sync.RWMutex // Logger uses the provided LogOutput @@ -120,7 +120,7 @@ type Server struct { // remoteConsuls is used to track the known consuls in // remote datacenters. Used to do DC forwarding. - remoteConsuls map[string][]*server_details.ServerDetails + remoteConsuls map[string][]*agent.Server remoteLock sync.RWMutex // rpcListener is used to listen for incoming connections @@ -217,10 +217,10 @@ func NewServer(config *Config) (*Server, error) { connPool: NewPool(config.LogOutput, serverRPCCache, serverMaxStreams, tlsWrap), eventChLAN: make(chan serf.Event, 256), eventChWAN: make(chan serf.Event, 256), - localConsuls: make(map[string]*server_details.ServerDetails), + localConsuls: make(map[string]*agent.Server), logger: logger, reconcileCh: make(chan serf.Member, 32), - remoteConsuls: make(map[string][]*server_details.ServerDetails), + remoteConsuls: make(map[string][]*agent.Server), rpcServer: rpc.NewServer(), rpcTLS: incomingTLS, tombstoneGC: gc, diff --git a/consul/servers/manager.go b/consul/servers/manager.go index c594427fe..15814697e 100644 --- a/consul/servers/manager.go +++ b/consul/servers/manager.go @@ -1,8 +1,8 @@ // Package servers provides a Manager interface for Manager managed -// server_details.ServerDetails objects. The servers package manages servers -// from a Consul client's perspective (i.e. a list of servers that a client -// talks with for RPCs). The servers package does not provide any API -// guarantees and should be called only by `hashicorp/consul`. +// agent.Server objects. The servers package manages servers from a Consul +// client's perspective (i.e. a list of servers that a client talks with for +// RPCs). The servers package does not provide any API guarantees and should +// be called only by `hashicorp/consul`. package servers import ( @@ -12,7 +12,7 @@ import ( "sync/atomic" "time" - "github.com/hashicorp/consul/consul/server_details" + "github.com/hashicorp/consul/consul/agent" "github.com/hashicorp/consul/lib" ) @@ -61,7 +61,7 @@ type ConsulClusterInfo interface { // Pinger is an interface wrapping client.ConnPool to prevent a // cyclic import dependency type Pinger interface { - PingConsulServer(server *server_details.ServerDetails) (bool, error) + PingConsulServer(s *agent.Server) (bool, error) } // serverList is a local copy of the struct used to maintain the list of @@ -72,7 +72,7 @@ type Pinger interface { type serverList struct { // servers tracks the locally known servers. List membership is // maintained by Serf. - servers []*server_details.ServerDetails + servers []*agent.Server } type Manager struct { @@ -108,7 +108,7 @@ type Manager struct { // begin seeing use after the rebalance timer fires or enough servers fail // organically. If the server is already known, merge the new server // details. -func (m *Manager) AddServer(server *server_details.ServerDetails) { +func (m *Manager) AddServer(s *agent.Server) { m.listLock.Lock() defer m.listLock.Unlock() l := m.getServerList() @@ -116,13 +116,13 @@ func (m *Manager) AddServer(server *server_details.ServerDetails) { // Check if this server is known found := false for idx, existing := range l.servers { - if existing.Name == server.Name { - newServers := make([]*server_details.ServerDetails, len(l.servers)) + if existing.Name == s.Name { + newServers := make([]*agent.Server, len(l.servers)) copy(newServers, l.servers) // Overwrite the existing server details in order to // possibly update metadata (e.g. server version) - newServers[idx] = server + newServers[idx] = s l.servers = newServers found = true @@ -132,9 +132,9 @@ func (m *Manager) AddServer(server *server_details.ServerDetails) { // Add to the list if not known if !found { - newServers := make([]*server_details.ServerDetails, len(l.servers), len(l.servers)+1) + newServers := make([]*agent.Server, len(l.servers), len(l.servers)+1) copy(newServers, l.servers) - newServers = append(newServers, server) + newServers = append(newServers, s) l.servers = newServers } @@ -149,13 +149,13 @@ func (m *Manager) AddServer(server *server_details.ServerDetails) { // less desirable than just returning the next server in the firing line. If // the next server fails, it will fail fast enough and cycleServer will be // called again. -func (l *serverList) cycleServer() (servers []*server_details.ServerDetails) { +func (l *serverList) cycleServer() (servers []*agent.Server) { numServers := len(l.servers) if numServers < 2 { return servers // No action required } - newServers := make([]*server_details.ServerDetails, 0, numServers) + newServers := make([]*agent.Server, 0, numServers) newServers = append(newServers, l.servers[1:]...) newServers = append(newServers, l.servers[0]) @@ -163,7 +163,7 @@ func (l *serverList) cycleServer() (servers []*server_details.ServerDetails) { } // removeServerByKey performs an inline removal of the first matching server -func (l *serverList) removeServerByKey(targetKey *server_details.Key) { +func (l *serverList) removeServerByKey(targetKey *agent.Key) { for i, s := range l.servers { if targetKey.Equal(s.Key()) { copy(l.servers[i:], l.servers[i+1:]) @@ -188,7 +188,7 @@ func (l *serverList) shuffleServers() { // server list. If the server at the front of the list has failed or fails // during an RPC call, it is rotated to the end of the list. If there are no // servers available, return nil. -func (m *Manager) FindServer() *server_details.ServerDetails { +func (m *Manager) FindServer() *agent.Server { l := m.getServerList() numServers := len(l.servers) if numServers == 0 { @@ -225,14 +225,14 @@ func New(logger *log.Logger, shutdownCh chan struct{}, clusterInfo ConsulCluster m.shutdownCh = shutdownCh l := serverList{} - l.servers = make([]*server_details.ServerDetails, 0) + l.servers = make([]*agent.Server, 0) m.saveServerList(l) return m } // NotifyFailedServer marks the passed in server as "failed" by rotating it // to the end of the server list. -func (m *Manager) NotifyFailedServer(server *server_details.ServerDetails) { +func (m *Manager) NotifyFailedServer(s *agent.Server) { l := m.getServerList() // If the server being failed is not the first server on the list, @@ -241,7 +241,7 @@ func (m *Manager) NotifyFailedServer(server *server_details.ServerDetails) { // the server to the end of the list. // Only rotate the server list when there is more than one server - if len(l.servers) > 1 && l.servers[0] == server && + if len(l.servers) > 1 && l.servers[0] == s && // Use atomic.CAS to emulate a TryLock(). atomic.CompareAndSwapInt32(&m.notifyFailedBarrier, 0, 1) { defer atomic.StoreInt32(&m.notifyFailedBarrier, 0) @@ -252,7 +252,7 @@ func (m *Manager) NotifyFailedServer(server *server_details.ServerDetails) { defer m.listLock.Unlock() l = m.getServerList() - if len(l.servers) > 1 && l.servers[0] == server { + if len(l.servers) > 1 && l.servers[0] == s { l.servers = l.cycleServer() m.saveServerList(l) } @@ -261,10 +261,9 @@ func (m *Manager) NotifyFailedServer(server *server_details.ServerDetails) { // NumServers takes out an internal "read lock" and returns the number of // servers. numServers includes both healthy and unhealthy servers. -func (m *Manager) NumServers() (numServers int) { +func (m *Manager) NumServers() int { l := m.getServerList() - numServers = len(l.servers) - return numServers + return len(l.servers) } // RebalanceServers shuffles the list of servers on this agent. The server @@ -355,14 +354,14 @@ func (m *Manager) reconcileServerList(l *serverList) bool { } type targetServer struct { - server *server_details.ServerDetails + server *agent.Server // 'b' == both // 'o' == original // 'n' == new state byte } - mergedList := make(map[server_details.Key]*targetServer, len(l.servers)) + mergedList := make(map[agent.Key]*targetServer, len(l.servers)) for _, s := range l.servers { mergedList[*s.Key()] = &targetServer{server: s, state: 'o'} } @@ -404,15 +403,15 @@ func (m *Manager) reconcileServerList(l *serverList) bool { // RemoveServer takes out an internal write lock and removes a server from // the server list. -func (m *Manager) RemoveServer(server *server_details.ServerDetails) { +func (m *Manager) RemoveServer(s *agent.Server) { m.listLock.Lock() defer m.listLock.Unlock() l := m.getServerList() // Remove the server if known for i, _ := range l.servers { - if l.servers[i].Name == server.Name { - newServers := make([]*server_details.ServerDetails, 0, len(l.servers)-1) + if l.servers[i].Name == s.Name { + newServers := make([]*agent.Server, 0, len(l.servers)-1) newServers = append(newServers, l.servers[:i]...) newServers = append(newServers, l.servers[i+1:]...) l.servers = newServers diff --git a/consul/servers/manager_internal_test.go b/consul/servers/manager_internal_test.go index eba5f32bc..6ca5759ce 100644 --- a/consul/servers/manager_internal_test.go +++ b/consul/servers/manager_internal_test.go @@ -9,7 +9,7 @@ import ( "testing" "time" - "github.com/hashicorp/consul/consul/server_details" + "github.com/hashicorp/consul/consul/agent" ) var ( @@ -31,7 +31,7 @@ type fauxConnPool struct { failPct float64 } -func (cp *fauxConnPool) PingConsulServer(server *server_details.ServerDetails) (bool, error) { +func (cp *fauxConnPool) PingConsulServer(server *agent.Server) (bool, error) { var success bool successProb := rand.Float64() if successProb > cp.failPct { @@ -63,14 +63,14 @@ func testManagerFailProb(failPct float64) (m *Manager) { return m } -// func (l *serverList) cycleServer() (servers []*server_details.ServerDetails) { +// func (l *serverList) cycleServer() (servers []*agent.Server) { func TestManagerInternal_cycleServer(t *testing.T) { m := testManager() l := m.getServerList() - server0 := &server_details.ServerDetails{Name: "server1"} - server1 := &server_details.ServerDetails{Name: "server2"} - server2 := &server_details.ServerDetails{Name: "server3"} + server0 := &agent.Server{Name: "server1"} + server1 := &agent.Server{Name: "server2"} + server2 := &agent.Server{Name: "server3"} l.servers = append(l.servers, server0, server1, server2) m.saveServerList(l) @@ -166,11 +166,11 @@ func test_reconcileServerList(maxServers int) (bool, error) { const failPct = 0.5 m := testManagerFailProb(failPct) - var failedServers, healthyServers []*server_details.ServerDetails + var failedServers, healthyServers []*agent.Server for i := 0; i < maxServers; i++ { nodeName := fmt.Sprintf("s%02d", i) - node := &server_details.ServerDetails{Name: nodeName} + node := &agent.Server{Name: nodeName} // Add 66% of servers to Manager if rand.Float64() > 0.33 { m.AddServer(node) @@ -230,7 +230,7 @@ func test_reconcileServerList(maxServers int) (bool, error) { return true, nil } - resultingServerMap := make(map[server_details.Key]bool) + resultingServerMap := make(map[agent.Key]bool) for _, s := range m.getServerList().servers { resultingServerMap[*s.Key()] = true } @@ -302,7 +302,7 @@ func TestManagerInternal_refreshServerRebalanceTimer(t *testing.T) { m := New(logger, shutdownCh, &fauxSerf{numNodes: s.numNodes}, &fauxConnPool{}) for i := 0; i < s.numServers; i++ { nodeName := fmt.Sprintf("s%02d", i) - m.AddServer(&server_details.ServerDetails{Name: nodeName}) + m.AddServer(&agent.Server{Name: nodeName}) } d := m.refreshServerRebalanceTimer() @@ -323,7 +323,7 @@ func TestManagerInternal_saveServerList(t *testing.T) { t.Fatalf("Manager.saveServerList failed to load init config") } - newServer := new(server_details.ServerDetails) + newServer := new(agent.Server) l.servers = append(l.servers, newServer) m.saveServerList(l) }() @@ -339,7 +339,7 @@ func TestManagerInternal_saveServerList(t *testing.T) { // Verify mutation w/o a save doesn't alter the original func() { - newServer := new(server_details.ServerDetails) + newServer := new(agent.Server) l := m.getServerList() l.servers = append(l.servers, newServer) diff --git a/consul/servers/manager_test.go b/consul/servers/manager_test.go index 150313247..5b467445d 100644 --- a/consul/servers/manager_test.go +++ b/consul/servers/manager_test.go @@ -9,7 +9,7 @@ import ( "strings" "testing" - "github.com/hashicorp/consul/consul/server_details" + "github.com/hashicorp/consul/consul/agent" "github.com/hashicorp/consul/consul/servers" ) @@ -32,7 +32,7 @@ type fauxConnPool struct { failPct float64 } -func (cp *fauxConnPool) PingConsulServer(server *server_details.ServerDetails) (bool, error) { +func (cp *fauxConnPool) PingConsulServer(server *agent.Server) (bool, error) { var success bool successProb := rand.Float64() if successProb > cp.failPct { @@ -64,7 +64,7 @@ func testManagerFailProb(failPct float64) (m *servers.Manager) { return m } -// func (m *Manager) AddServer(server *server_details.ServerDetails) { +// func (m *Manager) AddServer(server *agent.Server) { func TestServers_AddServer(t *testing.T) { m := testManager() var num int @@ -73,7 +73,7 @@ func TestServers_AddServer(t *testing.T) { t.Fatalf("Expected zero servers to start") } - s1 := &server_details.ServerDetails{Name: "s1"} + s1 := &agent.Server{Name: "s1"} m.AddServer(s1) num = m.NumServers() if num != 1 { @@ -86,7 +86,7 @@ func TestServers_AddServer(t *testing.T) { t.Fatalf("Expected one server (still)") } - s2 := &server_details.ServerDetails{Name: "s2"} + s2 := &agent.Server{Name: "s2"} m.AddServer(s2) num = m.NumServers() if num != 2 { @@ -94,7 +94,7 @@ func TestServers_AddServer(t *testing.T) { } } -// func (m *Manager) FindServer() (server *server_details.ServerDetails) { +// func (m *Manager) FindServer() (server *agent.Server) { func TestServers_FindServer(t *testing.T) { m := testManager() @@ -102,7 +102,7 @@ func TestServers_FindServer(t *testing.T) { t.Fatalf("Expected nil return") } - m.AddServer(&server_details.ServerDetails{Name: "s1"}) + m.AddServer(&agent.Server{Name: "s1"}) if m.NumServers() != 1 { t.Fatalf("Expected one server") } @@ -120,7 +120,7 @@ func TestServers_FindServer(t *testing.T) { t.Fatalf("Expected s1 server (still)") } - m.AddServer(&server_details.ServerDetails{Name: "s2"}) + m.AddServer(&agent.Server{Name: "s2"}) if m.NumServers() != 2 { t.Fatalf("Expected two servers") } @@ -153,7 +153,7 @@ func TestServers_New(t *testing.T) { } } -// func (m *Manager) NotifyFailedServer(server *server_details.ServerDetails) { +// func (m *Manager) NotifyFailedServer(server *agent.Server) { func TestServers_NotifyFailedServer(t *testing.T) { m := testManager() @@ -161,8 +161,8 @@ func TestServers_NotifyFailedServer(t *testing.T) { t.Fatalf("Expected zero servers to start") } - s1 := &server_details.ServerDetails{Name: "s1"} - s2 := &server_details.ServerDetails{Name: "s2"} + s1 := &agent.Server{Name: "s1"} + s2 := &agent.Server{Name: "s2"} // Try notifying for a server that is not managed by Manager m.NotifyFailedServer(s1) @@ -215,7 +215,7 @@ func TestServers_NumServers(t *testing.T) { t.Fatalf("Expected zero servers to start") } - s := &server_details.ServerDetails{} + s := &agent.Server{} m.AddServer(s) num = m.NumServers() if num != 1 { @@ -234,7 +234,7 @@ func TestServers_RebalanceServers(t *testing.T) { // Make a huge list of nodes. for i := 0; i < maxServers; i++ { nodeName := fmt.Sprintf("s%02d", i) - m.AddServer(&server_details.ServerDetails{Name: nodeName}) + m.AddServer(&agent.Server{Name: nodeName}) } // Keep track of how many unique shuffles we get. @@ -260,7 +260,7 @@ func TestServers_RebalanceServers(t *testing.T) { } } -// func (m *Manager) RemoveServer(server *server_details.ServerDetails) { +// func (m *Manager) RemoveServer(server *agent.Server) { func TestManager_RemoveServer(t *testing.T) { const nodeNameFmt = "s%02d" m := testManager() @@ -271,21 +271,21 @@ func TestManager_RemoveServer(t *testing.T) { // Test removing server before its added nodeName := fmt.Sprintf(nodeNameFmt, 1) - s1 := &server_details.ServerDetails{Name: nodeName} + s1 := &agent.Server{Name: nodeName} m.RemoveServer(s1) m.AddServer(s1) nodeName = fmt.Sprintf(nodeNameFmt, 2) - s2 := &server_details.ServerDetails{Name: nodeName} + s2 := &agent.Server{Name: nodeName} m.RemoveServer(s2) m.AddServer(s2) const maxServers = 19 - servers := make([]*server_details.ServerDetails, maxServers) + servers := make([]*agent.Server, maxServers) // Already added two servers above for i := maxServers; i > 2; i-- { nodeName := fmt.Sprintf(nodeNameFmt, i) - server := &server_details.ServerDetails{Name: nodeName} + server := &agent.Server{Name: nodeName} servers = append(servers, server) m.AddServer(server) } @@ -299,7 +299,7 @@ func TestManager_RemoveServer(t *testing.T) { t.Fatalf("Expected %d servers, received %d", maxServers, m.NumServers()) } - findServer := func(server *server_details.ServerDetails) bool { + findServer := func(server *agent.Server) bool { for i := m.NumServers(); i > 0; i-- { s := m.FindServer() if s == server { @@ -310,7 +310,7 @@ func TestManager_RemoveServer(t *testing.T) { } expectedNumServers := maxServers - removedServers := make([]*server_details.ServerDetails, 0, maxServers) + removedServers := make([]*agent.Server, 0, maxServers) // Remove servers from the front of the list for i := 3; i > 0; i-- {