open-nomad/nomad/client_stats_endpoint_test.go

227 lines
5.8 KiB
Go
Raw Normal View History

2018-01-12 22:09:01 +00:00
package nomad
import (
"testing"
"time"
2018-01-12 22:09:01 +00:00
msgpackrpc "github.com/hashicorp/net-rpc-msgpackrpc"
2018-01-31 20:13:57 +00:00
"github.com/hashicorp/nomad/acl"
"github.com/hashicorp/nomad/ci"
2018-01-12 22:09:01 +00:00
"github.com/hashicorp/nomad/client"
"github.com/hashicorp/nomad/client/config"
cstructs "github.com/hashicorp/nomad/client/structs"
2018-01-15 22:48:53 +00:00
"github.com/hashicorp/nomad/helper/uuid"
2018-01-31 20:13:57 +00:00
"github.com/hashicorp/nomad/nomad/mock"
2018-01-12 22:09:01 +00:00
"github.com/hashicorp/nomad/nomad/structs"
"github.com/hashicorp/nomad/testutil"
"github.com/stretchr/testify/require"
)
func TestClientStats_Stats_Local(t *testing.T) {
ci.Parallel(t)
2018-01-12 22:09:01 +00:00
require := require.New(t)
// Start a server and client
s, cleanupS := TestServer(t, nil)
defer cleanupS()
2018-01-12 22:09:01 +00:00
codec := rpcClient(t, s)
testutil.WaitForLeader(t, s.RPC)
c, cleanupC := client.TestClient(t, func(c *config.Config) {
2018-01-12 22:09:01 +00:00
c.Servers = []string{s.config.RPCAddr.String()}
})
defer cleanupC()
2018-01-12 22:09:01 +00:00
testutil.WaitForResult(func() (bool, error) {
nodes := s.connectedNodes()
return len(nodes) == 1, nil
}, func(err error) {
t.Fatalf("should have a clients")
})
// Make the request without having a node-id
2018-02-06 01:20:42 +00:00
req := &structs.NodeSpecificRequest{
2018-01-12 22:09:01 +00:00
QueryOptions: structs.QueryOptions{Region: "global"},
}
// Fetch the response
var resp cstructs.ClientStatsResponse
err := msgpackrpc.CallWithCodec(codec, "ClientStats.Stats", req, &resp)
require.NotNil(err)
require.Contains(err.Error(), "missing")
// Fetch the response setting the node id
req.NodeID = c.NodeID()
var resp2 cstructs.ClientStatsResponse
err = msgpackrpc.CallWithCodec(codec, "ClientStats.Stats", req, &resp2)
require.Nil(err)
require.NotNil(resp2.HostStats)
}
2018-01-15 22:48:53 +00:00
2018-01-31 20:13:57 +00:00
func TestClientStats_Stats_Local_ACL(t *testing.T) {
ci.Parallel(t)
2018-01-31 20:13:57 +00:00
require := require.New(t)
// Start a server
s, root, cleanupS := TestACLServer(t, nil)
defer cleanupS()
2018-01-31 20:13:57 +00:00
codec := rpcClient(t, s)
testutil.WaitForLeader(t, s.RPC)
// Create a bad token
policyBad := mock.NamespacePolicy("other", "", []string{acl.NamespaceCapabilityReadFS})
tokenBad := mock.CreatePolicyAndToken(t, s.State(), 1005, "invalid", policyBad)
policyGood := mock.NodePolicy(acl.PolicyRead)
tokenGood := mock.CreatePolicyAndToken(t, s.State(), 1009, "valid2", policyGood)
cases := []struct {
Name string
Token string
ExpectedError string
}{
{
Name: "bad token",
Token: tokenBad.SecretID,
ExpectedError: structs.ErrPermissionDenied.Error(),
},
{
Name: "good token",
Token: tokenGood.SecretID,
ExpectedError: "Unknown node",
},
{
Name: "root token",
Token: root.SecretID,
ExpectedError: "Unknown node",
},
}
for _, c := range cases {
t.Run(c.Name, func(t *testing.T) {
// Make the request without having a node-id
2018-02-06 01:20:42 +00:00
req := &structs.NodeSpecificRequest{
2018-01-31 20:13:57 +00:00
NodeID: uuid.Generate(),
QueryOptions: structs.QueryOptions{
AuthToken: c.Token,
Region: "global",
},
}
// Fetch the response
var resp cstructs.ClientStatsResponse
err := msgpackrpc.CallWithCodec(codec, "ClientStats.Stats", req, &resp)
require.NotNil(err)
require.Contains(err.Error(), c.ExpectedError)
})
}
}
2018-01-15 22:48:53 +00:00
func TestClientStats_Stats_NoNode(t *testing.T) {
ci.Parallel(t)
2018-01-15 22:48:53 +00:00
require := require.New(t)
// Start a server and client
s, cleanupS := TestServer(t, nil)
defer cleanupS()
2018-01-15 22:48:53 +00:00
codec := rpcClient(t, s)
testutil.WaitForLeader(t, s.RPC)
// Make the request with a nonexistent node-id
2018-02-06 01:20:42 +00:00
req := &structs.NodeSpecificRequest{
2018-01-15 22:48:53 +00:00
NodeID: uuid.Generate(),
QueryOptions: structs.QueryOptions{Region: "global"},
}
// Fetch the response
var resp cstructs.ClientStatsResponse
err := msgpackrpc.CallWithCodec(codec, "ClientStats.Stats", req, &resp)
require.Nil(resp.HostStats)
require.NotNil(err)
require.Contains(err.Error(), "Unknown node")
}
func TestClientStats_Stats_OldNode(t *testing.T) {
ci.Parallel(t)
require := require.New(t)
// Start a server
s, cleanupS := TestServer(t, nil)
defer cleanupS()
state := s.State()
codec := rpcClient(t, s)
testutil.WaitForLeader(t, s.RPC)
// Test for an old version error
node := mock.Node()
node.Attributes["nomad.version"] = "0.7.1"
require.Nil(state.UpsertNode(structs.MsgTypeTestSetup, 1005, node))
req := &structs.NodeSpecificRequest{
NodeID: node.ID,
QueryOptions: structs.QueryOptions{Region: "global"},
}
// Fetch the response
var resp cstructs.ClientStatsResponse
err := msgpackrpc.CallWithCodec(codec, "ClientStats.Stats", req, &resp)
require.True(structs.IsErrNodeLacksRpc(err), err.Error())
}
2018-01-15 22:48:53 +00:00
func TestClientStats_Stats_Remote(t *testing.T) {
ci.Parallel(t)
2018-01-15 22:48:53 +00:00
require := require.New(t)
// Start a server and client
Simplify Bootstrap logic in tests This change updates tests to honor `BootstrapExpect` exclusively when forming test clusters and removes test only knobs, e.g. `config.DevDisableBootstrap`. Background: Test cluster creation is fragile. Test servers don't follow the BootstapExpected route like production clusters. Instead they start as single node clusters and then get rejoin and may risk causing brain split or other test flakiness. The test framework expose few knobs to control those (e.g. `config.DevDisableBootstrap` and `config.Bootstrap`) that control whether a server should bootstrap the cluster. These flags are confusing and it's unclear when to use: their usage in multi-node cluster isn't properly documented. Furthermore, they have some bad side-effects as they don't control Raft library: If `config.DevDisableBootstrap` is true, the test server may not immediately attempt to bootstrap a cluster, but after an election timeout (~50ms), Raft may force a leadership election and win it (with only one vote) and cause a split brain. The knobs are also confusing as Bootstrap is an overloaded term. In BootstrapExpect, we refer to bootstrapping the cluster only after N servers are connected. But in tests and the knobs above, it refers to whether the server is a single node cluster and shouldn't wait for any other server. Changes: This commit makes two changes: First, it relies on `BootstrapExpected` instead of `Bootstrap` and/or `DevMode` flags. This change is relatively trivial. Introduce a `Bootstrapped` flag to track if the cluster is bootstrapped. This allows us to keep `BootstrapExpected` immutable. Previously, the flag was a config value but it gets set to 0 after cluster bootstrap completes.
2020-03-02 15:29:24 +00:00
s1, cleanupS1 := TestServer(t, func(c *Config) {
c.BootstrapExpect = 2
})
defer cleanupS1()
s2, cleanupS2 := TestServer(t, func(c *Config) {
Simplify Bootstrap logic in tests This change updates tests to honor `BootstrapExpect` exclusively when forming test clusters and removes test only knobs, e.g. `config.DevDisableBootstrap`. Background: Test cluster creation is fragile. Test servers don't follow the BootstapExpected route like production clusters. Instead they start as single node clusters and then get rejoin and may risk causing brain split or other test flakiness. The test framework expose few knobs to control those (e.g. `config.DevDisableBootstrap` and `config.Bootstrap`) that control whether a server should bootstrap the cluster. These flags are confusing and it's unclear when to use: their usage in multi-node cluster isn't properly documented. Furthermore, they have some bad side-effects as they don't control Raft library: If `config.DevDisableBootstrap` is true, the test server may not immediately attempt to bootstrap a cluster, but after an election timeout (~50ms), Raft may force a leadership election and win it (with only one vote) and cause a split brain. The knobs are also confusing as Bootstrap is an overloaded term. In BootstrapExpect, we refer to bootstrapping the cluster only after N servers are connected. But in tests and the knobs above, it refers to whether the server is a single node cluster and shouldn't wait for any other server. Changes: This commit makes two changes: First, it relies on `BootstrapExpected` instead of `Bootstrap` and/or `DevMode` flags. This change is relatively trivial. Introduce a `Bootstrapped` flag to track if the cluster is bootstrapped. This allows us to keep `BootstrapExpected` immutable. Previously, the flag was a config value but it gets set to 0 after cluster bootstrap completes.
2020-03-02 15:29:24 +00:00
c.BootstrapExpect = 2
2018-01-15 22:48:53 +00:00
})
defer cleanupS2()
2018-01-15 22:48:53 +00:00
TestJoin(t, s1, s2)
testutil.WaitForLeader(t, s1.RPC)
testutil.WaitForLeader(t, s2.RPC)
codec := rpcClient(t, s2)
c, cleanup := client.TestClient(t, func(c *config.Config) {
2018-01-15 22:48:53 +00:00
c.Servers = []string{s2.config.RPCAddr.String()}
})
defer cleanup()
2018-01-15 22:48:53 +00:00
// Wait for client initialization
select {
case <-c.Ready():
case <-time.After(10 * time.Second):
require.Fail("client timedout on initialize")
}
2018-01-15 22:48:53 +00:00
testutil.WaitForResult(func() (bool, error) {
nodes := s2.connectedNodes()
return len(nodes) == 1, nil
}, func(err error) {
t.Fatalf("should have a clients")
})
2018-02-06 01:20:42 +00:00
// Force remove the connection locally in case it exists
s1.nodeConnsLock.Lock()
delete(s1.nodeConns, c.NodeID())
s1.nodeConnsLock.Unlock()
2018-01-15 22:48:53 +00:00
// Make the request without having a node-id
2018-02-06 01:20:42 +00:00
req := &structs.NodeSpecificRequest{
2018-01-15 22:48:53 +00:00
NodeID: uuid.Generate(),
QueryOptions: structs.QueryOptions{Region: "global"},
}
// Fetch the response
req.NodeID = c.NodeID()
var resp cstructs.ClientStatsResponse
err := msgpackrpc.CallWithCodec(codec, "ClientStats.Stats", req, &resp)
require.Nil(err)
require.NotNil(resp.HostStats)
}