e4c8b59919
* scheduler: allow updates after alloc reconnects When an allocation reconnects to a cluster the scheduler needs to run special logic to handle the reconnection, check if a replacement was create and stop one of them. If the allocation kept running while the node was disconnected, it will be reconnected with `ClientStatus: running` and the node will have `Status: ready`. This combination is the same as the normal steady state of allocation, where everything is running as expected. In order to differentiate between the two states (an allocation that is reconnecting and one that is just running) the scheduler needs an extra piece of state. The current implementation uses the presence of a `TaskClientReconnected` task event to detect when the allocation has reconnected and thus must go through the reconnection process. But this event remains even after the allocation is reconnected, causing all future evals to consider the allocation as still reconnecting. This commit changes the reconnect logic to use an `AllocState` to register when the allocation was reconnected. This provides the following benefits: - Only a limited number of task states are kept, and they are used for many other events. It's possible that, upon reconnecting, several actions are triggered that could cause the `TaskClientReconnected` event to be dropped. - Task events are set by clients and so their timestamps are subject to time skew from servers. This prevents using time to determine if an allocation reconnected after a disconnect event. - Disconnect events are already stored as `AllocState` and so storing reconnects there as well makes it the only source of information required. With the new logic, the reconnection logic is only triggered if the last `AllocState` is a disconnect event, meaning that the allocation has not been reconnected yet. After the reconnection is handled, the new `ClientStatus` is store in `AllocState` allowing future evals to skip the reconnection logic. * scheduler: prevent spurious placement on reconnect When a client reconnects it makes two independent RPC calls: - `Node.UpdateStatus` to heartbeat and set its status as `ready`. - `Node.UpdateAlloc` to update the status of its allocations. These two calls can happen in any order, and in case the allocations are updated before a heartbeat it causes the state to be the same as a node being disconnected: the node status will still be `disconnected` while the allocation `ClientStatus` is set to `running`. The current implementation did not handle this order of events properly, and the scheduler would create an unnecessary placement since it considered the allocation was being disconnected. This extra allocation would then be quickly stopped by the heartbeat eval. This commit adds a new code path to handle this order of events. If the node is `disconnected` and the allocation `ClientStatus` is `running` the scheduler will check if the allocation is actually reconnecting using its `AllocState` events. * rpc: only allow alloc updates from `ready` nodes Clients interact with servers using three main RPC methods: - `Node.GetAllocs` reads allocation data from the server and writes it to the client. - `Node.UpdateAlloc` reads allocation from from the client and writes them to the server. - `Node.UpdateStatus` writes the client status to the server and is used as the heartbeat mechanism. These three methods are called periodically by the clients and are done so independently from each other, meaning that there can't be any assumptions in their ordering. This can generate scenarios that are hard to reason about and to code for. For example, when a client misses too many heartbeats it will be considered `down` or `disconnected` and the allocations it was running are set to `lost` or `unknown`. When connectivity is restored the to rest of the cluster, the natural mental model is to think that the client will heartbeat first and then update its allocations status into the servers. But since there's no inherit order in these calls the reverse is just as possible: the client updates the alloc status and then heartbeats. This results in a state where allocs are, for example, `running` while the client is still `disconnected`. This commit adds a new verification to the `Node.UpdateAlloc` method to reject updates from nodes that are not `ready`, forcing clients to heartbeat first. Since this check is done server-side there is no need to coordinate operations client-side: they can continue sending these requests independently and alloc update will succeed after the heartbeat is done. * chagelog: add entry for #15068 * code review * client: skip terminal allocations on reconnect When the client reconnects with the server it synchronizes the state of its allocations by sending data using the `Node.UpdateAlloc` RPC and fetching data using the `Node.GetClientAllocs` RPC. If the data fetch happens before the data write, `unknown` allocations will still be in this state and would trigger the `allocRunner.Reconnect` flow. But when the server `DesiredStatus` for the allocation is `stop` the client should not reconnect the allocation. * apply more code review changes * scheduler: persist changes to reconnected allocs Reconnected allocs have a new AllocState entry that must be persisted by the plan applier. * rpc: read node ID from allocs in UpdateAlloc The AllocUpdateRequest struct is used in three disjoint use cases: 1. Stripped allocs from clients Node.UpdateAlloc RPC using the Allocs, and WriteRequest fields 2. Raft log message using the Allocs, Evals, and WriteRequest fields 3. Plan updates using the AllocsStopped, AllocsUpdated, and Job fields Adding a new field that would only be used in one these cases (1) made things more confusing and error prone. While in theory an AllocUpdateRequest could send allocations from different nodes, in practice this never actually happens since only clients call this method with their own allocations. * scheduler: remove logic to handle exceptional case This condition could only be hit if, somehow, the allocation status was set to "running" while the client was "unknown". This was addressed by enforcing an order in "Node.UpdateStatus" and "Node.UpdateAlloc" RPC calls, so this scenario is not expected to happen. Adding unnecessary code to the scheduler makes it harder to read and reason about it. * more code review * remove another unused test
4229 lines
119 KiB
Go
4229 lines
119 KiB
Go
package nomad
|
|
|
|
import (
|
|
"errors"
|
|
"fmt"
|
|
"net"
|
|
"reflect"
|
|
"strings"
|
|
"testing"
|
|
"time"
|
|
|
|
memdb "github.com/hashicorp/go-memdb"
|
|
msgpackrpc "github.com/hashicorp/net-rpc-msgpackrpc"
|
|
"github.com/hashicorp/nomad/acl"
|
|
"github.com/hashicorp/nomad/ci"
|
|
"github.com/hashicorp/nomad/command/agent/consul"
|
|
"github.com/hashicorp/nomad/helper"
|
|
"github.com/hashicorp/nomad/helper/pointer"
|
|
"github.com/hashicorp/nomad/helper/uuid"
|
|
"github.com/hashicorp/nomad/nomad/mock"
|
|
"github.com/hashicorp/nomad/nomad/state"
|
|
"github.com/hashicorp/nomad/nomad/structs"
|
|
"github.com/hashicorp/nomad/testutil"
|
|
vapi "github.com/hashicorp/vault/api"
|
|
"github.com/kr/pretty"
|
|
"github.com/stretchr/testify/assert"
|
|
"github.com/stretchr/testify/require"
|
|
)
|
|
|
|
func TestClientEndpoint_Register(t *testing.T) {
|
|
ci.Parallel(t)
|
|
require := require.New(t)
|
|
|
|
s1, cleanupS1 := TestServer(t, nil)
|
|
defer cleanupS1()
|
|
codec := rpcClient(t, s1)
|
|
testutil.WaitForLeader(t, s1.RPC)
|
|
|
|
// Check that we have no client connections
|
|
require.Empty(s1.connectedNodes())
|
|
|
|
// Create the register request
|
|
node := mock.Node()
|
|
req := &structs.NodeRegisterRequest{
|
|
Node: node,
|
|
WriteRequest: structs.WriteRequest{Region: "global"},
|
|
}
|
|
|
|
// Fetch the response
|
|
var resp structs.GenericResponse
|
|
if err := msgpackrpc.CallWithCodec(codec, "Node.Register", req, &resp); err != nil {
|
|
t.Fatalf("err: %v", err)
|
|
}
|
|
if resp.Index == 0 {
|
|
t.Fatalf("bad index: %d", resp.Index)
|
|
}
|
|
|
|
// Check that we have the client connections
|
|
nodes := s1.connectedNodes()
|
|
require.Len(nodes, 1)
|
|
require.Contains(nodes, node.ID)
|
|
|
|
// Check for the node in the FSM
|
|
state := s1.fsm.State()
|
|
ws := memdb.NewWatchSet()
|
|
out, err := state.NodeByID(ws, node.ID)
|
|
if err != nil {
|
|
t.Fatalf("err: %v", err)
|
|
}
|
|
if out == nil {
|
|
t.Fatalf("expected node")
|
|
}
|
|
if out.CreateIndex != resp.Index {
|
|
t.Fatalf("index mis-match")
|
|
}
|
|
if out.ComputedClass == "" {
|
|
t.Fatal("ComputedClass not set")
|
|
}
|
|
|
|
// Close the connection and check that we remove the client connections
|
|
require.Nil(codec.Close())
|
|
testutil.WaitForResult(func() (bool, error) {
|
|
nodes := s1.connectedNodes()
|
|
return len(nodes) == 0, nil
|
|
}, func(err error) {
|
|
t.Fatalf("should have no clients")
|
|
})
|
|
}
|
|
|
|
// This test asserts that we only track node connections if they are not from
|
|
// forwarded RPCs. This is essential otherwise we will think a Yamux session to
|
|
// a Nomad server is actually the session to the node.
|
|
func TestClientEndpoint_Register_NodeConn_Forwarded(t *testing.T) {
|
|
ci.Parallel(t)
|
|
require := require.New(t)
|
|
|
|
s1, cleanupS1 := TestServer(t, func(c *Config) {
|
|
c.BootstrapExpect = 2
|
|
})
|
|
|
|
defer cleanupS1()
|
|
s2, cleanupS2 := TestServer(t, func(c *Config) {
|
|
c.BootstrapExpect = 2
|
|
})
|
|
defer cleanupS2()
|
|
TestJoin(t, s1, s2)
|
|
testutil.WaitForLeader(t, s1.RPC)
|
|
testutil.WaitForLeader(t, s2.RPC)
|
|
|
|
// Determine the non-leader server
|
|
var leader, nonLeader *Server
|
|
if s1.IsLeader() {
|
|
leader = s1
|
|
nonLeader = s2
|
|
} else {
|
|
leader = s2
|
|
nonLeader = s1
|
|
}
|
|
|
|
// Send the requests to the non-leader
|
|
codec := rpcClient(t, nonLeader)
|
|
|
|
// Check that we have no client connections
|
|
require.Empty(nonLeader.connectedNodes())
|
|
require.Empty(leader.connectedNodes())
|
|
|
|
// Create the register request
|
|
node := mock.Node()
|
|
req := &structs.NodeRegisterRequest{
|
|
Node: node,
|
|
WriteRequest: structs.WriteRequest{Region: "global"},
|
|
}
|
|
|
|
// Fetch the response
|
|
var resp structs.GenericResponse
|
|
if err := msgpackrpc.CallWithCodec(codec, "Node.Register", req, &resp); err != nil {
|
|
t.Fatalf("err: %v", err)
|
|
}
|
|
if resp.Index == 0 {
|
|
t.Fatalf("bad index: %d", resp.Index)
|
|
}
|
|
|
|
// Check that we have the client connections on the non leader
|
|
nodes := nonLeader.connectedNodes()
|
|
require.Len(nodes, 1)
|
|
require.Contains(nodes, node.ID)
|
|
|
|
// Check that we have no client connections on the leader
|
|
nodes = leader.connectedNodes()
|
|
require.Empty(nodes)
|
|
|
|
// Check for the node in the FSM
|
|
state := leader.State()
|
|
testutil.WaitForResult(func() (bool, error) {
|
|
out, err := state.NodeByID(nil, node.ID)
|
|
if err != nil {
|
|
return false, err
|
|
}
|
|
if out == nil {
|
|
return false, fmt.Errorf("expected node")
|
|
}
|
|
if out.CreateIndex != resp.Index {
|
|
return false, fmt.Errorf("index mis-match")
|
|
}
|
|
if out.ComputedClass == "" {
|
|
return false, fmt.Errorf("ComputedClass not set")
|
|
}
|
|
|
|
return true, nil
|
|
}, func(err error) {
|
|
t.Fatalf("err: %v", err)
|
|
})
|
|
|
|
// Close the connection and check that we remove the client connections
|
|
require.Nil(codec.Close())
|
|
testutil.WaitForResult(func() (bool, error) {
|
|
nodes := nonLeader.connectedNodes()
|
|
return len(nodes) == 0, nil
|
|
}, func(err error) {
|
|
t.Fatalf("should have no clients")
|
|
})
|
|
}
|
|
|
|
func TestClientEndpoint_Register_SecretMismatch(t *testing.T) {
|
|
ci.Parallel(t)
|
|
|
|
s1, cleanupS1 := TestServer(t, nil)
|
|
defer cleanupS1()
|
|
codec := rpcClient(t, s1)
|
|
testutil.WaitForLeader(t, s1.RPC)
|
|
|
|
// Create the register request
|
|
node := mock.Node()
|
|
req := &structs.NodeRegisterRequest{
|
|
Node: node,
|
|
WriteRequest: structs.WriteRequest{Region: "global"},
|
|
}
|
|
|
|
// Fetch the response
|
|
var resp structs.GenericResponse
|
|
if err := msgpackrpc.CallWithCodec(codec, "Node.Register", req, &resp); err != nil {
|
|
t.Fatalf("err: %v", err)
|
|
}
|
|
|
|
// Update the nodes SecretID
|
|
node.SecretID = uuid.Generate()
|
|
err := msgpackrpc.CallWithCodec(codec, "Node.Register", req, &resp)
|
|
if err == nil || !strings.Contains(err.Error(), "Not registering") {
|
|
t.Fatalf("Expecting error regarding mismatching secret id: %v", err)
|
|
}
|
|
}
|
|
|
|
// Test the deprecated single node deregistration path
|
|
func TestClientEndpoint_DeregisterOne(t *testing.T) {
|
|
ci.Parallel(t)
|
|
|
|
s1, cleanupS1 := TestServer(t, nil)
|
|
defer cleanupS1()
|
|
codec := rpcClient(t, s1)
|
|
testutil.WaitForLeader(t, s1.RPC)
|
|
|
|
// Create the register request
|
|
node := mock.Node()
|
|
reg := &structs.NodeRegisterRequest{
|
|
Node: node,
|
|
WriteRequest: structs.WriteRequest{Region: "global"},
|
|
}
|
|
|
|
// Fetch the response
|
|
var resp structs.GenericResponse
|
|
if err := msgpackrpc.CallWithCodec(codec, "Node.Register", reg, &resp); err != nil {
|
|
t.Fatalf("err: %v", err)
|
|
}
|
|
|
|
// Deregister
|
|
dereg := &structs.NodeDeregisterRequest{
|
|
NodeID: node.ID,
|
|
WriteRequest: structs.WriteRequest{Region: "global"},
|
|
}
|
|
var resp2 structs.GenericResponse
|
|
if err := msgpackrpc.CallWithCodec(codec, "Node.Deregister", dereg, &resp2); err != nil {
|
|
t.Fatalf("err: %v", err)
|
|
}
|
|
if resp2.Index == 0 {
|
|
t.Fatalf("bad index: %d", resp2.Index)
|
|
}
|
|
|
|
// Check for the node in the FSM
|
|
state := s1.fsm.State()
|
|
ws := memdb.NewWatchSet()
|
|
out, err := state.NodeByID(ws, node.ID)
|
|
if err != nil {
|
|
t.Fatalf("err: %v", err)
|
|
}
|
|
if out != nil {
|
|
t.Fatalf("unexpected node")
|
|
}
|
|
}
|
|
|
|
func TestClientEndpoint_Deregister_ACL(t *testing.T) {
|
|
ci.Parallel(t)
|
|
|
|
s1, root, cleanupS1 := TestACLServer(t, nil)
|
|
defer cleanupS1()
|
|
codec := rpcClient(t, s1)
|
|
testutil.WaitForLeader(t, s1.RPC)
|
|
|
|
// Create the node
|
|
node := mock.Node()
|
|
node1 := mock.Node()
|
|
state := s1.fsm.State()
|
|
if err := state.UpsertNode(structs.MsgTypeTestSetup, 1, node); err != nil {
|
|
t.Fatalf("err: %v", err)
|
|
}
|
|
if err := state.UpsertNode(structs.MsgTypeTestSetup, 2, node1); err != nil {
|
|
t.Fatalf("err: %v", err)
|
|
}
|
|
|
|
// Create the policy and tokens
|
|
validToken := mock.CreatePolicyAndToken(t, state, 1001, "test-valid", mock.NodePolicy(acl.PolicyWrite))
|
|
invalidToken := mock.CreatePolicyAndToken(t, state, 1003, "test-invalid", mock.NodePolicy(acl.PolicyRead))
|
|
|
|
// Deregister without any token and expect it to fail
|
|
dereg := &structs.NodeBatchDeregisterRequest{
|
|
NodeIDs: []string{node.ID},
|
|
WriteRequest: structs.WriteRequest{Region: "global"},
|
|
}
|
|
var resp structs.GenericResponse
|
|
if err := msgpackrpc.CallWithCodec(codec, "Node.BatchDeregister", dereg, &resp); err == nil {
|
|
t.Fatalf("node de-register succeeded")
|
|
}
|
|
|
|
// Deregister with a valid token
|
|
dereg.AuthToken = validToken.SecretID
|
|
if err := msgpackrpc.CallWithCodec(codec, "Node.BatchDeregister", dereg, &resp); err != nil {
|
|
t.Fatalf("err: %v", err)
|
|
}
|
|
|
|
// Check for the node in the FSM
|
|
ws := memdb.NewWatchSet()
|
|
out, err := state.NodeByID(ws, node.ID)
|
|
if err != nil {
|
|
t.Fatalf("err: %v", err)
|
|
}
|
|
if out != nil {
|
|
t.Fatalf("unexpected node")
|
|
}
|
|
|
|
// Deregister with an invalid token.
|
|
dereg1 := &structs.NodeBatchDeregisterRequest{
|
|
NodeIDs: []string{node1.ID},
|
|
WriteRequest: structs.WriteRequest{Region: "global"},
|
|
}
|
|
dereg1.AuthToken = invalidToken.SecretID
|
|
if err := msgpackrpc.CallWithCodec(codec, "Node.BatchDeregister", dereg1, &resp); err == nil {
|
|
t.Fatalf("rpc should not have succeeded")
|
|
}
|
|
|
|
// Try with a root token
|
|
dereg1.AuthToken = root.SecretID
|
|
if err := msgpackrpc.CallWithCodec(codec, "Node.BatchDeregister", dereg1, &resp); err != nil {
|
|
t.Fatalf("err: %v", err)
|
|
}
|
|
}
|
|
|
|
func TestClientEndpoint_Deregister_Vault(t *testing.T) {
|
|
ci.Parallel(t)
|
|
|
|
s1, cleanupS1 := TestServer(t, nil)
|
|
defer cleanupS1()
|
|
codec := rpcClient(t, s1)
|
|
testutil.WaitForLeader(t, s1.RPC)
|
|
|
|
// Create the register request
|
|
node := mock.Node()
|
|
reg := &structs.NodeRegisterRequest{
|
|
Node: node,
|
|
WriteRequest: structs.WriteRequest{Region: "global"},
|
|
}
|
|
|
|
// Fetch the response
|
|
var resp structs.GenericResponse
|
|
if err := msgpackrpc.CallWithCodec(codec, "Node.Register", reg, &resp); err != nil {
|
|
t.Fatalf("err: %v", err)
|
|
}
|
|
|
|
// Swap the servers Vault Client
|
|
tvc := &TestVaultClient{}
|
|
s1.vault = tvc
|
|
|
|
// Put some Vault accessors in the state store for that node
|
|
state := s1.fsm.State()
|
|
va1 := mock.VaultAccessor()
|
|
va1.NodeID = node.ID
|
|
va2 := mock.VaultAccessor()
|
|
va2.NodeID = node.ID
|
|
state.UpsertVaultAccessor(100, []*structs.VaultAccessor{va1, va2})
|
|
|
|
// Deregister
|
|
dereg := &structs.NodeBatchDeregisterRequest{
|
|
NodeIDs: []string{node.ID},
|
|
WriteRequest: structs.WriteRequest{Region: "global"},
|
|
}
|
|
var resp2 structs.GenericResponse
|
|
if err := msgpackrpc.CallWithCodec(codec, "Node.BatchDeregister", dereg, &resp2); err != nil {
|
|
t.Fatalf("err: %v", err)
|
|
}
|
|
if resp2.Index == 0 {
|
|
t.Fatalf("bad index: %d", resp2.Index)
|
|
}
|
|
|
|
// Check for the node in the FSM
|
|
ws := memdb.NewWatchSet()
|
|
out, err := state.NodeByID(ws, node.ID)
|
|
if err != nil {
|
|
t.Fatalf("err: %v", err)
|
|
}
|
|
if out != nil {
|
|
t.Fatalf("unexpected node")
|
|
}
|
|
|
|
// Check that the endpoint revoked the tokens
|
|
if l := len(tvc.RevokedTokens); l != 2 {
|
|
t.Fatalf("Deregister revoked %d tokens; want 2", l)
|
|
}
|
|
}
|
|
|
|
func TestClientEndpoint_UpdateStatus(t *testing.T) {
|
|
ci.Parallel(t)
|
|
require := require.New(t)
|
|
|
|
s1, cleanupS1 := TestServer(t, nil)
|
|
defer cleanupS1()
|
|
codec := rpcClient(t, s1)
|
|
testutil.WaitForLeader(t, s1.RPC)
|
|
|
|
// Check that we have no client connections
|
|
require.Empty(s1.connectedNodes())
|
|
|
|
// Create the register request
|
|
node := mock.Node()
|
|
reg := &structs.NodeRegisterRequest{
|
|
Node: node,
|
|
WriteRequest: structs.WriteRequest{Region: "global"},
|
|
}
|
|
|
|
// Fetch the response
|
|
var resp structs.NodeUpdateResponse
|
|
if err := msgpackrpc.CallWithCodec(codec, "Node.Register", reg, &resp); err != nil {
|
|
t.Fatalf("err: %v", err)
|
|
}
|
|
|
|
// Check for heartbeat interval
|
|
ttl := resp.HeartbeatTTL
|
|
if ttl < s1.config.MinHeartbeatTTL || ttl > 2*s1.config.MinHeartbeatTTL {
|
|
t.Fatalf("bad: %#v", ttl)
|
|
}
|
|
|
|
// Update the status
|
|
dereg := &structs.NodeUpdateStatusRequest{
|
|
NodeID: node.ID,
|
|
Status: structs.NodeStatusInit,
|
|
WriteRequest: structs.WriteRequest{Region: "global"},
|
|
}
|
|
var resp2 structs.NodeUpdateResponse
|
|
if err := msgpackrpc.CallWithCodec(codec, "Node.UpdateStatus", dereg, &resp2); err != nil {
|
|
t.Fatalf("err: %v", err)
|
|
}
|
|
if resp2.Index == 0 {
|
|
t.Fatalf("bad index: %d", resp2.Index)
|
|
}
|
|
|
|
// Check for heartbeat interval
|
|
ttl = resp2.HeartbeatTTL
|
|
if ttl < s1.config.MinHeartbeatTTL || ttl > 2*s1.config.MinHeartbeatTTL {
|
|
t.Fatalf("bad: %#v", ttl)
|
|
}
|
|
|
|
// Check that we have the client connections
|
|
nodes := s1.connectedNodes()
|
|
require.Len(nodes, 1)
|
|
require.Contains(nodes, node.ID)
|
|
|
|
// Check for the node in the FSM
|
|
state := s1.fsm.State()
|
|
ws := memdb.NewWatchSet()
|
|
out, err := state.NodeByID(ws, node.ID)
|
|
if err != nil {
|
|
t.Fatalf("err: %v", err)
|
|
}
|
|
if out == nil {
|
|
t.Fatalf("expected node")
|
|
}
|
|
if out.ModifyIndex != resp2.Index {
|
|
t.Fatalf("index mis-match")
|
|
}
|
|
|
|
// Close the connection and check that we remove the client connections
|
|
require.Nil(codec.Close())
|
|
testutil.WaitForResult(func() (bool, error) {
|
|
nodes := s1.connectedNodes()
|
|
return len(nodes) == 0, nil
|
|
}, func(err error) {
|
|
t.Fatalf("should have no clients")
|
|
})
|
|
}
|
|
|
|
func TestClientEndpoint_UpdateStatus_Vault(t *testing.T) {
|
|
ci.Parallel(t)
|
|
|
|
s1, cleanupS1 := TestServer(t, nil)
|
|
defer cleanupS1()
|
|
codec := rpcClient(t, s1)
|
|
testutil.WaitForLeader(t, s1.RPC)
|
|
|
|
// Create the register request
|
|
node := mock.Node()
|
|
reg := &structs.NodeRegisterRequest{
|
|
Node: node,
|
|
WriteRequest: structs.WriteRequest{Region: "global"},
|
|
}
|
|
|
|
// Fetch the response
|
|
var resp structs.NodeUpdateResponse
|
|
if err := msgpackrpc.CallWithCodec(codec, "Node.Register", reg, &resp); err != nil {
|
|
t.Fatalf("err: %v", err)
|
|
}
|
|
|
|
// Check for heartbeat interval
|
|
ttl := resp.HeartbeatTTL
|
|
if ttl < s1.config.MinHeartbeatTTL || ttl > 2*s1.config.MinHeartbeatTTL {
|
|
t.Fatalf("bad: %#v", ttl)
|
|
}
|
|
|
|
// Swap the servers Vault Client
|
|
tvc := &TestVaultClient{}
|
|
s1.vault = tvc
|
|
|
|
// Put some Vault accessors in the state store for that node
|
|
state := s1.fsm.State()
|
|
va1 := mock.VaultAccessor()
|
|
va1.NodeID = node.ID
|
|
va2 := mock.VaultAccessor()
|
|
va2.NodeID = node.ID
|
|
state.UpsertVaultAccessor(100, []*structs.VaultAccessor{va1, va2})
|
|
|
|
// Update the status to be down
|
|
dereg := &structs.NodeUpdateStatusRequest{
|
|
NodeID: node.ID,
|
|
Status: structs.NodeStatusDown,
|
|
WriteRequest: structs.WriteRequest{Region: "global"},
|
|
}
|
|
var resp2 structs.NodeUpdateResponse
|
|
if err := msgpackrpc.CallWithCodec(codec, "Node.UpdateStatus", dereg, &resp2); err != nil {
|
|
t.Fatalf("err: %v", err)
|
|
}
|
|
if resp2.Index == 0 {
|
|
t.Fatalf("bad index: %d", resp2.Index)
|
|
}
|
|
|
|
// Check that the endpoint revoked the tokens
|
|
if l := len(tvc.RevokedTokens); l != 2 {
|
|
t.Fatalf("Deregister revoked %d tokens; want 2", l)
|
|
}
|
|
}
|
|
|
|
func TestClientEndpoint_UpdateStatus_HeartbeatRecovery(t *testing.T) {
|
|
ci.Parallel(t)
|
|
require := require.New(t)
|
|
|
|
s1, cleanupS1 := TestServer(t, nil)
|
|
defer cleanupS1()
|
|
codec := rpcClient(t, s1)
|
|
testutil.WaitForLeader(t, s1.RPC)
|
|
|
|
// Check that we have no client connections
|
|
require.Empty(s1.connectedNodes())
|
|
|
|
// Create the register request but make the node down
|
|
node := mock.Node()
|
|
node.Status = structs.NodeStatusDown
|
|
reg := &structs.NodeRegisterRequest{
|
|
Node: node,
|
|
WriteRequest: structs.WriteRequest{Region: "global"},
|
|
}
|
|
|
|
// Fetch the response
|
|
var resp structs.NodeUpdateResponse
|
|
require.NoError(msgpackrpc.CallWithCodec(codec, "Node.Register", reg, &resp))
|
|
|
|
// Update the status
|
|
dereg := &structs.NodeUpdateStatusRequest{
|
|
NodeID: node.ID,
|
|
Status: structs.NodeStatusInit,
|
|
WriteRequest: structs.WriteRequest{Region: "global"},
|
|
}
|
|
var resp2 structs.NodeUpdateResponse
|
|
require.NoError(msgpackrpc.CallWithCodec(codec, "Node.UpdateStatus", dereg, &resp2))
|
|
require.NotZero(resp2.Index)
|
|
|
|
// Check for heartbeat interval
|
|
ttl := resp2.HeartbeatTTL
|
|
if ttl < s1.config.MinHeartbeatTTL || ttl > 2*s1.config.MinHeartbeatTTL {
|
|
t.Fatalf("bad: %#v", ttl)
|
|
}
|
|
|
|
// Check for the node in the FSM
|
|
state := s1.fsm.State()
|
|
ws := memdb.NewWatchSet()
|
|
out, err := state.NodeByID(ws, node.ID)
|
|
require.NoError(err)
|
|
require.NotNil(out)
|
|
require.EqualValues(resp2.Index, out.ModifyIndex)
|
|
require.Len(out.Events, 2)
|
|
require.Equal(NodeHeartbeatEventReregistered, out.Events[1].Message)
|
|
}
|
|
|
|
func TestClientEndpoint_Register_GetEvals(t *testing.T) {
|
|
ci.Parallel(t)
|
|
|
|
s1, cleanupS1 := TestServer(t, nil)
|
|
defer cleanupS1()
|
|
codec := rpcClient(t, s1)
|
|
testutil.WaitForLeader(t, s1.RPC)
|
|
|
|
// Register a system job.
|
|
job := mock.SystemJob()
|
|
state := s1.fsm.State()
|
|
if err := state.UpsertJob(structs.MsgTypeTestSetup, 1, job); err != nil {
|
|
t.Fatalf("err: %v", err)
|
|
}
|
|
|
|
// Create the register request going directly to ready
|
|
node := mock.Node()
|
|
node.Status = structs.NodeStatusReady
|
|
reg := &structs.NodeRegisterRequest{
|
|
Node: node,
|
|
WriteRequest: structs.WriteRequest{Region: "global"},
|
|
}
|
|
|
|
// Fetch the response
|
|
var resp structs.NodeUpdateResponse
|
|
if err := msgpackrpc.CallWithCodec(codec, "Node.Register", reg, &resp); err != nil {
|
|
t.Fatalf("err: %v", err)
|
|
}
|
|
|
|
// Check for heartbeat interval
|
|
ttl := resp.HeartbeatTTL
|
|
if ttl < s1.config.MinHeartbeatTTL || ttl > 2*s1.config.MinHeartbeatTTL {
|
|
t.Fatalf("bad: %#v", ttl)
|
|
}
|
|
|
|
// Check for an eval caused by the system job.
|
|
if len(resp.EvalIDs) != 1 {
|
|
t.Fatalf("expected one eval; got %#v", resp.EvalIDs)
|
|
}
|
|
|
|
evalID := resp.EvalIDs[0]
|
|
ws := memdb.NewWatchSet()
|
|
eval, err := state.EvalByID(ws, evalID)
|
|
if err != nil {
|
|
t.Fatalf("could not get eval %v", evalID)
|
|
}
|
|
|
|
if eval.Type != "system" {
|
|
t.Fatalf("unexpected eval type; got %v; want %q", eval.Type, "system")
|
|
}
|
|
|
|
// Check for the node in the FSM
|
|
out, err := state.NodeByID(ws, node.ID)
|
|
if err != nil {
|
|
t.Fatalf("err: %v", err)
|
|
}
|
|
if out == nil {
|
|
t.Fatalf("expected node")
|
|
}
|
|
if out.ModifyIndex != resp.Index {
|
|
t.Fatalf("index mis-match")
|
|
}
|
|
|
|
// Transition it to down and then ready
|
|
node.Status = structs.NodeStatusDown
|
|
reg = &structs.NodeRegisterRequest{
|
|
Node: node,
|
|
WriteRequest: structs.WriteRequest{Region: "global"},
|
|
}
|
|
|
|
// Fetch the response
|
|
if err := msgpackrpc.CallWithCodec(codec, "Node.Register", reg, &resp); err != nil {
|
|
t.Fatalf("err: %v", err)
|
|
}
|
|
|
|
if len(resp.EvalIDs) != 1 {
|
|
t.Fatalf("expected one eval; got %#v", resp.EvalIDs)
|
|
}
|
|
|
|
node.Status = structs.NodeStatusReady
|
|
reg = &structs.NodeRegisterRequest{
|
|
Node: node,
|
|
WriteRequest: structs.WriteRequest{Region: "global"},
|
|
}
|
|
|
|
// Fetch the response
|
|
if err := msgpackrpc.CallWithCodec(codec, "Node.Register", reg, &resp); err != nil {
|
|
t.Fatalf("err: %v", err)
|
|
}
|
|
|
|
if len(resp.EvalIDs) != 1 {
|
|
t.Fatalf("expected one eval; got %#v", resp.EvalIDs)
|
|
}
|
|
}
|
|
|
|
func TestClientEndpoint_UpdateStatus_GetEvals(t *testing.T) {
|
|
ci.Parallel(t)
|
|
|
|
s1, cleanupS1 := TestServer(t, nil)
|
|
defer cleanupS1()
|
|
codec := rpcClient(t, s1)
|
|
testutil.WaitForLeader(t, s1.RPC)
|
|
|
|
// Register a system job.
|
|
job := mock.SystemJob()
|
|
state := s1.fsm.State()
|
|
if err := state.UpsertJob(structs.MsgTypeTestSetup, 1, job); err != nil {
|
|
t.Fatalf("err: %v", err)
|
|
}
|
|
|
|
// Create the register request
|
|
node := mock.Node()
|
|
node.Status = structs.NodeStatusInit
|
|
reg := &structs.NodeRegisterRequest{
|
|
Node: node,
|
|
WriteRequest: structs.WriteRequest{Region: "global"},
|
|
}
|
|
|
|
// Fetch the response
|
|
var resp structs.NodeUpdateResponse
|
|
if err := msgpackrpc.CallWithCodec(codec, "Node.Register", reg, &resp); err != nil {
|
|
t.Fatalf("err: %v", err)
|
|
}
|
|
|
|
// Check for heartbeat interval
|
|
ttl := resp.HeartbeatTTL
|
|
if ttl < s1.config.MinHeartbeatTTL || ttl > 2*s1.config.MinHeartbeatTTL {
|
|
t.Fatalf("bad: %#v", ttl)
|
|
}
|
|
|
|
// Update the status
|
|
update := &structs.NodeUpdateStatusRequest{
|
|
NodeID: node.ID,
|
|
Status: structs.NodeStatusReady,
|
|
WriteRequest: structs.WriteRequest{Region: "global"},
|
|
}
|
|
var resp2 structs.NodeUpdateResponse
|
|
if err := msgpackrpc.CallWithCodec(codec, "Node.UpdateStatus", update, &resp2); err != nil {
|
|
t.Fatalf("err: %v", err)
|
|
}
|
|
if resp2.Index == 0 {
|
|
t.Fatalf("bad index: %d", resp2.Index)
|
|
}
|
|
|
|
// Check for an eval caused by the system job.
|
|
if len(resp2.EvalIDs) != 1 {
|
|
t.Fatalf("expected one eval; got %#v", resp2.EvalIDs)
|
|
}
|
|
|
|
evalID := resp2.EvalIDs[0]
|
|
ws := memdb.NewWatchSet()
|
|
eval, err := state.EvalByID(ws, evalID)
|
|
if err != nil {
|
|
t.Fatalf("could not get eval %v", evalID)
|
|
}
|
|
|
|
if eval.Type != "system" {
|
|
t.Fatalf("unexpected eval type; got %v; want %q", eval.Type, "system")
|
|
}
|
|
|
|
// Check for heartbeat interval
|
|
ttl = resp2.HeartbeatTTL
|
|
if ttl < s1.config.MinHeartbeatTTL || ttl > 2*s1.config.MinHeartbeatTTL {
|
|
t.Fatalf("bad: %#v", ttl)
|
|
}
|
|
|
|
// Check for the node in the FSM
|
|
out, err := state.NodeByID(ws, node.ID)
|
|
if err != nil {
|
|
t.Fatalf("err: %v", err)
|
|
}
|
|
if out == nil {
|
|
t.Fatalf("expected node")
|
|
}
|
|
if out.ModifyIndex != resp2.Index {
|
|
t.Fatalf("index mis-match")
|
|
}
|
|
}
|
|
|
|
func TestClientEndpoint_UpdateStatus_HeartbeatOnly(t *testing.T) {
|
|
ci.Parallel(t)
|
|
|
|
s1, cleanupS1 := TestServer(t, func(c *Config) {
|
|
c.BootstrapExpect = 3
|
|
})
|
|
defer cleanupS1()
|
|
|
|
s2, cleanupS2 := TestServer(t, func(c *Config) {
|
|
c.BootstrapExpect = 3
|
|
})
|
|
defer cleanupS2()
|
|
|
|
s3, cleanupS3 := TestServer(t, func(c *Config) {
|
|
c.BootstrapExpect = 3
|
|
})
|
|
defer cleanupS3()
|
|
servers := []*Server{s1, s2, s3}
|
|
TestJoin(t, s1, s2, s3)
|
|
|
|
for _, s := range servers {
|
|
testutil.WaitForResult(func() (bool, error) {
|
|
peers, _ := s.numPeers()
|
|
return peers == 3, nil
|
|
}, func(err error) {
|
|
t.Fatalf("should have 3 peers")
|
|
})
|
|
}
|
|
|
|
codec := rpcClient(t, s1)
|
|
testutil.WaitForLeader(t, s1.RPC)
|
|
|
|
// Create the register request
|
|
node := mock.Node()
|
|
reg := &structs.NodeRegisterRequest{
|
|
Node: node,
|
|
WriteRequest: structs.WriteRequest{Region: "global"},
|
|
}
|
|
|
|
// Fetch the response
|
|
var resp structs.NodeUpdateResponse
|
|
if err := msgpackrpc.CallWithCodec(codec, "Node.Register", reg, &resp); err != nil {
|
|
t.Fatalf("err: %v", err)
|
|
}
|
|
|
|
// Check for heartbeat interval
|
|
ttl := resp.HeartbeatTTL
|
|
if ttl < s1.config.MinHeartbeatTTL || ttl > 2*s1.config.MinHeartbeatTTL {
|
|
t.Fatalf("bad: %#v", ttl)
|
|
}
|
|
|
|
// Check for heartbeat servers
|
|
serverAddrs := resp.Servers
|
|
if len(serverAddrs) == 0 {
|
|
t.Fatalf("bad: %#v", serverAddrs)
|
|
}
|
|
|
|
// Update the status, static state
|
|
dereg := &structs.NodeUpdateStatusRequest{
|
|
NodeID: node.ID,
|
|
Status: node.Status,
|
|
WriteRequest: structs.WriteRequest{Region: "global"},
|
|
}
|
|
var resp2 structs.NodeUpdateResponse
|
|
if err := msgpackrpc.CallWithCodec(codec, "Node.UpdateStatus", dereg, &resp2); err != nil {
|
|
t.Fatalf("err: %v", err)
|
|
}
|
|
if resp2.Index != 0 {
|
|
t.Fatalf("bad index: %d", resp2.Index)
|
|
}
|
|
|
|
// Check for heartbeat interval
|
|
ttl = resp2.HeartbeatTTL
|
|
if ttl < s1.config.MinHeartbeatTTL || ttl > 2*s1.config.MinHeartbeatTTL {
|
|
t.Fatalf("bad: %#v", ttl)
|
|
}
|
|
}
|
|
|
|
func TestClientEndpoint_UpdateStatus_HeartbeatOnly_Advertise(t *testing.T) {
|
|
ci.Parallel(t)
|
|
require := require.New(t)
|
|
|
|
advAddr := "127.0.1.1:1234"
|
|
adv, err := net.ResolveTCPAddr("tcp", advAddr)
|
|
require.Nil(err)
|
|
|
|
s1, cleanupS1 := TestServer(t, func(c *Config) {
|
|
c.ClientRPCAdvertise = adv
|
|
})
|
|
defer cleanupS1()
|
|
codec := rpcClient(t, s1)
|
|
testutil.WaitForLeader(t, s1.RPC)
|
|
|
|
// Create the register request
|
|
node := mock.Node()
|
|
reg := &structs.NodeRegisterRequest{
|
|
Node: node,
|
|
WriteRequest: structs.WriteRequest{Region: "global"},
|
|
}
|
|
|
|
// Fetch the response
|
|
var resp structs.NodeUpdateResponse
|
|
if err := msgpackrpc.CallWithCodec(codec, "Node.Register", reg, &resp); err != nil {
|
|
t.Fatalf("err: %v", err)
|
|
}
|
|
|
|
// Check for heartbeat interval
|
|
ttl := resp.HeartbeatTTL
|
|
if ttl < s1.config.MinHeartbeatTTL || ttl > 2*s1.config.MinHeartbeatTTL {
|
|
t.Fatalf("bad: %#v", ttl)
|
|
}
|
|
|
|
// Check for heartbeat servers
|
|
require.Len(resp.Servers, 1)
|
|
require.Equal(resp.Servers[0].RPCAdvertiseAddr, advAddr)
|
|
}
|
|
|
|
func TestNode_UpdateStatus_ServiceRegistrations(t *testing.T) {
|
|
ci.Parallel(t)
|
|
|
|
testServer, serverCleanup := TestServer(t, nil)
|
|
defer serverCleanup()
|
|
testutil.WaitForLeader(t, testServer.RPC)
|
|
|
|
// Create a node and upsert this into state.
|
|
node := mock.Node()
|
|
require.NoError(t, testServer.State().UpsertNode(structs.MsgTypeTestSetup, 10, node))
|
|
|
|
// Generate service registrations, ensuring the nodeID is set to the
|
|
// generated node from above.
|
|
services := mock.ServiceRegistrations()
|
|
|
|
for _, s := range services {
|
|
s.NodeID = node.ID
|
|
}
|
|
|
|
// Upsert the service registrations into state.
|
|
require.NoError(t, testServer.State().UpsertServiceRegistrations(structs.MsgTypeTestSetup, 20, services))
|
|
|
|
// Check the service registrations are in state as we expect, so we can
|
|
// have confidence in the rest of the test.
|
|
ws := memdb.NewWatchSet()
|
|
nodeRegs, err := testServer.State().GetServiceRegistrationsByNodeID(ws, node.ID)
|
|
require.NoError(t, err)
|
|
require.Len(t, nodeRegs, 2)
|
|
require.Equal(t, nodeRegs[0].NodeID, node.ID)
|
|
require.Equal(t, nodeRegs[1].NodeID, node.ID)
|
|
|
|
// Generate and trigger a node down status update. This mimics what happens
|
|
// when the node fails its heart-beating.
|
|
args := structs.NodeUpdateStatusRequest{
|
|
NodeID: node.ID,
|
|
Status: structs.NodeStatusDown,
|
|
WriteRequest: structs.WriteRequest{Region: "global"},
|
|
}
|
|
|
|
var reply structs.NodeUpdateResponse
|
|
require.NoError(t, testServer.staticEndpoints.Node.UpdateStatus(&args, &reply))
|
|
|
|
// Query our state, to ensure the node service registrations have been
|
|
// removed.
|
|
nodeRegs, err = testServer.State().GetServiceRegistrationsByNodeID(ws, node.ID)
|
|
require.NoError(t, err)
|
|
require.Len(t, nodeRegs, 0)
|
|
}
|
|
|
|
// TestClientEndpoint_UpdateDrain asserts the ability to initiate drain
|
|
// against a node and cancel that drain. It also asserts:
|
|
// * an evaluation is created when the node becomes eligible
|
|
// * drain metadata is properly persisted in Node.LastDrain
|
|
func TestClientEndpoint_UpdateDrain(t *testing.T) {
|
|
ci.Parallel(t)
|
|
require := require.New(t)
|
|
|
|
s1, cleanupS1 := TestServer(t, nil)
|
|
defer cleanupS1()
|
|
codec := rpcClient(t, s1)
|
|
testutil.WaitForLeader(t, s1.RPC)
|
|
|
|
// Disable drainer to prevent drain from completing during test
|
|
s1.nodeDrainer.SetEnabled(false, nil)
|
|
|
|
// Create the register request
|
|
node := mock.Node()
|
|
reg := &structs.NodeRegisterRequest{
|
|
Node: node,
|
|
WriteRequest: structs.WriteRequest{Region: "global"},
|
|
}
|
|
|
|
// Fetch the response
|
|
var resp structs.NodeUpdateResponse
|
|
require.Nil(msgpackrpc.CallWithCodec(codec, "Node.Register", reg, &resp))
|
|
|
|
beforeUpdate := time.Now()
|
|
strategy := &structs.DrainStrategy{
|
|
DrainSpec: structs.DrainSpec{
|
|
Deadline: 10 * time.Second,
|
|
},
|
|
}
|
|
|
|
// Update the status
|
|
dereg := &structs.NodeUpdateDrainRequest{
|
|
NodeID: node.ID,
|
|
DrainStrategy: strategy,
|
|
Meta: map[string]string{"message": "this node is not needed"},
|
|
WriteRequest: structs.WriteRequest{Region: "global"},
|
|
}
|
|
var resp2 structs.NodeDrainUpdateResponse
|
|
require.Nil(msgpackrpc.CallWithCodec(codec, "Node.UpdateDrain", dereg, &resp2))
|
|
require.NotZero(resp2.Index)
|
|
|
|
// Check for the node in the FSM
|
|
state := s1.fsm.State()
|
|
ws := memdb.NewWatchSet()
|
|
out, err := state.NodeByID(ws, node.ID)
|
|
require.Nil(err)
|
|
require.NotNil(out.DrainStrategy)
|
|
require.Equal(strategy.Deadline, out.DrainStrategy.Deadline)
|
|
require.Len(out.Events, 2)
|
|
require.Equal(NodeDrainEventDrainSet, out.Events[1].Message)
|
|
require.NotNil(out.LastDrain)
|
|
require.Equal(structs.DrainMetadata{
|
|
StartedAt: out.LastDrain.UpdatedAt,
|
|
UpdatedAt: out.LastDrain.StartedAt,
|
|
Status: structs.DrainStatusDraining,
|
|
Meta: map[string]string{"message": "this node is not needed"},
|
|
}, *out.LastDrain)
|
|
|
|
// before+deadline should be before the forced deadline
|
|
require.True(beforeUpdate.Add(strategy.Deadline).Before(out.DrainStrategy.ForceDeadline))
|
|
|
|
// now+deadline should be after the forced deadline
|
|
require.True(time.Now().Add(strategy.Deadline).After(out.DrainStrategy.ForceDeadline))
|
|
|
|
drainStartedAt := out.DrainStrategy.StartedAt
|
|
// StartedAt should be close to the time the drain started
|
|
require.WithinDuration(beforeUpdate, drainStartedAt, 1*time.Second)
|
|
|
|
// StartedAt shouldn't change if a new request comes while still draining
|
|
require.Nil(msgpackrpc.CallWithCodec(codec, "Node.UpdateDrain", dereg, &resp2))
|
|
ws = memdb.NewWatchSet()
|
|
out, err = state.NodeByID(ws, node.ID)
|
|
require.NoError(err)
|
|
require.True(out.DrainStrategy.StartedAt.Equal(drainStartedAt))
|
|
|
|
// Register a system job
|
|
job := mock.SystemJob()
|
|
require.Nil(s1.State().UpsertJob(structs.MsgTypeTestSetup, 10, job))
|
|
|
|
// Update the eligibility and expect evals
|
|
dereg.DrainStrategy = nil
|
|
dereg.MarkEligible = true
|
|
dereg.Meta = map[string]string{"cancelled": "yes"}
|
|
var resp3 structs.NodeDrainUpdateResponse
|
|
require.Nil(msgpackrpc.CallWithCodec(codec, "Node.UpdateDrain", dereg, &resp3))
|
|
require.NotZero(resp3.Index)
|
|
require.NotZero(resp3.EvalCreateIndex)
|
|
require.Len(resp3.EvalIDs, 1)
|
|
|
|
// Check for updated node in the FSM
|
|
ws = memdb.NewWatchSet()
|
|
out, err = state.NodeByID(ws, node.ID)
|
|
require.NoError(err)
|
|
require.Len(out.Events, 4)
|
|
require.Equal(NodeDrainEventDrainDisabled, out.Events[3].Message)
|
|
require.NotNil(out.LastDrain)
|
|
require.NotNil(out.LastDrain)
|
|
require.False(out.LastDrain.UpdatedAt.Before(out.LastDrain.StartedAt))
|
|
require.Equal(structs.DrainMetadata{
|
|
StartedAt: out.LastDrain.StartedAt,
|
|
UpdatedAt: out.LastDrain.UpdatedAt,
|
|
Status: structs.DrainStatusCanceled,
|
|
Meta: map[string]string{"cancelled": "yes"},
|
|
}, *out.LastDrain)
|
|
|
|
// Check that calling UpdateDrain with the same DrainStrategy does not emit
|
|
// a node event.
|
|
require.Nil(msgpackrpc.CallWithCodec(codec, "Node.UpdateDrain", dereg, &resp3))
|
|
ws = memdb.NewWatchSet()
|
|
out, err = state.NodeByID(ws, node.ID)
|
|
require.NoError(err)
|
|
require.Len(out.Events, 4)
|
|
}
|
|
|
|
// TestClientEndpoint_UpdatedDrainAndCompleted asserts that drain metadata
|
|
// is properly persisted in Node.LastDrain as the node drain is updated and
|
|
// completes.
|
|
func TestClientEndpoint_UpdatedDrainAndCompleted(t *testing.T) {
|
|
ci.Parallel(t)
|
|
require := require.New(t)
|
|
|
|
s1, cleanupS1 := TestServer(t, nil)
|
|
defer cleanupS1()
|
|
codec := rpcClient(t, s1)
|
|
testutil.WaitForLeader(t, s1.RPC)
|
|
state := s1.fsm.State()
|
|
|
|
// Disable drainer for now
|
|
s1.nodeDrainer.SetEnabled(false, nil)
|
|
|
|
// Create the register request
|
|
node := mock.Node()
|
|
reg := &structs.NodeRegisterRequest{
|
|
Node: node,
|
|
WriteRequest: structs.WriteRequest{Region: "global"},
|
|
}
|
|
|
|
// Fetch the response
|
|
var resp structs.NodeUpdateResponse
|
|
require.Nil(msgpackrpc.CallWithCodec(codec, "Node.Register", reg, &resp))
|
|
|
|
strategy := &structs.DrainStrategy{
|
|
DrainSpec: structs.DrainSpec{
|
|
Deadline: 10 * time.Second,
|
|
},
|
|
}
|
|
|
|
// Update the status
|
|
dereg := &structs.NodeUpdateDrainRequest{
|
|
NodeID: node.ID,
|
|
DrainStrategy: strategy,
|
|
Meta: map[string]string{
|
|
"message": "first drain",
|
|
},
|
|
WriteRequest: structs.WriteRequest{Region: "global"},
|
|
}
|
|
var resp2 structs.NodeDrainUpdateResponse
|
|
require.Nil(msgpackrpc.CallWithCodec(codec, "Node.UpdateDrain", dereg, &resp2))
|
|
require.NotZero(resp2.Index)
|
|
|
|
// Check for the node in the FSM
|
|
out, err := state.NodeByID(nil, node.ID)
|
|
require.Nil(err)
|
|
require.NotNil(out.DrainStrategy)
|
|
require.NotNil(out.LastDrain)
|
|
firstDrainUpdate := out.LastDrain.UpdatedAt
|
|
require.Equal(structs.DrainMetadata{
|
|
StartedAt: firstDrainUpdate,
|
|
UpdatedAt: firstDrainUpdate,
|
|
Status: structs.DrainStatusDraining,
|
|
Meta: map[string]string{"message": "first drain"},
|
|
}, *out.LastDrain)
|
|
|
|
time.Sleep(1 * time.Second)
|
|
|
|
// Update the drain
|
|
dereg.DrainStrategy.DrainSpec.Deadline *= 2
|
|
dereg.Meta["message"] = "second drain"
|
|
require.Nil(msgpackrpc.CallWithCodec(codec, "Node.UpdateDrain", dereg, &resp2))
|
|
require.NotZero(resp2.Index)
|
|
|
|
out, err = state.NodeByID(nil, node.ID)
|
|
require.Nil(err)
|
|
require.NotNil(out.DrainStrategy)
|
|
require.NotNil(out.LastDrain)
|
|
secondDrainUpdate := out.LastDrain.UpdatedAt
|
|
require.True(secondDrainUpdate.After(firstDrainUpdate))
|
|
require.Equal(structs.DrainMetadata{
|
|
StartedAt: firstDrainUpdate,
|
|
UpdatedAt: secondDrainUpdate,
|
|
Status: structs.DrainStatusDraining,
|
|
Meta: map[string]string{"message": "second drain"},
|
|
}, *out.LastDrain)
|
|
|
|
time.Sleep(1 * time.Second)
|
|
|
|
// Enable the drainer, wait for completion
|
|
s1.nodeDrainer.SetEnabled(true, state)
|
|
|
|
testutil.WaitForResult(func() (bool, error) {
|
|
out, err = state.NodeByID(nil, node.ID)
|
|
if err != nil {
|
|
return false, err
|
|
}
|
|
if out == nil {
|
|
return false, fmt.Errorf("could not find node")
|
|
}
|
|
return out.DrainStrategy == nil, nil
|
|
}, func(err error) {
|
|
t.Fatalf("err: %v", err)
|
|
})
|
|
|
|
require.True(out.LastDrain.UpdatedAt.After(secondDrainUpdate))
|
|
require.Equal(structs.DrainMetadata{
|
|
StartedAt: firstDrainUpdate,
|
|
UpdatedAt: out.LastDrain.UpdatedAt,
|
|
Status: structs.DrainStatusComplete,
|
|
Meta: map[string]string{"message": "second drain"},
|
|
}, *out.LastDrain)
|
|
}
|
|
|
|
// TestClientEndpoint_UpdatedDrainNoop asserts that drain metadata is properly
|
|
// persisted in Node.LastDrain when calls to Node.UpdateDrain() don't affect
|
|
// the drain status.
|
|
func TestClientEndpoint_UpdatedDrainNoop(t *testing.T) {
|
|
ci.Parallel(t)
|
|
require := require.New(t)
|
|
|
|
s1, cleanupS1 := TestServer(t, nil)
|
|
defer cleanupS1()
|
|
codec := rpcClient(t, s1)
|
|
testutil.WaitForLeader(t, s1.RPC)
|
|
state := s1.fsm.State()
|
|
|
|
// Create the register request
|
|
node := mock.Node()
|
|
reg := &structs.NodeRegisterRequest{
|
|
Node: node,
|
|
WriteRequest: structs.WriteRequest{Region: "global"},
|
|
}
|
|
|
|
// Fetch the response
|
|
var resp structs.NodeUpdateResponse
|
|
require.Nil(msgpackrpc.CallWithCodec(codec, "Node.Register", reg, &resp))
|
|
|
|
// Update the status
|
|
dereg := &structs.NodeUpdateDrainRequest{
|
|
NodeID: node.ID,
|
|
DrainStrategy: &structs.DrainStrategy{
|
|
DrainSpec: structs.DrainSpec{
|
|
Deadline: 10 * time.Second,
|
|
},
|
|
},
|
|
Meta: map[string]string{
|
|
"message": "drain",
|
|
},
|
|
WriteRequest: structs.WriteRequest{Region: "global"},
|
|
}
|
|
var drainResp structs.NodeDrainUpdateResponse
|
|
require.Nil(msgpackrpc.CallWithCodec(codec, "Node.UpdateDrain", dereg, &drainResp))
|
|
require.NotZero(drainResp.Index)
|
|
|
|
var out *structs.Node
|
|
testutil.WaitForResult(func() (bool, error) {
|
|
var err error
|
|
out, err = state.NodeByID(nil, node.ID)
|
|
if err != nil {
|
|
return false, err
|
|
}
|
|
if out == nil {
|
|
return false, fmt.Errorf("could not find node")
|
|
}
|
|
return out.DrainStrategy == nil && out.SchedulingEligibility == structs.NodeSchedulingIneligible, nil
|
|
}, func(err error) {
|
|
t.Fatalf("err: %v", err)
|
|
})
|
|
|
|
require.Equal(structs.DrainStatusComplete, out.LastDrain.Status)
|
|
require.Equal(map[string]string{"message": "drain"}, out.LastDrain.Meta)
|
|
prevDrain := out.LastDrain
|
|
|
|
// call again with Drain Strategy nil; should be a no-op because drain is already complete
|
|
dereg.DrainStrategy = nil
|
|
dereg.Meta = map[string]string{
|
|
"new_message": "is new",
|
|
}
|
|
require.Nil(msgpackrpc.CallWithCodec(codec, "Node.UpdateDrain", dereg, &drainResp))
|
|
require.NotZero(drainResp.Index)
|
|
|
|
out, err := state.NodeByID(nil, node.ID)
|
|
require.Nil(err)
|
|
require.Nil(out.DrainStrategy)
|
|
require.NotNil(out.LastDrain)
|
|
require.Equal(prevDrain, out.LastDrain)
|
|
}
|
|
|
|
// TestClientEndpoint_UpdateDrain_ACL asserts that Node.UpdateDrain() enforces
|
|
// node.write ACLs, and that token accessor ID is properly persisted in
|
|
// Node.LastDrain.AccessorID
|
|
func TestClientEndpoint_UpdateDrain_ACL(t *testing.T) {
|
|
ci.Parallel(t)
|
|
|
|
s1, root, cleanupS1 := TestACLServer(t, nil)
|
|
defer cleanupS1()
|
|
codec := rpcClient(t, s1)
|
|
testutil.WaitForLeader(t, s1.RPC)
|
|
require := require.New(t)
|
|
|
|
// Create the node
|
|
node := mock.Node()
|
|
state := s1.fsm.State()
|
|
|
|
require.Nil(state.UpsertNode(structs.MsgTypeTestSetup, 1, node), "UpsertNode")
|
|
|
|
// Create the policy and tokens
|
|
validToken := mock.CreatePolicyAndToken(t, state, 1001, "test-valid", mock.NodePolicy(acl.PolicyWrite))
|
|
invalidToken := mock.CreatePolicyAndToken(t, state, 1003, "test-invalid", mock.NodePolicy(acl.PolicyRead))
|
|
|
|
// Update the status without a token and expect failure
|
|
dereg := &structs.NodeUpdateDrainRequest{
|
|
NodeID: node.ID,
|
|
DrainStrategy: &structs.DrainStrategy{
|
|
DrainSpec: structs.DrainSpec{
|
|
Deadline: 10 * time.Second,
|
|
},
|
|
},
|
|
WriteRequest: structs.WriteRequest{Region: "global"},
|
|
}
|
|
{
|
|
var resp structs.NodeDrainUpdateResponse
|
|
err := msgpackrpc.CallWithCodec(codec, "Node.UpdateDrain", dereg, &resp)
|
|
require.NotNil(err, "RPC")
|
|
require.Equal(err.Error(), structs.ErrPermissionDenied.Error())
|
|
}
|
|
|
|
// Try with a valid token
|
|
dereg.AuthToken = validToken.SecretID
|
|
{
|
|
var resp structs.NodeDrainUpdateResponse
|
|
require.Nil(msgpackrpc.CallWithCodec(codec, "Node.UpdateDrain", dereg, &resp), "RPC")
|
|
out, err := state.NodeByID(nil, node.ID)
|
|
require.NoError(err)
|
|
require.Equal(validToken.AccessorID, out.LastDrain.AccessorID)
|
|
}
|
|
|
|
// Try with a invalid token
|
|
dereg.AuthToken = invalidToken.SecretID
|
|
{
|
|
var resp structs.NodeDrainUpdateResponse
|
|
err := msgpackrpc.CallWithCodec(codec, "Node.UpdateDrain", dereg, &resp)
|
|
require.NotNil(err, "RPC")
|
|
require.Equal(err.Error(), structs.ErrPermissionDenied.Error())
|
|
}
|
|
|
|
// Try with a root token
|
|
dereg.DrainStrategy.DrainSpec.Deadline = 20 * time.Second
|
|
dereg.AuthToken = root.SecretID
|
|
{
|
|
var resp structs.NodeDrainUpdateResponse
|
|
require.Nil(msgpackrpc.CallWithCodec(codec, "Node.UpdateDrain", dereg, &resp), "RPC")
|
|
out, err := state.NodeByID(nil, node.ID)
|
|
require.NoError(err)
|
|
require.Equal(root.AccessorID, out.LastDrain.AccessorID)
|
|
}
|
|
}
|
|
|
|
// This test ensures that Nomad marks client state of allocations which are in
|
|
// pending/running state to lost when a node is marked as down.
|
|
func TestClientEndpoint_Drain_Down(t *testing.T) {
|
|
ci.Parallel(t)
|
|
|
|
s1, cleanupS1 := TestServer(t, nil)
|
|
defer cleanupS1()
|
|
codec := rpcClient(t, s1)
|
|
testutil.WaitForLeader(t, s1.RPC)
|
|
require := require.New(t)
|
|
|
|
// Register a node
|
|
node := mock.Node()
|
|
reg := &structs.NodeRegisterRequest{
|
|
Node: node,
|
|
WriteRequest: structs.WriteRequest{Region: "global"},
|
|
}
|
|
// Fetch the response
|
|
var resp structs.NodeUpdateResponse
|
|
require.Nil(msgpackrpc.CallWithCodec(codec, "Node.Register", reg, &resp))
|
|
|
|
// Register a service job
|
|
var jobResp structs.JobRegisterResponse
|
|
job := mock.Job()
|
|
job.TaskGroups[0].Count = 1
|
|
jobReq := &structs.JobRegisterRequest{
|
|
Job: job,
|
|
WriteRequest: structs.WriteRequest{
|
|
Region: "global",
|
|
Namespace: job.Namespace,
|
|
},
|
|
}
|
|
require.Nil(msgpackrpc.CallWithCodec(codec, "Job.Register", jobReq, &jobResp))
|
|
|
|
// Register a system job
|
|
var jobResp1 structs.JobRegisterResponse
|
|
job1 := mock.SystemJob()
|
|
job1.TaskGroups[0].Count = 1
|
|
jobReq1 := &structs.JobRegisterRequest{
|
|
Job: job1,
|
|
WriteRequest: structs.WriteRequest{
|
|
Region: "global",
|
|
Namespace: job1.Namespace,
|
|
},
|
|
}
|
|
require.Nil(msgpackrpc.CallWithCodec(codec, "Job.Register", jobReq1, &jobResp1))
|
|
|
|
// Wait for the scheduler to create an allocation
|
|
testutil.WaitForResult(func() (bool, error) {
|
|
ws := memdb.NewWatchSet()
|
|
allocs, err := s1.fsm.state.AllocsByJob(ws, job.Namespace, job.ID, true)
|
|
if err != nil {
|
|
return false, err
|
|
}
|
|
allocs1, err := s1.fsm.state.AllocsByJob(ws, job1.Namespace, job1.ID, true)
|
|
if err != nil {
|
|
return false, err
|
|
}
|
|
return len(allocs) > 0 && len(allocs1) > 0, nil
|
|
}, func(err error) {
|
|
t.Fatalf("err: %v", err)
|
|
})
|
|
|
|
// Drain the node
|
|
dereg := &structs.NodeUpdateDrainRequest{
|
|
NodeID: node.ID,
|
|
DrainStrategy: &structs.DrainStrategy{
|
|
DrainSpec: structs.DrainSpec{
|
|
Deadline: -1 * time.Second,
|
|
},
|
|
},
|
|
WriteRequest: structs.WriteRequest{Region: "global"},
|
|
}
|
|
var resp2 structs.NodeDrainUpdateResponse
|
|
require.Nil(msgpackrpc.CallWithCodec(codec, "Node.UpdateDrain", dereg, &resp2))
|
|
|
|
// Mark the node as down
|
|
node.Status = structs.NodeStatusDown
|
|
reg = &structs.NodeRegisterRequest{
|
|
Node: node,
|
|
WriteRequest: structs.WriteRequest{Region: "global"},
|
|
}
|
|
require.Nil(msgpackrpc.CallWithCodec(codec, "Node.Register", reg, &resp))
|
|
|
|
// Ensure that the allocation has transitioned to lost
|
|
testutil.WaitForResult(func() (bool, error) {
|
|
ws := memdb.NewWatchSet()
|
|
summary, err := s1.fsm.state.JobSummaryByID(ws, job.Namespace, job.ID)
|
|
if err != nil {
|
|
return false, err
|
|
}
|
|
expectedSummary := &structs.JobSummary{
|
|
JobID: job.ID,
|
|
Namespace: job.Namespace,
|
|
Summary: map[string]structs.TaskGroupSummary{
|
|
"web": {
|
|
Queued: 1,
|
|
Lost: 1,
|
|
},
|
|
},
|
|
Children: new(structs.JobChildrenSummary),
|
|
CreateIndex: jobResp.JobModifyIndex,
|
|
ModifyIndex: summary.ModifyIndex,
|
|
}
|
|
if !reflect.DeepEqual(summary, expectedSummary) {
|
|
return false, fmt.Errorf("Service: expected: %#v, actual: %#v", expectedSummary, summary)
|
|
}
|
|
|
|
summary1, err := s1.fsm.state.JobSummaryByID(ws, job1.Namespace, job1.ID)
|
|
if err != nil {
|
|
return false, err
|
|
}
|
|
expectedSummary1 := &structs.JobSummary{
|
|
JobID: job1.ID,
|
|
Namespace: job1.Namespace,
|
|
Summary: map[string]structs.TaskGroupSummary{
|
|
"web": {
|
|
Lost: 1,
|
|
},
|
|
},
|
|
Children: new(structs.JobChildrenSummary),
|
|
CreateIndex: jobResp1.JobModifyIndex,
|
|
ModifyIndex: summary1.ModifyIndex,
|
|
}
|
|
if !reflect.DeepEqual(summary1, expectedSummary1) {
|
|
return false, fmt.Errorf("System: expected: %#v, actual: %#v", expectedSummary1, summary1)
|
|
}
|
|
return true, nil
|
|
}, func(err error) {
|
|
t.Fatalf("err: %v", err)
|
|
})
|
|
}
|
|
|
|
func TestClientEndpoint_UpdateEligibility(t *testing.T) {
|
|
ci.Parallel(t)
|
|
require := require.New(t)
|
|
|
|
s1, cleanupS1 := TestServer(t, nil)
|
|
defer cleanupS1()
|
|
codec := rpcClient(t, s1)
|
|
testutil.WaitForLeader(t, s1.RPC)
|
|
|
|
// Create the register request
|
|
node := mock.Node()
|
|
reg := &structs.NodeRegisterRequest{
|
|
Node: node,
|
|
WriteRequest: structs.WriteRequest{Region: "global"},
|
|
}
|
|
|
|
// Fetch the response
|
|
var resp structs.NodeUpdateResponse
|
|
require.Nil(msgpackrpc.CallWithCodec(codec, "Node.Register", reg, &resp))
|
|
|
|
// Update the eligibility
|
|
elig := &structs.NodeUpdateEligibilityRequest{
|
|
NodeID: node.ID,
|
|
Eligibility: structs.NodeSchedulingIneligible,
|
|
WriteRequest: structs.WriteRequest{Region: "global"},
|
|
}
|
|
var resp2 structs.NodeEligibilityUpdateResponse
|
|
require.Nil(msgpackrpc.CallWithCodec(codec, "Node.UpdateEligibility", elig, &resp2))
|
|
require.NotZero(resp2.Index)
|
|
require.Zero(resp2.EvalCreateIndex)
|
|
require.Empty(resp2.EvalIDs)
|
|
|
|
// Check for the node in the FSM
|
|
state := s1.fsm.State()
|
|
out, err := state.NodeByID(nil, node.ID)
|
|
require.Nil(err)
|
|
require.Equal(out.SchedulingEligibility, structs.NodeSchedulingIneligible)
|
|
require.Len(out.Events, 2)
|
|
require.Equal(NodeEligibilityEventIneligible, out.Events[1].Message)
|
|
|
|
// Register a system job
|
|
job := mock.SystemJob()
|
|
require.Nil(s1.State().UpsertJob(structs.MsgTypeTestSetup, 10, job))
|
|
|
|
// Update the eligibility and expect evals
|
|
elig.Eligibility = structs.NodeSchedulingEligible
|
|
var resp3 structs.NodeEligibilityUpdateResponse
|
|
require.Nil(msgpackrpc.CallWithCodec(codec, "Node.UpdateEligibility", elig, &resp3))
|
|
require.NotZero(resp3.Index)
|
|
require.NotZero(resp3.EvalCreateIndex)
|
|
require.Len(resp3.EvalIDs, 1)
|
|
|
|
out, err = state.NodeByID(nil, node.ID)
|
|
require.Nil(err)
|
|
require.Len(out.Events, 3)
|
|
require.Equal(NodeEligibilityEventEligible, out.Events[2].Message)
|
|
}
|
|
|
|
func TestClientEndpoint_UpdateEligibility_ACL(t *testing.T) {
|
|
ci.Parallel(t)
|
|
|
|
s1, root, cleanupS1 := TestACLServer(t, nil)
|
|
defer cleanupS1()
|
|
codec := rpcClient(t, s1)
|
|
testutil.WaitForLeader(t, s1.RPC)
|
|
require := require.New(t)
|
|
|
|
// Create the node
|
|
node := mock.Node()
|
|
state := s1.fsm.State()
|
|
|
|
require.Nil(state.UpsertNode(structs.MsgTypeTestSetup, 1, node), "UpsertNode")
|
|
|
|
// Create the policy and tokens
|
|
validToken := mock.CreatePolicyAndToken(t, state, 1001, "test-valid", mock.NodePolicy(acl.PolicyWrite))
|
|
invalidToken := mock.CreatePolicyAndToken(t, state, 1003, "test-invalid", mock.NodePolicy(acl.PolicyRead))
|
|
|
|
// Update the status without a token and expect failure
|
|
dereg := &structs.NodeUpdateEligibilityRequest{
|
|
NodeID: node.ID,
|
|
Eligibility: structs.NodeSchedulingIneligible,
|
|
WriteRequest: structs.WriteRequest{Region: "global"},
|
|
}
|
|
{
|
|
var resp structs.NodeEligibilityUpdateResponse
|
|
err := msgpackrpc.CallWithCodec(codec, "Node.UpdateEligibility", dereg, &resp)
|
|
require.NotNil(err, "RPC")
|
|
require.Equal(err.Error(), structs.ErrPermissionDenied.Error())
|
|
}
|
|
|
|
// Try with a valid token
|
|
dereg.AuthToken = validToken.SecretID
|
|
{
|
|
var resp structs.NodeEligibilityUpdateResponse
|
|
require.Nil(msgpackrpc.CallWithCodec(codec, "Node.UpdateEligibility", dereg, &resp), "RPC")
|
|
}
|
|
|
|
// Try with a invalid token
|
|
dereg.AuthToken = invalidToken.SecretID
|
|
{
|
|
var resp structs.NodeEligibilityUpdateResponse
|
|
err := msgpackrpc.CallWithCodec(codec, "Node.UpdateEligibility", dereg, &resp)
|
|
require.NotNil(err, "RPC")
|
|
require.Equal(err.Error(), structs.ErrPermissionDenied.Error())
|
|
}
|
|
|
|
// Try with a root token
|
|
dereg.AuthToken = root.SecretID
|
|
{
|
|
var resp structs.NodeEligibilityUpdateResponse
|
|
require.Nil(msgpackrpc.CallWithCodec(codec, "Node.UpdateEligibility", dereg, &resp), "RPC")
|
|
}
|
|
}
|
|
|
|
func TestClientEndpoint_GetNode(t *testing.T) {
|
|
ci.Parallel(t)
|
|
|
|
s1, cleanupS1 := TestServer(t, nil)
|
|
defer cleanupS1()
|
|
codec := rpcClient(t, s1)
|
|
testutil.WaitForLeader(t, s1.RPC)
|
|
|
|
// Create the register request
|
|
node := mock.Node()
|
|
reg := &structs.NodeRegisterRequest{
|
|
Node: node,
|
|
WriteRequest: structs.WriteRequest{Region: "global"},
|
|
}
|
|
|
|
// Fetch the response
|
|
var resp structs.GenericResponse
|
|
if err := msgpackrpc.CallWithCodec(codec, "Node.Register", reg, &resp); err != nil {
|
|
t.Fatalf("err: %v", err)
|
|
}
|
|
node.CreateIndex = resp.Index
|
|
node.ModifyIndex = resp.Index
|
|
|
|
// Lookup the node
|
|
get := &structs.NodeSpecificRequest{
|
|
NodeID: node.ID,
|
|
QueryOptions: structs.QueryOptions{Region: "global"},
|
|
}
|
|
var resp2 structs.SingleNodeResponse
|
|
if err := msgpackrpc.CallWithCodec(codec, "Node.GetNode", get, &resp2); err != nil {
|
|
t.Fatalf("err: %v", err)
|
|
}
|
|
if resp2.Index != resp.Index {
|
|
t.Fatalf("Bad index: %d %d", resp2.Index, resp.Index)
|
|
}
|
|
|
|
if resp2.Node.ComputedClass == "" {
|
|
t.Fatalf("bad ComputedClass: %#v", resp2.Node)
|
|
}
|
|
|
|
// Update the status updated at value
|
|
node.StatusUpdatedAt = resp2.Node.StatusUpdatedAt
|
|
node.SecretID = ""
|
|
node.Events = resp2.Node.Events
|
|
require.Equal(t, node, resp2.Node)
|
|
|
|
// assert that the node register event was set correctly
|
|
if len(resp2.Node.Events) != 1 {
|
|
t.Fatalf("Did not set node events: %#v", resp2.Node)
|
|
}
|
|
if resp2.Node.Events[0].Message != state.NodeRegisterEventRegistered {
|
|
t.Fatalf("Did not set node register event correctly: %#v", resp2.Node)
|
|
}
|
|
|
|
// Lookup non-existing node
|
|
get.NodeID = "12345678-abcd-efab-cdef-123456789abc"
|
|
if err := msgpackrpc.CallWithCodec(codec, "Node.GetNode", get, &resp2); err != nil {
|
|
t.Fatalf("err: %v", err)
|
|
}
|
|
if resp2.Index != resp.Index {
|
|
t.Fatalf("Bad index: %d %d", resp2.Index, resp.Index)
|
|
}
|
|
if resp2.Node != nil {
|
|
t.Fatalf("unexpected node")
|
|
}
|
|
}
|
|
|
|
func TestClientEndpoint_GetNode_ACL(t *testing.T) {
|
|
ci.Parallel(t)
|
|
|
|
s1, root, cleanupS1 := TestACLServer(t, nil)
|
|
defer cleanupS1()
|
|
codec := rpcClient(t, s1)
|
|
testutil.WaitForLeader(t, s1.RPC)
|
|
assert := assert.New(t)
|
|
|
|
// Create the node
|
|
node := mock.Node()
|
|
state := s1.fsm.State()
|
|
assert.Nil(state.UpsertNode(structs.MsgTypeTestSetup, 1, node), "UpsertNode")
|
|
|
|
// Create the policy and tokens
|
|
validToken := mock.CreatePolicyAndToken(t, state, 1001, "test-valid", mock.NodePolicy(acl.PolicyRead))
|
|
invalidToken := mock.CreatePolicyAndToken(t, state, 1003, "test-invalid", mock.NodePolicy(acl.PolicyDeny))
|
|
|
|
// Lookup the node without a token and expect failure
|
|
req := &structs.NodeSpecificRequest{
|
|
NodeID: node.ID,
|
|
QueryOptions: structs.QueryOptions{Region: "global"},
|
|
}
|
|
{
|
|
var resp structs.SingleNodeResponse
|
|
err := msgpackrpc.CallWithCodec(codec, "Node.GetNode", req, &resp)
|
|
assert.NotNil(err, "RPC")
|
|
assert.Equal(err.Error(), structs.ErrPermissionDenied.Error())
|
|
}
|
|
|
|
// Try with a valid token
|
|
req.AuthToken = validToken.SecretID
|
|
{
|
|
var resp structs.SingleNodeResponse
|
|
assert.Nil(msgpackrpc.CallWithCodec(codec, "Node.GetNode", req, &resp), "RPC")
|
|
assert.Equal(node.ID, resp.Node.ID)
|
|
}
|
|
|
|
// Try with a Node.SecretID
|
|
req.AuthToken = node.SecretID
|
|
{
|
|
var resp structs.SingleNodeResponse
|
|
assert.Nil(msgpackrpc.CallWithCodec(codec, "Node.GetNode", req, &resp), "RPC")
|
|
assert.Equal(node.ID, resp.Node.ID)
|
|
}
|
|
|
|
// Try with a invalid token
|
|
req.AuthToken = invalidToken.SecretID
|
|
{
|
|
var resp structs.SingleNodeResponse
|
|
err := msgpackrpc.CallWithCodec(codec, "Node.GetNode", req, &resp)
|
|
assert.NotNil(err, "RPC")
|
|
assert.Equal(err.Error(), structs.ErrPermissionDenied.Error())
|
|
}
|
|
|
|
// Try with a root token
|
|
req.AuthToken = root.SecretID
|
|
{
|
|
var resp structs.SingleNodeResponse
|
|
assert.Nil(msgpackrpc.CallWithCodec(codec, "Node.GetNode", req, &resp), "RPC")
|
|
assert.Equal(node.ID, resp.Node.ID)
|
|
}
|
|
}
|
|
|
|
func TestClientEndpoint_GetNode_Blocking(t *testing.T) {
|
|
ci.Parallel(t)
|
|
|
|
s1, cleanupS1 := TestServer(t, nil)
|
|
defer cleanupS1()
|
|
state := s1.fsm.State()
|
|
codec := rpcClient(t, s1)
|
|
testutil.WaitForLeader(t, s1.RPC)
|
|
|
|
// Create the node
|
|
node1 := mock.Node()
|
|
node2 := mock.Node()
|
|
|
|
// First create an unrelated node.
|
|
time.AfterFunc(100*time.Millisecond, func() {
|
|
if err := state.UpsertNode(structs.MsgTypeTestSetup, 100, node1); err != nil {
|
|
t.Fatalf("err: %v", err)
|
|
}
|
|
})
|
|
|
|
// Upsert the node we are watching later
|
|
time.AfterFunc(200*time.Millisecond, func() {
|
|
if err := state.UpsertNode(structs.MsgTypeTestSetup, 200, node2); err != nil {
|
|
t.Fatalf("err: %v", err)
|
|
}
|
|
})
|
|
|
|
// Lookup the node
|
|
req := &structs.NodeSpecificRequest{
|
|
NodeID: node2.ID,
|
|
QueryOptions: structs.QueryOptions{
|
|
Region: "global",
|
|
MinQueryIndex: 150,
|
|
},
|
|
}
|
|
var resp structs.SingleNodeResponse
|
|
start := time.Now()
|
|
if err := msgpackrpc.CallWithCodec(codec, "Node.GetNode", req, &resp); err != nil {
|
|
t.Fatalf("err: %v", err)
|
|
}
|
|
|
|
if elapsed := time.Since(start); elapsed < 200*time.Millisecond {
|
|
t.Fatalf("should block (returned in %s) %#v", elapsed, resp)
|
|
}
|
|
if resp.Index != 200 {
|
|
t.Fatalf("Bad index: %d %d", resp.Index, 200)
|
|
}
|
|
if resp.Node == nil || resp.Node.ID != node2.ID {
|
|
t.Fatalf("bad: %#v", resp.Node)
|
|
}
|
|
|
|
// Node update triggers watches
|
|
time.AfterFunc(100*time.Millisecond, func() {
|
|
nodeUpdate := mock.Node()
|
|
nodeUpdate.ID = node2.ID
|
|
nodeUpdate.Status = structs.NodeStatusDown
|
|
if err := state.UpsertNode(structs.MsgTypeTestSetup, 300, nodeUpdate); err != nil {
|
|
t.Fatalf("err: %v", err)
|
|
}
|
|
})
|
|
|
|
req.QueryOptions.MinQueryIndex = 250
|
|
var resp2 structs.SingleNodeResponse
|
|
start = time.Now()
|
|
if err := msgpackrpc.CallWithCodec(codec, "Node.GetNode", req, &resp2); err != nil {
|
|
t.Fatalf("err: %v", err)
|
|
}
|
|
|
|
if elapsed := time.Since(start); elapsed < 100*time.Millisecond {
|
|
t.Fatalf("should block (returned in %s) %#v", elapsed, resp)
|
|
}
|
|
if resp2.Index != 300 {
|
|
t.Fatalf("Bad index: %d %d", resp2.Index, 300)
|
|
}
|
|
if resp2.Node == nil || resp2.Node.Status != structs.NodeStatusDown {
|
|
t.Fatalf("bad: %#v", resp2.Node)
|
|
}
|
|
|
|
// Node delete triggers watches
|
|
time.AfterFunc(100*time.Millisecond, func() {
|
|
if err := state.DeleteNode(structs.MsgTypeTestSetup, 400, []string{node2.ID}); err != nil {
|
|
t.Fatalf("err: %v", err)
|
|
}
|
|
})
|
|
|
|
req.QueryOptions.MinQueryIndex = 350
|
|
var resp3 structs.SingleNodeResponse
|
|
start = time.Now()
|
|
if err := msgpackrpc.CallWithCodec(codec, "Node.GetNode", req, &resp3); err != nil {
|
|
t.Fatalf("err: %v", err)
|
|
}
|
|
|
|
if elapsed := time.Since(start); elapsed < 100*time.Millisecond {
|
|
t.Fatalf("should block (returned in %s) %#v", elapsed, resp)
|
|
}
|
|
if resp3.Index != 400 {
|
|
t.Fatalf("Bad index: %d %d", resp2.Index, 400)
|
|
}
|
|
if resp3.Node != nil {
|
|
t.Fatalf("bad: %#v", resp3.Node)
|
|
}
|
|
}
|
|
|
|
func TestClientEndpoint_GetAllocs(t *testing.T) {
|
|
ci.Parallel(t)
|
|
|
|
s1, cleanupS1 := TestServer(t, nil)
|
|
defer cleanupS1()
|
|
codec := rpcClient(t, s1)
|
|
testutil.WaitForLeader(t, s1.RPC)
|
|
|
|
// Create the register request
|
|
node := mock.Node()
|
|
reg := &structs.NodeRegisterRequest{
|
|
Node: node,
|
|
WriteRequest: structs.WriteRequest{Region: "global"},
|
|
}
|
|
|
|
// Fetch the response
|
|
var resp structs.GenericResponse
|
|
if err := msgpackrpc.CallWithCodec(codec, "Node.Register", reg, &resp); err != nil {
|
|
t.Fatalf("err: %v", err)
|
|
}
|
|
node.CreateIndex = resp.Index
|
|
node.ModifyIndex = resp.Index
|
|
|
|
// Inject fake evaluations
|
|
alloc := mock.Alloc()
|
|
alloc.NodeID = node.ID
|
|
state := s1.fsm.State()
|
|
state.UpsertJobSummary(99, mock.JobSummary(alloc.JobID))
|
|
err := state.UpsertAllocs(structs.MsgTypeTestSetup, 100, []*structs.Allocation{alloc})
|
|
if err != nil {
|
|
t.Fatalf("err: %v", err)
|
|
}
|
|
|
|
// Lookup the allocs
|
|
get := &structs.NodeSpecificRequest{
|
|
NodeID: node.ID,
|
|
QueryOptions: structs.QueryOptions{Region: "global"},
|
|
}
|
|
var resp2 structs.NodeAllocsResponse
|
|
if err := msgpackrpc.CallWithCodec(codec, "Node.GetAllocs", get, &resp2); err != nil {
|
|
t.Fatalf("err: %v", err)
|
|
}
|
|
if resp2.Index != 100 {
|
|
t.Fatalf("Bad index: %d %d", resp2.Index, 100)
|
|
}
|
|
|
|
if len(resp2.Allocs) != 1 || resp2.Allocs[0].ID != alloc.ID {
|
|
t.Fatalf("bad: %#v", resp2.Allocs)
|
|
}
|
|
|
|
// Lookup non-existing node
|
|
get.NodeID = "foobarbaz"
|
|
if err := msgpackrpc.CallWithCodec(codec, "Node.GetAllocs", get, &resp2); err != nil {
|
|
t.Fatalf("err: %v", err)
|
|
}
|
|
if resp2.Index != 100 {
|
|
t.Fatalf("Bad index: %d %d", resp2.Index, 100)
|
|
}
|
|
if len(resp2.Allocs) != 0 {
|
|
t.Fatalf("unexpected node")
|
|
}
|
|
}
|
|
|
|
func TestClientEndpoint_GetAllocs_ACL_Basic(t *testing.T) {
|
|
ci.Parallel(t)
|
|
|
|
s1, root, cleanupS1 := TestACLServer(t, nil)
|
|
defer cleanupS1()
|
|
codec := rpcClient(t, s1)
|
|
testutil.WaitForLeader(t, s1.RPC)
|
|
assert := assert.New(t)
|
|
|
|
// Create the node
|
|
allocDefaultNS := mock.Alloc()
|
|
node := mock.Node()
|
|
allocDefaultNS.NodeID = node.ID
|
|
state := s1.fsm.State()
|
|
assert.Nil(state.UpsertNode(structs.MsgTypeTestSetup, 1, node), "UpsertNode")
|
|
assert.Nil(state.UpsertJobSummary(2, mock.JobSummary(allocDefaultNS.JobID)), "UpsertJobSummary")
|
|
allocs := []*structs.Allocation{allocDefaultNS}
|
|
assert.Nil(state.UpsertAllocs(structs.MsgTypeTestSetup, 5, allocs), "UpsertAllocs")
|
|
|
|
// Create the namespace policy and tokens
|
|
validDefaultToken := mock.CreatePolicyAndToken(t, state, 1001, "test-default-valid", mock.NodePolicy(acl.PolicyRead)+
|
|
mock.NamespacePolicy(structs.DefaultNamespace, "", []string{acl.NamespaceCapabilityReadJob}))
|
|
invalidToken := mock.CreatePolicyAndToken(t, state, 1004, "test-invalid",
|
|
mock.NamespacePolicy(structs.DefaultNamespace, "", []string{acl.NamespaceCapabilityReadJob}))
|
|
|
|
req := &structs.NodeSpecificRequest{
|
|
NodeID: node.ID,
|
|
QueryOptions: structs.QueryOptions{
|
|
Region: "global",
|
|
},
|
|
}
|
|
|
|
// Lookup the node without a token and expect failure
|
|
{
|
|
var resp structs.NodeAllocsResponse
|
|
err := msgpackrpc.CallWithCodec(codec, "Node.GetAllocs", req, &resp)
|
|
assert.NotNil(err, "RPC")
|
|
assert.Equal(err.Error(), structs.ErrPermissionDenied.Error())
|
|
}
|
|
|
|
// Try with a valid token for the default namespace
|
|
req.AuthToken = validDefaultToken.SecretID
|
|
{
|
|
var resp structs.NodeAllocsResponse
|
|
assert.Nil(msgpackrpc.CallWithCodec(codec, "Node.GetAllocs", req, &resp), "RPC")
|
|
assert.Len(resp.Allocs, 1)
|
|
assert.Equal(allocDefaultNS.ID, resp.Allocs[0].ID)
|
|
}
|
|
|
|
// Try with a invalid token
|
|
req.AuthToken = invalidToken.SecretID
|
|
{
|
|
var resp structs.NodeAllocsResponse
|
|
err := msgpackrpc.CallWithCodec(codec, "Node.GetAllocs", req, &resp)
|
|
assert.NotNil(err, "RPC")
|
|
assert.Equal(err.Error(), structs.ErrPermissionDenied.Error())
|
|
}
|
|
|
|
// Try with a root token
|
|
req.AuthToken = root.SecretID
|
|
{
|
|
var resp structs.NodeAllocsResponse
|
|
assert.Nil(msgpackrpc.CallWithCodec(codec, "Node.GetAllocs", req, &resp), "RPC")
|
|
assert.Len(resp.Allocs, 1)
|
|
for _, alloc := range resp.Allocs {
|
|
switch alloc.ID {
|
|
case allocDefaultNS.ID:
|
|
// expected
|
|
default:
|
|
t.Errorf("unexpected alloc %q for namespace %q", alloc.ID, alloc.Namespace)
|
|
}
|
|
}
|
|
}
|
|
}
|
|
|
|
func TestClientEndpoint_GetAllocs_ACL_Namespaces(t *testing.T) {
|
|
ci.Parallel(t)
|
|
s1, root, cleanupS1 := TestACLServer(t, nil)
|
|
defer cleanupS1()
|
|
codec := rpcClient(t, s1)
|
|
testutil.WaitForLeader(t, s1.RPC)
|
|
assert := assert.New(t)
|
|
|
|
// Create the namespaces
|
|
ns1 := mock.Namespace()
|
|
ns2 := mock.Namespace()
|
|
ns1.Name = "altnamespace"
|
|
ns2.Name = "should-only-be-displayed-for-root-ns"
|
|
|
|
// Create the allocs
|
|
allocDefaultNS := mock.Alloc()
|
|
allocAltNS := mock.Alloc()
|
|
allocAltNS.Namespace = ns1.Name
|
|
allocOtherNS := mock.Alloc()
|
|
allocOtherNS.Namespace = ns2.Name
|
|
|
|
node := mock.Node()
|
|
allocDefaultNS.NodeID = node.ID
|
|
allocAltNS.NodeID = node.ID
|
|
allocOtherNS.NodeID = node.ID
|
|
state := s1.fsm.State()
|
|
assert.Nil(state.UpsertNamespaces(1, []*structs.Namespace{ns1, ns2}), "UpsertNamespaces")
|
|
assert.Nil(state.UpsertNode(structs.MsgTypeTestSetup, 2, node), "UpsertNode")
|
|
assert.Nil(state.UpsertJobSummary(3, mock.JobSummary(allocDefaultNS.JobID)), "UpsertJobSummary")
|
|
assert.Nil(state.UpsertJobSummary(4, mock.JobSummary(allocAltNS.JobID)), "UpsertJobSummary")
|
|
assert.Nil(state.UpsertJobSummary(5, mock.JobSummary(allocOtherNS.JobID)), "UpsertJobSummary")
|
|
allocs := []*structs.Allocation{allocDefaultNS, allocAltNS, allocOtherNS}
|
|
assert.Nil(state.UpsertAllocs(structs.MsgTypeTestSetup, 6, allocs), "UpsertAllocs")
|
|
|
|
// Create the namespace policy and tokens
|
|
validDefaultToken := mock.CreatePolicyAndToken(t, state, 1001, "test-default-valid", mock.NodePolicy(acl.PolicyRead)+
|
|
mock.NamespacePolicy(structs.DefaultNamespace, "", []string{acl.NamespaceCapabilityReadJob}))
|
|
validNoNSToken := mock.CreatePolicyAndToken(t, state, 1003, "test-alt-valid", mock.NodePolicy(acl.PolicyRead))
|
|
invalidToken := mock.CreatePolicyAndToken(t, state, 1004, "test-invalid",
|
|
mock.NamespacePolicy(structs.DefaultNamespace, "", []string{acl.NamespaceCapabilityReadJob}))
|
|
|
|
// Lookup the node without a token and expect failure
|
|
req := &structs.NodeSpecificRequest{
|
|
NodeID: node.ID,
|
|
QueryOptions: structs.QueryOptions{Region: "global"},
|
|
}
|
|
{
|
|
var resp structs.NodeAllocsResponse
|
|
err := msgpackrpc.CallWithCodec(codec, "Node.GetAllocs", req, &resp)
|
|
assert.NotNil(err, "RPC")
|
|
assert.Equal(err.Error(), structs.ErrPermissionDenied.Error())
|
|
}
|
|
|
|
// Try with a valid token for the default namespace
|
|
req.AuthToken = validDefaultToken.SecretID
|
|
{
|
|
var resp structs.NodeAllocsResponse
|
|
assert.Nil(msgpackrpc.CallWithCodec(codec, "Node.GetAllocs", req, &resp), "RPC")
|
|
assert.Len(resp.Allocs, 1)
|
|
assert.Equal(allocDefaultNS.ID, resp.Allocs[0].ID)
|
|
}
|
|
|
|
// Try with a valid token for a namespace with no allocs on this node
|
|
req.AuthToken = validNoNSToken.SecretID
|
|
{
|
|
var resp structs.NodeAllocsResponse
|
|
assert.Nil(msgpackrpc.CallWithCodec(codec, "Node.GetAllocs", req, &resp), "RPC")
|
|
assert.Len(resp.Allocs, 0)
|
|
}
|
|
|
|
// Try with a invalid token
|
|
req.AuthToken = invalidToken.SecretID
|
|
{
|
|
var resp structs.NodeAllocsResponse
|
|
err := msgpackrpc.CallWithCodec(codec, "Node.GetAllocs", req, &resp)
|
|
assert.NotNil(err, "RPC")
|
|
assert.Equal(err.Error(), structs.ErrPermissionDenied.Error())
|
|
}
|
|
|
|
// Try with a root token
|
|
req.AuthToken = root.SecretID
|
|
{
|
|
var resp structs.NodeAllocsResponse
|
|
assert.Nil(msgpackrpc.CallWithCodec(codec, "Node.GetAllocs", req, &resp), "RPC")
|
|
assert.Len(resp.Allocs, 3)
|
|
for _, alloc := range resp.Allocs {
|
|
switch alloc.ID {
|
|
case allocDefaultNS.ID, allocAltNS.ID, allocOtherNS.ID:
|
|
// expected
|
|
default:
|
|
t.Errorf("unexpected alloc %q for namespace %q", alloc.ID, alloc.Namespace)
|
|
}
|
|
}
|
|
}
|
|
}
|
|
|
|
func TestClientEndpoint_GetClientAllocs(t *testing.T) {
|
|
ci.Parallel(t)
|
|
require := require.New(t)
|
|
|
|
s1, cleanupS1 := TestServer(t, nil)
|
|
defer cleanupS1()
|
|
codec := rpcClient(t, s1)
|
|
testutil.WaitForLeader(t, s1.RPC)
|
|
|
|
// Check that we have no client connections
|
|
require.Empty(s1.connectedNodes())
|
|
|
|
// Create the register request
|
|
node := mock.Node()
|
|
state := s1.fsm.State()
|
|
require.Nil(state.UpsertNode(structs.MsgTypeTestSetup, 98, node))
|
|
|
|
// Inject fake evaluations
|
|
alloc := mock.Alloc()
|
|
alloc.NodeID = node.ID
|
|
state.UpsertJobSummary(99, mock.JobSummary(alloc.JobID))
|
|
err := state.UpsertAllocs(structs.MsgTypeTestSetup, 100, []*structs.Allocation{alloc})
|
|
if err != nil {
|
|
t.Fatalf("err: %v", err)
|
|
}
|
|
|
|
// Lookup the allocs
|
|
get := &structs.NodeSpecificRequest{
|
|
NodeID: node.ID,
|
|
SecretID: node.SecretID,
|
|
QueryOptions: structs.QueryOptions{Region: "global"},
|
|
}
|
|
var resp2 structs.NodeClientAllocsResponse
|
|
if err := msgpackrpc.CallWithCodec(codec, "Node.GetClientAllocs", get, &resp2); err != nil {
|
|
t.Fatalf("err: %v", err)
|
|
}
|
|
if resp2.Index != 100 {
|
|
t.Fatalf("Bad index: %d %d", resp2.Index, 100)
|
|
}
|
|
|
|
if len(resp2.Allocs) != 1 || resp2.Allocs[alloc.ID] != 100 {
|
|
t.Fatalf("bad: %#v", resp2.Allocs)
|
|
}
|
|
|
|
// Check that we have the client connections
|
|
nodes := s1.connectedNodes()
|
|
require.Len(nodes, 1)
|
|
require.Contains(nodes, node.ID)
|
|
|
|
// Lookup node with bad SecretID
|
|
get.SecretID = "foobarbaz"
|
|
var resp3 structs.NodeClientAllocsResponse
|
|
err = msgpackrpc.CallWithCodec(codec, "Node.GetClientAllocs", get, &resp3)
|
|
if err == nil || !strings.Contains(err.Error(), "does not match") {
|
|
t.Fatalf("err: %v", err)
|
|
}
|
|
|
|
// Lookup non-existing node
|
|
get.NodeID = uuid.Generate()
|
|
var resp4 structs.NodeClientAllocsResponse
|
|
if err := msgpackrpc.CallWithCodec(codec, "Node.GetClientAllocs", get, &resp4); err != nil {
|
|
t.Fatalf("err: %v", err)
|
|
}
|
|
if resp4.Index != 100 {
|
|
t.Fatalf("Bad index: %d %d", resp3.Index, 100)
|
|
}
|
|
if len(resp4.Allocs) != 0 {
|
|
t.Fatalf("unexpected node %#v", resp3.Allocs)
|
|
}
|
|
|
|
// Close the connection and check that we remove the client connections
|
|
require.Nil(codec.Close())
|
|
testutil.WaitForResult(func() (bool, error) {
|
|
nodes := s1.connectedNodes()
|
|
return len(nodes) == 0, nil
|
|
}, func(err error) {
|
|
t.Fatalf("should have no clients")
|
|
})
|
|
}
|
|
|
|
func TestClientEndpoint_GetClientAllocs_Blocking(t *testing.T) {
|
|
ci.Parallel(t)
|
|
|
|
s1, cleanupS1 := TestServer(t, nil)
|
|
defer cleanupS1()
|
|
codec := rpcClient(t, s1)
|
|
testutil.WaitForLeader(t, s1.RPC)
|
|
|
|
// Create the register request
|
|
node := mock.Node()
|
|
reg := &structs.NodeRegisterRequest{
|
|
Node: node,
|
|
WriteRequest: structs.WriteRequest{Region: "global"},
|
|
}
|
|
|
|
// Fetch the response
|
|
var resp structs.GenericResponse
|
|
if err := msgpackrpc.CallWithCodec(codec, "Node.Register", reg, &resp); err != nil {
|
|
t.Fatalf("err: %v", err)
|
|
}
|
|
node.CreateIndex = resp.Index
|
|
node.ModifyIndex = resp.Index
|
|
|
|
// Inject fake evaluations async
|
|
now := time.Now().UTC().UnixNano()
|
|
alloc := mock.Alloc()
|
|
alloc.NodeID = node.ID
|
|
alloc.ModifyTime = now
|
|
store := s1.fsm.State()
|
|
store.UpsertJobSummary(99, mock.JobSummary(alloc.JobID))
|
|
start := time.Now()
|
|
time.AfterFunc(100*time.Millisecond, func() {
|
|
err := store.UpsertAllocs(structs.MsgTypeTestSetup, 100, []*structs.Allocation{alloc})
|
|
if err != nil {
|
|
t.Fatalf("err: %v", err)
|
|
}
|
|
})
|
|
|
|
// Lookup the allocs in a blocking query
|
|
req := &structs.NodeSpecificRequest{
|
|
NodeID: node.ID,
|
|
SecretID: node.SecretID,
|
|
QueryOptions: structs.QueryOptions{
|
|
Region: "global",
|
|
MinQueryIndex: 50,
|
|
MaxQueryTime: time.Second,
|
|
},
|
|
}
|
|
var resp2 structs.NodeClientAllocsResponse
|
|
if err := msgpackrpc.CallWithCodec(codec, "Node.GetClientAllocs", req, &resp2); err != nil {
|
|
t.Fatalf("err: %v", err)
|
|
}
|
|
|
|
// Should block at least 100ms
|
|
if time.Since(start) < 100*time.Millisecond {
|
|
t.Fatalf("too fast")
|
|
}
|
|
|
|
if resp2.Index != 100 {
|
|
t.Fatalf("Bad index: %d %d", resp2.Index, 100)
|
|
}
|
|
|
|
if len(resp2.Allocs) != 1 || resp2.Allocs[alloc.ID] != 100 {
|
|
t.Fatalf("bad: %#v", resp2.Allocs)
|
|
}
|
|
|
|
iter, err := store.AllocsByIDPrefix(nil, structs.DefaultNamespace, alloc.ID, state.SortDefault)
|
|
if err != nil {
|
|
t.Fatalf("err: %v", err)
|
|
}
|
|
|
|
getAllocs := func(iter memdb.ResultIterator) []*structs.Allocation {
|
|
var allocs []*structs.Allocation
|
|
for {
|
|
raw := iter.Next()
|
|
if raw == nil {
|
|
break
|
|
}
|
|
allocs = append(allocs, raw.(*structs.Allocation))
|
|
}
|
|
return allocs
|
|
}
|
|
out := getAllocs(iter)
|
|
|
|
if len(out) != 1 {
|
|
t.Fatalf("Expected to get one allocation but got:%v", out)
|
|
}
|
|
|
|
if out[0].ModifyTime != now {
|
|
t.Fatalf("Invalid modify time %v", out[0].ModifyTime)
|
|
}
|
|
|
|
// Alloc updates fire watches
|
|
time.AfterFunc(100*time.Millisecond, func() {
|
|
allocUpdate := mock.Alloc()
|
|
allocUpdate.NodeID = alloc.NodeID
|
|
allocUpdate.ID = alloc.ID
|
|
allocUpdate.ClientStatus = structs.AllocClientStatusRunning
|
|
store.UpsertJobSummary(199, mock.JobSummary(allocUpdate.JobID))
|
|
err := store.UpsertAllocs(structs.MsgTypeTestSetup, 200, []*structs.Allocation{allocUpdate})
|
|
if err != nil {
|
|
t.Fatalf("err: %v", err)
|
|
}
|
|
})
|
|
|
|
req.QueryOptions.MinQueryIndex = 150
|
|
var resp3 structs.NodeClientAllocsResponse
|
|
if err := msgpackrpc.CallWithCodec(codec, "Node.GetClientAllocs", req, &resp3); err != nil {
|
|
t.Fatalf("err: %v", err)
|
|
}
|
|
|
|
if time.Since(start) < 100*time.Millisecond {
|
|
t.Fatalf("too fast")
|
|
}
|
|
if resp3.Index != 200 {
|
|
t.Fatalf("Bad index: %d %d", resp3.Index, 200)
|
|
}
|
|
if len(resp3.Allocs) != 1 || resp3.Allocs[alloc.ID] != 200 {
|
|
t.Fatalf("bad: %#v", resp3.Allocs)
|
|
}
|
|
}
|
|
|
|
func TestClientEndpoint_GetClientAllocs_Blocking_GC(t *testing.T) {
|
|
ci.Parallel(t)
|
|
assert := assert.New(t)
|
|
|
|
s1, cleanupS1 := TestServer(t, nil)
|
|
defer cleanupS1()
|
|
codec := rpcClient(t, s1)
|
|
testutil.WaitForLeader(t, s1.RPC)
|
|
|
|
// Create the register request
|
|
node := mock.Node()
|
|
reg := &structs.NodeRegisterRequest{
|
|
Node: node,
|
|
WriteRequest: structs.WriteRequest{Region: "global"},
|
|
}
|
|
|
|
// Fetch the response
|
|
var resp structs.GenericResponse
|
|
assert.Nil(msgpackrpc.CallWithCodec(codec, "Node.Register", reg, &resp))
|
|
node.CreateIndex = resp.Index
|
|
node.ModifyIndex = resp.Index
|
|
|
|
// Inject fake allocations async
|
|
alloc1 := mock.Alloc()
|
|
alloc1.NodeID = node.ID
|
|
alloc2 := mock.Alloc()
|
|
alloc2.NodeID = node.ID
|
|
state := s1.fsm.State()
|
|
state.UpsertJobSummary(99, mock.JobSummary(alloc1.JobID))
|
|
start := time.Now()
|
|
time.AfterFunc(100*time.Millisecond, func() {
|
|
assert.Nil(state.UpsertAllocs(structs.MsgTypeTestSetup, 100, []*structs.Allocation{alloc1, alloc2}))
|
|
})
|
|
|
|
// Lookup the allocs in a blocking query
|
|
req := &structs.NodeSpecificRequest{
|
|
NodeID: node.ID,
|
|
SecretID: node.SecretID,
|
|
QueryOptions: structs.QueryOptions{
|
|
Region: "global",
|
|
MinQueryIndex: 50,
|
|
MaxQueryTime: time.Second,
|
|
},
|
|
}
|
|
var resp2 structs.NodeClientAllocsResponse
|
|
assert.Nil(msgpackrpc.CallWithCodec(codec, "Node.GetClientAllocs", req, &resp2))
|
|
|
|
// Should block at least 100ms
|
|
if time.Since(start) < 100*time.Millisecond {
|
|
t.Fatalf("too fast")
|
|
}
|
|
|
|
assert.EqualValues(100, resp2.Index)
|
|
if assert.Len(resp2.Allocs, 2) {
|
|
assert.EqualValues(100, resp2.Allocs[alloc1.ID])
|
|
}
|
|
|
|
// Delete an allocation
|
|
time.AfterFunc(100*time.Millisecond, func() {
|
|
assert.Nil(state.DeleteEval(200, nil, []string{alloc2.ID}, false))
|
|
})
|
|
|
|
req.QueryOptions.MinQueryIndex = 150
|
|
var resp3 structs.NodeClientAllocsResponse
|
|
assert.Nil(msgpackrpc.CallWithCodec(codec, "Node.GetClientAllocs", req, &resp3))
|
|
|
|
if time.Since(start) < 100*time.Millisecond {
|
|
t.Fatalf("too fast")
|
|
}
|
|
assert.EqualValues(200, resp3.Index)
|
|
if assert.Len(resp3.Allocs, 1) {
|
|
assert.EqualValues(100, resp3.Allocs[alloc1.ID])
|
|
}
|
|
}
|
|
|
|
// A MigrateToken should not be created if an allocation shares the same node
|
|
// with its previous allocation
|
|
func TestClientEndpoint_GetClientAllocs_WithoutMigrateTokens(t *testing.T) {
|
|
ci.Parallel(t)
|
|
assert := assert.New(t)
|
|
|
|
s1, cleanupS1 := TestServer(t, nil)
|
|
defer cleanupS1()
|
|
codec := rpcClient(t, s1)
|
|
testutil.WaitForLeader(t, s1.RPC)
|
|
|
|
// Create the register request
|
|
node := mock.Node()
|
|
reg := &structs.NodeRegisterRequest{
|
|
Node: node,
|
|
WriteRequest: structs.WriteRequest{Region: "global"},
|
|
}
|
|
|
|
// Fetch the response
|
|
var resp structs.GenericResponse
|
|
if err := msgpackrpc.CallWithCodec(codec, "Node.Register", reg, &resp); err != nil {
|
|
t.Fatalf("err: %v", err)
|
|
}
|
|
node.CreateIndex = resp.Index
|
|
node.ModifyIndex = resp.Index
|
|
|
|
// Inject fake evaluations
|
|
prevAlloc := mock.Alloc()
|
|
prevAlloc.NodeID = node.ID
|
|
alloc := mock.Alloc()
|
|
alloc.NodeID = node.ID
|
|
alloc.PreviousAllocation = prevAlloc.ID
|
|
alloc.DesiredStatus = structs.AllocClientStatusComplete
|
|
state := s1.fsm.State()
|
|
state.UpsertJobSummary(99, mock.JobSummary(alloc.JobID))
|
|
err := state.UpsertAllocs(structs.MsgTypeTestSetup, 100, []*structs.Allocation{prevAlloc, alloc})
|
|
assert.Nil(err)
|
|
|
|
// Lookup the allocs
|
|
get := &structs.NodeSpecificRequest{
|
|
NodeID: node.ID,
|
|
SecretID: node.SecretID,
|
|
QueryOptions: structs.QueryOptions{Region: "global"},
|
|
}
|
|
var resp2 structs.NodeClientAllocsResponse
|
|
|
|
err = msgpackrpc.CallWithCodec(codec, "Node.GetClientAllocs", get, &resp2)
|
|
assert.Nil(err)
|
|
|
|
assert.Equal(uint64(100), resp2.Index)
|
|
assert.Equal(2, len(resp2.Allocs))
|
|
assert.Equal(uint64(100), resp2.Allocs[alloc.ID])
|
|
assert.Equal(0, len(resp2.MigrateTokens))
|
|
}
|
|
|
|
func TestClientEndpoint_GetAllocs_Blocking(t *testing.T) {
|
|
ci.Parallel(t)
|
|
|
|
s1, cleanupS1 := TestServer(t, nil)
|
|
defer cleanupS1()
|
|
codec := rpcClient(t, s1)
|
|
testutil.WaitForLeader(t, s1.RPC)
|
|
|
|
// Create the register request
|
|
node := mock.Node()
|
|
reg := &structs.NodeRegisterRequest{
|
|
Node: node,
|
|
WriteRequest: structs.WriteRequest{Region: "global"},
|
|
}
|
|
|
|
// Fetch the response
|
|
var resp structs.GenericResponse
|
|
if err := msgpackrpc.CallWithCodec(codec, "Node.Register", reg, &resp); err != nil {
|
|
t.Fatalf("err: %v", err)
|
|
}
|
|
node.CreateIndex = resp.Index
|
|
node.ModifyIndex = resp.Index
|
|
|
|
// Inject fake evaluations async
|
|
alloc := mock.Alloc()
|
|
alloc.NodeID = node.ID
|
|
state := s1.fsm.State()
|
|
state.UpsertJobSummary(99, mock.JobSummary(alloc.JobID))
|
|
start := time.Now()
|
|
time.AfterFunc(100*time.Millisecond, func() {
|
|
err := state.UpsertAllocs(structs.MsgTypeTestSetup, 100, []*structs.Allocation{alloc})
|
|
if err != nil {
|
|
t.Fatalf("err: %v", err)
|
|
}
|
|
})
|
|
|
|
// Lookup the allocs in a blocking query
|
|
req := &structs.NodeSpecificRequest{
|
|
NodeID: node.ID,
|
|
QueryOptions: structs.QueryOptions{
|
|
Region: "global",
|
|
MinQueryIndex: 50,
|
|
MaxQueryTime: time.Second,
|
|
},
|
|
}
|
|
var resp2 structs.NodeAllocsResponse
|
|
if err := msgpackrpc.CallWithCodec(codec, "Node.GetAllocs", req, &resp2); err != nil {
|
|
t.Fatalf("err: %v", err)
|
|
}
|
|
|
|
// Should block at least 100ms
|
|
if time.Since(start) < 100*time.Millisecond {
|
|
t.Fatalf("too fast")
|
|
}
|
|
|
|
if resp2.Index != 100 {
|
|
t.Fatalf("Bad index: %d %d", resp2.Index, 100)
|
|
}
|
|
|
|
if len(resp2.Allocs) != 1 || resp2.Allocs[0].ID != alloc.ID {
|
|
t.Fatalf("bad: %#v", resp2.Allocs)
|
|
}
|
|
|
|
// Alloc updates fire watches
|
|
time.AfterFunc(100*time.Millisecond, func() {
|
|
allocUpdate := mock.Alloc()
|
|
allocUpdate.NodeID = alloc.NodeID
|
|
allocUpdate.ID = alloc.ID
|
|
allocUpdate.ClientStatus = structs.AllocClientStatusRunning
|
|
state.UpsertJobSummary(199, mock.JobSummary(allocUpdate.JobID))
|
|
err := state.UpdateAllocsFromClient(structs.MsgTypeTestSetup, 200, []*structs.Allocation{allocUpdate})
|
|
if err != nil {
|
|
t.Fatalf("err: %v", err)
|
|
}
|
|
})
|
|
|
|
req.QueryOptions.MinQueryIndex = 150
|
|
var resp3 structs.NodeAllocsResponse
|
|
if err := msgpackrpc.CallWithCodec(codec, "Node.GetAllocs", req, &resp3); err != nil {
|
|
t.Fatalf("err: %v", err)
|
|
}
|
|
|
|
if time.Since(start) < 100*time.Millisecond {
|
|
t.Fatalf("too fast")
|
|
}
|
|
if resp3.Index != 200 {
|
|
t.Fatalf("Bad index: %d %d", resp3.Index, 200)
|
|
}
|
|
if len(resp3.Allocs) != 1 || resp3.Allocs[0].ClientStatus != structs.AllocClientStatusRunning {
|
|
t.Fatalf("bad: %#v", resp3.Allocs[0])
|
|
}
|
|
}
|
|
|
|
func TestClientEndpoint_UpdateAlloc(t *testing.T) {
|
|
ci.Parallel(t)
|
|
|
|
s1, cleanupS1 := TestServer(t, func(c *Config) {
|
|
// Disabling scheduling in this test so that we can
|
|
// ensure that the state store doesn't accumulate more evals
|
|
// than what we expect the unit test to add
|
|
c.NumSchedulers = 0
|
|
})
|
|
|
|
defer cleanupS1()
|
|
codec := rpcClient(t, s1)
|
|
testutil.WaitForLeader(t, s1.RPC)
|
|
require := require.New(t)
|
|
|
|
// Create the register request
|
|
node := mock.Node()
|
|
reg := &structs.NodeRegisterRequest{
|
|
Node: node,
|
|
WriteRequest: structs.WriteRequest{Region: "global"},
|
|
}
|
|
|
|
// Fetch the response
|
|
var resp structs.GenericResponse
|
|
if err := msgpackrpc.CallWithCodec(codec, "Node.Register", reg, &resp); err != nil {
|
|
t.Fatalf("err: %v", err)
|
|
}
|
|
|
|
state := s1.fsm.State()
|
|
// Inject mock job
|
|
job := mock.Job()
|
|
job.ID = "mytestjob"
|
|
err := state.UpsertJob(structs.MsgTypeTestSetup, 101, job)
|
|
require.Nil(err)
|
|
|
|
// Inject fake allocations
|
|
alloc := mock.Alloc()
|
|
alloc.JobID = job.ID
|
|
alloc.NodeID = node.ID
|
|
err = state.UpsertJobSummary(99, mock.JobSummary(alloc.JobID))
|
|
require.Nil(err)
|
|
alloc.TaskGroup = job.TaskGroups[0].Name
|
|
|
|
alloc2 := mock.Alloc()
|
|
alloc2.JobID = job.ID
|
|
alloc2.NodeID = node.ID
|
|
err = state.UpsertJobSummary(99, mock.JobSummary(alloc2.JobID))
|
|
require.Nil(err)
|
|
alloc2.TaskGroup = job.TaskGroups[0].Name
|
|
|
|
err = state.UpsertAllocs(structs.MsgTypeTestSetup, 100, []*structs.Allocation{alloc, alloc2})
|
|
require.Nil(err)
|
|
|
|
// Attempt updates of more than one alloc for the same job
|
|
clientAlloc1 := new(structs.Allocation)
|
|
*clientAlloc1 = *alloc
|
|
clientAlloc1.ClientStatus = structs.AllocClientStatusFailed
|
|
|
|
clientAlloc2 := new(structs.Allocation)
|
|
*clientAlloc2 = *alloc2
|
|
clientAlloc2.ClientStatus = structs.AllocClientStatusFailed
|
|
|
|
// Update the alloc
|
|
update := &structs.AllocUpdateRequest{
|
|
Alloc: []*structs.Allocation{clientAlloc1, clientAlloc2},
|
|
WriteRequest: structs.WriteRequest{Region: "global"},
|
|
}
|
|
var resp2 structs.NodeAllocsResponse
|
|
start := time.Now()
|
|
err = msgpackrpc.CallWithCodec(codec, "Node.UpdateAlloc", update, &resp2)
|
|
require.Nil(err)
|
|
require.NotEqual(uint64(0), resp2.Index)
|
|
|
|
if diff := time.Since(start); diff < batchUpdateInterval {
|
|
t.Fatalf("too fast: %v", diff)
|
|
}
|
|
|
|
// Lookup the alloc
|
|
ws := memdb.NewWatchSet()
|
|
out, err := state.AllocByID(ws, alloc.ID)
|
|
require.Nil(err)
|
|
require.Equal(structs.AllocClientStatusFailed, out.ClientStatus)
|
|
require.True(out.ModifyTime > 0)
|
|
|
|
// Assert that exactly one eval with TriggeredBy EvalTriggerRetryFailedAlloc exists
|
|
evaluations, err := state.EvalsByJob(ws, job.Namespace, job.ID)
|
|
require.Nil(err)
|
|
require.True(len(evaluations) != 0)
|
|
foundCount := 0
|
|
for _, resultEval := range evaluations {
|
|
if resultEval.TriggeredBy == structs.EvalTriggerRetryFailedAlloc && resultEval.WaitUntil.IsZero() {
|
|
foundCount++
|
|
}
|
|
}
|
|
require.Equal(1, foundCount, "Should create exactly one eval for failed allocs")
|
|
|
|
}
|
|
|
|
func TestClientEndpoint_UpdateAlloc_NodeNotReady(t *testing.T) {
|
|
ci.Parallel(t)
|
|
|
|
s1, cleanupS1 := TestServer(t, nil)
|
|
defer cleanupS1()
|
|
codec := rpcClient(t, s1)
|
|
testutil.WaitForLeader(t, s1.RPC)
|
|
|
|
// Register node.
|
|
node := mock.Node()
|
|
reg := &structs.NodeRegisterRequest{
|
|
Node: node,
|
|
WriteRequest: structs.WriteRequest{Region: "global"},
|
|
}
|
|
var resp structs.GenericResponse
|
|
err := msgpackrpc.CallWithCodec(codec, "Node.Register", reg, &resp)
|
|
require.NoError(t, err)
|
|
|
|
// Inject mock job and allocation.
|
|
state := s1.fsm.State()
|
|
|
|
job := mock.Job()
|
|
err = state.UpsertJob(structs.MsgTypeTestSetup, 101, job)
|
|
require.NoError(t, err)
|
|
|
|
alloc := mock.Alloc()
|
|
alloc.JobID = job.ID
|
|
alloc.NodeID = node.ID
|
|
alloc.TaskGroup = job.TaskGroups[0].Name
|
|
alloc.ClientStatus = structs.AllocClientStatusRunning
|
|
|
|
err = state.UpsertJobSummary(99, mock.JobSummary(alloc.JobID))
|
|
require.NoError(t, err)
|
|
err = state.UpsertAllocs(structs.MsgTypeTestSetup, 100, []*structs.Allocation{alloc})
|
|
require.NoError(t, err)
|
|
|
|
// Mark node as down.
|
|
err = state.UpdateNodeStatus(structs.MsgTypeTestSetup, 101, node.ID, structs.NodeStatusDown, time.Now().UnixNano(), nil)
|
|
require.NoError(t, err)
|
|
|
|
// Try to update alloc.
|
|
updatedAlloc := new(structs.Allocation)
|
|
*updatedAlloc = *alloc
|
|
updatedAlloc.ClientStatus = structs.AllocClientStatusFailed
|
|
|
|
allocUpdateReq := &structs.AllocUpdateRequest{
|
|
Alloc: []*structs.Allocation{updatedAlloc},
|
|
WriteRequest: structs.WriteRequest{Region: "global"},
|
|
}
|
|
var allocUpdateResp structs.NodeAllocsResponse
|
|
err = msgpackrpc.CallWithCodec(codec, "Node.UpdateAlloc", allocUpdateReq, &allocUpdateResp)
|
|
require.ErrorContains(t, err, "not ready")
|
|
|
|
// Send request without an explicit node ID.
|
|
updatedAlloc.NodeID = ""
|
|
err = msgpackrpc.CallWithCodec(codec, "Node.UpdateAlloc", allocUpdateReq, &allocUpdateResp)
|
|
require.ErrorContains(t, err, "missing node ID")
|
|
|
|
// Send request with invalid node ID.
|
|
updatedAlloc.NodeID = "not-valid"
|
|
err = msgpackrpc.CallWithCodec(codec, "Node.UpdateAlloc", allocUpdateReq, &allocUpdateResp)
|
|
require.ErrorContains(t, err, "node lookup failed")
|
|
|
|
// Send request with non-existing node ID.
|
|
updatedAlloc.NodeID = uuid.Generate()
|
|
err = msgpackrpc.CallWithCodec(codec, "Node.UpdateAlloc", allocUpdateReq, &allocUpdateResp)
|
|
require.ErrorContains(t, err, "not found")
|
|
|
|
// Mark node as ready and try again.
|
|
err = state.UpdateNodeStatus(structs.MsgTypeTestSetup, 102, node.ID, structs.NodeStatusReady, time.Now().UnixNano(), nil)
|
|
require.NoError(t, err)
|
|
|
|
updatedAlloc.NodeID = node.ID
|
|
err = msgpackrpc.CallWithCodec(codec, "Node.UpdateAlloc", allocUpdateReq, &allocUpdateResp)
|
|
require.NoError(t, err)
|
|
}
|
|
|
|
func TestClientEndpoint_BatchUpdate(t *testing.T) {
|
|
ci.Parallel(t)
|
|
|
|
s1, cleanupS1 := TestServer(t, nil)
|
|
defer cleanupS1()
|
|
codec := rpcClient(t, s1)
|
|
testutil.WaitForLeader(t, s1.RPC)
|
|
|
|
// Create the register request
|
|
node := mock.Node()
|
|
reg := &structs.NodeRegisterRequest{
|
|
Node: node,
|
|
WriteRequest: structs.WriteRequest{Region: "global"},
|
|
}
|
|
|
|
// Fetch the response
|
|
var resp structs.GenericResponse
|
|
if err := msgpackrpc.CallWithCodec(codec, "Node.Register", reg, &resp); err != nil {
|
|
t.Fatalf("err: %v", err)
|
|
}
|
|
|
|
// Inject fake evaluations
|
|
alloc := mock.Alloc()
|
|
alloc.NodeID = node.ID
|
|
state := s1.fsm.State()
|
|
state.UpsertJobSummary(99, mock.JobSummary(alloc.JobID))
|
|
err := state.UpsertAllocs(structs.MsgTypeTestSetup, 100, []*structs.Allocation{alloc})
|
|
if err != nil {
|
|
t.Fatalf("err: %v", err)
|
|
}
|
|
|
|
// Attempt update
|
|
clientAlloc := new(structs.Allocation)
|
|
*clientAlloc = *alloc
|
|
clientAlloc.ClientStatus = structs.AllocClientStatusFailed
|
|
|
|
// Call to do the batch update
|
|
bf := structs.NewBatchFuture()
|
|
endpoint := s1.staticEndpoints.Node
|
|
endpoint.batchUpdate(bf, []*structs.Allocation{clientAlloc}, nil)
|
|
if err := bf.Wait(); err != nil {
|
|
t.Fatalf("err: %v", err)
|
|
}
|
|
if bf.Index() == 0 {
|
|
t.Fatalf("Bad index: %d", bf.Index())
|
|
}
|
|
|
|
// Lookup the alloc
|
|
ws := memdb.NewWatchSet()
|
|
out, err := state.AllocByID(ws, alloc.ID)
|
|
if err != nil {
|
|
t.Fatalf("err: %v", err)
|
|
}
|
|
if out.ClientStatus != structs.AllocClientStatusFailed {
|
|
t.Fatalf("Bad: %#v", out)
|
|
}
|
|
}
|
|
|
|
func TestClientEndpoint_UpdateAlloc_Vault(t *testing.T) {
|
|
ci.Parallel(t)
|
|
|
|
s1, cleanupS1 := TestServer(t, nil)
|
|
defer cleanupS1()
|
|
codec := rpcClient(t, s1)
|
|
testutil.WaitForLeader(t, s1.RPC)
|
|
|
|
// Create the register request
|
|
node := mock.Node()
|
|
reg := &structs.NodeRegisterRequest{
|
|
Node: node,
|
|
WriteRequest: structs.WriteRequest{Region: "global"},
|
|
}
|
|
|
|
// Fetch the response
|
|
var resp structs.GenericResponse
|
|
if err := msgpackrpc.CallWithCodec(codec, "Node.Register", reg, &resp); err != nil {
|
|
t.Fatalf("err: %v", err)
|
|
}
|
|
|
|
// Swap the servers Vault Client
|
|
tvc := &TestVaultClient{}
|
|
s1.vault = tvc
|
|
|
|
// Inject fake allocation and vault accessor
|
|
alloc := mock.Alloc()
|
|
alloc.NodeID = node.ID
|
|
state := s1.fsm.State()
|
|
state.UpsertJobSummary(99, mock.JobSummary(alloc.JobID))
|
|
if err := state.UpsertAllocs(structs.MsgTypeTestSetup, 100, []*structs.Allocation{alloc}); err != nil {
|
|
t.Fatalf("err: %v", err)
|
|
}
|
|
|
|
va := mock.VaultAccessor()
|
|
va.NodeID = node.ID
|
|
va.AllocID = alloc.ID
|
|
if err := state.UpsertVaultAccessor(101, []*structs.VaultAccessor{va}); err != nil {
|
|
t.Fatalf("err: %v", err)
|
|
}
|
|
|
|
// Inject mock job
|
|
job := mock.Job()
|
|
job.ID = alloc.JobID
|
|
err := state.UpsertJob(structs.MsgTypeTestSetup, 101, job)
|
|
if err != nil {
|
|
t.Fatalf("err: %v", err)
|
|
}
|
|
|
|
// Attempt update
|
|
clientAlloc := new(structs.Allocation)
|
|
*clientAlloc = *alloc
|
|
clientAlloc.ClientStatus = structs.AllocClientStatusFailed
|
|
|
|
// Update the alloc
|
|
update := &structs.AllocUpdateRequest{
|
|
Alloc: []*structs.Allocation{clientAlloc},
|
|
WriteRequest: structs.WriteRequest{Region: "global"},
|
|
}
|
|
var resp2 structs.NodeAllocsResponse
|
|
start := time.Now()
|
|
if err := msgpackrpc.CallWithCodec(codec, "Node.UpdateAlloc", update, &resp2); err != nil {
|
|
t.Fatalf("err: %v", err)
|
|
}
|
|
if resp2.Index == 0 {
|
|
t.Fatalf("Bad index: %d", resp2.Index)
|
|
}
|
|
if diff := time.Since(start); diff < batchUpdateInterval {
|
|
t.Fatalf("too fast: %v", diff)
|
|
}
|
|
|
|
// Lookup the alloc
|
|
ws := memdb.NewWatchSet()
|
|
out, err := state.AllocByID(ws, alloc.ID)
|
|
if err != nil {
|
|
t.Fatalf("err: %v", err)
|
|
}
|
|
if out.ClientStatus != structs.AllocClientStatusFailed {
|
|
t.Fatalf("Bad: %#v", out)
|
|
}
|
|
|
|
if l := len(tvc.RevokedTokens); l != 1 {
|
|
t.Fatalf("Deregister revoked %d tokens; want 1", l)
|
|
}
|
|
}
|
|
|
|
func TestClientEndpoint_CreateNodeEvals(t *testing.T) {
|
|
ci.Parallel(t)
|
|
|
|
s1, cleanupS1 := TestServer(t, nil)
|
|
defer cleanupS1()
|
|
testutil.WaitForLeader(t, s1.RPC)
|
|
state := s1.fsm.State()
|
|
|
|
idx, err := state.LatestIndex()
|
|
require.NoError(t, err)
|
|
|
|
node := mock.Node()
|
|
err = state.UpsertNode(structs.MsgTypeTestSetup, idx, node)
|
|
require.NoError(t, err)
|
|
idx++
|
|
|
|
// Inject fake evaluations
|
|
alloc := mock.Alloc()
|
|
alloc.NodeID = node.ID
|
|
state.UpsertJobSummary(1, mock.JobSummary(alloc.JobID))
|
|
require.NoError(t, state.UpsertAllocs(structs.MsgTypeTestSetup, idx, []*structs.Allocation{alloc}))
|
|
idx++
|
|
|
|
// Inject a fake system job.
|
|
job := mock.SystemJob()
|
|
if err := state.UpsertJob(structs.MsgTypeTestSetup, idx, job); err != nil {
|
|
t.Fatalf("err: %v", err)
|
|
}
|
|
idx++
|
|
|
|
// Create some evaluations
|
|
ids, index, err := s1.staticEndpoints.Node.createNodeEvals(node, 1)
|
|
if err != nil {
|
|
t.Fatalf("err: %v", err)
|
|
}
|
|
if index == 0 {
|
|
t.Fatalf("bad: %d", index)
|
|
}
|
|
if len(ids) != 2 {
|
|
t.Fatalf("bad: %s", ids)
|
|
}
|
|
|
|
// Lookup the evaluations
|
|
ws := memdb.NewWatchSet()
|
|
evalByType := make(map[string]*structs.Evaluation, 2)
|
|
for _, id := range ids {
|
|
eval, err := state.EvalByID(ws, id)
|
|
if err != nil {
|
|
t.Fatalf("err: %v", err)
|
|
}
|
|
if eval == nil {
|
|
t.Fatalf("expected eval")
|
|
}
|
|
|
|
if old, ok := evalByType[eval.Type]; ok {
|
|
t.Fatalf("multiple evals of the same type: %v and %v", old, eval)
|
|
}
|
|
|
|
evalByType[eval.Type] = eval
|
|
}
|
|
|
|
if len(evalByType) != 2 {
|
|
t.Fatalf("Expected a service and system job; got %#v", evalByType)
|
|
}
|
|
|
|
// Ensure the evals are correct.
|
|
for schedType, eval := range evalByType {
|
|
expPriority := alloc.Job.Priority
|
|
expJobID := alloc.JobID
|
|
if schedType == "system" {
|
|
expPriority = job.Priority
|
|
expJobID = job.ID
|
|
}
|
|
|
|
t.Logf("checking eval: %v", pretty.Sprint(eval))
|
|
require.Equal(t, index, eval.CreateIndex)
|
|
require.Equal(t, structs.EvalTriggerNodeUpdate, eval.TriggeredBy)
|
|
require.Equal(t, alloc.NodeID, eval.NodeID)
|
|
require.Equal(t, uint64(1), eval.NodeModifyIndex)
|
|
switch eval.Status {
|
|
case structs.EvalStatusPending, structs.EvalStatusComplete:
|
|
// success
|
|
default:
|
|
t.Fatalf("expected pending or complete, found %v", eval.Status)
|
|
}
|
|
require.Equal(t, expPriority, eval.Priority)
|
|
require.Equal(t, expJobID, eval.JobID)
|
|
require.NotZero(t, eval.CreateTime)
|
|
require.NotZero(t, eval.ModifyTime)
|
|
}
|
|
}
|
|
|
|
// TestClientEndpoint_CreateNodeEvals_MultipleNSes asserts that evals are made
|
|
// for all jobs across namespaces
|
|
func TestClientEndpoint_CreateNodeEvals_MultipleNSes(t *testing.T) {
|
|
ci.Parallel(t)
|
|
|
|
s1, cleanupS1 := TestServer(t, nil)
|
|
defer cleanupS1()
|
|
testutil.WaitForLeader(t, s1.RPC)
|
|
|
|
state := s1.fsm.State()
|
|
|
|
idx := uint64(3)
|
|
ns1 := mock.Namespace()
|
|
err := state.UpsertNamespaces(idx, []*structs.Namespace{ns1})
|
|
require.NoError(t, err)
|
|
idx++
|
|
|
|
node := mock.Node()
|
|
err = state.UpsertNode(structs.MsgTypeTestSetup, idx, node)
|
|
require.NoError(t, err)
|
|
idx++
|
|
|
|
// Inject a fake system job.
|
|
defaultJob := mock.SystemJob()
|
|
err = state.UpsertJob(structs.MsgTypeTestSetup, idx, defaultJob)
|
|
require.NoError(t, err)
|
|
idx++
|
|
|
|
nsJob := mock.SystemJob()
|
|
nsJob.ID = defaultJob.ID
|
|
nsJob.Namespace = ns1.Name
|
|
err = state.UpsertJob(structs.MsgTypeTestSetup, idx, nsJob)
|
|
require.NoError(t, err)
|
|
idx++
|
|
|
|
// Create some evaluations
|
|
evalIDs, index, err := s1.staticEndpoints.Node.createNodeEvals(node, 1)
|
|
require.NoError(t, err)
|
|
require.NotZero(t, index)
|
|
require.Len(t, evalIDs, 2)
|
|
|
|
byNS := map[string]*structs.Evaluation{}
|
|
for _, evalID := range evalIDs {
|
|
eval, err := state.EvalByID(nil, evalID)
|
|
require.NoError(t, err)
|
|
byNS[eval.Namespace] = eval
|
|
}
|
|
|
|
require.Len(t, byNS, 2)
|
|
|
|
defaultNSEval := byNS[defaultJob.Namespace]
|
|
require.NotNil(t, defaultNSEval)
|
|
require.Equal(t, defaultJob.ID, defaultNSEval.JobID)
|
|
require.Equal(t, defaultJob.Namespace, defaultNSEval.Namespace)
|
|
|
|
otherNSEval := byNS[nsJob.Namespace]
|
|
require.NotNil(t, otherNSEval)
|
|
require.Equal(t, nsJob.ID, otherNSEval.JobID)
|
|
require.Equal(t, nsJob.Namespace, otherNSEval.Namespace)
|
|
}
|
|
|
|
// TestClientEndpoint_CreateNodeEvals_MultipleDCes asserts that evals are made
|
|
// only for the DC the node is in.
|
|
func TestClientEndpoint_CreateNodeEvals_MultipleDCes(t *testing.T) {
|
|
ci.Parallel(t)
|
|
|
|
s1, cleanupS1 := TestServer(t, nil)
|
|
defer cleanupS1()
|
|
testutil.WaitForLeader(t, s1.RPC)
|
|
|
|
state := s1.fsm.State()
|
|
|
|
idx, err := state.LatestIndex()
|
|
require.NoError(t, err)
|
|
|
|
node := mock.Node()
|
|
node.Datacenter = "test1"
|
|
err = state.UpsertNode(structs.MsgTypeTestSetup, idx, node)
|
|
require.NoError(t, err)
|
|
idx++
|
|
|
|
// Inject a fake system job in the same dc
|
|
defaultJob := mock.SystemJob()
|
|
defaultJob.Datacenters = []string{"test1", "test2"}
|
|
err = state.UpsertJob(structs.MsgTypeTestSetup, idx, defaultJob)
|
|
require.NoError(t, err)
|
|
idx++
|
|
|
|
// Inject a fake system job in a different dc
|
|
nsJob := mock.SystemJob()
|
|
nsJob.Datacenters = []string{"test2", "test3"}
|
|
err = state.UpsertJob(structs.MsgTypeTestSetup, idx, nsJob)
|
|
require.NoError(t, err)
|
|
idx++
|
|
|
|
// Create evaluations
|
|
evalIDs, index, err := s1.staticEndpoints.Node.createNodeEvals(node, 1)
|
|
require.NoError(t, err)
|
|
require.NotZero(t, index)
|
|
require.Len(t, evalIDs, 1)
|
|
|
|
eval, err := state.EvalByID(nil, evalIDs[0])
|
|
require.NoError(t, err)
|
|
require.Equal(t, defaultJob.ID, eval.JobID)
|
|
}
|
|
|
|
func TestClientEndpoint_Evaluate(t *testing.T) {
|
|
ci.Parallel(t)
|
|
|
|
s1, cleanupS1 := TestServer(t, func(c *Config) {
|
|
c.NumSchedulers = 0 // Prevent automatic dequeue
|
|
})
|
|
defer cleanupS1()
|
|
codec := rpcClient(t, s1)
|
|
testutil.WaitForLeader(t, s1.RPC)
|
|
|
|
// Inject fake evaluations
|
|
alloc := mock.Alloc()
|
|
node := mock.Node()
|
|
node.ID = alloc.NodeID
|
|
state := s1.fsm.State()
|
|
err := state.UpsertNode(structs.MsgTypeTestSetup, 1, node)
|
|
if err != nil {
|
|
t.Fatalf("err: %v", err)
|
|
}
|
|
state.UpsertJobSummary(2, mock.JobSummary(alloc.JobID))
|
|
err = state.UpsertAllocs(structs.MsgTypeTestSetup, 3, []*structs.Allocation{alloc})
|
|
if err != nil {
|
|
t.Fatalf("err: %v", err)
|
|
}
|
|
|
|
// Re-evaluate
|
|
req := &structs.NodeEvaluateRequest{
|
|
NodeID: alloc.NodeID,
|
|
WriteRequest: structs.WriteRequest{Region: "global"},
|
|
}
|
|
|
|
// Fetch the response
|
|
var resp structs.NodeUpdateResponse
|
|
if err := msgpackrpc.CallWithCodec(codec, "Node.Evaluate", req, &resp); err != nil {
|
|
t.Fatalf("err: %v", err)
|
|
}
|
|
if resp.Index == 0 {
|
|
t.Fatalf("bad index: %d", resp.Index)
|
|
}
|
|
|
|
// Create some evaluations
|
|
ids := resp.EvalIDs
|
|
if len(ids) != 1 {
|
|
t.Fatalf("bad: %s", ids)
|
|
}
|
|
|
|
// Lookup the evaluation
|
|
ws := memdb.NewWatchSet()
|
|
eval, err := state.EvalByID(ws, ids[0])
|
|
if err != nil {
|
|
t.Fatalf("err: %v", err)
|
|
}
|
|
if eval == nil {
|
|
t.Fatalf("expected eval")
|
|
}
|
|
if eval.CreateIndex != resp.Index {
|
|
t.Fatalf("index mis-match")
|
|
}
|
|
|
|
if eval.Priority != alloc.Job.Priority {
|
|
t.Fatalf("bad: %#v", eval)
|
|
}
|
|
if eval.Type != alloc.Job.Type {
|
|
t.Fatalf("bad: %#v", eval)
|
|
}
|
|
if eval.TriggeredBy != structs.EvalTriggerNodeUpdate {
|
|
t.Fatalf("bad: %#v", eval)
|
|
}
|
|
if eval.JobID != alloc.JobID {
|
|
t.Fatalf("bad: %#v", eval)
|
|
}
|
|
if eval.NodeID != alloc.NodeID {
|
|
t.Fatalf("bad: %#v", eval)
|
|
}
|
|
if eval.NodeModifyIndex != 1 {
|
|
t.Fatalf("bad: %#v", eval)
|
|
}
|
|
if eval.Status != structs.EvalStatusPending {
|
|
t.Fatalf("bad: %#v", eval)
|
|
}
|
|
if eval.CreateTime == 0 {
|
|
t.Fatalf("CreateTime is unset: %#v", eval)
|
|
}
|
|
if eval.ModifyTime == 0 {
|
|
t.Fatalf("ModifyTime is unset: %#v", eval)
|
|
}
|
|
}
|
|
|
|
func TestClientEndpoint_Evaluate_ACL(t *testing.T) {
|
|
ci.Parallel(t)
|
|
|
|
s1, root, cleanupS1 := TestACLServer(t, nil)
|
|
defer cleanupS1()
|
|
codec := rpcClient(t, s1)
|
|
testutil.WaitForLeader(t, s1.RPC)
|
|
assert := assert.New(t)
|
|
|
|
// Create the node with an alloc
|
|
alloc := mock.Alloc()
|
|
node := mock.Node()
|
|
node.ID = alloc.NodeID
|
|
state := s1.fsm.State()
|
|
|
|
assert.Nil(state.UpsertNode(structs.MsgTypeTestSetup, 1, node), "UpsertNode")
|
|
assert.Nil(state.UpsertJobSummary(2, mock.JobSummary(alloc.JobID)), "UpsertJobSummary")
|
|
assert.Nil(state.UpsertAllocs(structs.MsgTypeTestSetup, 3, []*structs.Allocation{alloc}), "UpsertAllocs")
|
|
|
|
// Create the policy and tokens
|
|
validToken := mock.CreatePolicyAndToken(t, state, 1001, "test-valid", mock.NodePolicy(acl.PolicyWrite))
|
|
invalidToken := mock.CreatePolicyAndToken(t, state, 1003, "test-invalid", mock.NodePolicy(acl.PolicyRead))
|
|
|
|
// Re-evaluate without a token and expect failure
|
|
req := &structs.NodeEvaluateRequest{
|
|
NodeID: alloc.NodeID,
|
|
WriteRequest: structs.WriteRequest{Region: "global"},
|
|
}
|
|
{
|
|
var resp structs.NodeUpdateResponse
|
|
err := msgpackrpc.CallWithCodec(codec, "Node.Evaluate", req, &resp)
|
|
assert.NotNil(err, "RPC")
|
|
assert.Equal(err.Error(), structs.ErrPermissionDenied.Error())
|
|
}
|
|
|
|
// Try with a valid token
|
|
req.AuthToken = validToken.SecretID
|
|
{
|
|
var resp structs.NodeUpdateResponse
|
|
assert.Nil(msgpackrpc.CallWithCodec(codec, "Node.Evaluate", req, &resp), "RPC")
|
|
}
|
|
|
|
// Try with a invalid token
|
|
req.AuthToken = invalidToken.SecretID
|
|
{
|
|
var resp structs.NodeUpdateResponse
|
|
err := msgpackrpc.CallWithCodec(codec, "Node.Evaluate", req, &resp)
|
|
assert.NotNil(err, "RPC")
|
|
assert.Equal(err.Error(), structs.ErrPermissionDenied.Error())
|
|
}
|
|
|
|
// Try with a root token
|
|
req.AuthToken = root.SecretID
|
|
{
|
|
var resp structs.NodeUpdateResponse
|
|
assert.Nil(msgpackrpc.CallWithCodec(codec, "Node.Evaluate", req, &resp), "RPC")
|
|
}
|
|
}
|
|
|
|
func TestClientEndpoint_ListNodes(t *testing.T) {
|
|
ci.Parallel(t)
|
|
|
|
s1, cleanupS1 := TestServer(t, nil)
|
|
defer cleanupS1()
|
|
codec := rpcClient(t, s1)
|
|
testutil.WaitForLeader(t, s1.RPC)
|
|
|
|
// Create the register request
|
|
node := mock.Node()
|
|
node.HostVolumes = map[string]*structs.ClientHostVolumeConfig{
|
|
"foo": {
|
|
Name: "foo",
|
|
Path: "/",
|
|
ReadOnly: true,
|
|
},
|
|
}
|
|
reg := &structs.NodeRegisterRequest{
|
|
Node: node,
|
|
WriteRequest: structs.WriteRequest{Region: "global"},
|
|
}
|
|
|
|
// Fetch the response
|
|
var resp structs.GenericResponse
|
|
if err := msgpackrpc.CallWithCodec(codec, "Node.Register", reg, &resp); err != nil {
|
|
t.Fatalf("err: %v", err)
|
|
}
|
|
node.CreateIndex = resp.Index
|
|
node.ModifyIndex = resp.Index
|
|
|
|
// Lookup the node
|
|
get := &structs.NodeListRequest{
|
|
QueryOptions: structs.QueryOptions{Region: "global"},
|
|
}
|
|
var resp2 structs.NodeListResponse
|
|
if err := msgpackrpc.CallWithCodec(codec, "Node.List", get, &resp2); err != nil {
|
|
t.Fatalf("err: %v", err)
|
|
}
|
|
if resp2.Index != resp.Index {
|
|
t.Fatalf("Bad index: %d %d", resp2.Index, resp.Index)
|
|
}
|
|
|
|
require.Len(t, resp2.Nodes, 1)
|
|
require.Equal(t, node.ID, resp2.Nodes[0].ID)
|
|
|
|
// #7344 - Assert HostVolumes are included in stub
|
|
require.Equal(t, node.HostVolumes, resp2.Nodes[0].HostVolumes)
|
|
|
|
// #9055 - Assert Resources are *not* included by default
|
|
require.Nil(t, resp2.Nodes[0].NodeResources)
|
|
require.Nil(t, resp2.Nodes[0].ReservedResources)
|
|
|
|
// Lookup the node with prefix
|
|
get = &structs.NodeListRequest{
|
|
QueryOptions: structs.QueryOptions{Region: "global", Prefix: node.ID[:4]},
|
|
}
|
|
var resp3 structs.NodeListResponse
|
|
if err := msgpackrpc.CallWithCodec(codec, "Node.List", get, &resp3); err != nil {
|
|
t.Fatalf("err: %v", err)
|
|
}
|
|
if resp3.Index != resp.Index {
|
|
t.Fatalf("Bad index: %d %d", resp3.Index, resp2.Index)
|
|
}
|
|
|
|
if len(resp3.Nodes) != 1 {
|
|
t.Fatalf("bad: %#v", resp3.Nodes)
|
|
}
|
|
if resp3.Nodes[0].ID != node.ID {
|
|
t.Fatalf("bad: %#v", resp3.Nodes[0])
|
|
}
|
|
}
|
|
|
|
func TestClientEndpoint_ListNodes_Fields(t *testing.T) {
|
|
ci.Parallel(t)
|
|
|
|
s1, cleanupS1 := TestServer(t, nil)
|
|
defer cleanupS1()
|
|
codec := rpcClient(t, s1)
|
|
testutil.WaitForLeader(t, s1.RPC)
|
|
|
|
// Create the register request
|
|
node := mock.Node()
|
|
reg := &structs.NodeRegisterRequest{
|
|
Node: node,
|
|
WriteRequest: structs.WriteRequest{Region: "global"},
|
|
}
|
|
|
|
// Fetch the response
|
|
var resp structs.GenericResponse
|
|
require.NoError(t, msgpackrpc.CallWithCodec(codec, "Node.Register", reg, &resp))
|
|
node.CreateIndex = resp.Index
|
|
node.ModifyIndex = resp.Index
|
|
|
|
// Lookup the node with fields
|
|
get := &structs.NodeListRequest{
|
|
QueryOptions: structs.QueryOptions{Region: "global"},
|
|
Fields: &structs.NodeStubFields{
|
|
Resources: true,
|
|
},
|
|
}
|
|
var resp2 structs.NodeListResponse
|
|
require.NoError(t, msgpackrpc.CallWithCodec(codec, "Node.List", get, &resp2))
|
|
require.Equal(t, resp.Index, resp2.Index)
|
|
require.Len(t, resp2.Nodes, 1)
|
|
require.Equal(t, node.ID, resp2.Nodes[0].ID)
|
|
require.NotNil(t, resp2.Nodes[0].NodeResources)
|
|
require.NotNil(t, resp2.Nodes[0].ReservedResources)
|
|
}
|
|
|
|
func TestClientEndpoint_ListNodes_ACL(t *testing.T) {
|
|
ci.Parallel(t)
|
|
|
|
s1, root, cleanupS1 := TestACLServer(t, nil)
|
|
defer cleanupS1()
|
|
codec := rpcClient(t, s1)
|
|
testutil.WaitForLeader(t, s1.RPC)
|
|
assert := assert.New(t)
|
|
|
|
// Create the node
|
|
node := mock.Node()
|
|
state := s1.fsm.State()
|
|
assert.Nil(state.UpsertNode(structs.MsgTypeTestSetup, 1, node), "UpsertNode")
|
|
|
|
// Create the namespace policy and tokens
|
|
validToken := mock.CreatePolicyAndToken(t, state, 1001, "test-valid", mock.NodePolicy(acl.PolicyRead))
|
|
invalidToken := mock.CreatePolicyAndToken(t, state, 1003, "test-invalid", mock.NodePolicy(acl.PolicyDeny))
|
|
|
|
// Lookup the node without a token and expect failure
|
|
req := &structs.NodeListRequest{
|
|
QueryOptions: structs.QueryOptions{Region: "global"},
|
|
}
|
|
{
|
|
var resp structs.NodeListResponse
|
|
err := msgpackrpc.CallWithCodec(codec, "Node.List", req, &resp)
|
|
assert.NotNil(err, "RPC")
|
|
assert.Equal(err.Error(), structs.ErrPermissionDenied.Error())
|
|
}
|
|
|
|
// Try with a valid token
|
|
req.AuthToken = validToken.SecretID
|
|
{
|
|
var resp structs.NodeListResponse
|
|
assert.Nil(msgpackrpc.CallWithCodec(codec, "Node.List", req, &resp), "RPC")
|
|
assert.Equal(node.ID, resp.Nodes[0].ID)
|
|
}
|
|
|
|
// Try with a invalid token
|
|
req.AuthToken = invalidToken.SecretID
|
|
{
|
|
var resp structs.NodeListResponse
|
|
err := msgpackrpc.CallWithCodec(codec, "Node.List", req, &resp)
|
|
assert.NotNil(err, "RPC")
|
|
assert.Equal(err.Error(), structs.ErrPermissionDenied.Error())
|
|
}
|
|
|
|
// Try with a root token
|
|
req.AuthToken = root.SecretID
|
|
{
|
|
var resp structs.NodeListResponse
|
|
assert.Nil(msgpackrpc.CallWithCodec(codec, "Node.List", req, &resp), "RPC")
|
|
assert.Equal(node.ID, resp.Nodes[0].ID)
|
|
}
|
|
}
|
|
|
|
func TestClientEndpoint_ListNodes_Blocking(t *testing.T) {
|
|
ci.Parallel(t)
|
|
|
|
s1, cleanupS1 := TestServer(t, nil)
|
|
defer cleanupS1()
|
|
state := s1.fsm.State()
|
|
codec := rpcClient(t, s1)
|
|
testutil.WaitForLeader(t, s1.RPC)
|
|
|
|
// Disable drainer to prevent drain from completing during test
|
|
s1.nodeDrainer.SetEnabled(false, nil)
|
|
|
|
// Create the node
|
|
node := mock.Node()
|
|
|
|
// Node upsert triggers watches
|
|
errCh := make(chan error, 1)
|
|
timer := time.AfterFunc(100*time.Millisecond, func() {
|
|
errCh <- state.UpsertNode(structs.MsgTypeTestSetup, 2, node)
|
|
})
|
|
defer timer.Stop()
|
|
|
|
req := &structs.NodeListRequest{
|
|
QueryOptions: structs.QueryOptions{
|
|
Region: "global",
|
|
MinQueryIndex: 1,
|
|
},
|
|
}
|
|
start := time.Now()
|
|
var resp structs.NodeListResponse
|
|
if err := msgpackrpc.CallWithCodec(codec, "Node.List", req, &resp); err != nil {
|
|
t.Fatalf("err: %v", err)
|
|
}
|
|
|
|
if err := <-errCh; err != nil {
|
|
t.Fatalf("error from timer: %v", err)
|
|
}
|
|
|
|
if elapsed := time.Since(start); elapsed < 100*time.Millisecond {
|
|
t.Fatalf("should block (returned in %s) %#v", elapsed, resp)
|
|
}
|
|
if resp.Index != 2 {
|
|
t.Fatalf("Bad index: %d %d", resp.Index, 2)
|
|
}
|
|
if len(resp.Nodes) != 1 || resp.Nodes[0].ID != node.ID {
|
|
t.Fatalf("bad: %#v", resp.Nodes)
|
|
}
|
|
|
|
// Node drain updates trigger watches.
|
|
time.AfterFunc(100*time.Millisecond, func() {
|
|
s := &structs.DrainStrategy{
|
|
DrainSpec: structs.DrainSpec{
|
|
Deadline: 10 * time.Second,
|
|
},
|
|
}
|
|
errCh <- state.UpdateNodeDrain(structs.MsgTypeTestSetup, 3, node.ID, s, false, 0, nil, nil, "")
|
|
})
|
|
|
|
req.MinQueryIndex = 2
|
|
var resp2 structs.NodeListResponse
|
|
start = time.Now()
|
|
if err := msgpackrpc.CallWithCodec(codec, "Node.List", req, &resp2); err != nil {
|
|
t.Fatalf("err: %v", err)
|
|
}
|
|
|
|
if err := <-errCh; err != nil {
|
|
t.Fatalf("error from timer: %v", err)
|
|
}
|
|
|
|
if elapsed := time.Since(start); elapsed < 100*time.Millisecond {
|
|
t.Fatalf("should block (returned in %s) %#v", elapsed, resp2)
|
|
}
|
|
if resp2.Index != 3 {
|
|
t.Fatalf("Bad index: %d %d", resp2.Index, 3)
|
|
}
|
|
if len(resp2.Nodes) != 1 || !resp2.Nodes[0].Drain {
|
|
t.Fatalf("bad: %#v", resp2.Nodes)
|
|
}
|
|
|
|
// Node status update triggers watches
|
|
time.AfterFunc(100*time.Millisecond, func() {
|
|
errCh <- state.UpdateNodeStatus(structs.MsgTypeTestSetup, 40, node.ID, structs.NodeStatusDown, 0, nil)
|
|
})
|
|
|
|
req.MinQueryIndex = 38
|
|
var resp3 structs.NodeListResponse
|
|
start = time.Now()
|
|
if err := msgpackrpc.CallWithCodec(codec, "Node.List", req, &resp3); err != nil {
|
|
t.Fatalf("err: %v", err)
|
|
}
|
|
|
|
if err := <-errCh; err != nil {
|
|
t.Fatalf("error from timer: %v", err)
|
|
}
|
|
|
|
if elapsed := time.Since(start); elapsed < 100*time.Millisecond {
|
|
t.Fatalf("should block (returned in %s) %#v", elapsed, resp3)
|
|
}
|
|
if resp3.Index != 40 {
|
|
t.Fatalf("Bad index: %d %d", resp3.Index, 40)
|
|
}
|
|
if len(resp3.Nodes) != 1 || resp3.Nodes[0].Status != structs.NodeStatusDown {
|
|
t.Fatalf("bad: %#v", resp3.Nodes)
|
|
}
|
|
|
|
// Node delete triggers watches.
|
|
time.AfterFunc(100*time.Millisecond, func() {
|
|
errCh <- state.DeleteNode(structs.MsgTypeTestSetup, 50, []string{node.ID})
|
|
})
|
|
|
|
req.MinQueryIndex = 45
|
|
var resp4 structs.NodeListResponse
|
|
start = time.Now()
|
|
if err := msgpackrpc.CallWithCodec(codec, "Node.List", req, &resp4); err != nil {
|
|
t.Fatalf("err: %v", err)
|
|
}
|
|
|
|
if err := <-errCh; err != nil {
|
|
t.Fatalf("error from timer: %v", err)
|
|
}
|
|
|
|
if elapsed := time.Since(start); elapsed < 100*time.Millisecond {
|
|
t.Fatalf("should block (returned in %s) %#v", elapsed, resp4)
|
|
}
|
|
if resp4.Index != 50 {
|
|
t.Fatalf("Bad index: %d %d", resp4.Index, 50)
|
|
}
|
|
if len(resp4.Nodes) != 0 {
|
|
t.Fatalf("bad: %#v", resp4.Nodes)
|
|
}
|
|
}
|
|
|
|
func TestClientEndpoint_DeriveVaultToken_Bad(t *testing.T) {
|
|
ci.Parallel(t)
|
|
|
|
s1, cleanupS1 := TestServer(t, nil)
|
|
defer cleanupS1()
|
|
state := s1.fsm.State()
|
|
codec := rpcClient(t, s1)
|
|
testutil.WaitForLeader(t, s1.RPC)
|
|
|
|
// Create the node
|
|
node := mock.Node()
|
|
if err := state.UpsertNode(structs.MsgTypeTestSetup, 2, node); err != nil {
|
|
t.Fatalf("err: %v", err)
|
|
}
|
|
|
|
// Create an alloc
|
|
alloc := mock.Alloc()
|
|
task := alloc.Job.TaskGroups[0].Tasks[0]
|
|
tasks := []string{task.Name}
|
|
if err := state.UpsertAllocs(structs.MsgTypeTestSetup, 3, []*structs.Allocation{alloc}); err != nil {
|
|
t.Fatalf("err: %v", err)
|
|
}
|
|
|
|
req := &structs.DeriveVaultTokenRequest{
|
|
NodeID: node.ID,
|
|
SecretID: uuid.Generate(),
|
|
AllocID: alloc.ID,
|
|
Tasks: tasks,
|
|
QueryOptions: structs.QueryOptions{
|
|
Region: "global",
|
|
},
|
|
}
|
|
|
|
var resp structs.DeriveVaultTokenResponse
|
|
if err := msgpackrpc.CallWithCodec(codec, "Node.DeriveVaultToken", req, &resp); err != nil {
|
|
t.Fatalf("bad: %v", err)
|
|
}
|
|
|
|
if resp.Error == nil || !strings.Contains(resp.Error.Error(), "SecretID mismatch") {
|
|
t.Fatalf("Expected SecretID mismatch: %v", resp.Error)
|
|
}
|
|
|
|
// Put the correct SecretID
|
|
req.SecretID = node.SecretID
|
|
|
|
// Now we should get an error about the allocation not running on the node
|
|
if err := msgpackrpc.CallWithCodec(codec, "Node.DeriveVaultToken", req, &resp); err != nil {
|
|
t.Fatalf("bad: %v", err)
|
|
}
|
|
if resp.Error == nil || !strings.Contains(resp.Error.Error(), "not running on Node") {
|
|
t.Fatalf("Expected not running on node error: %v", resp.Error)
|
|
}
|
|
|
|
// Update to be running on the node
|
|
alloc.NodeID = node.ID
|
|
if err := state.UpsertAllocs(structs.MsgTypeTestSetup, 4, []*structs.Allocation{alloc}); err != nil {
|
|
t.Fatalf("err: %v", err)
|
|
}
|
|
|
|
// Now we should get an error about the job not needing any Vault secrets
|
|
if err := msgpackrpc.CallWithCodec(codec, "Node.DeriveVaultToken", req, &resp); err != nil {
|
|
t.Fatalf("bad: %v", err)
|
|
}
|
|
if resp.Error == nil || !strings.Contains(resp.Error.Error(), "does not require") {
|
|
t.Fatalf("Expected no policies error: %v", resp.Error)
|
|
}
|
|
|
|
// Update to be terminal
|
|
alloc.DesiredStatus = structs.AllocDesiredStatusStop
|
|
if err := state.UpsertAllocs(structs.MsgTypeTestSetup, 5, []*structs.Allocation{alloc}); err != nil {
|
|
t.Fatalf("err: %v", err)
|
|
}
|
|
|
|
// Now we should get an error about the job not needing any Vault secrets
|
|
if err := msgpackrpc.CallWithCodec(codec, "Node.DeriveVaultToken", req, &resp); err != nil {
|
|
t.Fatalf("bad: %v", err)
|
|
}
|
|
if resp.Error == nil || !strings.Contains(resp.Error.Error(), "terminal") {
|
|
t.Fatalf("Expected terminal allocation error: %v", resp.Error)
|
|
}
|
|
}
|
|
|
|
func TestClientEndpoint_DeriveVaultToken(t *testing.T) {
|
|
ci.Parallel(t)
|
|
|
|
s1, cleanupS1 := TestServer(t, nil)
|
|
defer cleanupS1()
|
|
state := s1.fsm.State()
|
|
codec := rpcClient(t, s1)
|
|
testutil.WaitForLeader(t, s1.RPC)
|
|
|
|
// Enable vault and allow authenticated
|
|
tr := true
|
|
s1.config.VaultConfig.Enabled = &tr
|
|
s1.config.VaultConfig.AllowUnauthenticated = &tr
|
|
|
|
// Replace the Vault Client on the server
|
|
tvc := &TestVaultClient{}
|
|
s1.vault = tvc
|
|
|
|
// Create the node
|
|
node := mock.Node()
|
|
if err := state.UpsertNode(structs.MsgTypeTestSetup, 2, node); err != nil {
|
|
t.Fatalf("err: %v", err)
|
|
}
|
|
|
|
// Create an alloc an allocation that has vault policies required
|
|
alloc := mock.Alloc()
|
|
alloc.NodeID = node.ID
|
|
task := alloc.Job.TaskGroups[0].Tasks[0]
|
|
tasks := []string{task.Name}
|
|
task.Vault = &structs.Vault{Policies: []string{"a", "b"}}
|
|
if err := state.UpsertAllocs(structs.MsgTypeTestSetup, 3, []*structs.Allocation{alloc}); err != nil {
|
|
t.Fatalf("err: %v", err)
|
|
}
|
|
|
|
// Return a secret for the task
|
|
token := uuid.Generate()
|
|
accessor := uuid.Generate()
|
|
ttl := 10
|
|
secret := &vapi.Secret{
|
|
WrapInfo: &vapi.SecretWrapInfo{
|
|
Token: token,
|
|
WrappedAccessor: accessor,
|
|
TTL: ttl,
|
|
},
|
|
}
|
|
tvc.SetCreateTokenSecret(alloc.ID, task.Name, secret)
|
|
|
|
req := &structs.DeriveVaultTokenRequest{
|
|
NodeID: node.ID,
|
|
SecretID: node.SecretID,
|
|
AllocID: alloc.ID,
|
|
Tasks: tasks,
|
|
QueryOptions: structs.QueryOptions{
|
|
Region: "global",
|
|
},
|
|
}
|
|
|
|
var resp structs.DeriveVaultTokenResponse
|
|
if err := msgpackrpc.CallWithCodec(codec, "Node.DeriveVaultToken", req, &resp); err != nil {
|
|
t.Fatalf("bad: %v", err)
|
|
}
|
|
if resp.Error != nil {
|
|
t.Fatalf("bad: %v", resp.Error)
|
|
}
|
|
|
|
// Check the state store and ensure that we created a VaultAccessor
|
|
ws := memdb.NewWatchSet()
|
|
va, err := state.VaultAccessor(ws, accessor)
|
|
if err != nil {
|
|
t.Fatalf("bad: %v", err)
|
|
}
|
|
if va == nil {
|
|
t.Fatalf("bad: %v", va)
|
|
}
|
|
|
|
if va.CreateIndex == 0 {
|
|
t.Fatalf("bad: %v", va)
|
|
}
|
|
|
|
va.CreateIndex = 0
|
|
expected := &structs.VaultAccessor{
|
|
AllocID: alloc.ID,
|
|
Task: task.Name,
|
|
NodeID: alloc.NodeID,
|
|
Accessor: accessor,
|
|
CreationTTL: ttl,
|
|
}
|
|
|
|
if !reflect.DeepEqual(expected, va) {
|
|
t.Fatalf("Got %#v; want %#v", va, expected)
|
|
}
|
|
}
|
|
|
|
func TestClientEndpoint_DeriveVaultToken_VaultError(t *testing.T) {
|
|
ci.Parallel(t)
|
|
|
|
s1, cleanupS1 := TestServer(t, nil)
|
|
defer cleanupS1()
|
|
state := s1.fsm.State()
|
|
codec := rpcClient(t, s1)
|
|
testutil.WaitForLeader(t, s1.RPC)
|
|
|
|
// Enable vault and allow authenticated
|
|
tr := true
|
|
s1.config.VaultConfig.Enabled = &tr
|
|
s1.config.VaultConfig.AllowUnauthenticated = &tr
|
|
|
|
// Replace the Vault Client on the server
|
|
tvc := &TestVaultClient{}
|
|
s1.vault = tvc
|
|
|
|
// Create the node
|
|
node := mock.Node()
|
|
if err := state.UpsertNode(structs.MsgTypeTestSetup, 2, node); err != nil {
|
|
t.Fatalf("err: %v", err)
|
|
}
|
|
|
|
// Create an alloc an allocation that has vault policies required
|
|
alloc := mock.Alloc()
|
|
alloc.NodeID = node.ID
|
|
task := alloc.Job.TaskGroups[0].Tasks[0]
|
|
tasks := []string{task.Name}
|
|
task.Vault = &structs.Vault{Policies: []string{"a", "b"}}
|
|
if err := state.UpsertAllocs(structs.MsgTypeTestSetup, 3, []*structs.Allocation{alloc}); err != nil {
|
|
t.Fatalf("err: %v", err)
|
|
}
|
|
|
|
// Return an error when creating the token
|
|
tvc.SetCreateTokenError(alloc.ID, task.Name,
|
|
structs.NewRecoverableError(fmt.Errorf("recover"), true))
|
|
|
|
req := &structs.DeriveVaultTokenRequest{
|
|
NodeID: node.ID,
|
|
SecretID: node.SecretID,
|
|
AllocID: alloc.ID,
|
|
Tasks: tasks,
|
|
QueryOptions: structs.QueryOptions{
|
|
Region: "global",
|
|
},
|
|
}
|
|
|
|
var resp structs.DeriveVaultTokenResponse
|
|
err := msgpackrpc.CallWithCodec(codec, "Node.DeriveVaultToken", req, &resp)
|
|
if err != nil {
|
|
t.Fatalf("bad: %v", err)
|
|
}
|
|
if resp.Error == nil || !resp.Error.IsRecoverable() {
|
|
t.Fatalf("bad: %+v", resp.Error)
|
|
}
|
|
}
|
|
|
|
func TestClientEndpoint_taskUsesConnect(t *testing.T) {
|
|
ci.Parallel(t)
|
|
|
|
try := func(t *testing.T, task *structs.Task, exp bool) {
|
|
result := taskUsesConnect(task)
|
|
require.Equal(t, exp, result)
|
|
}
|
|
|
|
t.Run("task uses connect", func(t *testing.T) {
|
|
try(t, &structs.Task{
|
|
// see nomad.newConnectSidecarTask for how this works
|
|
Name: "connect-proxy-myservice",
|
|
Kind: "connect-proxy:myservice",
|
|
}, true)
|
|
})
|
|
|
|
t.Run("task does not use connect", func(t *testing.T) {
|
|
try(t, &structs.Task{
|
|
Name: "mytask",
|
|
Kind: "incorrect:mytask",
|
|
}, false)
|
|
})
|
|
|
|
t.Run("task does not exist", func(t *testing.T) {
|
|
try(t, nil, false)
|
|
})
|
|
}
|
|
|
|
func TestClientEndpoint_tasksNotUsingConnect(t *testing.T) {
|
|
ci.Parallel(t)
|
|
|
|
taskGroup := &structs.TaskGroup{
|
|
Name: "testgroup",
|
|
Tasks: []*structs.Task{{
|
|
Name: "connect-proxy-service1",
|
|
Kind: structs.NewTaskKind(structs.ConnectProxyPrefix, "service1"),
|
|
}, {
|
|
Name: "incorrect-task3",
|
|
Kind: "incorrect:task3",
|
|
}, {
|
|
Name: "connect-proxy-service4",
|
|
Kind: structs.NewTaskKind(structs.ConnectProxyPrefix, "service4"),
|
|
}, {
|
|
Name: "incorrect-task5",
|
|
Kind: "incorrect:task5",
|
|
}, {
|
|
Name: "task6",
|
|
Kind: structs.NewTaskKind(structs.ConnectNativePrefix, "service6"),
|
|
}},
|
|
}
|
|
|
|
requestingTasks := []string{
|
|
"connect-proxy-service1", // yes
|
|
"task2", // does not exist
|
|
"task3", // no
|
|
"connect-proxy-service4", // yes
|
|
"task5", // no
|
|
"task6", // yes, native
|
|
}
|
|
|
|
notConnect, usingConnect := connectTasks(taskGroup, requestingTasks)
|
|
|
|
notConnectExp := []string{"task2", "task3", "task5"}
|
|
usingConnectExp := []connectTask{
|
|
{TaskName: "connect-proxy-service1", TaskKind: "connect-proxy:service1"},
|
|
{TaskName: "connect-proxy-service4", TaskKind: "connect-proxy:service4"},
|
|
{TaskName: "task6", TaskKind: "connect-native:service6"},
|
|
}
|
|
|
|
require.Equal(t, notConnectExp, notConnect)
|
|
require.Equal(t, usingConnectExp, usingConnect)
|
|
}
|
|
|
|
func mutateConnectJob(t *testing.T, job *structs.Job) {
|
|
var jch jobConnectHook
|
|
_, warnings, err := jch.Mutate(job)
|
|
require.Empty(t, warnings)
|
|
require.NoError(t, err)
|
|
}
|
|
|
|
func TestClientEndpoint_DeriveSIToken(t *testing.T) {
|
|
ci.Parallel(t)
|
|
r := require.New(t)
|
|
|
|
s1, cleanupS1 := TestServer(t, nil) // already sets consul mocks
|
|
defer cleanupS1()
|
|
|
|
state := s1.fsm.State()
|
|
codec := rpcClient(t, s1)
|
|
testutil.WaitForLeader(t, s1.RPC)
|
|
|
|
// Set allow unauthenticated (no operator token required)
|
|
s1.config.ConsulConfig.AllowUnauthenticated = pointer.Of(true)
|
|
|
|
// Create the node
|
|
node := mock.Node()
|
|
err := state.UpsertNode(structs.MsgTypeTestSetup, 2, node)
|
|
r.NoError(err)
|
|
|
|
// Create an alloc with a typical connect service (sidecar) defined
|
|
alloc := mock.ConnectAlloc()
|
|
alloc.NodeID = node.ID
|
|
mutateConnectJob(t, alloc.Job) // appends sidecar task
|
|
sidecarTask := alloc.Job.TaskGroups[0].Tasks[1]
|
|
|
|
err = state.UpsertAllocs(structs.MsgTypeTestSetup, 3, []*structs.Allocation{alloc})
|
|
r.NoError(err)
|
|
|
|
request := &structs.DeriveSITokenRequest{
|
|
NodeID: node.ID,
|
|
SecretID: node.SecretID,
|
|
AllocID: alloc.ID,
|
|
Tasks: []string{sidecarTask.Name},
|
|
QueryOptions: structs.QueryOptions{Region: "global"},
|
|
}
|
|
|
|
var response structs.DeriveSITokenResponse
|
|
err = msgpackrpc.CallWithCodec(codec, "Node.DeriveSIToken", request, &response)
|
|
r.NoError(err)
|
|
r.Nil(response.Error)
|
|
|
|
// Check the state store and ensure we created a Consul SI Token Accessor
|
|
ws := memdb.NewWatchSet()
|
|
accessors, err := state.SITokenAccessorsByNode(ws, node.ID)
|
|
r.NoError(err)
|
|
r.Equal(1, len(accessors)) // only asked for one
|
|
r.Equal("connect-proxy-testconnect", accessors[0].TaskName) // set by the mock
|
|
r.Equal(node.ID, accessors[0].NodeID) // should match
|
|
r.Equal(alloc.ID, accessors[0].AllocID) // should match
|
|
r.True(helper.IsUUID(accessors[0].AccessorID)) // should be set
|
|
r.Greater(accessors[0].CreateIndex, uint64(3)) // more than 3rd
|
|
}
|
|
|
|
func TestClientEndpoint_DeriveSIToken_ConsulError(t *testing.T) {
|
|
ci.Parallel(t)
|
|
r := require.New(t)
|
|
|
|
s1, cleanupS1 := TestServer(t, nil)
|
|
defer cleanupS1()
|
|
state := s1.fsm.State()
|
|
codec := rpcClient(t, s1)
|
|
testutil.WaitForLeader(t, s1.RPC)
|
|
|
|
// Set allow unauthenticated (no operator token required)
|
|
s1.config.ConsulConfig.AllowUnauthenticated = pointer.Of(true)
|
|
|
|
// Create the node
|
|
node := mock.Node()
|
|
err := state.UpsertNode(structs.MsgTypeTestSetup, 2, node)
|
|
r.NoError(err)
|
|
|
|
// Create an alloc with a typical connect service (sidecar) defined
|
|
alloc := mock.ConnectAlloc()
|
|
alloc.NodeID = node.ID
|
|
mutateConnectJob(t, alloc.Job) // appends sidecar task
|
|
sidecarTask := alloc.Job.TaskGroups[0].Tasks[1]
|
|
|
|
// rejigger the server to use a broken mock consul
|
|
mockACLsAPI := consul.NewMockACLsAPI(s1.logger)
|
|
mockACLsAPI.SetError(structs.NewRecoverableError(errors.New("consul recoverable error"), true))
|
|
m := NewConsulACLsAPI(mockACLsAPI, s1.logger, nil)
|
|
s1.consulACLs = m
|
|
|
|
err = state.UpsertAllocs(structs.MsgTypeTestSetup, 3, []*structs.Allocation{alloc})
|
|
r.NoError(err)
|
|
|
|
request := &structs.DeriveSITokenRequest{
|
|
NodeID: node.ID,
|
|
SecretID: node.SecretID,
|
|
AllocID: alloc.ID,
|
|
Tasks: []string{sidecarTask.Name},
|
|
QueryOptions: structs.QueryOptions{Region: "global"},
|
|
}
|
|
|
|
var response structs.DeriveSITokenResponse
|
|
err = msgpackrpc.CallWithCodec(codec, "Node.DeriveSIToken", request, &response)
|
|
r.NoError(err)
|
|
r.NotNil(response.Error) // error should be set
|
|
r.True(response.Error.IsRecoverable()) // and is recoverable
|
|
}
|
|
|
|
func TestClientEndpoint_EmitEvents(t *testing.T) {
|
|
ci.Parallel(t)
|
|
require := require.New(t)
|
|
|
|
s1, cleanupS1 := TestServer(t, nil)
|
|
defer cleanupS1()
|
|
state := s1.fsm.State()
|
|
codec := rpcClient(t, s1)
|
|
testutil.WaitForLeader(t, s1.RPC)
|
|
|
|
// create a node that we can register our event to
|
|
node := mock.Node()
|
|
err := state.UpsertNode(structs.MsgTypeTestSetup, 2, node)
|
|
require.Nil(err)
|
|
|
|
nodeEvent := &structs.NodeEvent{
|
|
Message: "Registration failed",
|
|
Subsystem: "Server",
|
|
Timestamp: time.Now(),
|
|
}
|
|
|
|
nodeEvents := map[string][]*structs.NodeEvent{node.ID: {nodeEvent}}
|
|
req := structs.EmitNodeEventsRequest{
|
|
NodeEvents: nodeEvents,
|
|
WriteRequest: structs.WriteRequest{Region: "global"},
|
|
}
|
|
|
|
var resp structs.GenericResponse
|
|
err = msgpackrpc.CallWithCodec(codec, "Node.EmitEvents", &req, &resp)
|
|
require.Nil(err)
|
|
require.NotEqual(uint64(0), resp.Index)
|
|
|
|
// Check for the node in the FSM
|
|
ws := memdb.NewWatchSet()
|
|
out, err := state.NodeByID(ws, node.ID)
|
|
require.Nil(err)
|
|
require.False(len(out.Events) < 2)
|
|
}
|
|
|
|
func TestClientEndpoint_ShouldCreateNodeEval(t *testing.T) {
|
|
ci.Parallel(t)
|
|
|
|
t.Run("spurious changes don't require eval", func(t *testing.T) {
|
|
n1 := mock.Node()
|
|
n2 := n1.Copy()
|
|
n2.SecretID = uuid.Generate()
|
|
n2.Links["vault"] = "links don't get interpolated"
|
|
n2.ModifyIndex++
|
|
|
|
require.False(t, shouldCreateNodeEval(n1, n2))
|
|
})
|
|
|
|
positiveCases := []struct {
|
|
name string
|
|
updateFn func(n *structs.Node)
|
|
}{
|
|
{
|
|
"data center changes",
|
|
func(n *structs.Node) { n.Datacenter += "u" },
|
|
},
|
|
{
|
|
"attribute change",
|
|
func(n *structs.Node) { n.Attributes["test.attribute"] = "something" },
|
|
},
|
|
{
|
|
"meta change",
|
|
func(n *structs.Node) { n.Meta["test.meta"] = "something" },
|
|
},
|
|
{
|
|
"drivers health changed",
|
|
func(n *structs.Node) { n.Drivers["exec"].Detected = false },
|
|
},
|
|
{
|
|
"new drivers",
|
|
func(n *structs.Node) {
|
|
n.Drivers["newdriver"] = &structs.DriverInfo{
|
|
Detected: true,
|
|
Healthy: true,
|
|
}
|
|
},
|
|
},
|
|
}
|
|
|
|
for _, c := range positiveCases {
|
|
t.Run(c.name, func(t *testing.T) {
|
|
n1 := mock.Node()
|
|
n2 := n1.Copy()
|
|
c.updateFn(n2)
|
|
|
|
require.Truef(t, shouldCreateNodeEval(n1, n2), "node changed but without node eval: %v", pretty.Diff(n1, n2))
|
|
})
|
|
}
|
|
}
|
|
|
|
func TestClientEndpoint_UpdateAlloc_Evals_ByTrigger(t *testing.T) {
|
|
t.Parallel()
|
|
|
|
type testCase struct {
|
|
name string
|
|
clientStatus string
|
|
serverClientStatus string
|
|
triggerBy string
|
|
missingJob bool
|
|
missingAlloc bool
|
|
invalidTaskGroup bool
|
|
}
|
|
|
|
testCases := []testCase{
|
|
{
|
|
name: "failed-alloc",
|
|
clientStatus: structs.AllocClientStatusFailed,
|
|
serverClientStatus: structs.AllocClientStatusRunning,
|
|
triggerBy: structs.EvalTriggerRetryFailedAlloc,
|
|
missingJob: false,
|
|
missingAlloc: false,
|
|
invalidTaskGroup: false,
|
|
},
|
|
{
|
|
name: "unknown-alloc",
|
|
clientStatus: structs.AllocClientStatusRunning,
|
|
serverClientStatus: structs.AllocClientStatusUnknown,
|
|
triggerBy: structs.EvalTriggerReconnect,
|
|
missingJob: false,
|
|
missingAlloc: false,
|
|
invalidTaskGroup: false,
|
|
},
|
|
{
|
|
name: "orphaned-unknown-alloc",
|
|
clientStatus: structs.AllocClientStatusRunning,
|
|
serverClientStatus: structs.AllocClientStatusUnknown,
|
|
triggerBy: structs.EvalTriggerJobDeregister,
|
|
missingJob: true,
|
|
missingAlloc: false,
|
|
invalidTaskGroup: false,
|
|
},
|
|
{
|
|
name: "running-job",
|
|
clientStatus: structs.AllocClientStatusRunning,
|
|
serverClientStatus: structs.AllocClientStatusRunning,
|
|
triggerBy: "",
|
|
missingJob: false,
|
|
missingAlloc: false,
|
|
invalidTaskGroup: false,
|
|
},
|
|
{
|
|
name: "complete-job",
|
|
clientStatus: structs.AllocClientStatusComplete,
|
|
serverClientStatus: structs.AllocClientStatusComplete,
|
|
triggerBy: "",
|
|
missingJob: false,
|
|
missingAlloc: false,
|
|
invalidTaskGroup: false,
|
|
},
|
|
{
|
|
name: "no-alloc-at-server",
|
|
clientStatus: structs.AllocClientStatusUnknown,
|
|
serverClientStatus: "",
|
|
triggerBy: "",
|
|
missingJob: false,
|
|
missingAlloc: true,
|
|
invalidTaskGroup: false,
|
|
},
|
|
{
|
|
name: "invalid-task-group",
|
|
clientStatus: structs.AllocClientStatusUnknown,
|
|
serverClientStatus: structs.AllocClientStatusRunning,
|
|
triggerBy: "",
|
|
missingJob: false,
|
|
missingAlloc: false,
|
|
invalidTaskGroup: true,
|
|
},
|
|
}
|
|
|
|
for _, tc := range testCases {
|
|
t.Run(tc.name, func(t *testing.T) {
|
|
s1, cleanupS1 := TestServer(t, func(c *Config) {
|
|
// Disabling scheduling in this test so that we can
|
|
// ensure that the state store doesn't accumulate more evals
|
|
// than what we expect the unit test to add
|
|
c.NumSchedulers = 0
|
|
})
|
|
|
|
defer cleanupS1()
|
|
codec := rpcClient(t, s1)
|
|
testutil.WaitForLeader(t, s1.RPC)
|
|
|
|
// Create the register request
|
|
node := mock.Node()
|
|
reg := &structs.NodeRegisterRequest{
|
|
Node: node,
|
|
WriteRequest: structs.WriteRequest{Region: "global"},
|
|
}
|
|
|
|
// Fetch the response
|
|
var nodeResp structs.GenericResponse
|
|
err := msgpackrpc.CallWithCodec(codec, "Node.Register", reg, &nodeResp)
|
|
require.NoError(t, err)
|
|
|
|
fsmState := s1.fsm.State()
|
|
|
|
job := mock.Job()
|
|
job.ID = tc.name + "-test-job"
|
|
|
|
if !tc.missingJob {
|
|
err = fsmState.UpsertJob(structs.MsgTypeTestSetup, 101, job)
|
|
require.NoError(t, err)
|
|
}
|
|
|
|
serverAlloc := mock.Alloc()
|
|
serverAlloc.JobID = job.ID
|
|
serverAlloc.NodeID = node.ID
|
|
serverAlloc.ClientStatus = tc.serverClientStatus
|
|
serverAlloc.TaskGroup = job.TaskGroups[0].Name
|
|
|
|
// Create the incoming client alloc.
|
|
clientAlloc := serverAlloc.Copy()
|
|
clientAlloc.ClientStatus = tc.clientStatus
|
|
|
|
err = fsmState.UpsertJobSummary(99, mock.JobSummary(serverAlloc.JobID))
|
|
require.NoError(t, err)
|
|
|
|
if tc.invalidTaskGroup {
|
|
serverAlloc.TaskGroup = "invalid"
|
|
}
|
|
|
|
if !tc.missingAlloc {
|
|
err = fsmState.UpsertAllocs(structs.MsgTypeTestSetup, 100, []*structs.Allocation{serverAlloc})
|
|
require.NoError(t, err)
|
|
}
|
|
|
|
updateReq := &structs.AllocUpdateRequest{
|
|
Alloc: []*structs.Allocation{clientAlloc},
|
|
WriteRequest: structs.WriteRequest{Region: "global"},
|
|
}
|
|
|
|
var nodeAllocResp structs.NodeAllocsResponse
|
|
err = msgpackrpc.CallWithCodec(codec, "Node.UpdateAlloc", updateReq, &nodeAllocResp)
|
|
require.NoError(t, err)
|
|
require.NotEqual(t, uint64(0), nodeAllocResp.Index)
|
|
|
|
// If no eval should be created validate, none were and return.
|
|
if tc.triggerBy == "" {
|
|
evaluations, err := fsmState.EvalsByJob(nil, job.Namespace, job.ID)
|
|
require.NoError(t, err)
|
|
require.Len(t, evaluations, 0)
|
|
return
|
|
}
|
|
|
|
// Lookup the alloc
|
|
updatedAlloc, err := fsmState.AllocByID(nil, serverAlloc.ID)
|
|
require.NoError(t, err)
|
|
require.Equal(t, tc.clientStatus, updatedAlloc.ClientStatus)
|
|
|
|
// Assert that exactly one eval with test case TriggeredBy exists
|
|
evaluations, err := fsmState.EvalsByJob(nil, job.Namespace, job.ID)
|
|
require.NoError(t, err)
|
|
require.Equal(t, 1, len(evaluations))
|
|
foundCount := 0
|
|
for _, resultEval := range evaluations {
|
|
if resultEval.TriggeredBy == tc.triggerBy && resultEval.WaitUntil.IsZero() {
|
|
foundCount++
|
|
}
|
|
}
|
|
require.Equal(t, 1, foundCount, "Should create exactly one eval for trigger by", tc.triggerBy)
|
|
})
|
|
}
|
|
|
|
}
|
|
|
|
func TestNode_List_PaginationFiltering(t *testing.T) {
|
|
ci.Parallel(t)
|
|
|
|
s1, _, cleanupS1 := TestACLServer(t, nil)
|
|
defer cleanupS1()
|
|
codec := rpcClient(t, s1)
|
|
testutil.WaitForLeader(t, s1.RPC)
|
|
|
|
// Build a set of nodes in various datacenters and states. This allows us
|
|
// to test different filter queries along with pagination.
|
|
mocks := []struct {
|
|
id string
|
|
dc string
|
|
status string
|
|
}{
|
|
{
|
|
id: "aaaa1111-3350-4b4b-d185-0e1992ed43e9",
|
|
dc: "dc2",
|
|
status: structs.NodeStatusDisconnected,
|
|
},
|
|
{
|
|
id: "aaaaaa22-3350-4b4b-d185-0e1992ed43e9",
|
|
dc: "dc1",
|
|
status: structs.NodeStatusReady,
|
|
},
|
|
{
|
|
id: "aaaaaa33-3350-4b4b-d185-0e1992ed43e9",
|
|
dc: "dc3",
|
|
status: structs.NodeStatusReady,
|
|
},
|
|
{
|
|
id: "aaaaaaaa-3350-4b4b-d185-0e1992ed43e9",
|
|
dc: "dc2",
|
|
status: structs.NodeStatusDown,
|
|
},
|
|
{
|
|
id: "aaaaaabb-3350-4b4b-d185-0e1992ed43e9",
|
|
dc: "dc3",
|
|
status: structs.NodeStatusDown,
|
|
},
|
|
{
|
|
id: "aaaaaacc-3350-4b4b-d185-0e1992ed43e9",
|
|
dc: "dc1",
|
|
status: structs.NodeStatusReady,
|
|
},
|
|
}
|
|
|
|
testState := s1.fsm.State()
|
|
|
|
for i, m := range mocks {
|
|
index := 1000 + uint64(i)
|
|
mockNode := mock.Node()
|
|
mockNode.ID = m.id
|
|
mockNode.Datacenter = m.dc
|
|
mockNode.Status = m.status
|
|
mockNode.CreateIndex = index
|
|
require.NoError(t, testState.UpsertNode(structs.MsgTypeTestSetup, index, mockNode))
|
|
}
|
|
|
|
// The server is running with ACLs enabled, so generate an adequate token
|
|
// to use.
|
|
aclToken := mock.CreatePolicyAndToken(t, testState, 1100, "test-valid-read",
|
|
mock.NodePolicy(acl.PolicyRead)).SecretID
|
|
|
|
cases := []struct {
|
|
name string
|
|
filter string
|
|
nextToken string
|
|
pageSize int32
|
|
expectedNextToken string
|
|
expectedIDs []string
|
|
expectedError string
|
|
}{
|
|
{
|
|
name: "pagination no filter",
|
|
pageSize: 2,
|
|
expectedNextToken: "aaaaaa33-3350-4b4b-d185-0e1992ed43e9",
|
|
expectedIDs: []string{
|
|
"aaaa1111-3350-4b4b-d185-0e1992ed43e9",
|
|
"aaaaaa22-3350-4b4b-d185-0e1992ed43e9",
|
|
},
|
|
},
|
|
{
|
|
name: "pagination no filter with next token",
|
|
pageSize: 2,
|
|
nextToken: "aaaaaa33-3350-4b4b-d185-0e1992ed43e9",
|
|
expectedNextToken: "aaaaaabb-3350-4b4b-d185-0e1992ed43e9",
|
|
expectedIDs: []string{
|
|
"aaaaaa33-3350-4b4b-d185-0e1992ed43e9",
|
|
"aaaaaaaa-3350-4b4b-d185-0e1992ed43e9",
|
|
},
|
|
},
|
|
{
|
|
name: "pagination no filter with next token end of pages",
|
|
pageSize: 2,
|
|
nextToken: "aaaaaabb-3350-4b4b-d185-0e1992ed43e9",
|
|
expectedNextToken: "",
|
|
expectedIDs: []string{
|
|
"aaaaaabb-3350-4b4b-d185-0e1992ed43e9",
|
|
"aaaaaacc-3350-4b4b-d185-0e1992ed43e9",
|
|
},
|
|
},
|
|
{
|
|
name: "filter no pagination",
|
|
filter: `Datacenter == "dc3"`,
|
|
expectedIDs: []string{
|
|
"aaaaaa33-3350-4b4b-d185-0e1992ed43e9",
|
|
"aaaaaabb-3350-4b4b-d185-0e1992ed43e9",
|
|
},
|
|
},
|
|
{
|
|
name: "filter and pagination",
|
|
filter: `Status != "ready"`,
|
|
pageSize: 2,
|
|
expectedNextToken: "aaaaaabb-3350-4b4b-d185-0e1992ed43e9",
|
|
expectedIDs: []string{
|
|
"aaaa1111-3350-4b4b-d185-0e1992ed43e9",
|
|
"aaaaaaaa-3350-4b4b-d185-0e1992ed43e9",
|
|
},
|
|
},
|
|
}
|
|
|
|
for _, tc := range cases {
|
|
t.Run(tc.name, func(t *testing.T) {
|
|
req := &structs.NodeListRequest{
|
|
QueryOptions: structs.QueryOptions{
|
|
Region: "global",
|
|
Filter: tc.filter,
|
|
PerPage: tc.pageSize,
|
|
NextToken: tc.nextToken,
|
|
},
|
|
}
|
|
req.AuthToken = aclToken
|
|
var resp structs.NodeListResponse
|
|
err := msgpackrpc.CallWithCodec(codec, "Node.List", req, &resp)
|
|
if tc.expectedError == "" {
|
|
require.NoError(t, err)
|
|
} else {
|
|
require.Error(t, err)
|
|
require.Contains(t, err.Error(), tc.expectedError)
|
|
return
|
|
}
|
|
|
|
actualIDs := []string{}
|
|
|
|
for _, node := range resp.Nodes {
|
|
actualIDs = append(actualIDs, node.ID)
|
|
}
|
|
require.Equal(t, tc.expectedIDs, actualIDs, "unexpected page of nodes")
|
|
require.Equal(t, tc.expectedNextToken, resp.QueryMeta.NextToken, "unexpected NextToken")
|
|
})
|
|
}
|
|
}
|