2016-07-11 23:31:32 +00:00
|
|
|
package consul
|
|
|
|
|
|
|
|
import (
|
2017-01-26 03:00:32 +00:00
|
|
|
"bytes"
|
2016-07-11 23:31:32 +00:00
|
|
|
"os"
|
|
|
|
"testing"
|
|
|
|
"time"
|
|
|
|
|
pkg refactor
command/agent/* -> agent/*
command/consul/* -> agent/consul/*
command/agent/command{,_test}.go -> command/agent{,_test}.go
command/base/command.go -> command/base.go
command/base/* -> command/*
commands.go -> command/commands.go
The script which did the refactor is:
(
cd $GOPATH/src/github.com/hashicorp/consul
git mv command/agent/command.go command/agent.go
git mv command/agent/command_test.go command/agent_test.go
git mv command/agent/flag_slice_value{,_test}.go command/
git mv command/agent .
git mv command/base/command.go command/base.go
git mv command/base/config_util{,_test}.go command/
git mv commands.go command/
git mv consul agent
rmdir command/base/
gsed -i -e 's|package agent|package command|' command/agent{,_test}.go
gsed -i -e 's|package agent|package command|' command/flag_slice_value{,_test}.go
gsed -i -e 's|package base|package command|' command/base.go command/config_util{,_test}.go
gsed -i -e 's|package main|package command|' command/commands.go
gsed -i -e 's|base.Command|BaseCommand|' command/commands.go
gsed -i -e 's|agent.Command|AgentCommand|' command/commands.go
gsed -i -e 's|\tCommand:|\tBaseCommand:|' command/commands.go
gsed -i -e 's|base\.||' command/commands.go
gsed -i -e 's|command\.||' command/commands.go
gsed -i -e 's|command|c|' main.go
gsed -i -e 's|range Commands|range command.Commands|' main.go
gsed -i -e 's|Commands: Commands|Commands: command.Commands|' main.go
gsed -i -e 's|base\.BoolValue|BoolValue|' command/operator_autopilot_set.go
gsed -i -e 's|base\.DurationValue|DurationValue|' command/operator_autopilot_set.go
gsed -i -e 's|base\.StringValue|StringValue|' command/operator_autopilot_set.go
gsed -i -e 's|base\.UintValue|UintValue|' command/operator_autopilot_set.go
gsed -i -e 's|\bCommand\b|BaseCommand|' command/base.go
gsed -i -e 's|BaseCommand Options|Command Options|' command/base.go
gsed -i -e 's|base.Command|BaseCommand|' command/*.go
gsed -i -e 's|c\.Command|c.BaseCommand|g' command/*.go
gsed -i -e 's|\tCommand:|\tBaseCommand:|' command/*_test.go
gsed -i -e 's|base\.||' command/*_test.go
gsed -i -e 's|\bCommand\b|AgentCommand|' command/agent{,_test}.go
gsed -i -e 's|cmd.AgentCommand|cmd.BaseCommand|' command/agent.go
gsed -i -e 's|cli.AgentCommand = new(Command)|cli.Command = new(AgentCommand)|' command/agent_test.go
gsed -i -e 's|exec.AgentCommand|exec.Command|' command/agent_test.go
gsed -i -e 's|exec.BaseCommand|exec.Command|' command/agent_test.go
gsed -i -e 's|NewTestAgent|agent.NewTestAgent|' command/agent_test.go
gsed -i -e 's|= TestConfig|= agent.TestConfig|' command/agent_test.go
gsed -i -e 's|: RetryJoin|: agent.RetryJoin|' command/agent_test.go
gsed -i -e 's|\.\./\.\./|../|' command/config_util_test.go
gsed -i -e 's|\bverifyUniqueListeners|VerifyUniqueListeners|' agent/config{,_test}.go command/agent.go
gsed -i -e 's|\bserfLANKeyring\b|SerfLANKeyring|g' agent/{agent,keyring,testagent}.go command/agent.go
gsed -i -e 's|\bserfWANKeyring\b|SerfWANKeyring|g' agent/{agent,keyring,testagent}.go command/agent.go
gsed -i -e 's|\bNewAgent\b|agent.New|g' command/agent{,_test}.go
gsed -i -e 's|\bNewAgent|New|' agent/{acl_test,agent,testagent}.go
gsed -i -e 's|\bAgent\b|agent.&|g' command/agent{,_test}.go
gsed -i -e 's|\bBool\b|agent.&|g' command/agent{,_test}.go
gsed -i -e 's|\bConfig\b|agent.&|g' command/agent{,_test}.go
gsed -i -e 's|\bDefaultConfig\b|agent.&|g' command/agent{,_test}.go
gsed -i -e 's|\bDevConfig\b|agent.&|g' command/agent{,_test}.go
gsed -i -e 's|\bMergeConfig\b|agent.&|g' command/agent{,_test}.go
gsed -i -e 's|\bReadConfigPaths\b|agent.&|g' command/agent{,_test}.go
gsed -i -e 's|\bParseMetaPair\b|agent.&|g' command/agent{,_test}.go
gsed -i -e 's|\bSerfLANKeyring\b|agent.&|g' command/agent{,_test}.go
gsed -i -e 's|\bSerfWANKeyring\b|agent.&|g' command/agent{,_test}.go
gsed -i -e 's|circonus\.agent|circonus|g' command/agent{,_test}.go
gsed -i -e 's|logger\.agent|logger|g' command/agent{,_test}.go
gsed -i -e 's|metrics\.agent|metrics|g' command/agent{,_test}.go
gsed -i -e 's|// agent.Agent|// agent|' command/agent{,_test}.go
gsed -i -e 's|a\.agent\.Config|a.Config|' command/agent{,_test}.go
gsed -i -e 's|agent\.AppendSliceValue|AppendSliceValue|' command/{configtest,validate}.go
gsed -i -e 's|consul/consul|agent/consul|' GNUmakefile
gsed -i -e 's|\.\./test|../../test|' agent/consul/server_test.go
# fix imports
f=$(grep -rl 'github.com/hashicorp/consul/command/agent' * | grep '\.go')
gsed -i -e 's|github.com/hashicorp/consul/command/agent|github.com/hashicorp/consul/agent|' $f
goimports -w $f
f=$(grep -rl 'github.com/hashicorp/consul/consul' * | grep '\.go')
gsed -i -e 's|github.com/hashicorp/consul/consul|github.com/hashicorp/consul/agent/consul|' $f
goimports -w $f
goimports -w command/*.go main.go
)
2017-06-09 22:28:28 +00:00
|
|
|
"github.com/hashicorp/consul/agent/consul/state"
|
2017-07-06 10:34:00 +00:00
|
|
|
"github.com/hashicorp/consul/agent/structs"
|
2017-04-19 23:00:11 +00:00
|
|
|
"github.com/hashicorp/consul/testrpc"
|
2017-06-26 11:35:24 +00:00
|
|
|
"github.com/hashicorp/consul/testutil/retry"
|
2017-01-26 03:00:32 +00:00
|
|
|
"github.com/hashicorp/go-memdb"
|
2016-07-11 23:31:32 +00:00
|
|
|
"github.com/hashicorp/net-rpc-msgpackrpc"
|
2018-07-25 19:26:27 +00:00
|
|
|
"github.com/stretchr/testify/assert"
|
|
|
|
"github.com/stretchr/testify/require"
|
2016-07-11 23:31:32 +00:00
|
|
|
)
|
|
|
|
|
|
|
|
func TestRPC_NoLeader_Fail(t *testing.T) {
|
2017-06-27 13:22:18 +00:00
|
|
|
t.Parallel()
|
2016-07-11 23:31:32 +00:00
|
|
|
dir1, s1 := testServerWithConfig(t, func(c *Config) {
|
|
|
|
c.RPCHoldTimeout = 1 * time.Millisecond
|
|
|
|
})
|
|
|
|
defer os.RemoveAll(dir1)
|
|
|
|
defer s1.Shutdown()
|
|
|
|
codec := rpcClient(t, s1)
|
|
|
|
defer codec.Close()
|
|
|
|
|
|
|
|
arg := structs.RegisterRequest{
|
|
|
|
Datacenter: "dc1",
|
|
|
|
Node: "foo",
|
|
|
|
Address: "127.0.0.1",
|
|
|
|
}
|
|
|
|
var out struct{}
|
|
|
|
|
|
|
|
// Make sure we eventually fail with a no leader error, which we should
|
|
|
|
// see given the short timeout.
|
|
|
|
err := msgpackrpc.CallWithCodec(codec, "Catalog.Register", &arg, &out)
|
2017-05-31 06:59:41 +00:00
|
|
|
if err == nil || err.Error() != structs.ErrNoLeader.Error() {
|
2016-07-11 23:31:32 +00:00
|
|
|
t.Fatalf("bad: %v", err)
|
|
|
|
}
|
|
|
|
|
|
|
|
// Now make sure it goes through.
|
2017-04-19 23:00:11 +00:00
|
|
|
testrpc.WaitForLeader(t, s1.RPC, "dc1")
|
2016-07-11 23:31:32 +00:00
|
|
|
err = msgpackrpc.CallWithCodec(codec, "Catalog.Register", &arg, &out)
|
|
|
|
if err != nil {
|
|
|
|
t.Fatalf("bad: %v", err)
|
|
|
|
}
|
|
|
|
}
|
|
|
|
|
|
|
|
func TestRPC_NoLeader_Retry(t *testing.T) {
|
2017-06-27 13:22:18 +00:00
|
|
|
t.Parallel()
|
2016-07-11 23:31:32 +00:00
|
|
|
dir1, s1 := testServerWithConfig(t, func(c *Config) {
|
|
|
|
c.RPCHoldTimeout = 10 * time.Second
|
|
|
|
})
|
|
|
|
defer os.RemoveAll(dir1)
|
|
|
|
defer s1.Shutdown()
|
|
|
|
codec := rpcClient(t, s1)
|
|
|
|
defer codec.Close()
|
|
|
|
|
|
|
|
arg := structs.RegisterRequest{
|
|
|
|
Datacenter: "dc1",
|
|
|
|
Node: "foo",
|
|
|
|
Address: "127.0.0.1",
|
|
|
|
}
|
|
|
|
var out struct{}
|
|
|
|
|
|
|
|
// This isn't sure-fire but tries to check that we don't have a
|
|
|
|
// leader going into the RPC, so we exercise the retry logic.
|
|
|
|
if ok, _ := s1.getLeader(); ok {
|
|
|
|
t.Fatalf("should not have a leader yet")
|
|
|
|
}
|
|
|
|
|
|
|
|
// The timeout is long enough to ride out any reasonable leader
|
|
|
|
// election.
|
|
|
|
err := msgpackrpc.CallWithCodec(codec, "Catalog.Register", &arg, &out)
|
|
|
|
if err != nil {
|
|
|
|
t.Fatalf("bad: %v", err)
|
|
|
|
}
|
|
|
|
}
|
2017-01-26 03:00:32 +00:00
|
|
|
|
2017-11-29 02:01:17 +00:00
|
|
|
type MockSink struct {
|
|
|
|
*bytes.Buffer
|
|
|
|
cancel bool
|
|
|
|
}
|
|
|
|
|
|
|
|
func (m *MockSink) ID() string {
|
|
|
|
return "Mock"
|
|
|
|
}
|
|
|
|
|
|
|
|
func (m *MockSink) Cancel() error {
|
|
|
|
m.cancel = true
|
|
|
|
return nil
|
|
|
|
}
|
|
|
|
|
|
|
|
func (m *MockSink) Close() error {
|
|
|
|
return nil
|
|
|
|
}
|
|
|
|
|
2017-01-26 03:00:32 +00:00
|
|
|
func TestRPC_blockingQuery(t *testing.T) {
|
2017-06-27 13:22:18 +00:00
|
|
|
t.Parallel()
|
2017-01-26 03:00:32 +00:00
|
|
|
dir, s := testServer(t)
|
|
|
|
defer os.RemoveAll(dir)
|
|
|
|
defer s.Shutdown()
|
|
|
|
|
2018-07-25 19:26:27 +00:00
|
|
|
require := require.New(t)
|
|
|
|
assert := assert.New(t)
|
|
|
|
|
|
|
|
// Perform a non-blocking query. Note that it's significant that the meta has
|
|
|
|
// a zero index in response - the implied opts.MinQueryIndex is also zero but
|
|
|
|
// this should not block still.
|
2017-01-26 03:00:32 +00:00
|
|
|
{
|
|
|
|
var opts structs.QueryOptions
|
|
|
|
var meta structs.QueryMeta
|
|
|
|
var calls int
|
2017-04-21 00:46:29 +00:00
|
|
|
fn := func(ws memdb.WatchSet, state *state.Store) error {
|
2017-01-26 03:00:32 +00:00
|
|
|
calls++
|
|
|
|
return nil
|
|
|
|
}
|
|
|
|
if err := s.blockingQuery(&opts, &meta, fn); err != nil {
|
|
|
|
t.Fatalf("err: %v", err)
|
|
|
|
}
|
|
|
|
if calls != 1 {
|
|
|
|
t.Fatalf("bad: %d", calls)
|
|
|
|
}
|
|
|
|
}
|
|
|
|
|
|
|
|
// Perform a blocking query that gets woken up and loops around once.
|
|
|
|
{
|
|
|
|
opts := structs.QueryOptions{
|
|
|
|
MinQueryIndex: 3,
|
|
|
|
}
|
|
|
|
var meta structs.QueryMeta
|
|
|
|
var calls int
|
2017-04-21 00:46:29 +00:00
|
|
|
fn := func(ws memdb.WatchSet, state *state.Store) error {
|
2017-01-26 03:00:32 +00:00
|
|
|
if calls == 0 {
|
|
|
|
meta.Index = 3
|
|
|
|
|
|
|
|
fakeCh := make(chan struct{})
|
|
|
|
close(fakeCh)
|
|
|
|
ws.Add(fakeCh)
|
|
|
|
} else {
|
|
|
|
meta.Index = 4
|
|
|
|
}
|
|
|
|
calls++
|
|
|
|
return nil
|
|
|
|
}
|
|
|
|
if err := s.blockingQuery(&opts, &meta, fn); err != nil {
|
|
|
|
t.Fatalf("err: %v", err)
|
|
|
|
}
|
|
|
|
if calls != 2 {
|
|
|
|
t.Fatalf("bad: %d", calls)
|
|
|
|
}
|
|
|
|
}
|
|
|
|
|
2018-07-25 19:26:27 +00:00
|
|
|
// Perform a blocking query that returns a zero index from blocking func (e.g.
|
|
|
|
// no state yet). This should still return an empty response immediately, but
|
|
|
|
// with index of 1 and then block on the next attempt. In one sense zero index
|
|
|
|
// is not really a valid response from a state method that is not an error but
|
|
|
|
// in practice a lot of state store operations do return it unless they
|
|
|
|
// explicitly special checks to turn 0 into 1. Often this is not caught or
|
|
|
|
// covered by tests but eventually when hit in the wild causes blocking
|
|
|
|
// clients to busy loop and burn CPU. This test ensure that blockingQuery
|
|
|
|
// systematically does the right thing to prevent future bugs like that.
|
|
|
|
{
|
|
|
|
opts := structs.QueryOptions{
|
|
|
|
MinQueryIndex: 0,
|
|
|
|
}
|
|
|
|
var meta structs.QueryMeta
|
|
|
|
var calls int
|
|
|
|
fn := func(ws memdb.WatchSet, state *state.Store) error {
|
|
|
|
if opts.MinQueryIndex > 0 {
|
|
|
|
// If client requested blocking, block forever. This is simulating
|
|
|
|
// waiting for the watched resource to be initialized/written to giving
|
|
|
|
// it a non-zero index. Note the timeout on the query options is relied
|
|
|
|
// on to stop the test taking forever.
|
|
|
|
fakeCh := make(chan struct{})
|
|
|
|
ws.Add(fakeCh)
|
|
|
|
}
|
|
|
|
meta.Index = 0
|
|
|
|
calls++
|
|
|
|
return nil
|
|
|
|
}
|
|
|
|
require.NoError(s.blockingQuery(&opts, &meta, fn))
|
|
|
|
assert.Equal(1, calls)
|
|
|
|
assert.Equal(uint64(1), meta.Index,
|
|
|
|
"expect fake index of 1 to force client to block on next update")
|
|
|
|
|
|
|
|
// Simulate client making next request
|
|
|
|
opts.MinQueryIndex = 1
|
|
|
|
opts.MaxQueryTime = 20 * time.Millisecond // Don't wait too long
|
|
|
|
|
|
|
|
// This time we should block even though the func returns index 0 still
|
|
|
|
t0 := time.Now()
|
|
|
|
require.NoError(s.blockingQuery(&opts, &meta, fn))
|
|
|
|
t1 := time.Now()
|
|
|
|
assert.Equal(2, calls)
|
|
|
|
assert.Equal(uint64(1), meta.Index,
|
|
|
|
"expect fake index of 1 to force client to block on next update")
|
|
|
|
assert.True(t1.Sub(t0) > 20*time.Millisecond,
|
|
|
|
"should have actually blocked waiting for timeout")
|
|
|
|
|
|
|
|
}
|
|
|
|
|
2017-01-26 03:00:32 +00:00
|
|
|
// Perform a query that blocks and gets interrupted when the state store
|
|
|
|
// is abandoned.
|
|
|
|
{
|
|
|
|
opts := structs.QueryOptions{
|
|
|
|
MinQueryIndex: 3,
|
|
|
|
}
|
|
|
|
var meta structs.QueryMeta
|
|
|
|
var calls int
|
2017-04-21 00:46:29 +00:00
|
|
|
fn := func(ws memdb.WatchSet, state *state.Store) error {
|
2017-01-26 03:00:32 +00:00
|
|
|
if calls == 0 {
|
|
|
|
meta.Index = 3
|
|
|
|
|
|
|
|
snap, err := s.fsm.Snapshot()
|
|
|
|
if err != nil {
|
|
|
|
t.Fatalf("err: %v", err)
|
|
|
|
}
|
|
|
|
defer snap.Release()
|
|
|
|
|
|
|
|
buf := bytes.NewBuffer(nil)
|
|
|
|
sink := &MockSink{buf, false}
|
|
|
|
if err := snap.Persist(sink); err != nil {
|
|
|
|
t.Fatalf("err: %v", err)
|
|
|
|
}
|
|
|
|
|
|
|
|
if err := s.fsm.Restore(sink); err != nil {
|
|
|
|
t.Fatalf("err: %v", err)
|
|
|
|
}
|
|
|
|
}
|
|
|
|
calls++
|
|
|
|
return nil
|
|
|
|
}
|
|
|
|
if err := s.blockingQuery(&opts, &meta, fn); err != nil {
|
|
|
|
t.Fatalf("err: %v", err)
|
|
|
|
}
|
|
|
|
if calls != 1 {
|
|
|
|
t.Fatalf("bad: %d", calls)
|
|
|
|
}
|
|
|
|
}
|
|
|
|
}
|
2017-06-16 16:58:12 +00:00
|
|
|
|
2017-06-26 11:35:24 +00:00
|
|
|
func TestRPC_ReadyForConsistentReads(t *testing.T) {
|
|
|
|
t.Parallel()
|
2017-06-16 16:58:12 +00:00
|
|
|
dir, s := testServerWithConfig(t, func(c *Config) {
|
|
|
|
c.RPCHoldTimeout = 2 * time.Millisecond
|
|
|
|
})
|
|
|
|
defer os.RemoveAll(dir)
|
|
|
|
defer s.Shutdown()
|
|
|
|
|
|
|
|
testrpc.WaitForLeader(t, s.RPC, "dc1")
|
|
|
|
|
|
|
|
if !s.isReadyForConsistentReads() {
|
|
|
|
t.Fatal("Server should be ready for consistent reads")
|
|
|
|
}
|
|
|
|
|
|
|
|
s.resetConsistentReadReady()
|
|
|
|
err := s.consistentRead()
|
2017-06-21 00:43:07 +00:00
|
|
|
if err.Error() != "Not ready to serve consistent reads" {
|
|
|
|
t.Fatal("Server should NOT be ready for consistent reads")
|
|
|
|
}
|
2017-06-16 16:58:12 +00:00
|
|
|
|
2017-06-26 11:35:24 +00:00
|
|
|
go func() {
|
|
|
|
time.Sleep(100 * time.Millisecond)
|
|
|
|
s.setConsistentReadReady()
|
|
|
|
}()
|
2017-06-16 16:58:12 +00:00
|
|
|
|
2017-06-26 11:35:24 +00:00
|
|
|
retry.Run(t, func(r *retry.R) {
|
|
|
|
if err := s.consistentRead(); err != nil {
|
|
|
|
r.Fatalf("Expected server to be ready for consistent reads, got error %v", err)
|
|
|
|
}
|
|
|
|
})
|
2017-06-16 16:58:12 +00:00
|
|
|
}
|