2020-09-28 14:13:10 +00:00
|
|
|
package nomad
|
|
|
|
|
|
|
|
import (
|
2020-12-01 16:11:34 +00:00
|
|
|
"context"
|
2020-09-28 14:13:10 +00:00
|
|
|
"encoding/json"
|
|
|
|
"fmt"
|
|
|
|
"io"
|
|
|
|
"net"
|
|
|
|
"strings"
|
|
|
|
"testing"
|
|
|
|
"time"
|
|
|
|
|
|
|
|
"github.com/hashicorp/go-msgpack/codec"
|
2020-12-01 16:11:34 +00:00
|
|
|
msgpackrpc "github.com/hashicorp/net-rpc-msgpackrpc"
|
2020-10-06 20:21:58 +00:00
|
|
|
"github.com/hashicorp/nomad/acl"
|
2020-09-28 14:13:10 +00:00
|
|
|
"github.com/hashicorp/nomad/nomad/mock"
|
|
|
|
"github.com/hashicorp/nomad/nomad/stream"
|
|
|
|
"github.com/hashicorp/nomad/nomad/structs"
|
|
|
|
"github.com/hashicorp/nomad/testutil"
|
|
|
|
"github.com/mitchellh/mapstructure"
|
|
|
|
"github.com/stretchr/testify/require"
|
|
|
|
)
|
|
|
|
|
|
|
|
func TestEventStream(t *testing.T) {
|
|
|
|
t.Parallel()
|
|
|
|
|
|
|
|
s1, cleanupS1 := TestServer(t, func(c *Config) {
|
2020-10-08 18:27:52 +00:00
|
|
|
c.EnableEventBroker = true
|
2020-09-28 14:13:10 +00:00
|
|
|
})
|
|
|
|
defer cleanupS1()
|
|
|
|
|
|
|
|
// Create request for all topics and keys
|
|
|
|
req := structs.EventStreamRequest{
|
2020-10-06 20:21:58 +00:00
|
|
|
Topics: map[structs.Topic][]string{"*": {"*"}},
|
2020-09-28 14:13:10 +00:00
|
|
|
QueryOptions: structs.QueryOptions{
|
|
|
|
Region: s1.Region(),
|
|
|
|
},
|
|
|
|
}
|
|
|
|
|
|
|
|
handler, err := s1.StreamingRpcHandler("Event.Stream")
|
|
|
|
require.Nil(t, err)
|
|
|
|
|
|
|
|
p1, p2 := net.Pipe()
|
|
|
|
defer p1.Close()
|
|
|
|
defer p2.Close()
|
|
|
|
|
|
|
|
errCh := make(chan error)
|
|
|
|
streamMsg := make(chan *structs.EventStreamWrapper)
|
|
|
|
|
|
|
|
// invoke handler
|
|
|
|
go handler(p2)
|
|
|
|
|
2020-10-06 20:21:58 +00:00
|
|
|
// decode request responses
|
2020-09-28 14:13:10 +00:00
|
|
|
go func() {
|
|
|
|
decoder := codec.NewDecoder(p1, structs.MsgpackHandle)
|
|
|
|
for {
|
|
|
|
var msg structs.EventStreamWrapper
|
|
|
|
if err := decoder.Decode(&msg); err != nil {
|
|
|
|
if err == io.EOF || strings.Contains(err.Error(), "closed") {
|
|
|
|
return
|
|
|
|
}
|
|
|
|
errCh <- fmt.Errorf("error decoding: %w", err)
|
|
|
|
}
|
|
|
|
|
|
|
|
streamMsg <- &msg
|
|
|
|
}
|
|
|
|
}()
|
|
|
|
|
|
|
|
// retrieve publisher for server, send event
|
2020-10-08 18:27:52 +00:00
|
|
|
publisher, err := s1.State().EventBroker()
|
2020-09-28 14:13:10 +00:00
|
|
|
require.NoError(t, err)
|
|
|
|
|
|
|
|
node := mock.Node()
|
2020-10-06 20:21:58 +00:00
|
|
|
publisher.Publish(&structs.Events{Index: uint64(1), Events: []structs.Event{{Topic: "test", Payload: node}}})
|
2020-09-28 14:13:10 +00:00
|
|
|
|
2020-10-06 20:21:58 +00:00
|
|
|
// Send request
|
2020-09-28 14:13:10 +00:00
|
|
|
encoder := codec.NewEncoder(p1, structs.MsgpackHandle)
|
|
|
|
require.Nil(t, encoder.Encode(req))
|
|
|
|
|
2020-10-08 18:27:52 +00:00
|
|
|
publisher.Publish(&structs.Events{Index: uint64(2), Events: []structs.Event{{Topic: "test", Payload: node}}})
|
|
|
|
publisher.Publish(&structs.Events{Index: uint64(3), Events: []structs.Event{{Topic: "test", Payload: node}}})
|
|
|
|
|
2020-09-28 14:13:10 +00:00
|
|
|
timeout := time.After(3 * time.Second)
|
2020-10-08 18:27:52 +00:00
|
|
|
got := 0
|
|
|
|
want := 3
|
2020-09-28 14:13:10 +00:00
|
|
|
OUTER:
|
|
|
|
for {
|
|
|
|
select {
|
|
|
|
case <-timeout:
|
|
|
|
t.Fatal("timeout waiting for event stream")
|
|
|
|
case err := <-errCh:
|
|
|
|
t.Fatal(err)
|
|
|
|
case msg := <-streamMsg:
|
|
|
|
if msg.Error != nil {
|
|
|
|
t.Fatalf("Got error: %v", msg.Error.Error())
|
|
|
|
}
|
|
|
|
|
|
|
|
// ignore heartbeat
|
2020-10-08 18:27:52 +00:00
|
|
|
if msg.Event == stream.JsonHeartbeat {
|
2020-09-28 14:13:10 +00:00
|
|
|
continue
|
|
|
|
}
|
|
|
|
|
2020-10-04 19:12:35 +00:00
|
|
|
var event structs.Events
|
2020-09-28 14:13:10 +00:00
|
|
|
err = json.Unmarshal(msg.Event.Data, &event)
|
|
|
|
require.NoError(t, err)
|
|
|
|
|
|
|
|
// decode fully to ensure we received expected out
|
|
|
|
var out structs.Node
|
|
|
|
cfg := &mapstructure.DecoderConfig{
|
|
|
|
Metadata: nil,
|
|
|
|
Result: &out,
|
|
|
|
}
|
|
|
|
dec, err := mapstructure.NewDecoder(cfg)
|
2020-10-01 18:43:28 +00:00
|
|
|
dec.Decode(event.Events[0].Payload)
|
2020-09-28 14:13:10 +00:00
|
|
|
require.NoError(t, err)
|
|
|
|
require.Equal(t, node.ID, out.ID)
|
2020-10-08 18:27:52 +00:00
|
|
|
|
|
|
|
got++
|
|
|
|
if got == want {
|
|
|
|
break OUTER
|
|
|
|
}
|
2020-09-28 14:13:10 +00:00
|
|
|
}
|
|
|
|
}
|
|
|
|
}
|
|
|
|
|
|
|
|
// TestEventStream_StreamErr asserts an error is returned when an event publisher
|
|
|
|
// closes its subscriptions
|
|
|
|
func TestEventStream_StreamErr(t *testing.T) {
|
|
|
|
t.Parallel()
|
|
|
|
|
|
|
|
s1, cleanupS1 := TestServer(t, func(c *Config) {
|
2020-10-08 18:27:52 +00:00
|
|
|
c.EnableEventBroker = true
|
2020-09-28 14:13:10 +00:00
|
|
|
})
|
|
|
|
defer cleanupS1()
|
|
|
|
|
|
|
|
testutil.WaitForLeader(t, s1.RPC)
|
|
|
|
|
|
|
|
req := structs.EventStreamRequest{
|
2020-10-04 19:12:35 +00:00
|
|
|
Topics: map[structs.Topic][]string{"*": {"*"}},
|
2020-09-28 14:13:10 +00:00
|
|
|
QueryOptions: structs.QueryOptions{
|
|
|
|
Region: s1.Region(),
|
|
|
|
},
|
|
|
|
}
|
|
|
|
|
|
|
|
handler, err := s1.StreamingRpcHandler("Event.Stream")
|
|
|
|
require.Nil(t, err)
|
|
|
|
|
|
|
|
p1, p2 := net.Pipe()
|
|
|
|
defer p1.Close()
|
|
|
|
defer p2.Close()
|
|
|
|
|
|
|
|
errCh := make(chan error)
|
|
|
|
streamMsg := make(chan *structs.EventStreamWrapper)
|
|
|
|
|
|
|
|
go handler(p2)
|
|
|
|
|
|
|
|
go func() {
|
|
|
|
decoder := codec.NewDecoder(p1, structs.MsgpackHandle)
|
|
|
|
for {
|
|
|
|
var msg structs.EventStreamWrapper
|
|
|
|
if err := decoder.Decode(&msg); err != nil {
|
|
|
|
if err == io.EOF || strings.Contains(err.Error(), "closed") {
|
|
|
|
return
|
|
|
|
}
|
|
|
|
errCh <- fmt.Errorf("error decoding: %w", err)
|
|
|
|
}
|
|
|
|
|
|
|
|
streamMsg <- &msg
|
|
|
|
}
|
|
|
|
}()
|
|
|
|
|
2020-10-08 18:27:52 +00:00
|
|
|
publisher, err := s1.State().EventBroker()
|
2020-09-28 14:13:10 +00:00
|
|
|
require.NoError(t, err)
|
|
|
|
|
|
|
|
node := mock.Node()
|
|
|
|
|
|
|
|
// send req
|
|
|
|
encoder := codec.NewEncoder(p1, structs.MsgpackHandle)
|
|
|
|
require.Nil(t, encoder.Encode(req))
|
|
|
|
|
2020-10-06 20:21:58 +00:00
|
|
|
// publish some events
|
|
|
|
publisher.Publish(&structs.Events{Index: uint64(1), Events: []structs.Event{{Topic: "test", Payload: node}}})
|
|
|
|
publisher.Publish(&structs.Events{Index: uint64(2), Events: []structs.Event{{Topic: "test", Payload: node}}})
|
2020-09-28 14:13:10 +00:00
|
|
|
|
|
|
|
timeout := time.After(5 * time.Second)
|
|
|
|
OUTER:
|
|
|
|
for {
|
|
|
|
select {
|
|
|
|
case <-timeout:
|
|
|
|
t.Fatal("timeout waiting for event stream")
|
|
|
|
case err := <-errCh:
|
|
|
|
t.Fatal(err)
|
|
|
|
case msg := <-streamMsg:
|
2020-10-06 20:21:58 +00:00
|
|
|
// close the publishers subscriptions forcing an error
|
|
|
|
// after an initial event is received
|
|
|
|
publisher.CloseAll()
|
2020-09-28 14:13:10 +00:00
|
|
|
if msg.Error == nil {
|
|
|
|
// continue trying for error
|
|
|
|
continue
|
|
|
|
}
|
|
|
|
require.NotNil(t, msg.Error)
|
|
|
|
require.Contains(t, msg.Error.Error(), "subscription closed by server")
|
|
|
|
break OUTER
|
|
|
|
}
|
|
|
|
}
|
|
|
|
}
|
|
|
|
|
|
|
|
// TestEventStream_RegionForward tests event streaming from one server
|
|
|
|
// to another in a different region
|
|
|
|
func TestEventStream_RegionForward(t *testing.T) {
|
|
|
|
t.Parallel()
|
|
|
|
|
|
|
|
s1, cleanupS1 := TestServer(t, func(c *Config) {
|
2020-10-08 18:27:52 +00:00
|
|
|
c.EnableEventBroker = true
|
2020-09-28 14:13:10 +00:00
|
|
|
})
|
|
|
|
defer cleanupS1()
|
|
|
|
|
|
|
|
s2, cleanupS2 := TestServer(t, func(c *Config) {
|
2020-10-08 18:27:52 +00:00
|
|
|
c.EnableEventBroker = true
|
2020-09-28 14:13:10 +00:00
|
|
|
c.Region = "foo"
|
|
|
|
})
|
|
|
|
defer cleanupS2()
|
|
|
|
|
|
|
|
TestJoin(t, s1, s2)
|
|
|
|
|
|
|
|
// Create request targed for region foo
|
|
|
|
req := structs.EventStreamRequest{
|
2020-10-04 19:12:35 +00:00
|
|
|
Topics: map[structs.Topic][]string{"*": {"*"}},
|
2020-09-28 14:13:10 +00:00
|
|
|
QueryOptions: structs.QueryOptions{
|
|
|
|
Region: "foo",
|
|
|
|
},
|
|
|
|
}
|
|
|
|
|
|
|
|
// Query s1 handler
|
|
|
|
handler, err := s1.StreamingRpcHandler("Event.Stream")
|
|
|
|
require.Nil(t, err)
|
|
|
|
|
|
|
|
p1, p2 := net.Pipe()
|
|
|
|
defer p1.Close()
|
|
|
|
defer p2.Close()
|
|
|
|
|
|
|
|
errCh := make(chan error)
|
|
|
|
streamMsg := make(chan *structs.EventStreamWrapper)
|
|
|
|
|
|
|
|
go handler(p2)
|
|
|
|
|
|
|
|
go func() {
|
|
|
|
decoder := codec.NewDecoder(p1, structs.MsgpackHandle)
|
|
|
|
for {
|
|
|
|
var msg structs.EventStreamWrapper
|
|
|
|
if err := decoder.Decode(&msg); err != nil {
|
|
|
|
if err == io.EOF || strings.Contains(err.Error(), "closed") {
|
|
|
|
return
|
|
|
|
}
|
|
|
|
errCh <- fmt.Errorf("error decoding: %w", err)
|
|
|
|
}
|
|
|
|
|
|
|
|
streamMsg <- &msg
|
|
|
|
}
|
|
|
|
}()
|
|
|
|
|
|
|
|
// publish with server 2
|
2020-10-08 18:27:52 +00:00
|
|
|
publisher, err := s2.State().EventBroker()
|
2020-09-28 14:13:10 +00:00
|
|
|
require.NoError(t, err)
|
|
|
|
|
|
|
|
node := mock.Node()
|
2020-10-06 20:21:58 +00:00
|
|
|
publisher.Publish(&structs.Events{Index: uint64(1), Events: []structs.Event{{Topic: "test", Payload: node}}})
|
2020-09-28 14:13:10 +00:00
|
|
|
|
|
|
|
// send req
|
|
|
|
encoder := codec.NewEncoder(p1, structs.MsgpackHandle)
|
|
|
|
require.Nil(t, encoder.Encode(req))
|
|
|
|
|
|
|
|
timeout := time.After(3 * time.Second)
|
|
|
|
OUTER:
|
|
|
|
for {
|
|
|
|
select {
|
|
|
|
case <-timeout:
|
|
|
|
t.Fatal("timeout waiting for event stream")
|
|
|
|
case err := <-errCh:
|
|
|
|
t.Fatal(err)
|
|
|
|
case msg := <-streamMsg:
|
|
|
|
if msg.Error != nil {
|
|
|
|
t.Fatalf("Got error: %v", msg.Error.Error())
|
|
|
|
}
|
|
|
|
|
2020-10-08 18:27:52 +00:00
|
|
|
if msg.Event == stream.JsonHeartbeat {
|
2020-09-28 14:13:10 +00:00
|
|
|
continue
|
|
|
|
}
|
|
|
|
|
2020-10-04 19:12:35 +00:00
|
|
|
var event structs.Events
|
2020-09-28 14:13:10 +00:00
|
|
|
err = json.Unmarshal(msg.Event.Data, &event)
|
|
|
|
require.NoError(t, err)
|
|
|
|
|
|
|
|
var out structs.Node
|
|
|
|
cfg := &mapstructure.DecoderConfig{
|
|
|
|
Metadata: nil,
|
|
|
|
Result: &out,
|
|
|
|
}
|
|
|
|
dec, err := mapstructure.NewDecoder(cfg)
|
2020-10-01 18:43:28 +00:00
|
|
|
dec.Decode(event.Events[0].Payload)
|
2020-09-28 14:13:10 +00:00
|
|
|
require.NoError(t, err)
|
|
|
|
require.Equal(t, node.ID, out.ID)
|
|
|
|
break OUTER
|
|
|
|
}
|
|
|
|
}
|
|
|
|
}
|
|
|
|
|
|
|
|
func TestEventStream_ACL(t *testing.T) {
|
2020-10-06 20:21:58 +00:00
|
|
|
t.Parallel()
|
|
|
|
require := require.New(t)
|
|
|
|
|
|
|
|
// start server
|
2020-12-01 16:11:34 +00:00
|
|
|
s, _, cleanupS := TestACLServer(t, nil)
|
2020-10-06 20:21:58 +00:00
|
|
|
defer cleanupS()
|
|
|
|
testutil.WaitForLeader(t, s.RPC)
|
|
|
|
|
|
|
|
policyBad := mock.NamespacePolicy("other", "", []string{acl.NamespaceCapabilityReadFS})
|
|
|
|
tokenBad := mock.CreatePolicyAndToken(t, s.State(), 1005, "invalid", policyBad)
|
|
|
|
|
2020-10-08 18:27:52 +00:00
|
|
|
policyNsGood := mock.NamespacePolicy("foo", "", []string{acl.NamespaceCapabilityReadJob})
|
|
|
|
tokenNsFoo := mock.CreatePolicyAndToken(t, s.State(), 1006, "valid", policyNsGood)
|
|
|
|
|
|
|
|
policyNsNode := mock.NamespacePolicy("foo", "", []string{acl.NamespaceCapabilityReadJob})
|
|
|
|
policyNsNode += "\n" + mock.NodePolicy("read")
|
|
|
|
tokenNsNode := mock.CreatePolicyAndToken(t, s.State(), 1007, "validnNsNode", policyNsNode)
|
|
|
|
|
2020-10-06 20:21:58 +00:00
|
|
|
cases := []struct {
|
|
|
|
Name string
|
|
|
|
Token string
|
2020-10-08 18:27:52 +00:00
|
|
|
Topics map[structs.Topic][]string
|
|
|
|
Namespace string
|
2020-10-06 20:21:58 +00:00
|
|
|
ExpectedErr string
|
2020-10-08 18:27:52 +00:00
|
|
|
PublishFn func(p *stream.EventBroker)
|
2020-10-06 20:21:58 +00:00
|
|
|
}{
|
|
|
|
{
|
2020-10-08 18:27:52 +00:00
|
|
|
Name: "no token",
|
|
|
|
Token: "",
|
|
|
|
Topics: map[structs.Topic][]string{
|
2020-12-03 16:48:18 +00:00
|
|
|
structs.TopicAll: {"*"},
|
2020-10-08 18:27:52 +00:00
|
|
|
},
|
|
|
|
ExpectedErr: structs.ErrPermissionDenied.Error(),
|
|
|
|
},
|
|
|
|
{
|
|
|
|
Name: "bad token",
|
|
|
|
Token: tokenBad.SecretID,
|
|
|
|
Topics: map[structs.Topic][]string{
|
2020-12-03 16:48:18 +00:00
|
|
|
structs.TopicAll: {"*"},
|
2020-10-08 18:27:52 +00:00
|
|
|
},
|
|
|
|
ExpectedErr: structs.ErrPermissionDenied.Error(),
|
|
|
|
},
|
|
|
|
{
|
|
|
|
Name: "job namespace token - correct ns",
|
|
|
|
Token: tokenNsFoo.SecretID,
|
|
|
|
Topics: map[structs.Topic][]string{
|
2020-12-03 16:48:18 +00:00
|
|
|
structs.TopicJob: {"*"},
|
|
|
|
structs.TopicEvaluation: {"*"},
|
|
|
|
structs.TopicAllocation: {"*"},
|
|
|
|
structs.TopicDeployment: {"*"},
|
2020-10-08 18:27:52 +00:00
|
|
|
},
|
|
|
|
Namespace: "foo",
|
|
|
|
ExpectedErr: "subscription closed by server",
|
|
|
|
PublishFn: func(p *stream.EventBroker) {
|
|
|
|
p.Publish(&structs.Events{Index: uint64(1000), Events: []structs.Event{{Topic: "Job", Namespace: "foo", Payload: mock.Job()}}})
|
|
|
|
},
|
|
|
|
},
|
|
|
|
{
|
|
|
|
Name: "job namespace token - incorrect ns",
|
|
|
|
Token: tokenNsFoo.SecretID,
|
|
|
|
Topics: map[structs.Topic][]string{
|
2020-12-03 16:48:18 +00:00
|
|
|
structs.TopicJob: {"*"}, // good
|
2020-10-08 18:27:52 +00:00
|
|
|
},
|
|
|
|
Namespace: "bar", // bad
|
2020-10-06 20:21:58 +00:00
|
|
|
ExpectedErr: structs.ErrPermissionDenied.Error(),
|
2020-10-08 18:27:52 +00:00
|
|
|
PublishFn: func(p *stream.EventBroker) {
|
|
|
|
p.Publish(&structs.Events{Index: uint64(1000), Events: []structs.Event{{Topic: "Job", Namespace: "foo", Payload: mock.Job()}}})
|
|
|
|
},
|
2020-10-06 20:21:58 +00:00
|
|
|
},
|
|
|
|
{
|
2020-10-08 18:27:52 +00:00
|
|
|
Name: "job namespace token - request management topic",
|
|
|
|
Token: tokenNsFoo.SecretID,
|
|
|
|
Topics: map[structs.Topic][]string{
|
2020-12-03 16:48:18 +00:00
|
|
|
structs.TopicAll: {"*"}, // bad
|
2020-10-08 18:27:52 +00:00
|
|
|
},
|
|
|
|
Namespace: "foo",
|
2020-10-06 20:21:58 +00:00
|
|
|
ExpectedErr: structs.ErrPermissionDenied.Error(),
|
2020-10-08 18:27:52 +00:00
|
|
|
PublishFn: func(p *stream.EventBroker) {
|
|
|
|
p.Publish(&structs.Events{Index: uint64(1000), Events: []structs.Event{{Topic: "Job", Namespace: "foo", Payload: mock.Job()}}})
|
|
|
|
},
|
2020-10-06 20:21:58 +00:00
|
|
|
},
|
|
|
|
{
|
2020-10-08 18:27:52 +00:00
|
|
|
Name: "job namespace token - request invalid node topic",
|
|
|
|
Token: tokenNsFoo.SecretID,
|
|
|
|
Topics: map[structs.Topic][]string{
|
2020-12-03 16:48:18 +00:00
|
|
|
structs.TopicEvaluation: {"*"}, // good
|
|
|
|
structs.TopicNode: {"*"}, // bad
|
2020-10-08 18:27:52 +00:00
|
|
|
},
|
|
|
|
Namespace: "foo",
|
|
|
|
ExpectedErr: structs.ErrPermissionDenied.Error(),
|
|
|
|
PublishFn: func(p *stream.EventBroker) {
|
|
|
|
p.Publish(&structs.Events{Index: uint64(1000), Events: []structs.Event{{Topic: "Job", Namespace: "foo", Payload: mock.Job()}}})
|
|
|
|
},
|
|
|
|
},
|
|
|
|
{
|
|
|
|
Name: "job+node namespace token, valid",
|
|
|
|
Token: tokenNsNode.SecretID,
|
|
|
|
Topics: map[structs.Topic][]string{
|
2020-12-03 16:48:18 +00:00
|
|
|
structs.TopicEvaluation: {"*"}, // good
|
|
|
|
structs.TopicNode: {"*"}, // good
|
2020-10-08 18:27:52 +00:00
|
|
|
},
|
|
|
|
Namespace: "foo",
|
2020-10-06 20:21:58 +00:00
|
|
|
ExpectedErr: "subscription closed by server",
|
2020-10-08 18:27:52 +00:00
|
|
|
PublishFn: func(p *stream.EventBroker) {
|
|
|
|
p.Publish(&structs.Events{Index: uint64(1000), Events: []structs.Event{{Topic: "Node", Payload: mock.Node()}}})
|
|
|
|
},
|
2020-10-06 20:21:58 +00:00
|
|
|
},
|
|
|
|
}
|
|
|
|
|
|
|
|
for _, tc := range cases {
|
|
|
|
t.Run(tc.Name, func(t *testing.T) {
|
2020-10-08 18:27:52 +00:00
|
|
|
var ns string
|
|
|
|
if tc.Namespace != "" {
|
|
|
|
ns = tc.Namespace
|
|
|
|
}
|
2020-10-06 20:21:58 +00:00
|
|
|
// Create request for all topics and keys
|
|
|
|
req := structs.EventStreamRequest{
|
2020-10-08 18:27:52 +00:00
|
|
|
Topics: tc.Topics,
|
2020-10-06 20:21:58 +00:00
|
|
|
QueryOptions: structs.QueryOptions{
|
|
|
|
Region: s.Region(),
|
2020-10-08 18:27:52 +00:00
|
|
|
Namespace: ns,
|
2020-10-06 20:21:58 +00:00
|
|
|
AuthToken: tc.Token,
|
|
|
|
},
|
|
|
|
}
|
|
|
|
|
|
|
|
handler, err := s.StreamingRpcHandler("Event.Stream")
|
|
|
|
require.Nil(err)
|
|
|
|
|
|
|
|
// create pipe
|
|
|
|
p1, p2 := net.Pipe()
|
|
|
|
defer p1.Close()
|
|
|
|
defer p2.Close()
|
|
|
|
|
|
|
|
errCh := make(chan error)
|
|
|
|
streamMsg := make(chan *structs.EventStreamWrapper)
|
|
|
|
|
|
|
|
go handler(p2)
|
|
|
|
|
|
|
|
// Start decoder
|
|
|
|
go func() {
|
|
|
|
decoder := codec.NewDecoder(p1, structs.MsgpackHandle)
|
|
|
|
for {
|
|
|
|
var msg structs.EventStreamWrapper
|
|
|
|
if err := decoder.Decode(&msg); err != nil {
|
|
|
|
if err == io.EOF || strings.Contains(err.Error(), "closed") {
|
|
|
|
return
|
|
|
|
}
|
|
|
|
errCh <- fmt.Errorf("error decoding: %w", err)
|
|
|
|
}
|
|
|
|
|
|
|
|
streamMsg <- &msg
|
|
|
|
}
|
|
|
|
}()
|
|
|
|
|
|
|
|
// send request
|
|
|
|
encoder := codec.NewEncoder(p1, structs.MsgpackHandle)
|
|
|
|
require.Nil(encoder.Encode(req))
|
|
|
|
|
2020-10-08 18:27:52 +00:00
|
|
|
publisher, err := s.State().EventBroker()
|
2020-10-06 20:21:58 +00:00
|
|
|
require.NoError(err)
|
|
|
|
|
|
|
|
// publish some events
|
|
|
|
node := mock.Node()
|
2020-10-08 18:27:52 +00:00
|
|
|
|
2020-10-06 20:21:58 +00:00
|
|
|
publisher.Publish(&structs.Events{Index: uint64(1), Events: []structs.Event{{Topic: "test", Payload: node}}})
|
|
|
|
publisher.Publish(&structs.Events{Index: uint64(2), Events: []structs.Event{{Topic: "test", Payload: node}}})
|
|
|
|
|
2020-10-08 18:27:52 +00:00
|
|
|
if tc.PublishFn != nil {
|
|
|
|
tc.PublishFn(publisher)
|
|
|
|
}
|
|
|
|
|
2020-10-06 20:21:58 +00:00
|
|
|
timeout := time.After(5 * time.Second)
|
|
|
|
OUTER:
|
|
|
|
for {
|
|
|
|
select {
|
|
|
|
case <-timeout:
|
2020-10-08 18:27:52 +00:00
|
|
|
t.Fatal("timeout waiting for events")
|
2020-10-06 20:21:58 +00:00
|
|
|
case err := <-errCh:
|
|
|
|
t.Fatal(err)
|
|
|
|
case msg := <-streamMsg:
|
|
|
|
// force error by closing all subscriptions
|
|
|
|
publisher.CloseAll()
|
|
|
|
if msg.Error == nil {
|
|
|
|
continue
|
|
|
|
}
|
|
|
|
|
|
|
|
if strings.Contains(msg.Error.Error(), tc.ExpectedErr) {
|
|
|
|
break OUTER
|
|
|
|
} else {
|
2020-10-08 18:27:52 +00:00
|
|
|
t.Fatalf("unexpected error %v", msg.Error)
|
2020-10-06 20:21:58 +00:00
|
|
|
}
|
|
|
|
}
|
|
|
|
}
|
|
|
|
})
|
|
|
|
}
|
2020-09-28 14:13:10 +00:00
|
|
|
}
|
2020-12-01 16:11:34 +00:00
|
|
|
|
|
|
|
// TestEventStream_ACL_Update_Close_Stream asserts that an active subscription
|
|
|
|
// is closed after the token is no longer valid
|
|
|
|
func TestEventStream_ACL_Update_Close_Stream(t *testing.T) {
|
|
|
|
t.Parallel()
|
|
|
|
|
|
|
|
// start server
|
|
|
|
s1, root, cleanupS := TestACLServer(t, nil)
|
|
|
|
defer cleanupS()
|
|
|
|
testutil.WaitForLeader(t, s1.RPC)
|
|
|
|
|
|
|
|
policyNsGood := mock.NamespacePolicy(structs.DefaultNamespace, "", []string{acl.NamespaceCapabilityReadJob})
|
|
|
|
tokenNsFoo := mock.CreatePolicyAndToken(t, s1.State(), 1006, "valid", policyNsGood)
|
|
|
|
|
|
|
|
req := structs.EventStreamRequest{
|
|
|
|
Topics: map[structs.Topic][]string{"Job": {"*"}},
|
|
|
|
QueryOptions: structs.QueryOptions{
|
|
|
|
Region: s1.Region(),
|
|
|
|
Namespace: structs.DefaultNamespace,
|
|
|
|
AuthToken: tokenNsFoo.SecretID,
|
|
|
|
},
|
|
|
|
}
|
|
|
|
|
|
|
|
handler, err := s1.StreamingRpcHandler("Event.Stream")
|
|
|
|
require.Nil(t, err)
|
|
|
|
|
|
|
|
p1, p2 := net.Pipe()
|
|
|
|
defer p1.Close()
|
|
|
|
defer p2.Close()
|
|
|
|
|
|
|
|
errCh := make(chan error)
|
|
|
|
streamMsg := make(chan *structs.EventStreamWrapper)
|
|
|
|
|
|
|
|
go handler(p2)
|
|
|
|
|
|
|
|
go func() {
|
|
|
|
decoder := codec.NewDecoder(p1, structs.MsgpackHandle)
|
|
|
|
for {
|
|
|
|
var msg structs.EventStreamWrapper
|
|
|
|
if err := decoder.Decode(&msg); err != nil {
|
|
|
|
if err == io.EOF || strings.Contains(err.Error(), "closed") {
|
|
|
|
return
|
|
|
|
}
|
|
|
|
errCh <- fmt.Errorf("error decoding: %w", err)
|
|
|
|
}
|
|
|
|
|
|
|
|
streamMsg <- &msg
|
|
|
|
}
|
|
|
|
}()
|
|
|
|
|
|
|
|
publisher, err := s1.State().EventBroker()
|
|
|
|
require.NoError(t, err)
|
|
|
|
|
|
|
|
job := mock.Job()
|
|
|
|
jobEvent := structs.JobEvent{
|
|
|
|
Job: job,
|
|
|
|
}
|
|
|
|
|
|
|
|
// send req
|
|
|
|
encoder := codec.NewEncoder(p1, structs.MsgpackHandle)
|
|
|
|
require.Nil(t, encoder.Encode(req))
|
|
|
|
|
|
|
|
// publish some events
|
|
|
|
publisher.Publish(&structs.Events{Index: uint64(1), Events: []structs.Event{{Topic: structs.TopicJob, Payload: jobEvent}}})
|
|
|
|
publisher.Publish(&structs.Events{Index: uint64(2), Events: []structs.Event{{Topic: structs.TopicJob, Payload: jobEvent}}})
|
|
|
|
|
|
|
|
// RPC to delete token
|
|
|
|
aclDelReq := &structs.ACLTokenDeleteRequest{
|
|
|
|
AccessorIDs: []string{tokenNsFoo.AccessorID},
|
|
|
|
WriteRequest: structs.WriteRequest{
|
|
|
|
Region: s1.Region(),
|
|
|
|
Namespace: structs.DefaultNamespace,
|
|
|
|
AuthToken: root.SecretID,
|
|
|
|
},
|
|
|
|
}
|
|
|
|
var aclResp structs.GenericResponse
|
|
|
|
|
|
|
|
ctx, cancel := context.WithDeadline(context.Background(), time.Now().Add(5*time.Second))
|
|
|
|
defer cancel()
|
|
|
|
|
|
|
|
codec := rpcClient(t, s1)
|
|
|
|
errChStream := make(chan error, 1)
|
|
|
|
go func() {
|
|
|
|
for {
|
|
|
|
select {
|
|
|
|
case <-ctx.Done():
|
|
|
|
errChStream <- ctx.Err()
|
|
|
|
return
|
|
|
|
case err := <-errCh:
|
|
|
|
errChStream <- err
|
|
|
|
return
|
|
|
|
case msg := <-streamMsg:
|
|
|
|
if msg.Error == nil {
|
|
|
|
// received a valid event, make RPC to delete token
|
|
|
|
// continue trying for error
|
|
|
|
continue
|
|
|
|
}
|
|
|
|
|
|
|
|
errChStream <- msg.Error
|
|
|
|
return
|
|
|
|
}
|
|
|
|
}
|
|
|
|
}()
|
|
|
|
|
|
|
|
// Delete the token used to create the stream
|
|
|
|
require.NoError(t, msgpackrpc.CallWithCodec(codec, "ACL.DeleteTokens", aclDelReq, &aclResp))
|
|
|
|
timeout := time.After(5 * time.Second)
|
|
|
|
OUTER:
|
|
|
|
for {
|
|
|
|
select {
|
|
|
|
case <-timeout:
|
|
|
|
t.Fatal("timeout waiting for event stream")
|
|
|
|
case err := <-errCh:
|
|
|
|
t.Fatal(err)
|
|
|
|
case err := <-errChStream:
|
|
|
|
// Success
|
|
|
|
require.Contains(t, err.Error(), stream.ErrSubscriptionClosed.Error())
|
|
|
|
break OUTER
|
|
|
|
}
|
|
|
|
}
|
|
|
|
}
|