package watch_test import ( "bytes" "crypto/ecdsa" "crypto/elliptic" "crypto/rand" "crypto/sha256" "crypto/x509" "crypto/x509/pkix" "encoding/json" "encoding/pem" "fmt" "math/big" "net/url" "sync" "testing" "time" "github.com/hashicorp/consul/api" "github.com/hashicorp/consul/api/watch" "github.com/hashicorp/consul/sdk/testutil" "github.com/stretchr/testify/require" ) func makeClient(t *testing.T) (*api.Client, *testutil.TestServer) { // Make client config conf := api.DefaultConfig() // Create server server, err := testutil.NewTestServerConfigT(t, nil) require.NoError(t, err) conf.Address = server.HTTPAddr server.WaitForLeader(t) // Create client client, err := api.NewClient(conf) if err != nil { server.Stop() // guaranteed to fail but will be a nice error message require.NoError(t, err) } return client, server } func updateConnectCA(t *testing.T, client *api.Client) { t.Helper() connect := client.Connect() cfg, _, err := connect.CAGetConfig(nil) require.NoError(t, err) require.NotNil(t, cfg.Config) // update the cert // Create the private key we'll use for this CA cert. pk, err := ecdsa.GenerateKey(elliptic.P256(), rand.Reader) require.NoError(t, err) bs, err := x509.MarshalECPrivateKey(pk) require.NoError(t, err) var buf bytes.Buffer require.NoError(t, pem.Encode(&buf, &pem.Block{Type: "EC PRIVATE KEY", Bytes: bs})) cfg.Config["PrivateKey"] = buf.String() bs, err = x509.MarshalPKIXPublicKey(pk.Public()) require.NoError(t, err) hash := sha256.Sum256(bs) kID := hash[:] // Create the CA cert template := x509.Certificate{ SerialNumber: big.NewInt(42), Subject: pkix.Name{CommonName: "CA Modified"}, URIs: []*url.URL{&url.URL{Scheme: "spiffe", Host: fmt.Sprintf("11111111-2222-3333-4444-555555555555.%s", "consul")}}, BasicConstraintsValid: true, KeyUsage: x509.KeyUsageCertSign | x509.KeyUsageCRLSign | x509.KeyUsageDigitalSignature, IsCA: true, NotAfter: time.Now().AddDate(10, 0, 0), NotBefore: time.Now(), AuthorityKeyId: kID, SubjectKeyId: kID, } bs, err = x509.CreateCertificate(rand.Reader, &template, &template, pk.Public(), pk) require.NoError(t, err) buf.Reset() require.NoError(t, pem.Encode(&buf, &pem.Block{Type: "CERTIFICATE", Bytes: bs})) cfg.Config["RootCert"] = buf.String() _, err = connect.CASetConfig(cfg, nil) require.NoError(t, err) } func TestKeyWatch(t *testing.T) { t.Parallel() c, s := makeClient(t) defer s.Stop() var ( wakeups []*api.KVPair notifyCh = make(chan struct{}) ) plan := mustParse(t, `{"type":"key", "key":"foo/bar/baz"}`) plan.Handler = func(idx uint64, raw interface{}) { var v *api.KVPair if raw == nil { // nil is a valid return value v = nil } else { var ok bool if v, ok = raw.(*api.KVPair); !ok { return // ignore } } wakeups = append(wakeups, v) notifyCh <- struct{}{} } var wg sync.WaitGroup wg.Add(1) go func() { defer wg.Done() if err := plan.Run(s.HTTPAddr); err != nil { t.Fatalf("err: %v", err) } }() defer plan.Stop() // Wait for first wakeup. <-notifyCh { kv := c.KV() pair := &api.KVPair{ Key: "foo/bar/baz", Value: []byte("test"), } if _, err := kv.Put(pair, nil); err != nil { t.Fatalf("err: %v", err) } } // Wait for second wakeup. <-notifyCh plan.Stop() wg.Wait() require.Len(t, wakeups, 2) { v := wakeups[0] require.Nil(t, v) } { v := wakeups[1] require.Equal(t, "test", string(v.Value)) } } func TestKeyWatch_With_PrefixDelete(t *testing.T) { t.Parallel() c, s := makeClient(t) defer s.Stop() var ( wakeups []*api.KVPair notifyCh = make(chan struct{}) ) plan := mustParse(t, `{"type":"key", "key":"foo/bar/baz"}`) plan.Handler = func(idx uint64, raw interface{}) { var v *api.KVPair if raw == nil { // nil is a valid return value v = nil } else { var ok bool if v, ok = raw.(*api.KVPair); !ok { return // ignore } } wakeups = append(wakeups, v) notifyCh <- struct{}{} } var wg sync.WaitGroup wg.Add(1) go func() { defer wg.Done() if err := plan.Run(s.HTTPAddr); err != nil { t.Fatalf("err: %v", err) } }() defer plan.Stop() // Wait for first wakeup. <-notifyCh { kv := c.KV() pair := &api.KVPair{ Key: "foo/bar/baz", Value: []byte("test"), } if _, err := kv.Put(pair, nil); err != nil { t.Fatalf("err: %v", err) } } // Wait for second wakeup. <-notifyCh plan.Stop() wg.Wait() require.Len(t, wakeups, 2) { v := wakeups[0] require.Nil(t, v) } { v := wakeups[1] require.Equal(t, "test", string(v.Value)) } } func TestKeyPrefixWatch(t *testing.T) { t.Parallel() c, s := makeClient(t) defer s.Stop() var ( wakeups []api.KVPairs notifyCh = make(chan struct{}) ) plan := mustParse(t, `{"type":"keyprefix", "prefix":"foo/"}`) plan.Handler = func(idx uint64, raw interface{}) { if raw == nil { return // ignore } v, ok := raw.(api.KVPairs) if !ok { return } wakeups = append(wakeups, v) notifyCh <- struct{}{} } var wg sync.WaitGroup wg.Add(1) go func() { defer wg.Done() if err := plan.Run(s.HTTPAddr); err != nil { t.Fatalf("err: %v", err) } }() defer plan.Stop() // Wait for first wakeup. <-notifyCh { kv := c.KV() pair := &api.KVPair{ Key: "foo/bar", } if _, err := kv.Put(pair, nil); err != nil { t.Fatalf("err: %v", err) } } // Wait for second wakeup. <-notifyCh plan.Stop() wg.Wait() require.Len(t, wakeups, 2) { v := wakeups[0] require.Len(t, v, 0) } { v := wakeups[1] require.Len(t, v, 1) require.Equal(t, "foo/bar", v[0].Key) } } func TestServicesWatch(t *testing.T) { t.Parallel() c, s := makeClient(t) defer s.Stop() var ( wakeups []map[string][]string notifyCh = make(chan struct{}) ) plan := mustParse(t, `{"type":"services"}`) plan.Handler = func(idx uint64, raw interface{}) { if raw == nil { return // ignore } v, ok := raw.(map[string][]string) if !ok { return // ignore } wakeups = append(wakeups, v) notifyCh <- struct{}{} } var wg sync.WaitGroup wg.Add(1) go func() { defer wg.Done() if err := plan.Run(s.HTTPAddr); err != nil { t.Fatalf("err: %v", err) } }() defer plan.Stop() // Wait for first wakeup. <-notifyCh { agent := c.Agent() reg := &api.AgentServiceRegistration{ ID: "foo", Name: "foo", } if err := agent.ServiceRegister(reg); err != nil { t.Fatalf("err: %v", err) } } // Wait for second wakeup. <-notifyCh plan.Stop() wg.Wait() require.Len(t, wakeups, 2) { v := wakeups[0] require.Len(t, v, 1) _, ok := v["consul"] require.True(t, ok) } { v := wakeups[1] require.Len(t, v, 2) _, ok := v["consul"] require.True(t, ok) _, ok = v["foo"] require.True(t, ok) } } func TestNodesWatch(t *testing.T) { t.Parallel() c, s := makeClient(t) defer s.Stop() s.WaitForSerfCheck(t) // wait for AE to sync var ( wakeups [][]*api.Node notifyCh = make(chan struct{}) ) plan := mustParse(t, `{"type":"nodes"}`) plan.Handler = func(idx uint64, raw interface{}) { if raw == nil { return // ignore } v, ok := raw.([]*api.Node) if !ok { return // ignore } wakeups = append(wakeups, v) notifyCh <- struct{}{} } var wg sync.WaitGroup wg.Add(1) go func() { defer wg.Done() if err := plan.Run(s.HTTPAddr); err != nil { t.Fatalf("err: %v", err) } }() defer plan.Stop() // Wait for first wakeup. <-notifyCh { catalog := c.Catalog() reg := &api.CatalogRegistration{ Node: "foobar", Address: "1.1.1.1", Datacenter: "dc1", } if _, err := catalog.Register(reg, nil); err != nil { t.Fatalf("err: %v", err) } } // Wait for second wakeup. <-notifyCh plan.Stop() wg.Wait() require.Len(t, wakeups, 2) var originalNodeName string { v := wakeups[0] require.Len(t, v, 1) originalNodeName = v[0].Node } { v := wakeups[1] require.Len(t, v, 2) if v[0].Node == originalNodeName { require.Equal(t, "foobar", v[1].Node) } else { require.Equal(t, "foobar", v[0].Node) } } } func TestServiceWatch(t *testing.T) { t.Parallel() c, s := makeClient(t) defer s.Stop() var ( wakeups [][]*api.ServiceEntry notifyCh = make(chan struct{}) ) plan := mustParse(t, `{"type":"service", "service":"foo", "tag":"bar", "passingonly":true}`) plan.Handler = func(idx uint64, raw interface{}) { if raw == nil { return // ignore } v, ok := raw.([]*api.ServiceEntry) if !ok { return // ignore } wakeups = append(wakeups, v) notifyCh <- struct{}{} } var wg sync.WaitGroup wg.Add(1) go func() { defer wg.Done() if err := plan.Run(s.HTTPAddr); err != nil { t.Fatalf("err: %v", err) } }() defer plan.Stop() // Wait for first wakeup. <-notifyCh { agent := c.Agent() reg := &api.AgentServiceRegistration{ ID: "foo", Name: "foo", Tags: []string{"bar"}, } if err := agent.ServiceRegister(reg); err != nil { t.Fatalf("err: %v", err) } } // Wait for second wakeup. <-notifyCh plan.Stop() wg.Wait() require.Len(t, wakeups, 2) { v := wakeups[0] require.Len(t, v, 0) } { v := wakeups[1] require.Len(t, v, 1) require.Equal(t, "foo", v[0].Service.ID) } } func TestServiceMultipleTagsWatch(t *testing.T) { t.Parallel() c, s := makeClient(t) defer s.Stop() var ( wakeups [][]*api.ServiceEntry notifyCh = make(chan struct{}) ) plan := mustParse(t, `{"type":"service", "service":"foo", "tag":["bar","buzz"], "passingonly":true}`) plan.Handler = func(idx uint64, raw interface{}) { if raw == nil { return // ignore } v, ok := raw.([]*api.ServiceEntry) if !ok { return // ignore } wakeups = append(wakeups, v) notifyCh <- struct{}{} } var wg sync.WaitGroup wg.Add(1) go func() { defer wg.Done() if err := plan.Run(s.HTTPAddr); err != nil { t.Fatalf("err: %v", err) } }() defer plan.Stop() // Wait for first wakeup. <-notifyCh { agent := c.Agent() // we do not want to find this one. reg := &api.AgentServiceRegistration{ ID: "foobarbiff", Name: "foo", Tags: []string{"bar", "biff"}, } if err := agent.ServiceRegister(reg); err != nil { t.Fatalf("err: %v", err) } // we do not want to find this one. reg = &api.AgentServiceRegistration{ ID: "foobuzzbiff", Name: "foo", Tags: []string{"buzz", "biff"}, } if err := agent.ServiceRegister(reg); err != nil { t.Fatalf("err: %v", err) } // we want to find this one reg = &api.AgentServiceRegistration{ ID: "foobarbuzzbiff", Name: "foo", Tags: []string{"bar", "buzz", "biff"}, } if err := agent.ServiceRegister(reg); err != nil { t.Fatalf("err: %v", err) } } // Wait for second wakeup. <-notifyCh plan.Stop() wg.Wait() require.Len(t, wakeups, 2) { v := wakeups[0] require.Len(t, v, 0) } { v := wakeups[1] require.Len(t, v, 1) require.Equal(t, "foobarbuzzbiff", v[0].Service.ID) require.ElementsMatch(t, []string{"bar", "buzz", "biff"}, v[0].Service.Tags) } } func TestChecksWatch_State(t *testing.T) { t.Parallel() c, s := makeClient(t) defer s.Stop() var ( wakeups [][]*api.HealthCheck notifyCh = make(chan struct{}) ) plan := mustParse(t, `{"type":"checks", "state":"warning"}`) plan.Handler = func(idx uint64, raw interface{}) { if raw == nil { return // ignore } v, ok := raw.([]*api.HealthCheck) if !ok { return // ignore } wakeups = append(wakeups, v) notifyCh <- struct{}{} } var wg sync.WaitGroup wg.Add(1) go func() { defer wg.Done() if err := plan.Run(s.HTTPAddr); err != nil { t.Fatalf("err: %v", err) } }() defer plan.Stop() // Wait for first wakeup. <-notifyCh { catalog := c.Catalog() reg := &api.CatalogRegistration{ Node: "foobar", Address: "1.1.1.1", Datacenter: "dc1", Check: &api.AgentCheck{ Node: "foobar", CheckID: "foobar", Name: "foobar", Status: api.HealthWarning, }, } if _, err := catalog.Register(reg, nil); err != nil { t.Fatalf("err: %v", err) } } // Wait for second wakeup. <-notifyCh plan.Stop() wg.Wait() require.Len(t, wakeups, 2) { v := wakeups[0] require.Len(t, v, 0) } { v := wakeups[1] require.Len(t, v, 1) require.Equal(t, "foobar", v[0].CheckID) require.Equal(t, "warning", v[0].Status) } } func TestChecksWatch_Service(t *testing.T) { t.Parallel() c, s := makeClient(t) defer s.Stop() var ( wakeups [][]*api.HealthCheck notifyCh = make(chan struct{}) ) plan := mustParse(t, `{"type":"checks", "service":"foobar"}`) plan.Handler = func(idx uint64, raw interface{}) { if raw == nil { return // ignore } v, ok := raw.([]*api.HealthCheck) if !ok { return // ignore } wakeups = append(wakeups, v) notifyCh <- struct{}{} } var wg sync.WaitGroup wg.Add(1) go func() { defer wg.Done() if err := plan.Run(s.HTTPAddr); err != nil { t.Fatalf("err: %v", err) } }() defer plan.Stop() // Wait for first wakeup. <-notifyCh { catalog := c.Catalog() reg := &api.CatalogRegistration{ Node: "foobar", Address: "1.1.1.1", Datacenter: "dc1", Service: &api.AgentService{ ID: "foobar", Service: "foobar", }, Check: &api.AgentCheck{ Node: "foobar", CheckID: "foobar", Name: "foobar", Status: api.HealthPassing, ServiceID: "foobar", }, } if _, err := catalog.Register(reg, nil); err != nil { t.Fatalf("err: %v", err) } } // Wait for second wakeup. <-notifyCh plan.Stop() wg.Wait() require.Len(t, wakeups, 2) { v := wakeups[0] require.Len(t, v, 0) } { v := wakeups[1] require.Len(t, v, 1) require.Equal(t, "foobar", v[0].CheckID) } } func TestEventWatch(t *testing.T) { t.Parallel() c, s := makeClient(t) defer s.Stop() var ( wakeups [][]*api.UserEvent notifyCh = make(chan struct{}) ) plan := mustParse(t, `{"type":"event", "name": "foo"}`) plan.Handler = func(idx uint64, raw interface{}) { if raw == nil { return } v, ok := raw.([]*api.UserEvent) if !ok { return // ignore } wakeups = append(wakeups, v) notifyCh <- struct{}{} } var wg sync.WaitGroup wg.Add(1) go func() { defer wg.Done() if err := plan.Run(s.HTTPAddr); err != nil { t.Fatalf("err: %v", err) } }() defer plan.Stop() // Wait for first wakeup. <-notifyCh { event := c.Event() params := &api.UserEvent{Name: "foo"} if _, _, err := event.Fire(params, nil); err != nil { t.Fatalf("err: %v", err) } } // Wait for second wakeup. <-notifyCh plan.Stop() wg.Wait() require.Len(t, wakeups, 2) { v := wakeups[0] require.Len(t, v, 0) } { v := wakeups[1] require.Len(t, v, 1) require.Equal(t, "foo", v[0].Name) } } func TestConnectRootsWatch(t *testing.T) { t.Parallel() // makeClient will bootstrap a CA c, s := makeClient(t) defer s.Stop() var ( wakeups []*api.CARootList notifyCh = make(chan struct{}) ) plan := mustParse(t, `{"type":"connect_roots"}`) plan.Handler = func(idx uint64, raw interface{}) { if raw == nil { return // ignore } v, ok := raw.(*api.CARootList) if !ok || v == nil || len(v.Roots) == 0 { return // ignore } wakeups = append(wakeups, v) notifyCh <- struct{}{} } var wg sync.WaitGroup wg.Add(1) go func() { defer wg.Done() if err := plan.Run(s.HTTPAddr); err != nil { t.Fatalf("err: %v", err) } }() defer plan.Stop() // Wait for first wakeup. <-notifyCh updateConnectCA(t, c) // Wait for second wakeup. <-notifyCh plan.Stop() wg.Wait() require.Len(t, wakeups, 2) var originalCAID string { // Only 1 CA is the bootstrapped state (i.e. first response). v := wakeups[0] require.Len(t, v.Roots, 1) originalCAID = v.ActiveRootID require.NotEmpty(t, originalCAID) } { // This is the new CA showing up. v := wakeups[1] require.Len(t, v.Roots, 2) require.NotEqual(t, originalCAID, v.ActiveRootID) } } func TestConnectLeafWatch(t *testing.T) { t.Parallel() // makeClient will bootstrap a CA c, s := makeClient(t) defer s.Stop() // Register a web service to get certs for { agent := c.Agent() reg := api.AgentServiceRegistration{ ID: "web", Name: "web", Port: 9090, } err := agent.ServiceRegister(®) require.Nil(t, err) } var ( wakeups []*api.LeafCert notifyCh = make(chan struct{}) ) plan := mustParse(t, `{"type":"connect_leaf", "service":"web"}`) plan.Handler = func(idx uint64, raw interface{}) { if raw == nil { return // ignore } v, ok := raw.(*api.LeafCert) if !ok || v == nil { return // ignore } wakeups = append(wakeups, v) notifyCh <- struct{}{} } var wg sync.WaitGroup wg.Add(1) go func() { defer wg.Done() if err := plan.Run(s.HTTPAddr); err != nil { t.Fatalf("err: %v", err) } }() defer plan.Stop() // Wait for first wakeup. <-notifyCh start := time.Now() updateConnectCA(t, c) // Wait for second wakeup. Note due to the 30s random jitter for this endpoint // this may take up to 30s. <-notifyCh t.Logf("leaf cert regen took %s with jitter", time.Since(start)) plan.Stop() wg.Wait() require.Len(t, wakeups, 2) var lastCert *api.LeafCert { // Initial fetch, just store the cert and return v := wakeups[0] require.NotNil(t, v) lastCert = v } { // Rotation completed. v := wakeups[1] require.NotNil(t, v) // TODO(banks): right now the root rotation actually causes Serial numbers // to reset so these end up all being the same. That needs fixing but it's // a bigger task than I want to bite off for this PR. //require.NotEqual(t, lastCert.SerialNumber, v.SerialNumber) require.NotEqual(t, lastCert.CertPEM, v.CertPEM) require.NotEqual(t, lastCert.PrivateKeyPEM, v.PrivateKeyPEM) require.NotEqual(t, lastCert.ModifyIndex, v.ModifyIndex) } } func TestAgentServiceWatch(t *testing.T) { t.Parallel() c, s := makeClient(t) defer s.Stop() var ( wakeups []*api.AgentService notifyCh = make(chan struct{}) ) // Register a local agent service reg := &api.AgentServiceRegistration{ Name: "web", Port: 8080, } client := c agent := client.Agent() err := agent.ServiceRegister(reg) require.NoError(t, err) plan := mustParse(t, `{"type":"agent_service", "service_id":"web"}`) plan.HybridHandler = func(blockParamVal watch.BlockingParamVal, raw interface{}) { if raw == nil { return // ignore } v, ok := raw.(*api.AgentService) if !ok || v == nil { return // ignore } wakeups = append(wakeups, v) notifyCh <- struct{}{} } var wg sync.WaitGroup wg.Add(1) go func() { defer wg.Done() if err := plan.Run(s.HTTPAddr); err != nil { t.Fatalf("err: %v", err) } }() defer plan.Stop() // Wait for first wakeup. <-notifyCh { // Change the service definition reg.Port = 9090 err := agent.ServiceRegister(reg) require.NoError(t, err) } // Wait for second wakeup. <-notifyCh plan.Stop() wg.Wait() require.Len(t, wakeups, 2) { v := wakeups[0] require.Equal(t, 8080, v.Port) } { v := wakeups[1] require.Equal(t, 9090, v.Port) } } func mustParse(t *testing.T, q string) *watch.Plan { t.Helper() var params map[string]interface{} if err := json.Unmarshal([]byte(q), ¶ms); err != nil { t.Fatal(err) } plan, err := watch.Parse(params) if err != nil { t.Fatalf("err: %v", err) } return plan }