115 lines
3.2 KiB
Go
115 lines
3.2 KiB
Go
package metrics
|
|
|
|
import (
|
|
"context"
|
|
"fmt"
|
|
"io/ioutil"
|
|
"net/http"
|
|
"net/url"
|
|
"strings"
|
|
"testing"
|
|
"time"
|
|
|
|
"github.com/stretchr/testify/require"
|
|
|
|
libcluster "github.com/hashicorp/consul/integration/consul-container/libs/cluster"
|
|
"github.com/hashicorp/consul/integration/consul-container/libs/node"
|
|
"github.com/hashicorp/consul/integration/consul-container/libs/utils"
|
|
"github.com/hashicorp/consul/lib/retry"
|
|
)
|
|
|
|
// Given a 3-server cluster, when the leader is elected, then leader's isLeader is 1 and non-leader's 0
|
|
func TestLeadershipMetrics(t *testing.T) {
|
|
var configs []node.Config
|
|
configs = append(configs,
|
|
node.Config{
|
|
HCL: `node_name="` + utils.RandName("consul-server") + `"
|
|
log_level="TRACE"
|
|
server=true
|
|
telemetry {
|
|
statsite_address = "127.0.0.1:2180"
|
|
}`,
|
|
Cmd: []string{"agent", "-client=0.0.0.0"},
|
|
Version: *utils.TargetImage,
|
|
})
|
|
|
|
numServer := 3
|
|
for i := 1; i < numServer; i++ {
|
|
configs = append(configs,
|
|
node.Config{
|
|
HCL: `node_name="` + utils.RandName("consul-server") + `"
|
|
log_level="TRACE"
|
|
bootstrap_expect=3
|
|
server=true`,
|
|
Cmd: []string{"agent", "-client=0.0.0.0"},
|
|
Version: *utils.TargetImage,
|
|
})
|
|
|
|
}
|
|
|
|
cluster, err := libcluster.New(configs)
|
|
require.NoError(t, err)
|
|
defer terminate(t, cluster)
|
|
|
|
svrCli := cluster.Nodes[0].GetClient()
|
|
libcluster.WaitForLeader(t, cluster, svrCli)
|
|
libcluster.WaitForMembers(t, svrCli, 3)
|
|
|
|
retryWithBackoff := func(agentNode node.Node, expectedStr string) error {
|
|
waiter := &retry.Waiter{
|
|
MaxWait: 5 * time.Minute,
|
|
}
|
|
_, port := agentNode.GetAddr()
|
|
ctx := context.Background()
|
|
for {
|
|
if waiter.Failures() > 5 {
|
|
return fmt.Errorf("reach max failure: %d", waiter.Failures())
|
|
}
|
|
|
|
metricsStr, err := getMetrics(t, "127.0.0.1", port, "/v1/agent/metrics")
|
|
if err != nil {
|
|
return fmt.Errorf("error get metrics: %v", err)
|
|
}
|
|
if strings.Contains(metricsStr, expectedStr) {
|
|
return nil
|
|
}
|
|
waiter.Wait(ctx)
|
|
}
|
|
}
|
|
|
|
leaderNode, err := cluster.Leader()
|
|
require.NoError(t, err)
|
|
leadAddr, leaderPort := leaderNode.GetAddr()
|
|
|
|
for i, n := range cluster.Nodes {
|
|
addr, port := n.GetAddr()
|
|
if addr == leadAddr && port == leaderPort {
|
|
err = retryWithBackoff(leaderNode, ".server.isLeader\",\"Value\":1,")
|
|
require.NoError(t, err, "%dth node(leader): could not find the metric %q in the /v1/agent/metrics response", i, ".server.isLeader\",\"Value\":1,")
|
|
} else {
|
|
err = retryWithBackoff(n, ".server.isLeader\",\"Value\":0,")
|
|
require.NoError(t, err, "%dth node(non-leader): could not find the metric %q in the /v1/agent/metrics response", i, ".server.isLeader\",\"Value\":0,")
|
|
}
|
|
}
|
|
}
|
|
|
|
func getMetrics(t *testing.T, addr string, port int, path string) (string, error) {
|
|
u, err := url.Parse(fmt.Sprintf("http://%s:%d", addr, port))
|
|
require.NoError(t, err)
|
|
u.Path = path
|
|
resp, err := http.Get(u.String())
|
|
if err != nil {
|
|
return "", fmt.Errorf("error get metrics: %v", err)
|
|
}
|
|
body, err := ioutil.ReadAll(resp.Body)
|
|
if err != nil {
|
|
return "nil", fmt.Errorf("error read metrics: %v", err)
|
|
}
|
|
return string(body), nil
|
|
}
|
|
|
|
func terminate(t *testing.T, cluster *libcluster.Cluster) {
|
|
err := cluster.Terminate()
|
|
require.NoError(t, err)
|
|
}
|