Backport of Fix more test flakes into release/1.16.x (#19545)
* backport of commit 5cb614ed7323303a5e55a2648410cbd9ec5099bd * Fix flaky metrics tests. --------- Co-authored-by: Derek Menteer <derek.menteer@hashicorp.com>
This commit is contained in:
parent
d7a81cb144
commit
fa78010b04
|
@ -12,26 +12,37 @@ import (
|
||||||
"os"
|
"os"
|
||||||
"path/filepath"
|
"path/filepath"
|
||||||
"strings"
|
"strings"
|
||||||
|
"sync/atomic"
|
||||||
"testing"
|
"testing"
|
||||||
"time"
|
|
||||||
|
|
||||||
"github.com/stretchr/testify/require"
|
"github.com/stretchr/testify/require"
|
||||||
|
|
||||||
"github.com/hashicorp/consul/agent/rpc/middleware"
|
"github.com/hashicorp/consul/agent/rpc/middleware"
|
||||||
"github.com/hashicorp/consul/lib/retry"
|
|
||||||
"github.com/hashicorp/consul/sdk/testutil"
|
"github.com/hashicorp/consul/sdk/testutil"
|
||||||
|
"github.com/hashicorp/consul/sdk/testutil/retry"
|
||||||
|
testretry "github.com/hashicorp/consul/sdk/testutil/retry"
|
||||||
"github.com/hashicorp/consul/testrpc"
|
"github.com/hashicorp/consul/testrpc"
|
||||||
"github.com/hashicorp/consul/tlsutil"
|
"github.com/hashicorp/consul/tlsutil"
|
||||||
)
|
)
|
||||||
|
|
||||||
|
var metricsPrefixCounter atomic.Uint64
|
||||||
|
|
||||||
|
// getUniqueMetricsPrefix generates a unique ID for each test to use as a metrics prefix.
|
||||||
|
// This is needed because go-metrics is effectively a global variable.
|
||||||
|
func getUniqueMetricsPrefix() string {
|
||||||
|
return fmt.Sprint("metrics_", metricsPrefixCounter.Add(1))
|
||||||
|
}
|
||||||
|
|
||||||
func skipIfShortTesting(t *testing.T) {
|
func skipIfShortTesting(t *testing.T) {
|
||||||
if testing.Short() {
|
if testing.Short() {
|
||||||
t.Skip("too slow for testing.Short")
|
t.Skip("too slow for testing.Short")
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
func recordPromMetrics(t *testing.T, a *TestAgent, respRec *httptest.ResponseRecorder) {
|
func recordPromMetrics(t require.TestingT, a *TestAgent, respRec *httptest.ResponseRecorder) {
|
||||||
t.Helper()
|
if tt, ok := t.(*testing.T); ok {
|
||||||
|
tt.Helper()
|
||||||
|
}
|
||||||
req, err := http.NewRequest("GET", "/v1/agent/metrics?format=prometheus", nil)
|
req, err := http.NewRequest("GET", "/v1/agent/metrics?format=prometheus", nil)
|
||||||
require.NoError(t, err, "Failed to generate new http request.")
|
require.NoError(t, err, "Failed to generate new http request.")
|
||||||
|
|
||||||
|
@ -178,7 +189,7 @@ func TestAgent_OneTwelveRPCMetrics(t *testing.T) {
|
||||||
// This test cannot use t.Parallel() since we modify global state, ie the global metrics instance
|
// This test cannot use t.Parallel() since we modify global state, ie the global metrics instance
|
||||||
|
|
||||||
t.Run("Check that 1.12 rpc metrics are not emitted by default.", func(t *testing.T) {
|
t.Run("Check that 1.12 rpc metrics are not emitted by default.", func(t *testing.T) {
|
||||||
metricsPrefix := "new_rpc_metrics"
|
metricsPrefix := getUniqueMetricsPrefix()
|
||||||
hcl := fmt.Sprintf(`
|
hcl := fmt.Sprintf(`
|
||||||
telemetry = {
|
telemetry = {
|
||||||
prometheus_retention_time = "5s"
|
prometheus_retention_time = "5s"
|
||||||
|
@ -201,7 +212,7 @@ func TestAgent_OneTwelveRPCMetrics(t *testing.T) {
|
||||||
})
|
})
|
||||||
|
|
||||||
t.Run("Check that 1.12 rpc metrics are emitted when specified by operator.", func(t *testing.T) {
|
t.Run("Check that 1.12 rpc metrics are emitted when specified by operator.", func(t *testing.T) {
|
||||||
metricsPrefix := "new_rpc_metrics_2"
|
metricsPrefix := getUniqueMetricsPrefix()
|
||||||
allowRPCMetricRule := metricsPrefix + "." + strings.Join(middleware.OneTwelveRPCSummary[0].Name, ".")
|
allowRPCMetricRule := metricsPrefix + "." + strings.Join(middleware.OneTwelveRPCSummary[0].Name, ".")
|
||||||
hcl := fmt.Sprintf(`
|
hcl := fmt.Sprintf(`
|
||||||
telemetry = {
|
telemetry = {
|
||||||
|
@ -240,44 +251,62 @@ func TestHTTPHandlers_AgentMetrics_LeaderShipMetrics(t *testing.T) {
|
||||||
// This test cannot use t.Parallel() since we modify global state, ie the global metrics instance
|
// This test cannot use t.Parallel() since we modify global state, ie the global metrics instance
|
||||||
|
|
||||||
t.Run("check that metric isLeader is set properly on server", func(t *testing.T) {
|
t.Run("check that metric isLeader is set properly on server", func(t *testing.T) {
|
||||||
hcl := `
|
metricsPrefix1 := getUniqueMetricsPrefix()
|
||||||
|
metricsPrefix2 := getUniqueMetricsPrefix()
|
||||||
|
|
||||||
|
hcl1 := fmt.Sprintf(`
|
||||||
|
server = true
|
||||||
telemetry = {
|
telemetry = {
|
||||||
prometheus_retention_time = "5s",
|
prometheus_retention_time = "25s",
|
||||||
metrics_prefix = "agent_is_leader"
|
disable_hostname = true
|
||||||
|
metrics_prefix = "%s"
|
||||||
}
|
}
|
||||||
|
`, metricsPrefix1)
|
||||||
|
|
||||||
|
hcl2 := fmt.Sprintf(`
|
||||||
|
server = true
|
||||||
|
telemetry = {
|
||||||
|
prometheus_retention_time = "25s",
|
||||||
|
disable_hostname = true
|
||||||
|
metrics_prefix = "%s"
|
||||||
|
}
|
||||||
|
`, metricsPrefix2)
|
||||||
|
|
||||||
|
overrides := `
|
||||||
|
bootstrap = false
|
||||||
|
bootstrap_expect = 2
|
||||||
`
|
`
|
||||||
|
|
||||||
a := StartTestAgent(t, TestAgent{HCL: hcl})
|
s1 := StartTestAgent(t, TestAgent{Name: "s1", HCL: hcl1, Overrides: overrides})
|
||||||
defer a.Shutdown()
|
s2 := StartTestAgent(t, TestAgent{Name: "s2", HCL: hcl2, Overrides: overrides})
|
||||||
|
defer s1.Shutdown()
|
||||||
|
defer s2.Shutdown()
|
||||||
|
|
||||||
retryWithBackoff := func(expectedStr string) error {
|
|
||||||
waiter := &retry.Waiter{
|
|
||||||
MaxWait: 1 * time.Minute,
|
|
||||||
}
|
|
||||||
ctx := context.Background()
|
|
||||||
for {
|
|
||||||
if waiter.Failures() > 7 {
|
|
||||||
return fmt.Errorf("reach max failure: %d", waiter.Failures())
|
|
||||||
}
|
|
||||||
respRec := httptest.NewRecorder()
|
|
||||||
recordPromMetrics(t, a, respRec)
|
|
||||||
|
|
||||||
out := respRec.Body.String()
|
|
||||||
if strings.Contains(out, expectedStr) {
|
|
||||||
return nil
|
|
||||||
}
|
|
||||||
waiter.Wait(ctx)
|
|
||||||
}
|
|
||||||
}
|
|
||||||
// agent hasn't become a leader
|
// agent hasn't become a leader
|
||||||
err := retryWithBackoff("isLeader 0")
|
retry.RunWith(retry.ThirtySeconds(), t, func(r *testretry.R) {
|
||||||
require.NoError(t, err, "non-leader server should have isLeader 0")
|
respRec := httptest.NewRecorder()
|
||||||
|
recordPromMetrics(r, s1, respRec)
|
||||||
|
found := strings.Contains(respRec.Body.String(), metricsPrefix1+"_server_isLeader 0")
|
||||||
|
require.True(r, found, "non-leader server should have isLeader 0")
|
||||||
|
})
|
||||||
|
|
||||||
testrpc.WaitForLeader(t, a.RPC, "dc1")
|
_, err := s2.JoinLAN([]string{s1.Config.SerfBindAddrLAN.String()}, nil)
|
||||||
|
require.NoError(t, err)
|
||||||
|
testrpc.WaitForLeader(t, s1.RPC, "dc1")
|
||||||
|
testrpc.WaitForLeader(t, s2.RPC, "dc1")
|
||||||
|
|
||||||
// Verify agent's isLeader metrics is 1
|
// Verify agent's isLeader metrics is 1
|
||||||
err = retryWithBackoff("isLeader 1")
|
retry.RunWith(retry.ThirtySeconds(), t, func(r *testretry.R) {
|
||||||
require.NoError(t, err, "leader should have isLeader 1")
|
respRec1 := httptest.NewRecorder()
|
||||||
|
recordPromMetrics(r, s1, respRec1)
|
||||||
|
found1 := strings.Contains(respRec1.Body.String(), metricsPrefix1+"_server_isLeader 1")
|
||||||
|
|
||||||
|
respRec2 := httptest.NewRecorder()
|
||||||
|
recordPromMetrics(r, s2, respRec2)
|
||||||
|
found2 := strings.Contains(respRec2.Body.String(), metricsPrefix2+"_server_isLeader 1")
|
||||||
|
|
||||||
|
require.True(r, found1 || found2, "leader server should have isLeader 1")
|
||||||
|
})
|
||||||
})
|
})
|
||||||
}
|
}
|
||||||
|
|
||||||
|
@ -288,15 +317,16 @@ func TestHTTPHandlers_AgentMetrics_ConsulAutopilot_Prometheus(t *testing.T) {
|
||||||
// This test cannot use t.Parallel() since we modify global state, ie the global metrics instance
|
// This test cannot use t.Parallel() since we modify global state, ie the global metrics instance
|
||||||
|
|
||||||
t.Run("Check consul_autopilot_* are not emitted metrics on clients", func(t *testing.T) {
|
t.Run("Check consul_autopilot_* are not emitted metrics on clients", func(t *testing.T) {
|
||||||
hcl := `
|
metricsPrefix := getUniqueMetricsPrefix()
|
||||||
|
hcl := fmt.Sprintf(`
|
||||||
telemetry = {
|
telemetry = {
|
||||||
prometheus_retention_time = "5s"
|
prometheus_retention_time = "5s"
|
||||||
disable_hostname = true
|
disable_hostname = true
|
||||||
metrics_prefix = "agent_1"
|
metrics_prefix = "%s"
|
||||||
}
|
}
|
||||||
bootstrap = false
|
bootstrap = false
|
||||||
server = false
|
server = false
|
||||||
`
|
`, metricsPrefix)
|
||||||
|
|
||||||
a := StartTestAgent(t, TestAgent{HCL: hcl})
|
a := StartTestAgent(t, TestAgent{HCL: hcl})
|
||||||
defer a.Shutdown()
|
defer a.Shutdown()
|
||||||
|
@ -304,21 +334,22 @@ func TestHTTPHandlers_AgentMetrics_ConsulAutopilot_Prometheus(t *testing.T) {
|
||||||
respRec := httptest.NewRecorder()
|
respRec := httptest.NewRecorder()
|
||||||
recordPromMetrics(t, a, respRec)
|
recordPromMetrics(t, a, respRec)
|
||||||
|
|
||||||
assertMetricNotExists(t, respRec, "agent_1_autopilot_healthy")
|
assertMetricNotExists(t, respRec, metricsPrefix+"_autopilot_healthy")
|
||||||
assertMetricNotExists(t, respRec, "agent_1_autopilot_failure_tolerance")
|
assertMetricNotExists(t, respRec, metricsPrefix+"_autopilot_failure_tolerance")
|
||||||
})
|
})
|
||||||
|
|
||||||
t.Run("Check consul_autopilot_healthy metric value on startup", func(t *testing.T) {
|
t.Run("Check consul_autopilot_healthy metric value on startup", func(t *testing.T) {
|
||||||
|
metricsPrefix := getUniqueMetricsPrefix()
|
||||||
|
hcl := fmt.Sprintf(`
|
||||||
// don't bootstrap agent so as not to
|
// don't bootstrap agent so as not to
|
||||||
// become a leader
|
// become a leader
|
||||||
hcl := `
|
|
||||||
telemetry = {
|
telemetry = {
|
||||||
prometheus_retention_time = "5s",
|
prometheus_retention_time = "5s",
|
||||||
disable_hostname = true
|
disable_hostname = true
|
||||||
metrics_prefix = "agent_2"
|
metrics_prefix = "%s"
|
||||||
}
|
}
|
||||||
bootstrap = false
|
bootstrap = false
|
||||||
`
|
`, metricsPrefix)
|
||||||
|
|
||||||
a := StartTestAgent(t, TestAgent{HCL: hcl})
|
a := StartTestAgent(t, TestAgent{HCL: hcl})
|
||||||
defer a.Shutdown()
|
defer a.Shutdown()
|
||||||
|
@ -326,8 +357,8 @@ func TestHTTPHandlers_AgentMetrics_ConsulAutopilot_Prometheus(t *testing.T) {
|
||||||
respRec := httptest.NewRecorder()
|
respRec := httptest.NewRecorder()
|
||||||
recordPromMetrics(t, a, respRec)
|
recordPromMetrics(t, a, respRec)
|
||||||
|
|
||||||
assertMetricExistsWithValue(t, respRec, "agent_2_autopilot_healthy", "1")
|
assertMetricExistsWithValue(t, respRec, metricsPrefix+"_autopilot_healthy", "1")
|
||||||
assertMetricExistsWithValue(t, respRec, "agent_2_autopilot_failure_tolerance", "0")
|
assertMetricExistsWithValue(t, respRec, metricsPrefix+"_autopilot_failure_tolerance", "0")
|
||||||
})
|
})
|
||||||
}
|
}
|
||||||
|
|
||||||
|
@ -363,16 +394,17 @@ func TestHTTPHandlers_AgentMetrics_TLSCertExpiry_Prometheus(t *testing.T) {
|
||||||
err = os.WriteFile(keyPath, []byte(key), 0600)
|
err = os.WriteFile(keyPath, []byte(key), 0600)
|
||||||
require.NoError(t, err)
|
require.NoError(t, err)
|
||||||
|
|
||||||
|
metricsPrefix := getUniqueMetricsPrefix()
|
||||||
hcl := fmt.Sprintf(`
|
hcl := fmt.Sprintf(`
|
||||||
telemetry = {
|
telemetry = {
|
||||||
prometheus_retention_time = "5s",
|
prometheus_retention_time = "5s",
|
||||||
disable_hostname = true
|
disable_hostname = true
|
||||||
metrics_prefix = "agent_3"
|
metrics_prefix = "%s"
|
||||||
}
|
}
|
||||||
ca_file = "%s"
|
ca_file = "%s"
|
||||||
cert_file = "%s"
|
cert_file = "%s"
|
||||||
key_file = "%s"
|
key_file = "%s"
|
||||||
`, caPath, certPath, keyPath)
|
`, metricsPrefix, caPath, certPath, keyPath)
|
||||||
|
|
||||||
a := StartTestAgent(t, TestAgent{HCL: hcl})
|
a := StartTestAgent(t, TestAgent{HCL: hcl})
|
||||||
defer a.Shutdown()
|
defer a.Shutdown()
|
||||||
|
@ -380,7 +412,7 @@ func TestHTTPHandlers_AgentMetrics_TLSCertExpiry_Prometheus(t *testing.T) {
|
||||||
respRec := httptest.NewRecorder()
|
respRec := httptest.NewRecorder()
|
||||||
recordPromMetrics(t, a, respRec)
|
recordPromMetrics(t, a, respRec)
|
||||||
|
|
||||||
require.Contains(t, respRec.Body.String(), "agent_3_agent_tls_cert_expiry 1.7")
|
require.Contains(t, respRec.Body.String(), metricsPrefix+"_agent_tls_cert_expiry 1.7")
|
||||||
}
|
}
|
||||||
|
|
||||||
func TestHTTPHandlers_AgentMetrics_CACertExpiry_Prometheus(t *testing.T) {
|
func TestHTTPHandlers_AgentMetrics_CACertExpiry_Prometheus(t *testing.T) {
|
||||||
|
@ -388,17 +420,18 @@ func TestHTTPHandlers_AgentMetrics_CACertExpiry_Prometheus(t *testing.T) {
|
||||||
// This test cannot use t.Parallel() since we modify global state, ie the global metrics instance
|
// This test cannot use t.Parallel() since we modify global state, ie the global metrics instance
|
||||||
|
|
||||||
t.Run("non-leader emits NaN", func(t *testing.T) {
|
t.Run("non-leader emits NaN", func(t *testing.T) {
|
||||||
hcl := `
|
metricsPrefix := getUniqueMetricsPrefix()
|
||||||
|
hcl := fmt.Sprintf(`
|
||||||
telemetry = {
|
telemetry = {
|
||||||
prometheus_retention_time = "5s",
|
prometheus_retention_time = "5s",
|
||||||
disable_hostname = true
|
disable_hostname = true
|
||||||
metrics_prefix = "agent_4"
|
metrics_prefix = "%s"
|
||||||
}
|
}
|
||||||
connect {
|
connect {
|
||||||
enabled = true
|
enabled = true
|
||||||
}
|
}
|
||||||
bootstrap = false
|
bootstrap = false
|
||||||
`
|
`, metricsPrefix)
|
||||||
|
|
||||||
a := StartTestAgent(t, TestAgent{HCL: hcl})
|
a := StartTestAgent(t, TestAgent{HCL: hcl})
|
||||||
defer a.Shutdown()
|
defer a.Shutdown()
|
||||||
|
@ -406,21 +439,22 @@ func TestHTTPHandlers_AgentMetrics_CACertExpiry_Prometheus(t *testing.T) {
|
||||||
respRec := httptest.NewRecorder()
|
respRec := httptest.NewRecorder()
|
||||||
recordPromMetrics(t, a, respRec)
|
recordPromMetrics(t, a, respRec)
|
||||||
|
|
||||||
require.Contains(t, respRec.Body.String(), "agent_4_mesh_active_root_ca_expiry NaN")
|
require.Contains(t, respRec.Body.String(), metricsPrefix+"_mesh_active_root_ca_expiry NaN")
|
||||||
require.Contains(t, respRec.Body.String(), "agent_4_mesh_active_signing_ca_expiry NaN")
|
require.Contains(t, respRec.Body.String(), metricsPrefix+"_mesh_active_signing_ca_expiry NaN")
|
||||||
})
|
})
|
||||||
|
|
||||||
t.Run("leader emits a value", func(t *testing.T) {
|
t.Run("leader emits a value", func(t *testing.T) {
|
||||||
hcl := `
|
metricsPrefix := getUniqueMetricsPrefix()
|
||||||
|
hcl := fmt.Sprintf(`
|
||||||
telemetry = {
|
telemetry = {
|
||||||
prometheus_retention_time = "5s",
|
prometheus_retention_time = "5s",
|
||||||
disable_hostname = true
|
disable_hostname = true
|
||||||
metrics_prefix = "agent_5"
|
metrics_prefix = "%s"
|
||||||
}
|
}
|
||||||
connect {
|
connect {
|
||||||
enabled = true
|
enabled = true
|
||||||
}
|
}
|
||||||
`
|
`, metricsPrefix)
|
||||||
|
|
||||||
a := StartTestAgent(t, TestAgent{HCL: hcl})
|
a := StartTestAgent(t, TestAgent{HCL: hcl})
|
||||||
defer a.Shutdown()
|
defer a.Shutdown()
|
||||||
|
@ -430,8 +464,8 @@ func TestHTTPHandlers_AgentMetrics_CACertExpiry_Prometheus(t *testing.T) {
|
||||||
recordPromMetrics(t, a, respRec)
|
recordPromMetrics(t, a, respRec)
|
||||||
|
|
||||||
out := respRec.Body.String()
|
out := respRec.Body.String()
|
||||||
require.Contains(t, out, "agent_5_mesh_active_root_ca_expiry 3.15")
|
require.Contains(t, out, metricsPrefix+"_mesh_active_root_ca_expiry 3.15")
|
||||||
require.Contains(t, out, "agent_5_mesh_active_signing_ca_expiry 3.15")
|
require.Contains(t, out, metricsPrefix+"_mesh_active_signing_ca_expiry 3.15")
|
||||||
})
|
})
|
||||||
|
|
||||||
}
|
}
|
||||||
|
@ -441,18 +475,19 @@ func TestHTTPHandlers_AgentMetrics_WAL_Prometheus(t *testing.T) {
|
||||||
// This test cannot use t.Parallel() since we modify global state, ie the global metrics instance
|
// This test cannot use t.Parallel() since we modify global state, ie the global metrics instance
|
||||||
|
|
||||||
t.Run("client agent emits nothing", func(t *testing.T) {
|
t.Run("client agent emits nothing", func(t *testing.T) {
|
||||||
hcl := `
|
metricsPrefix := getUniqueMetricsPrefix()
|
||||||
|
hcl := fmt.Sprintf(`
|
||||||
server = false
|
server = false
|
||||||
telemetry = {
|
telemetry = {
|
||||||
prometheus_retention_time = "5s",
|
prometheus_retention_time = "5s",
|
||||||
disable_hostname = true
|
disable_hostname = true
|
||||||
metrics_prefix = "agent_4"
|
metrics_prefix = "%s"
|
||||||
}
|
}
|
||||||
raft_logstore {
|
raft_logstore {
|
||||||
backend = "wal"
|
backend = "wal"
|
||||||
}
|
}
|
||||||
bootstrap = false
|
bootstrap = false
|
||||||
`
|
`, metricsPrefix)
|
||||||
|
|
||||||
a := StartTestAgent(t, TestAgent{HCL: hcl})
|
a := StartTestAgent(t, TestAgent{HCL: hcl})
|
||||||
defer a.Shutdown()
|
defer a.Shutdown()
|
||||||
|
@ -460,17 +495,18 @@ func TestHTTPHandlers_AgentMetrics_WAL_Prometheus(t *testing.T) {
|
||||||
respRec := httptest.NewRecorder()
|
respRec := httptest.NewRecorder()
|
||||||
recordPromMetrics(t, a, respRec)
|
recordPromMetrics(t, a, respRec)
|
||||||
|
|
||||||
require.NotContains(t, respRec.Body.String(), "agent_4_raft_wal")
|
require.NotContains(t, respRec.Body.String(), metricsPrefix+"_raft_wal")
|
||||||
})
|
})
|
||||||
|
|
||||||
t.Run("server with WAL enabled emits WAL metrics", func(t *testing.T) {
|
t.Run("server with WAL enabled emits WAL metrics", func(t *testing.T) {
|
||||||
hcl := `
|
metricsPrefix := getUniqueMetricsPrefix()
|
||||||
|
hcl := fmt.Sprintf(`
|
||||||
server = true
|
server = true
|
||||||
bootstrap = true
|
bootstrap = true
|
||||||
telemetry = {
|
telemetry = {
|
||||||
prometheus_retention_time = "5s",
|
prometheus_retention_time = "5s",
|
||||||
disable_hostname = true
|
disable_hostname = true
|
||||||
metrics_prefix = "agent_5"
|
metrics_prefix = "%s"
|
||||||
}
|
}
|
||||||
connect {
|
connect {
|
||||||
enabled = true
|
enabled = true
|
||||||
|
@ -478,44 +514,40 @@ func TestHTTPHandlers_AgentMetrics_WAL_Prometheus(t *testing.T) {
|
||||||
raft_logstore {
|
raft_logstore {
|
||||||
backend = "wal"
|
backend = "wal"
|
||||||
}
|
}
|
||||||
`
|
`, metricsPrefix)
|
||||||
|
|
||||||
a := StartTestAgent(t, TestAgent{HCL: hcl})
|
a := StartTestAgent(t, TestAgent{HCL: hcl})
|
||||||
defer a.Shutdown()
|
defer a.Shutdown()
|
||||||
testrpc.WaitForLeader(t, a.RPC, "dc1")
|
testrpc.WaitForLeader(t, a.RPC, "dc1")
|
||||||
|
|
||||||
|
testretry.Run(t, func(r *testretry.R) {
|
||||||
respRec := httptest.NewRecorder()
|
respRec := httptest.NewRecorder()
|
||||||
recordPromMetrics(t, a, respRec)
|
recordPromMetrics(r, a, respRec)
|
||||||
|
|
||||||
out := respRec.Body.String()
|
out := respRec.Body.String()
|
||||||
defer func() {
|
require.Contains(r, out, metricsPrefix+"_raft_wal_head_truncations")
|
||||||
if t.Failed() {
|
require.Contains(r, out, metricsPrefix+"_raft_wal_last_segment_age_seconds")
|
||||||
t.Log("--- Failed output START ---")
|
require.Contains(r, out, metricsPrefix+"_raft_wal_log_appends")
|
||||||
t.Log(out)
|
require.Contains(r, out, metricsPrefix+"_raft_wal_log_entries_read")
|
||||||
t.Log("--- Failed output END ---")
|
require.Contains(r, out, metricsPrefix+"_raft_wal_log_entries_written")
|
||||||
}
|
require.Contains(r, out, metricsPrefix+"_raft_wal_log_entry_bytes_read")
|
||||||
}()
|
require.Contains(r, out, metricsPrefix+"_raft_wal_log_entry_bytes_written")
|
||||||
require.Contains(t, out, "agent_5_raft_wal_head_truncations")
|
require.Contains(r, out, metricsPrefix+"_raft_wal_segment_rotations")
|
||||||
require.Contains(t, out, "agent_5_raft_wal_last_segment_age_seconds")
|
require.Contains(r, out, metricsPrefix+"_raft_wal_stable_gets")
|
||||||
require.Contains(t, out, "agent_5_raft_wal_log_appends")
|
require.Contains(r, out, metricsPrefix+"_raft_wal_stable_sets")
|
||||||
require.Contains(t, out, "agent_5_raft_wal_log_entries_read")
|
require.Contains(r, out, metricsPrefix+"_raft_wal_tail_truncations")
|
||||||
require.Contains(t, out, "agent_5_raft_wal_log_entries_written")
|
})
|
||||||
require.Contains(t, out, "agent_5_raft_wal_log_entry_bytes_read")
|
|
||||||
require.Contains(t, out, "agent_5_raft_wal_log_entry_bytes_written")
|
|
||||||
require.Contains(t, out, "agent_5_raft_wal_segment_rotations")
|
|
||||||
require.Contains(t, out, "agent_5_raft_wal_stable_gets")
|
|
||||||
require.Contains(t, out, "agent_5_raft_wal_stable_sets")
|
|
||||||
require.Contains(t, out, "agent_5_raft_wal_tail_truncations")
|
|
||||||
})
|
})
|
||||||
|
|
||||||
t.Run("server without WAL enabled emits no WAL metrics", func(t *testing.T) {
|
t.Run("server without WAL enabled emits no WAL metrics", func(t *testing.T) {
|
||||||
hcl := `
|
metricsPrefix := getUniqueMetricsPrefix()
|
||||||
|
hcl := fmt.Sprintf(`
|
||||||
server = true
|
server = true
|
||||||
bootstrap = true
|
bootstrap = true
|
||||||
telemetry = {
|
telemetry = {
|
||||||
prometheus_retention_time = "5s",
|
prometheus_retention_time = "5s",
|
||||||
disable_hostname = true
|
disable_hostname = true
|
||||||
metrics_prefix = "agent_6"
|
metrics_prefix = "%s"
|
||||||
}
|
}
|
||||||
connect {
|
connect {
|
||||||
enabled = true
|
enabled = true
|
||||||
|
@ -523,7 +555,7 @@ func TestHTTPHandlers_AgentMetrics_WAL_Prometheus(t *testing.T) {
|
||||||
raft_logstore {
|
raft_logstore {
|
||||||
backend = "boltdb"
|
backend = "boltdb"
|
||||||
}
|
}
|
||||||
`
|
`, metricsPrefix)
|
||||||
|
|
||||||
a := StartTestAgent(t, TestAgent{HCL: hcl})
|
a := StartTestAgent(t, TestAgent{HCL: hcl})
|
||||||
defer a.Shutdown()
|
defer a.Shutdown()
|
||||||
|
@ -532,7 +564,7 @@ func TestHTTPHandlers_AgentMetrics_WAL_Prometheus(t *testing.T) {
|
||||||
respRec := httptest.NewRecorder()
|
respRec := httptest.NewRecorder()
|
||||||
recordPromMetrics(t, a, respRec)
|
recordPromMetrics(t, a, respRec)
|
||||||
|
|
||||||
require.NotContains(t, respRec.Body.String(), "agent_6_raft_wal")
|
require.NotContains(t, respRec.Body.String(), metricsPrefix+"_raft_wal")
|
||||||
})
|
})
|
||||||
|
|
||||||
}
|
}
|
||||||
|
@ -542,12 +574,13 @@ func TestHTTPHandlers_AgentMetrics_LogVerifier_Prometheus(t *testing.T) {
|
||||||
// This test cannot use t.Parallel() since we modify global state, ie the global metrics instance
|
// This test cannot use t.Parallel() since we modify global state, ie the global metrics instance
|
||||||
|
|
||||||
t.Run("client agent emits nothing", func(t *testing.T) {
|
t.Run("client agent emits nothing", func(t *testing.T) {
|
||||||
hcl := `
|
metricsPrefix := getUniqueMetricsPrefix()
|
||||||
|
hcl := fmt.Sprintf(`
|
||||||
server = false
|
server = false
|
||||||
telemetry = {
|
telemetry = {
|
||||||
prometheus_retention_time = "5s",
|
prometheus_retention_time = "5s",
|
||||||
disable_hostname = true
|
disable_hostname = true
|
||||||
metrics_prefix = "agent_4"
|
metrics_prefix = "%s"
|
||||||
}
|
}
|
||||||
raft_logstore {
|
raft_logstore {
|
||||||
verification {
|
verification {
|
||||||
|
@ -556,7 +589,7 @@ func TestHTTPHandlers_AgentMetrics_LogVerifier_Prometheus(t *testing.T) {
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
bootstrap = false
|
bootstrap = false
|
||||||
`
|
`, metricsPrefix)
|
||||||
|
|
||||||
a := StartTestAgent(t, TestAgent{HCL: hcl})
|
a := StartTestAgent(t, TestAgent{HCL: hcl})
|
||||||
defer a.Shutdown()
|
defer a.Shutdown()
|
||||||
|
@ -564,17 +597,18 @@ func TestHTTPHandlers_AgentMetrics_LogVerifier_Prometheus(t *testing.T) {
|
||||||
respRec := httptest.NewRecorder()
|
respRec := httptest.NewRecorder()
|
||||||
recordPromMetrics(t, a, respRec)
|
recordPromMetrics(t, a, respRec)
|
||||||
|
|
||||||
require.NotContains(t, respRec.Body.String(), "agent_4_raft_logstore_verifier")
|
require.NotContains(t, respRec.Body.String(), metricsPrefix+"_raft_logstore_verifier")
|
||||||
})
|
})
|
||||||
|
|
||||||
t.Run("server with verifier enabled emits all metrics", func(t *testing.T) {
|
t.Run("server with verifier enabled emits all metrics", func(t *testing.T) {
|
||||||
hcl := `
|
metricsPrefix := getUniqueMetricsPrefix()
|
||||||
|
hcl := fmt.Sprintf(`
|
||||||
server = true
|
server = true
|
||||||
bootstrap = true
|
bootstrap = true
|
||||||
telemetry = {
|
telemetry = {
|
||||||
prometheus_retention_time = "5s",
|
prometheus_retention_time = "5s",
|
||||||
disable_hostname = true
|
disable_hostname = true
|
||||||
metrics_prefix = "agent_5"
|
metrics_prefix = "%s"
|
||||||
}
|
}
|
||||||
connect {
|
connect {
|
||||||
enabled = true
|
enabled = true
|
||||||
|
@ -585,38 +619,33 @@ func TestHTTPHandlers_AgentMetrics_LogVerifier_Prometheus(t *testing.T) {
|
||||||
interval = "1s"
|
interval = "1s"
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
`
|
`, metricsPrefix)
|
||||||
|
|
||||||
a := StartTestAgent(t, TestAgent{HCL: hcl})
|
a := StartTestAgent(t, TestAgent{HCL: hcl})
|
||||||
defer a.Shutdown()
|
defer a.Shutdown()
|
||||||
testrpc.WaitForLeader(t, a.RPC, "dc1")
|
testrpc.WaitForLeader(t, a.RPC, "dc1")
|
||||||
|
|
||||||
|
testretry.Run(t, func(r *testretry.R) {
|
||||||
respRec := httptest.NewRecorder()
|
respRec := httptest.NewRecorder()
|
||||||
recordPromMetrics(t, a, respRec)
|
recordPromMetrics(r, a, respRec)
|
||||||
|
|
||||||
out := respRec.Body.String()
|
out := respRec.Body.String()
|
||||||
defer func() {
|
require.Contains(r, out, metricsPrefix+"_raft_logstore_verifier_checkpoints_written")
|
||||||
if t.Failed() {
|
require.Contains(r, out, metricsPrefix+"_raft_logstore_verifier_dropped_reports")
|
||||||
t.Log("--- Failed output START ---")
|
require.Contains(r, out, metricsPrefix+"_raft_logstore_verifier_ranges_verified")
|
||||||
t.Log(out)
|
require.Contains(r, out, metricsPrefix+"_raft_logstore_verifier_read_checksum_failures")
|
||||||
t.Log("--- Failed output END ---")
|
require.Contains(r, out, metricsPrefix+"_raft_logstore_verifier_write_checksum_failures")
|
||||||
}
|
})
|
||||||
}()
|
|
||||||
require.Contains(t, out, "agent_5_raft_logstore_verifier_checkpoints_written")
|
|
||||||
require.Contains(t, out, "agent_5_raft_logstore_verifier_dropped_reports")
|
|
||||||
require.Contains(t, out, "agent_5_raft_logstore_verifier_ranges_verified")
|
|
||||||
require.Contains(t, out, "agent_5_raft_logstore_verifier_read_checksum_failures")
|
|
||||||
require.Contains(t, out, "agent_5_raft_logstore_verifier_write_checksum_failures")
|
|
||||||
})
|
})
|
||||||
|
|
||||||
t.Run("server with verifier disabled emits no extra metrics", func(t *testing.T) {
|
t.Run("server with verifier disabled emits no extra metrics", func(t *testing.T) {
|
||||||
hcl := `
|
metricsPrefix := getUniqueMetricsPrefix()
|
||||||
|
hcl := fmt.Sprintf(`
|
||||||
server = true
|
server = true
|
||||||
bootstrap = true
|
bootstrap = true
|
||||||
telemetry = {
|
telemetry = {
|
||||||
prometheus_retention_time = "5s",
|
prometheus_retention_time = "5s",
|
||||||
disable_hostname = true
|
disable_hostname = true
|
||||||
metrics_prefix = "agent_6"
|
metrics_prefix = "%s"
|
||||||
}
|
}
|
||||||
connect {
|
connect {
|
||||||
enabled = true
|
enabled = true
|
||||||
|
@ -626,7 +655,7 @@ func TestHTTPHandlers_AgentMetrics_LogVerifier_Prometheus(t *testing.T) {
|
||||||
enabled = false
|
enabled = false
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
`
|
`, metricsPrefix)
|
||||||
|
|
||||||
a := StartTestAgent(t, TestAgent{HCL: hcl})
|
a := StartTestAgent(t, TestAgent{HCL: hcl})
|
||||||
defer a.Shutdown()
|
defer a.Shutdown()
|
||||||
|
@ -635,7 +664,7 @@ func TestHTTPHandlers_AgentMetrics_LogVerifier_Prometheus(t *testing.T) {
|
||||||
respRec := httptest.NewRecorder()
|
respRec := httptest.NewRecorder()
|
||||||
recordPromMetrics(t, a, respRec)
|
recordPromMetrics(t, a, respRec)
|
||||||
|
|
||||||
require.NotContains(t, respRec.Body.String(), "agent_6_raft_logstore_verifier")
|
require.NotContains(t, respRec.Body.String(), metricsPrefix+"_raft_logstore_verifier")
|
||||||
})
|
})
|
||||||
|
|
||||||
}
|
}
|
||||||
|
|
|
@ -74,6 +74,7 @@ func TestSnapshot_Options(t *testing.T) {
|
||||||
t.Run(method, func(t *testing.T) {
|
t.Run(method, func(t *testing.T) {
|
||||||
a := NewTestAgent(t, TestACLConfig())
|
a := NewTestAgent(t, TestACLConfig())
|
||||||
defer a.Shutdown()
|
defer a.Shutdown()
|
||||||
|
testrpc.WaitForLeader(t, a.RPC, "dc1")
|
||||||
|
|
||||||
body := bytes.NewBuffer(nil)
|
body := bytes.NewBuffer(nil)
|
||||||
req, _ := http.NewRequest(method, "/v1/snapshot", body)
|
req, _ := http.NewRequest(method, "/v1/snapshot", body)
|
||||||
|
@ -88,6 +89,7 @@ func TestSnapshot_Options(t *testing.T) {
|
||||||
t.Run(method, func(t *testing.T) {
|
t.Run(method, func(t *testing.T) {
|
||||||
a := NewTestAgent(t, TestACLConfig())
|
a := NewTestAgent(t, TestACLConfig())
|
||||||
defer a.Shutdown()
|
defer a.Shutdown()
|
||||||
|
testrpc.WaitForLeader(t, a.RPC, "dc1")
|
||||||
|
|
||||||
body := bytes.NewBuffer(nil)
|
body := bytes.NewBuffer(nil)
|
||||||
req, _ := http.NewRequest(method, "/v1/snapshot?dc=nope", body)
|
req, _ := http.NewRequest(method, "/v1/snapshot?dc=nope", body)
|
||||||
|
@ -101,6 +103,7 @@ func TestSnapshot_Options(t *testing.T) {
|
||||||
t.Run(method, func(t *testing.T) {
|
t.Run(method, func(t *testing.T) {
|
||||||
a := NewTestAgent(t, TestACLConfig())
|
a := NewTestAgent(t, TestACLConfig())
|
||||||
defer a.Shutdown()
|
defer a.Shutdown()
|
||||||
|
testrpc.WaitForLeader(t, a.RPC, "dc1")
|
||||||
|
|
||||||
body := bytes.NewBuffer(nil)
|
body := bytes.NewBuffer(nil)
|
||||||
req, _ := http.NewRequest(method, "/v1/snapshot?stale", body)
|
req, _ := http.NewRequest(method, "/v1/snapshot?stale", body)
|
||||||
|
|
Loading…
Reference in New Issue