#!/usr/bin/env bats load helpers @test "s1 proxy is running correct version" { assert_envoy_version 19000 } @test "s1 proxy admin is up on :19000" { retry_default curl -f -s localhost:19000/stats -o /dev/null } @test "s2 proxy admin is up on :19001" { retry_default curl -f -s localhost:19001/stats -o /dev/null } @test "s1 proxy listener should be up and have right cert" { assert_proxy_presents_cert_uri localhost:21000 s1 } @test "s2 proxy listener should be up and have right cert" { assert_proxy_presents_cert_uri localhost:21001 s2 } @test "s2 proxies should be healthy in primary" { assert_service_has_healthy_instances s2 1 primary } @test "s2 proxies should be healthy in secondary" { assert_service_has_healthy_instances s2 1 secondary } @test "gateway-secondary should be up and listening" { retry_long nc -z consul-secondary-client:4432 } ################ # PHASE 1: we show that by default requests are served from the primary # Note: when failover is configured the cluster is named for the original # service not any destination related to failover. @test "s1 upstream should have healthy endpoints for s2 in both primary and failover" { assert_upstream_has_endpoints_in_status 127.0.0.1:19000 failover-target~0~s2.default.primary HEALTHY 1 assert_upstream_has_endpoints_in_status 127.0.0.1:19000 failover-target~1~s2.default.primary HEALTHY 1 } @test "s1 upstream should be able to connect to s2 via upstream s2 to start" { assert_expected_fortio_name s2 } @test "s1 upstream made 1 connection" { assert_envoy_metric_at_least 127.0.0.1:19000 "cluster.failover-target~0~s2.default.primary.*cx_total" 1 } ################ # PHASE 2: we show that in failover requests are served from the secondary # @test "terminate instance of s2 primary envoy which should trigger failover to s2 secondary when tcp check fails" { kill_envoy s2 primary } @test "s2 proxies should be unhealthy in primary" { assert_service_has_healthy_instances s2 0 primary } @test "s1 upstream should have healthy endpoints for s2 secondary and unhealthy endpoints for s2 primary" { assert_upstream_has_endpoints_in_status 127.0.0.1:19000 failover-target~0~s2.default.primary UNHEALTHY 1 assert_upstream_has_endpoints_in_status 127.0.0.1:19000 failover-target~1~s2.default.primary HEALTHY 1 } @test "reset envoy statistics" { reset_envoy_metrics 127.0.0.1:19000 } @test "s1 upstream should be able to connect to s2 in secondary now" { assert_expected_fortio_name s2-secondary } @test "s1 upstream made 1 connection again" { assert_envoy_metric_at_least 127.0.0.1:19000 "cluster.failover-target~1~s2.default.primary.*cx_total" 1 }