Implement the service endpoints controller (#17216)
* Add a ReplaceType dep mapper and move them into their own file * Implement the service endpoints controller * Implement a Catalog Controllers Integration Test
This commit is contained in:
parent
378a15af32
commit
08f9cd1678
2
go.mod
2
go.mod
|
@ -103,6 +103,7 @@ require (
|
|||
go.opentelemetry.io/proto/otlp v0.19.0
|
||||
go.uber.org/goleak v1.1.10
|
||||
golang.org/x/crypto v0.1.0
|
||||
golang.org/x/exp v0.0.0-20230321023759-10a507213a29
|
||||
golang.org/x/net v0.10.0
|
||||
golang.org/x/oauth2 v0.6.0
|
||||
golang.org/x/sync v0.2.0
|
||||
|
@ -240,7 +241,6 @@ require (
|
|||
go.opencensus.io v0.24.0 // indirect
|
||||
go.opentelemetry.io/otel/trace v1.16.0 // indirect
|
||||
go.uber.org/atomic v1.9.0 // indirect
|
||||
golang.org/x/exp v0.0.0-20230321023759-10a507213a29 // indirect
|
||||
golang.org/x/lint v0.0.0-20210508222113-6edffad5e616 // indirect
|
||||
golang.org/x/mod v0.10.0 // indirect
|
||||
golang.org/x/term v0.8.0 // indirect
|
||||
|
|
|
@ -0,0 +1,37 @@
|
|||
{
|
||||
"id": {
|
||||
"type": {
|
||||
"group": "catalog",
|
||||
"group_version": "v1alpha1",
|
||||
"kind": "Service"
|
||||
},
|
||||
"tenancy": {
|
||||
"partition": "default",
|
||||
"namespace": "default",
|
||||
"peer_name": "local"
|
||||
},
|
||||
"name": "api"
|
||||
},
|
||||
"data": {
|
||||
"@type": "hashicorp.consul.catalog.v1alpha1.Service",
|
||||
"workloads": {
|
||||
"prefixes": [
|
||||
"api-"
|
||||
]
|
||||
},
|
||||
"ports": [
|
||||
{
|
||||
"target_port": "http",
|
||||
"protocol": "PROTOCOL_HTTP"
|
||||
},
|
||||
{
|
||||
"target_port": "grpc",
|
||||
"protocol": "PROTOCOL_GRPC"
|
||||
},
|
||||
{
|
||||
"target_port": "mesh",
|
||||
"protocol": "PROTOCOL_MESH"
|
||||
}
|
||||
]
|
||||
}
|
||||
}
|
|
@ -0,0 +1,33 @@
|
|||
{
|
||||
"id": {
|
||||
"type": {
|
||||
"group": "catalog",
|
||||
"group_version": "v1alpha1",
|
||||
"kind": "HealthStatus"
|
||||
},
|
||||
"tenancy": {
|
||||
"partition": "default",
|
||||
"namespace": "default",
|
||||
"peer_name": "local"
|
||||
},
|
||||
"name": "api-1-health"
|
||||
},
|
||||
"owner": {
|
||||
"type": {
|
||||
"group": "catalog",
|
||||
"group_version": "v1alpha1",
|
||||
"kind": "Workload"
|
||||
},
|
||||
"tenancy": {
|
||||
"partition": "default",
|
||||
"namespace": "default",
|
||||
"peer_name": "local"
|
||||
},
|
||||
"name": "api-1"
|
||||
},
|
||||
"data": {
|
||||
"@type": "hashicorp.consul.catalog.v1alpha1.HealthStatus",
|
||||
"type": "synthetic",
|
||||
"status": "HEALTH_PASSING"
|
||||
}
|
||||
}
|
|
@ -0,0 +1,46 @@
|
|||
{
|
||||
"id": {
|
||||
"type": {
|
||||
"group": "catalog",
|
||||
"group_version": "v1alpha1",
|
||||
"kind": "Workload"
|
||||
},
|
||||
"tenancy": {
|
||||
"partition": "default",
|
||||
"namespace": "default",
|
||||
"peer_name": "local"
|
||||
},
|
||||
"name": "api-1"
|
||||
},
|
||||
"data": {
|
||||
"@type": "hashicorp.consul.catalog.v1alpha1.Workload",
|
||||
"addresses": [
|
||||
{
|
||||
"host": "172.16.1.1"
|
||||
},
|
||||
{
|
||||
"host": "198.18.2.1",
|
||||
"external": true,
|
||||
"ports": [
|
||||
"mesh"
|
||||
]
|
||||
}
|
||||
],
|
||||
"ports": {
|
||||
"http": {
|
||||
"port": 8080,
|
||||
"protocol": "PROTOCOL_HTTP"
|
||||
},
|
||||
"grpc": {
|
||||
"port": 9090,
|
||||
"protocol": "PROTOCOL_GRPC"
|
||||
},
|
||||
"mesh": {
|
||||
"port": 10000,
|
||||
"protocol": "PROTOCOL_MESH"
|
||||
}
|
||||
},
|
||||
"identity": "api",
|
||||
"node_name": "node-1"
|
||||
}
|
||||
}
|
|
@ -0,0 +1,33 @@
|
|||
{
|
||||
"id": {
|
||||
"type": {
|
||||
"group": "catalog",
|
||||
"group_version": "v1alpha1",
|
||||
"kind": "HealthStatus"
|
||||
},
|
||||
"tenancy": {
|
||||
"partition": "default",
|
||||
"namespace": "default",
|
||||
"peer_name": "local"
|
||||
},
|
||||
"name": "api-10-health"
|
||||
},
|
||||
"owner": {
|
||||
"type": {
|
||||
"group": "catalog",
|
||||
"group_version": "v1alpha1",
|
||||
"kind": "Workload"
|
||||
},
|
||||
"tenancy": {
|
||||
"partition": "default",
|
||||
"namespace": "default",
|
||||
"peer_name": "local"
|
||||
},
|
||||
"name": "api-10"
|
||||
},
|
||||
"data": {
|
||||
"@type": "hashicorp.consul.catalog.v1alpha1.HealthStatus",
|
||||
"type": "synthetic",
|
||||
"status": "HEALTH_WARNING"
|
||||
}
|
||||
}
|
|
@ -0,0 +1,46 @@
|
|||
{
|
||||
"id": {
|
||||
"type": {
|
||||
"group": "catalog",
|
||||
"group_version": "v1alpha1",
|
||||
"kind": "Workload"
|
||||
},
|
||||
"tenancy": {
|
||||
"partition": "default",
|
||||
"namespace": "default",
|
||||
"peer_name": "local"
|
||||
},
|
||||
"name": "api-10"
|
||||
},
|
||||
"data": {
|
||||
"@type": "hashicorp.consul.catalog.v1alpha1.Workload",
|
||||
"addresses": [
|
||||
{
|
||||
"host": "172.16.1.10"
|
||||
},
|
||||
{
|
||||
"host": "198.18.2.10",
|
||||
"external": true,
|
||||
"ports": [
|
||||
"mesh"
|
||||
]
|
||||
}
|
||||
],
|
||||
"ports": {
|
||||
"http": {
|
||||
"port": 8080,
|
||||
"protocol": "PROTOCOL_HTTP"
|
||||
},
|
||||
"grpc": {
|
||||
"port": 9090,
|
||||
"protocol": "PROTOCOL_GRPC"
|
||||
},
|
||||
"mesh": {
|
||||
"port": 10000,
|
||||
"protocol": "PROTOCOL_MESH"
|
||||
}
|
||||
},
|
||||
"identity": "api",
|
||||
"node_name": "node-3"
|
||||
}
|
||||
}
|
|
@ -0,0 +1,33 @@
|
|||
{
|
||||
"id": {
|
||||
"type": {
|
||||
"group": "catalog",
|
||||
"group_version": "v1alpha1",
|
||||
"kind": "HealthStatus"
|
||||
},
|
||||
"tenancy": {
|
||||
"partition": "default",
|
||||
"namespace": "default",
|
||||
"peer_name": "local"
|
||||
},
|
||||
"name": "api-11-health"
|
||||
},
|
||||
"owner": {
|
||||
"type": {
|
||||
"group": "catalog",
|
||||
"group_version": "v1alpha1",
|
||||
"kind": "Workload"
|
||||
},
|
||||
"tenancy": {
|
||||
"partition": "default",
|
||||
"namespace": "default",
|
||||
"peer_name": "local"
|
||||
},
|
||||
"name": "api-11"
|
||||
},
|
||||
"data": {
|
||||
"@type": "hashicorp.consul.catalog.v1alpha1.HealthStatus",
|
||||
"type": "synthetic",
|
||||
"status": "HEALTH_CRITICAL"
|
||||
}
|
||||
}
|
|
@ -0,0 +1,46 @@
|
|||
{
|
||||
"id": {
|
||||
"type": {
|
||||
"group": "catalog",
|
||||
"group_version": "v1alpha1",
|
||||
"kind": "Workload"
|
||||
},
|
||||
"tenancy": {
|
||||
"partition": "default",
|
||||
"namespace": "default",
|
||||
"peer_name": "local"
|
||||
},
|
||||
"name": "api-11"
|
||||
},
|
||||
"data": {
|
||||
"@type": "hashicorp.consul.catalog.v1alpha1.Workload",
|
||||
"addresses": [
|
||||
{
|
||||
"host": "172.16.1.11"
|
||||
},
|
||||
{
|
||||
"host": "198.18.2.11",
|
||||
"external": true,
|
||||
"ports": [
|
||||
"mesh"
|
||||
]
|
||||
}
|
||||
],
|
||||
"ports": {
|
||||
"http": {
|
||||
"port": 8080,
|
||||
"protocol": "PROTOCOL_HTTP"
|
||||
},
|
||||
"grpc": {
|
||||
"port": 9090,
|
||||
"protocol": "PROTOCOL_GRPC"
|
||||
},
|
||||
"mesh": {
|
||||
"port": 10000,
|
||||
"protocol": "PROTOCOL_MESH"
|
||||
}
|
||||
},
|
||||
"identity": "api",
|
||||
"node_name": "node-3"
|
||||
}
|
||||
}
|
|
@ -0,0 +1,33 @@
|
|||
{
|
||||
"id": {
|
||||
"type": {
|
||||
"group": "catalog",
|
||||
"group_version": "v1alpha1",
|
||||
"kind": "HealthStatus"
|
||||
},
|
||||
"tenancy": {
|
||||
"partition": "default",
|
||||
"namespace": "default",
|
||||
"peer_name": "local"
|
||||
},
|
||||
"name": "api-12-health"
|
||||
},
|
||||
"owner": {
|
||||
"type": {
|
||||
"group": "catalog",
|
||||
"group_version": "v1alpha1",
|
||||
"kind": "Workload"
|
||||
},
|
||||
"tenancy": {
|
||||
"partition": "default",
|
||||
"namespace": "default",
|
||||
"peer_name": "local"
|
||||
},
|
||||
"name": "api-12"
|
||||
},
|
||||
"data": {
|
||||
"@type": "hashicorp.consul.catalog.v1alpha1.HealthStatus",
|
||||
"type": "synthetic",
|
||||
"status": "HEALTH_MAINTENANCE"
|
||||
}
|
||||
}
|
|
@ -0,0 +1,46 @@
|
|||
{
|
||||
"id": {
|
||||
"type": {
|
||||
"group": "catalog",
|
||||
"group_version": "v1alpha1",
|
||||
"kind": "Workload"
|
||||
},
|
||||
"tenancy": {
|
||||
"partition": "default",
|
||||
"namespace": "default",
|
||||
"peer_name": "local"
|
||||
},
|
||||
"name": "api-12"
|
||||
},
|
||||
"data": {
|
||||
"@type": "hashicorp.consul.catalog.v1alpha1.Workload",
|
||||
"addresses": [
|
||||
{
|
||||
"host": "172.16.1.12"
|
||||
},
|
||||
{
|
||||
"host": "198.18.2.12",
|
||||
"external": true,
|
||||
"ports": [
|
||||
"mesh"
|
||||
]
|
||||
}
|
||||
],
|
||||
"ports": {
|
||||
"http": {
|
||||
"port": 8080,
|
||||
"protocol": "PROTOCOL_HTTP"
|
||||
},
|
||||
"grpc": {
|
||||
"port": 9090,
|
||||
"protocol": "PROTOCOL_GRPC"
|
||||
},
|
||||
"mesh": {
|
||||
"port": 10000,
|
||||
"protocol": "PROTOCOL_MESH"
|
||||
}
|
||||
},
|
||||
"identity": "api",
|
||||
"node_name": "node-3"
|
||||
}
|
||||
}
|
|
@ -0,0 +1,33 @@
|
|||
{
|
||||
"id": {
|
||||
"type": {
|
||||
"group": "catalog",
|
||||
"group_version": "v1alpha1",
|
||||
"kind": "HealthStatus"
|
||||
},
|
||||
"tenancy": {
|
||||
"partition": "default",
|
||||
"namespace": "default",
|
||||
"peer_name": "local"
|
||||
},
|
||||
"name": "api-13-health"
|
||||
},
|
||||
"owner": {
|
||||
"type": {
|
||||
"group": "catalog",
|
||||
"group_version": "v1alpha1",
|
||||
"kind": "Workload"
|
||||
},
|
||||
"tenancy": {
|
||||
"partition": "default",
|
||||
"namespace": "default",
|
||||
"peer_name": "local"
|
||||
},
|
||||
"name": "api-13"
|
||||
},
|
||||
"data": {
|
||||
"@type": "hashicorp.consul.catalog.v1alpha1.HealthStatus",
|
||||
"type": "synthetic",
|
||||
"status": "HEALTH_PASSING"
|
||||
}
|
||||
}
|
|
@ -0,0 +1,46 @@
|
|||
{
|
||||
"id": {
|
||||
"type": {
|
||||
"group": "catalog",
|
||||
"group_version": "v1alpha1",
|
||||
"kind": "Workload"
|
||||
},
|
||||
"tenancy": {
|
||||
"partition": "default",
|
||||
"namespace": "default",
|
||||
"peer_name": "local"
|
||||
},
|
||||
"name": "api-13"
|
||||
},
|
||||
"data": {
|
||||
"@type": "hashicorp.consul.catalog.v1alpha1.Workload",
|
||||
"addresses": [
|
||||
{
|
||||
"host": "172.16.1.13"
|
||||
},
|
||||
{
|
||||
"host": "198.18.2.13",
|
||||
"external": true,
|
||||
"ports": [
|
||||
"mesh"
|
||||
]
|
||||
}
|
||||
],
|
||||
"ports": {
|
||||
"http": {
|
||||
"port": 8080,
|
||||
"protocol": "PROTOCOL_HTTP"
|
||||
},
|
||||
"grpc": {
|
||||
"port": 9090,
|
||||
"protocol": "PROTOCOL_GRPC"
|
||||
},
|
||||
"mesh": {
|
||||
"port": 10000,
|
||||
"protocol": "PROTOCOL_MESH"
|
||||
}
|
||||
},
|
||||
"identity": "api",
|
||||
"node_name": "node-4"
|
||||
}
|
||||
}
|
|
@ -0,0 +1,33 @@
|
|||
{
|
||||
"id": {
|
||||
"type": {
|
||||
"group": "catalog",
|
||||
"group_version": "v1alpha1",
|
||||
"kind": "HealthStatus"
|
||||
},
|
||||
"tenancy": {
|
||||
"partition": "default",
|
||||
"namespace": "default",
|
||||
"peer_name": "local"
|
||||
},
|
||||
"name": "api-14-health"
|
||||
},
|
||||
"owner": {
|
||||
"type": {
|
||||
"group": "catalog",
|
||||
"group_version": "v1alpha1",
|
||||
"kind": "Workload"
|
||||
},
|
||||
"tenancy": {
|
||||
"partition": "default",
|
||||
"namespace": "default",
|
||||
"peer_name": "local"
|
||||
},
|
||||
"name": "api-14"
|
||||
},
|
||||
"data": {
|
||||
"@type": "hashicorp.consul.catalog.v1alpha1.HealthStatus",
|
||||
"type": "synthetic",
|
||||
"status": "HEALTH_WARNING"
|
||||
}
|
||||
}
|
|
@ -0,0 +1,46 @@
|
|||
{
|
||||
"id": {
|
||||
"type": {
|
||||
"group": "catalog",
|
||||
"group_version": "v1alpha1",
|
||||
"kind": "Workload"
|
||||
},
|
||||
"tenancy": {
|
||||
"partition": "default",
|
||||
"namespace": "default",
|
||||
"peer_name": "local"
|
||||
},
|
||||
"name": "api-14"
|
||||
},
|
||||
"data": {
|
||||
"@type": "hashicorp.consul.catalog.v1alpha1.Workload",
|
||||
"addresses": [
|
||||
{
|
||||
"host": "172.16.1.14"
|
||||
},
|
||||
{
|
||||
"host": "198.18.2.14",
|
||||
"external": true,
|
||||
"ports": [
|
||||
"mesh"
|
||||
]
|
||||
}
|
||||
],
|
||||
"ports": {
|
||||
"http": {
|
||||
"port": 8080,
|
||||
"protocol": "PROTOCOL_HTTP"
|
||||
},
|
||||
"grpc": {
|
||||
"port": 9090,
|
||||
"protocol": "PROTOCOL_GRPC"
|
||||
},
|
||||
"mesh": {
|
||||
"port": 10000,
|
||||
"protocol": "PROTOCOL_MESH"
|
||||
}
|
||||
},
|
||||
"identity": "api",
|
||||
"node_name": "node-4"
|
||||
}
|
||||
}
|
|
@ -0,0 +1,33 @@
|
|||
{
|
||||
"id": {
|
||||
"type": {
|
||||
"group": "catalog",
|
||||
"group_version": "v1alpha1",
|
||||
"kind": "HealthStatus"
|
||||
},
|
||||
"tenancy": {
|
||||
"partition": "default",
|
||||
"namespace": "default",
|
||||
"peer_name": "local"
|
||||
},
|
||||
"name": "api-15-health"
|
||||
},
|
||||
"owner": {
|
||||
"type": {
|
||||
"group": "catalog",
|
||||
"group_version": "v1alpha1",
|
||||
"kind": "Workload"
|
||||
},
|
||||
"tenancy": {
|
||||
"partition": "default",
|
||||
"namespace": "default",
|
||||
"peer_name": "local"
|
||||
},
|
||||
"name": "api-15"
|
||||
},
|
||||
"data": {
|
||||
"@type": "hashicorp.consul.catalog.v1alpha1.HealthStatus",
|
||||
"type": "synthetic",
|
||||
"status": "HEALTH_CRITICAL"
|
||||
}
|
||||
}
|
|
@ -0,0 +1,46 @@
|
|||
{
|
||||
"id": {
|
||||
"type": {
|
||||
"group": "catalog",
|
||||
"group_version": "v1alpha1",
|
||||
"kind": "Workload"
|
||||
},
|
||||
"tenancy": {
|
||||
"partition": "default",
|
||||
"namespace": "default",
|
||||
"peer_name": "local"
|
||||
},
|
||||
"name": "api-15"
|
||||
},
|
||||
"data": {
|
||||
"@type": "hashicorp.consul.catalog.v1alpha1.Workload",
|
||||
"addresses": [
|
||||
{
|
||||
"host": "172.16.1.15"
|
||||
},
|
||||
{
|
||||
"host": "198.18.2.15",
|
||||
"external": true,
|
||||
"ports": [
|
||||
"mesh"
|
||||
]
|
||||
}
|
||||
],
|
||||
"ports": {
|
||||
"http": {
|
||||
"port": 8080,
|
||||
"protocol": "PROTOCOL_HTTP"
|
||||
},
|
||||
"grpc": {
|
||||
"port": 9090,
|
||||
"protocol": "PROTOCOL_GRPC"
|
||||
},
|
||||
"mesh": {
|
||||
"port": 10000,
|
||||
"protocol": "PROTOCOL_MESH"
|
||||
}
|
||||
},
|
||||
"identity": "api",
|
||||
"node_name": "node-4"
|
||||
}
|
||||
}
|
|
@ -0,0 +1,33 @@
|
|||
{
|
||||
"id": {
|
||||
"type": {
|
||||
"group": "catalog",
|
||||
"group_version": "v1alpha1",
|
||||
"kind": "HealthStatus"
|
||||
},
|
||||
"tenancy": {
|
||||
"partition": "default",
|
||||
"namespace": "default",
|
||||
"peer_name": "local"
|
||||
},
|
||||
"name": "api-16-health"
|
||||
},
|
||||
"owner": {
|
||||
"type": {
|
||||
"group": "catalog",
|
||||
"group_version": "v1alpha1",
|
||||
"kind": "Workload"
|
||||
},
|
||||
"tenancy": {
|
||||
"partition": "default",
|
||||
"namespace": "default",
|
||||
"peer_name": "local"
|
||||
},
|
||||
"name": "api-16"
|
||||
},
|
||||
"data": {
|
||||
"@type": "hashicorp.consul.catalog.v1alpha1.HealthStatus",
|
||||
"type": "synthetic",
|
||||
"status": "HEALTH_MAINTENANCE"
|
||||
}
|
||||
}
|
|
@ -0,0 +1,46 @@
|
|||
{
|
||||
"id": {
|
||||
"type": {
|
||||
"group": "catalog",
|
||||
"group_version": "v1alpha1",
|
||||
"kind": "Workload"
|
||||
},
|
||||
"tenancy": {
|
||||
"partition": "default",
|
||||
"namespace": "default",
|
||||
"peer_name": "local"
|
||||
},
|
||||
"name": "api-16"
|
||||
},
|
||||
"data": {
|
||||
"@type": "hashicorp.consul.catalog.v1alpha1.Workload",
|
||||
"addresses": [
|
||||
{
|
||||
"host": "172.16.1.16"
|
||||
},
|
||||
{
|
||||
"host": "198.18.2.16",
|
||||
"external": true,
|
||||
"ports": [
|
||||
"mesh"
|
||||
]
|
||||
}
|
||||
],
|
||||
"ports": {
|
||||
"http": {
|
||||
"port": 8080,
|
||||
"protocol": "PROTOCOL_HTTP"
|
||||
},
|
||||
"grpc": {
|
||||
"port": 9090,
|
||||
"protocol": "PROTOCOL_GRPC"
|
||||
},
|
||||
"mesh": {
|
||||
"port": 10000,
|
||||
"protocol": "PROTOCOL_MESH"
|
||||
}
|
||||
},
|
||||
"identity": "api",
|
||||
"node_name": "node-4"
|
||||
}
|
||||
}
|
|
@ -0,0 +1,33 @@
|
|||
{
|
||||
"id": {
|
||||
"type": {
|
||||
"group": "catalog",
|
||||
"group_version": "v1alpha1",
|
||||
"kind": "HealthStatus"
|
||||
},
|
||||
"tenancy": {
|
||||
"partition": "default",
|
||||
"namespace": "default",
|
||||
"peer_name": "local"
|
||||
},
|
||||
"name": "api-17-health"
|
||||
},
|
||||
"owner": {
|
||||
"type": {
|
||||
"group": "catalog",
|
||||
"group_version": "v1alpha1",
|
||||
"kind": "Workload"
|
||||
},
|
||||
"tenancy": {
|
||||
"partition": "default",
|
||||
"namespace": "default",
|
||||
"peer_name": "local"
|
||||
},
|
||||
"name": "api-17"
|
||||
},
|
||||
"data": {
|
||||
"@type": "hashicorp.consul.catalog.v1alpha1.HealthStatus",
|
||||
"type": "synthetic",
|
||||
"status": "HEALTH_PASSING"
|
||||
}
|
||||
}
|
|
@ -0,0 +1,45 @@
|
|||
{
|
||||
"id": {
|
||||
"type": {
|
||||
"group": "catalog",
|
||||
"group_version": "v1alpha1",
|
||||
"kind": "Workload"
|
||||
},
|
||||
"tenancy": {
|
||||
"partition": "default",
|
||||
"namespace": "default",
|
||||
"peer_name": "local"
|
||||
},
|
||||
"name": "api-17"
|
||||
},
|
||||
"data": {
|
||||
"@type": "hashicorp.consul.catalog.v1alpha1.Workload",
|
||||
"addresses": [
|
||||
{
|
||||
"host": "172.16.1.17"
|
||||
},
|
||||
{
|
||||
"host": "198.18.2.17",
|
||||
"external": true,
|
||||
"ports": [
|
||||
"mesh"
|
||||
]
|
||||
}
|
||||
],
|
||||
"ports": {
|
||||
"http": {
|
||||
"port": 8080,
|
||||
"protocol": "PROTOCOL_HTTP"
|
||||
},
|
||||
"grpc": {
|
||||
"port": 9090,
|
||||
"protocol": "PROTOCOL_GRPC"
|
||||
},
|
||||
"mesh": {
|
||||
"port": 10000,
|
||||
"protocol": "PROTOCOL_MESH"
|
||||
}
|
||||
},
|
||||
"identity": "api"
|
||||
}
|
||||
}
|
|
@ -0,0 +1,33 @@
|
|||
{
|
||||
"id": {
|
||||
"type": {
|
||||
"group": "catalog",
|
||||
"group_version": "v1alpha1",
|
||||
"kind": "HealthStatus"
|
||||
},
|
||||
"tenancy": {
|
||||
"partition": "default",
|
||||
"namespace": "default",
|
||||
"peer_name": "local"
|
||||
},
|
||||
"name": "api-18-health"
|
||||
},
|
||||
"owner": {
|
||||
"type": {
|
||||
"group": "catalog",
|
||||
"group_version": "v1alpha1",
|
||||
"kind": "Workload"
|
||||
},
|
||||
"tenancy": {
|
||||
"partition": "default",
|
||||
"namespace": "default",
|
||||
"peer_name": "local"
|
||||
},
|
||||
"name": "api-18"
|
||||
},
|
||||
"data": {
|
||||
"@type": "hashicorp.consul.catalog.v1alpha1.HealthStatus",
|
||||
"type": "synthetic",
|
||||
"status": "HEALTH_WARNING"
|
||||
}
|
||||
}
|
|
@ -0,0 +1,45 @@
|
|||
{
|
||||
"id": {
|
||||
"type": {
|
||||
"group": "catalog",
|
||||
"group_version": "v1alpha1",
|
||||
"kind": "Workload"
|
||||
},
|
||||
"tenancy": {
|
||||
"partition": "default",
|
||||
"namespace": "default",
|
||||
"peer_name": "local"
|
||||
},
|
||||
"name": "api-18"
|
||||
},
|
||||
"data": {
|
||||
"@type": "hashicorp.consul.catalog.v1alpha1.Workload",
|
||||
"addresses": [
|
||||
{
|
||||
"host": "172.16.1.18"
|
||||
},
|
||||
{
|
||||
"host": "198.18.2.18",
|
||||
"external": true,
|
||||
"ports": [
|
||||
"mesh"
|
||||
]
|
||||
}
|
||||
],
|
||||
"ports": {
|
||||
"http": {
|
||||
"port": 8080,
|
||||
"protocol": "PROTOCOL_HTTP"
|
||||
},
|
||||
"grpc": {
|
||||
"port": 9090,
|
||||
"protocol": "PROTOCOL_GRPC"
|
||||
},
|
||||
"mesh": {
|
||||
"port": 10000,
|
||||
"protocol": "PROTOCOL_MESH"
|
||||
}
|
||||
},
|
||||
"identity": "api"
|
||||
}
|
||||
}
|
|
@ -0,0 +1,33 @@
|
|||
{
|
||||
"id": {
|
||||
"type": {
|
||||
"group": "catalog",
|
||||
"group_version": "v1alpha1",
|
||||
"kind": "HealthStatus"
|
||||
},
|
||||
"tenancy": {
|
||||
"partition": "default",
|
||||
"namespace": "default",
|
||||
"peer_name": "local"
|
||||
},
|
||||
"name": "api-19-health"
|
||||
},
|
||||
"owner": {
|
||||
"type": {
|
||||
"group": "catalog",
|
||||
"group_version": "v1alpha1",
|
||||
"kind": "Workload"
|
||||
},
|
||||
"tenancy": {
|
||||
"partition": "default",
|
||||
"namespace": "default",
|
||||
"peer_name": "local"
|
||||
},
|
||||
"name": "api-19"
|
||||
},
|
||||
"data": {
|
||||
"@type": "hashicorp.consul.catalog.v1alpha1.HealthStatus",
|
||||
"type": "synthetic",
|
||||
"status": "HEALTH_CRITICAL"
|
||||
}
|
||||
}
|
|
@ -0,0 +1,45 @@
|
|||
{
|
||||
"id": {
|
||||
"type": {
|
||||
"group": "catalog",
|
||||
"group_version": "v1alpha1",
|
||||
"kind": "Workload"
|
||||
},
|
||||
"tenancy": {
|
||||
"partition": "default",
|
||||
"namespace": "default",
|
||||
"peer_name": "local"
|
||||
},
|
||||
"name": "api-19"
|
||||
},
|
||||
"data": {
|
||||
"@type": "hashicorp.consul.catalog.v1alpha1.Workload",
|
||||
"addresses": [
|
||||
{
|
||||
"host": "172.16.1.19"
|
||||
},
|
||||
{
|
||||
"host": "198.18.2.19",
|
||||
"external": true,
|
||||
"ports": [
|
||||
"mesh"
|
||||
]
|
||||
}
|
||||
],
|
||||
"ports": {
|
||||
"http": {
|
||||
"port": 8080,
|
||||
"protocol": "PROTOCOL_HTTP"
|
||||
},
|
||||
"grpc": {
|
||||
"port": 9090,
|
||||
"protocol": "PROTOCOL_GRPC"
|
||||
},
|
||||
"mesh": {
|
||||
"port": 10000,
|
||||
"protocol": "PROTOCOL_MESH"
|
||||
}
|
||||
},
|
||||
"identity": "api"
|
||||
}
|
||||
}
|
|
@ -0,0 +1,33 @@
|
|||
{
|
||||
"id": {
|
||||
"type": {
|
||||
"group": "catalog",
|
||||
"group_version": "v1alpha1",
|
||||
"kind": "HealthStatus"
|
||||
},
|
||||
"tenancy": {
|
||||
"partition": "default",
|
||||
"namespace": "default",
|
||||
"peer_name": "local"
|
||||
},
|
||||
"name": "api-2-health"
|
||||
},
|
||||
"owner": {
|
||||
"type": {
|
||||
"group": "catalog",
|
||||
"group_version": "v1alpha1",
|
||||
"kind": "Workload"
|
||||
},
|
||||
"tenancy": {
|
||||
"partition": "default",
|
||||
"namespace": "default",
|
||||
"peer_name": "local"
|
||||
},
|
||||
"name": "api-2"
|
||||
},
|
||||
"data": {
|
||||
"@type": "hashicorp.consul.catalog.v1alpha1.HealthStatus",
|
||||
"type": "synthetic",
|
||||
"status": "HEALTH_WARNING"
|
||||
}
|
||||
}
|
|
@ -0,0 +1,46 @@
|
|||
{
|
||||
"id": {
|
||||
"type": {
|
||||
"group": "catalog",
|
||||
"group_version": "v1alpha1",
|
||||
"kind": "Workload"
|
||||
},
|
||||
"tenancy": {
|
||||
"partition": "default",
|
||||
"namespace": "default",
|
||||
"peer_name": "local"
|
||||
},
|
||||
"name": "api-2"
|
||||
},
|
||||
"data": {
|
||||
"@type": "hashicorp.consul.catalog.v1alpha1.Workload",
|
||||
"addresses": [
|
||||
{
|
||||
"host": "172.16.1.2"
|
||||
},
|
||||
{
|
||||
"host": "198.18.2.2",
|
||||
"external": true,
|
||||
"ports": [
|
||||
"mesh"
|
||||
]
|
||||
}
|
||||
],
|
||||
"ports": {
|
||||
"http": {
|
||||
"port": 8080,
|
||||
"protocol": "PROTOCOL_HTTP"
|
||||
},
|
||||
"grpc": {
|
||||
"port": 9090,
|
||||
"protocol": "PROTOCOL_GRPC"
|
||||
},
|
||||
"mesh": {
|
||||
"port": 10000,
|
||||
"protocol": "PROTOCOL_MESH"
|
||||
}
|
||||
},
|
||||
"identity": "api",
|
||||
"node_name": "node-1"
|
||||
}
|
||||
}
|
|
@ -0,0 +1,33 @@
|
|||
{
|
||||
"id": {
|
||||
"type": {
|
||||
"group": "catalog",
|
||||
"group_version": "v1alpha1",
|
||||
"kind": "HealthStatus"
|
||||
},
|
||||
"tenancy": {
|
||||
"partition": "default",
|
||||
"namespace": "default",
|
||||
"peer_name": "local"
|
||||
},
|
||||
"name": "api-20-health"
|
||||
},
|
||||
"owner": {
|
||||
"type": {
|
||||
"group": "catalog",
|
||||
"group_version": "v1alpha1",
|
||||
"kind": "Workload"
|
||||
},
|
||||
"tenancy": {
|
||||
"partition": "default",
|
||||
"namespace": "default",
|
||||
"peer_name": "local"
|
||||
},
|
||||
"name": "api-20"
|
||||
},
|
||||
"data": {
|
||||
"@type": "hashicorp.consul.catalog.v1alpha1.HealthStatus",
|
||||
"type": "synthetic",
|
||||
"status": "HEALTH_MAINTENANCE"
|
||||
}
|
||||
}
|
|
@ -0,0 +1,45 @@
|
|||
{
|
||||
"id": {
|
||||
"type": {
|
||||
"group": "catalog",
|
||||
"group_version": "v1alpha1",
|
||||
"kind": "Workload"
|
||||
},
|
||||
"tenancy": {
|
||||
"partition": "default",
|
||||
"namespace": "default",
|
||||
"peer_name": "local"
|
||||
},
|
||||
"name": "api-20"
|
||||
},
|
||||
"data": {
|
||||
"@type": "hashicorp.consul.catalog.v1alpha1.Workload",
|
||||
"addresses": [
|
||||
{
|
||||
"host": "172.16.1.20"
|
||||
},
|
||||
{
|
||||
"host": "198.18.2.20",
|
||||
"external": true,
|
||||
"ports": [
|
||||
"mesh"
|
||||
]
|
||||
}
|
||||
],
|
||||
"ports": {
|
||||
"http": {
|
||||
"port": 8080,
|
||||
"protocol": "PROTOCOL_HTTP"
|
||||
},
|
||||
"grpc": {
|
||||
"port": 9090,
|
||||
"protocol": "PROTOCOL_GRPC"
|
||||
},
|
||||
"mesh": {
|
||||
"port": 10000,
|
||||
"protocol": "PROTOCOL_MESH"
|
||||
}
|
||||
},
|
||||
"identity": "api"
|
||||
}
|
||||
}
|
|
@ -0,0 +1,33 @@
|
|||
{
|
||||
"id": {
|
||||
"type": {
|
||||
"group": "catalog",
|
||||
"group_version": "v1alpha1",
|
||||
"kind": "HealthStatus"
|
||||
},
|
||||
"tenancy": {
|
||||
"partition": "default",
|
||||
"namespace": "default",
|
||||
"peer_name": "local"
|
||||
},
|
||||
"name": "api-3-health"
|
||||
},
|
||||
"owner": {
|
||||
"type": {
|
||||
"group": "catalog",
|
||||
"group_version": "v1alpha1",
|
||||
"kind": "Workload"
|
||||
},
|
||||
"tenancy": {
|
||||
"partition": "default",
|
||||
"namespace": "default",
|
||||
"peer_name": "local"
|
||||
},
|
||||
"name": "api-3"
|
||||
},
|
||||
"data": {
|
||||
"@type": "hashicorp.consul.catalog.v1alpha1.HealthStatus",
|
||||
"type": "synthetic",
|
||||
"status": "HEALTH_CRITICAL"
|
||||
}
|
||||
}
|
|
@ -0,0 +1,46 @@
|
|||
{
|
||||
"id": {
|
||||
"type": {
|
||||
"group": "catalog",
|
||||
"group_version": "v1alpha1",
|
||||
"kind": "Workload"
|
||||
},
|
||||
"tenancy": {
|
||||
"partition": "default",
|
||||
"namespace": "default",
|
||||
"peer_name": "local"
|
||||
},
|
||||
"name": "api-3"
|
||||
},
|
||||
"data": {
|
||||
"@type": "hashicorp.consul.catalog.v1alpha1.Workload",
|
||||
"addresses": [
|
||||
{
|
||||
"host": "172.16.1.3"
|
||||
},
|
||||
{
|
||||
"host": "198.18.2.3",
|
||||
"external": true,
|
||||
"ports": [
|
||||
"mesh"
|
||||
]
|
||||
}
|
||||
],
|
||||
"ports": {
|
||||
"http": {
|
||||
"port": 8080,
|
||||
"protocol": "PROTOCOL_HTTP"
|
||||
},
|
||||
"grpc": {
|
||||
"port": 9090,
|
||||
"protocol": "PROTOCOL_GRPC"
|
||||
},
|
||||
"mesh": {
|
||||
"port": 10000,
|
||||
"protocol": "PROTOCOL_MESH"
|
||||
}
|
||||
},
|
||||
"identity": "api",
|
||||
"node_name": "node-1"
|
||||
}
|
||||
}
|
|
@ -0,0 +1,33 @@
|
|||
{
|
||||
"id": {
|
||||
"type": {
|
||||
"group": "catalog",
|
||||
"group_version": "v1alpha1",
|
||||
"kind": "HealthStatus"
|
||||
},
|
||||
"tenancy": {
|
||||
"partition": "default",
|
||||
"namespace": "default",
|
||||
"peer_name": "local"
|
||||
},
|
||||
"name": "api-4-health"
|
||||
},
|
||||
"owner": {
|
||||
"type": {
|
||||
"group": "catalog",
|
||||
"group_version": "v1alpha1",
|
||||
"kind": "Workload"
|
||||
},
|
||||
"tenancy": {
|
||||
"partition": "default",
|
||||
"namespace": "default",
|
||||
"peer_name": "local"
|
||||
},
|
||||
"name": "api-4"
|
||||
},
|
||||
"data": {
|
||||
"@type": "hashicorp.consul.catalog.v1alpha1.HealthStatus",
|
||||
"type": "synthetic",
|
||||
"status": "HEALTH_MAINTENANCE"
|
||||
}
|
||||
}
|
|
@ -0,0 +1,46 @@
|
|||
{
|
||||
"id": {
|
||||
"type": {
|
||||
"group": "catalog",
|
||||
"group_version": "v1alpha1",
|
||||
"kind": "Workload"
|
||||
},
|
||||
"tenancy": {
|
||||
"partition": "default",
|
||||
"namespace": "default",
|
||||
"peer_name": "local"
|
||||
},
|
||||
"name": "api-4"
|
||||
},
|
||||
"data": {
|
||||
"@type": "hashicorp.consul.catalog.v1alpha1.Workload",
|
||||
"addresses": [
|
||||
{
|
||||
"host": "172.16.1.4"
|
||||
},
|
||||
{
|
||||
"host": "198.18.2.4",
|
||||
"external": true,
|
||||
"ports": [
|
||||
"mesh"
|
||||
]
|
||||
}
|
||||
],
|
||||
"ports": {
|
||||
"http": {
|
||||
"port": 8080,
|
||||
"protocol": "PROTOCOL_HTTP"
|
||||
},
|
||||
"grpc": {
|
||||
"port": 9090,
|
||||
"protocol": "PROTOCOL_GRPC"
|
||||
},
|
||||
"mesh": {
|
||||
"port": 10000,
|
||||
"protocol": "PROTOCOL_MESH"
|
||||
}
|
||||
},
|
||||
"identity": "api",
|
||||
"node_name": "node-1"
|
||||
}
|
||||
}
|
|
@ -0,0 +1,33 @@
|
|||
{
|
||||
"id": {
|
||||
"type": {
|
||||
"group": "catalog",
|
||||
"group_version": "v1alpha1",
|
||||
"kind": "HealthStatus"
|
||||
},
|
||||
"tenancy": {
|
||||
"partition": "default",
|
||||
"namespace": "default",
|
||||
"peer_name": "local"
|
||||
},
|
||||
"name": "api-5-health"
|
||||
},
|
||||
"owner": {
|
||||
"type": {
|
||||
"group": "catalog",
|
||||
"group_version": "v1alpha1",
|
||||
"kind": "Workload"
|
||||
},
|
||||
"tenancy": {
|
||||
"partition": "default",
|
||||
"namespace": "default",
|
||||
"peer_name": "local"
|
||||
},
|
||||
"name": "api-5"
|
||||
},
|
||||
"data": {
|
||||
"@type": "hashicorp.consul.catalog.v1alpha1.HealthStatus",
|
||||
"type": "synthetic",
|
||||
"status": "HEALTH_PASSING"
|
||||
}
|
||||
}
|
|
@ -0,0 +1,46 @@
|
|||
{
|
||||
"id": {
|
||||
"type": {
|
||||
"group": "catalog",
|
||||
"group_version": "v1alpha1",
|
||||
"kind": "Workload"
|
||||
},
|
||||
"tenancy": {
|
||||
"partition": "default",
|
||||
"namespace": "default",
|
||||
"peer_name": "local"
|
||||
},
|
||||
"name": "api-5"
|
||||
},
|
||||
"data": {
|
||||
"@type": "hashicorp.consul.catalog.v1alpha1.Workload",
|
||||
"addresses": [
|
||||
{
|
||||
"host": "172.16.1.5"
|
||||
},
|
||||
{
|
||||
"host": "198.18.2.5",
|
||||
"external": true,
|
||||
"ports": [
|
||||
"mesh"
|
||||
]
|
||||
}
|
||||
],
|
||||
"ports": {
|
||||
"http": {
|
||||
"port": 8080,
|
||||
"protocol": "PROTOCOL_HTTP"
|
||||
},
|
||||
"grpc": {
|
||||
"port": 9090,
|
||||
"protocol": "PROTOCOL_GRPC"
|
||||
},
|
||||
"mesh": {
|
||||
"port": 10000,
|
||||
"protocol": "PROTOCOL_MESH"
|
||||
}
|
||||
},
|
||||
"identity": "api",
|
||||
"node_name": "node-2"
|
||||
}
|
||||
}
|
|
@ -0,0 +1,33 @@
|
|||
{
|
||||
"id": {
|
||||
"type": {
|
||||
"group": "catalog",
|
||||
"group_version": "v1alpha1",
|
||||
"kind": "HealthStatus"
|
||||
},
|
||||
"tenancy": {
|
||||
"partition": "default",
|
||||
"namespace": "default",
|
||||
"peer_name": "local"
|
||||
},
|
||||
"name": "api-6-health"
|
||||
},
|
||||
"owner": {
|
||||
"type": {
|
||||
"group": "catalog",
|
||||
"group_version": "v1alpha1",
|
||||
"kind": "Workload"
|
||||
},
|
||||
"tenancy": {
|
||||
"partition": "default",
|
||||
"namespace": "default",
|
||||
"peer_name": "local"
|
||||
},
|
||||
"name": "api-6"
|
||||
},
|
||||
"data": {
|
||||
"@type": "hashicorp.consul.catalog.v1alpha1.HealthStatus",
|
||||
"type": "synthetic",
|
||||
"status": "HEALTH_WARNING"
|
||||
}
|
||||
}
|
|
@ -0,0 +1,46 @@
|
|||
{
|
||||
"id": {
|
||||
"type": {
|
||||
"group": "catalog",
|
||||
"group_version": "v1alpha1",
|
||||
"kind": "Workload"
|
||||
},
|
||||
"tenancy": {
|
||||
"partition": "default",
|
||||
"namespace": "default",
|
||||
"peer_name": "local"
|
||||
},
|
||||
"name": "api-6"
|
||||
},
|
||||
"data": {
|
||||
"@type": "hashicorp.consul.catalog.v1alpha1.Workload",
|
||||
"addresses": [
|
||||
{
|
||||
"host": "172.16.1.6"
|
||||
},
|
||||
{
|
||||
"host": "198.18.2.6",
|
||||
"external": true,
|
||||
"ports": [
|
||||
"mesh"
|
||||
]
|
||||
}
|
||||
],
|
||||
"ports": {
|
||||
"http": {
|
||||
"port": 8080,
|
||||
"protocol": "PROTOCOL_HTTP"
|
||||
},
|
||||
"grpc": {
|
||||
"port": 9090,
|
||||
"protocol": "PROTOCOL_GRPC"
|
||||
},
|
||||
"mesh": {
|
||||
"port": 10000,
|
||||
"protocol": "PROTOCOL_MESH"
|
||||
}
|
||||
},
|
||||
"identity": "api",
|
||||
"node_name": "node-2"
|
||||
}
|
||||
}
|
|
@ -0,0 +1,33 @@
|
|||
{
|
||||
"id": {
|
||||
"type": {
|
||||
"group": "catalog",
|
||||
"group_version": "v1alpha1",
|
||||
"kind": "HealthStatus"
|
||||
},
|
||||
"tenancy": {
|
||||
"partition": "default",
|
||||
"namespace": "default",
|
||||
"peer_name": "local"
|
||||
},
|
||||
"name": "api-7-health"
|
||||
},
|
||||
"owner": {
|
||||
"type": {
|
||||
"group": "catalog",
|
||||
"group_version": "v1alpha1",
|
||||
"kind": "Workload"
|
||||
},
|
||||
"tenancy": {
|
||||
"partition": "default",
|
||||
"namespace": "default",
|
||||
"peer_name": "local"
|
||||
},
|
||||
"name": "api-7"
|
||||
},
|
||||
"data": {
|
||||
"@type": "hashicorp.consul.catalog.v1alpha1.HealthStatus",
|
||||
"type": "synthetic",
|
||||
"status": "HEALTH_CRITICAL"
|
||||
}
|
||||
}
|
|
@ -0,0 +1,46 @@
|
|||
{
|
||||
"id": {
|
||||
"type": {
|
||||
"group": "catalog",
|
||||
"group_version": "v1alpha1",
|
||||
"kind": "Workload"
|
||||
},
|
||||
"tenancy": {
|
||||
"partition": "default",
|
||||
"namespace": "default",
|
||||
"peer_name": "local"
|
||||
},
|
||||
"name": "api-7"
|
||||
},
|
||||
"data": {
|
||||
"@type": "hashicorp.consul.catalog.v1alpha1.Workload",
|
||||
"addresses": [
|
||||
{
|
||||
"host": "172.16.1.7"
|
||||
},
|
||||
{
|
||||
"host": "198.18.2.7",
|
||||
"external": true,
|
||||
"ports": [
|
||||
"mesh"
|
||||
]
|
||||
}
|
||||
],
|
||||
"ports": {
|
||||
"http": {
|
||||
"port": 8080,
|
||||
"protocol": "PROTOCOL_HTTP"
|
||||
},
|
||||
"grpc": {
|
||||
"port": 9090,
|
||||
"protocol": "PROTOCOL_GRPC"
|
||||
},
|
||||
"mesh": {
|
||||
"port": 10000,
|
||||
"protocol": "PROTOCOL_MESH"
|
||||
}
|
||||
},
|
||||
"identity": "api",
|
||||
"node_name": "node-2"
|
||||
}
|
||||
}
|
|
@ -0,0 +1,33 @@
|
|||
{
|
||||
"id": {
|
||||
"type": {
|
||||
"group": "catalog",
|
||||
"group_version": "v1alpha1",
|
||||
"kind": "HealthStatus"
|
||||
},
|
||||
"tenancy": {
|
||||
"partition": "default",
|
||||
"namespace": "default",
|
||||
"peer_name": "local"
|
||||
},
|
||||
"name": "api-8-health"
|
||||
},
|
||||
"owner": {
|
||||
"type": {
|
||||
"group": "catalog",
|
||||
"group_version": "v1alpha1",
|
||||
"kind": "Workload"
|
||||
},
|
||||
"tenancy": {
|
||||
"partition": "default",
|
||||
"namespace": "default",
|
||||
"peer_name": "local"
|
||||
},
|
||||
"name": "api-8"
|
||||
},
|
||||
"data": {
|
||||
"@type": "hashicorp.consul.catalog.v1alpha1.HealthStatus",
|
||||
"type": "synthetic",
|
||||
"status": "HEALTH_MAINTENANCE"
|
||||
}
|
||||
}
|
|
@ -0,0 +1,46 @@
|
|||
{
|
||||
"id": {
|
||||
"type": {
|
||||
"group": "catalog",
|
||||
"group_version": "v1alpha1",
|
||||
"kind": "Workload"
|
||||
},
|
||||
"tenancy": {
|
||||
"partition": "default",
|
||||
"namespace": "default",
|
||||
"peer_name": "local"
|
||||
},
|
||||
"name": "api-8"
|
||||
},
|
||||
"data": {
|
||||
"@type": "hashicorp.consul.catalog.v1alpha1.Workload",
|
||||
"addresses": [
|
||||
{
|
||||
"host": "172.16.1.8"
|
||||
},
|
||||
{
|
||||
"host": "198.18.2.8",
|
||||
"external": true,
|
||||
"ports": [
|
||||
"mesh"
|
||||
]
|
||||
}
|
||||
],
|
||||
"ports": {
|
||||
"http": {
|
||||
"port": 8080,
|
||||
"protocol": "PROTOCOL_HTTP"
|
||||
},
|
||||
"grpc": {
|
||||
"port": 9090,
|
||||
"protocol": "PROTOCOL_GRPC"
|
||||
},
|
||||
"mesh": {
|
||||
"port": 10000,
|
||||
"protocol": "PROTOCOL_MESH"
|
||||
}
|
||||
},
|
||||
"identity": "api",
|
||||
"node_name": "node-2"
|
||||
}
|
||||
}
|
|
@ -0,0 +1,33 @@
|
|||
{
|
||||
"id": {
|
||||
"type": {
|
||||
"group": "catalog",
|
||||
"group_version": "v1alpha1",
|
||||
"kind": "HealthStatus"
|
||||
},
|
||||
"tenancy": {
|
||||
"partition": "default",
|
||||
"namespace": "default",
|
||||
"peer_name": "local"
|
||||
},
|
||||
"name": "api-9-health"
|
||||
},
|
||||
"owner": {
|
||||
"type": {
|
||||
"group": "catalog",
|
||||
"group_version": "v1alpha1",
|
||||
"kind": "Workload"
|
||||
},
|
||||
"tenancy": {
|
||||
"partition": "default",
|
||||
"namespace": "default",
|
||||
"peer_name": "local"
|
||||
},
|
||||
"name": "api-9"
|
||||
},
|
||||
"data": {
|
||||
"@type": "hashicorp.consul.catalog.v1alpha1.HealthStatus",
|
||||
"type": "synthetic",
|
||||
"status": "HEALTH_PASSING"
|
||||
}
|
||||
}
|
|
@ -0,0 +1,46 @@
|
|||
{
|
||||
"id": {
|
||||
"type": {
|
||||
"group": "catalog",
|
||||
"group_version": "v1alpha1",
|
||||
"kind": "Workload"
|
||||
},
|
||||
"tenancy": {
|
||||
"partition": "default",
|
||||
"namespace": "default",
|
||||
"peer_name": "local"
|
||||
},
|
||||
"name": "api-9"
|
||||
},
|
||||
"data": {
|
||||
"@type": "hashicorp.consul.catalog.v1alpha1.Workload",
|
||||
"addresses": [
|
||||
{
|
||||
"host": "172.16.1.9"
|
||||
},
|
||||
{
|
||||
"host": "198.18.2.9",
|
||||
"external": true,
|
||||
"ports": [
|
||||
"mesh"
|
||||
]
|
||||
}
|
||||
],
|
||||
"ports": {
|
||||
"http": {
|
||||
"port": 8080,
|
||||
"protocol": "PROTOCOL_HTTP"
|
||||
},
|
||||
"grpc": {
|
||||
"port": 9090,
|
||||
"protocol": "PROTOCOL_GRPC"
|
||||
},
|
||||
"mesh": {
|
||||
"port": 10000,
|
||||
"protocol": "PROTOCOL_MESH"
|
||||
}
|
||||
},
|
||||
"identity": "api",
|
||||
"node_name": "node-3"
|
||||
}
|
||||
}
|
|
@ -0,0 +1,47 @@
|
|||
{
|
||||
"id": {
|
||||
"type": {
|
||||
"group": "catalog",
|
||||
"group_version": "v1alpha1",
|
||||
"kind": "ServiceEndpoints"
|
||||
},
|
||||
"tenancy": {
|
||||
"partition": "default",
|
||||
"namespace": "default",
|
||||
"peer_name": "local"
|
||||
},
|
||||
"name": "foo"
|
||||
},
|
||||
"owner": {
|
||||
"type": {
|
||||
"group": "catalog",
|
||||
"group_version": "v1alpha1",
|
||||
"kind": "Service"
|
||||
},
|
||||
"tenancy": {
|
||||
"partition": "default",
|
||||
"namespace": "default",
|
||||
"peer_name": "local"
|
||||
},
|
||||
"name": "foo"
|
||||
},
|
||||
"data": {
|
||||
"@type": "hashicorp.consul.catalog.v1alpha1.ServiceEndpoints",
|
||||
"endpoints": [
|
||||
{
|
||||
"addresses": [
|
||||
{
|
||||
"host": "198.18.0.1"
|
||||
}
|
||||
],
|
||||
"ports": {
|
||||
"external-service-port": {
|
||||
"port": 9876,
|
||||
"protocol": "PROTOCOL_HTTP2"
|
||||
}
|
||||
},
|
||||
"health_status": "HEALTH_PASSING"
|
||||
}
|
||||
]
|
||||
}
|
||||
}
|
|
@ -0,0 +1,24 @@
|
|||
{
|
||||
"id": {
|
||||
"type": {
|
||||
"group": "catalog",
|
||||
"group_version": "v1alpha1",
|
||||
"kind": "Service"
|
||||
},
|
||||
"tenancy": {
|
||||
"partition": "default",
|
||||
"namespace": "default",
|
||||
"peer_name": "local"
|
||||
},
|
||||
"name": "foo"
|
||||
},
|
||||
"data": {
|
||||
"@type": "hashicorp.consul.catalog.v1alpha1.Service",
|
||||
"ports": [
|
||||
{
|
||||
"target_port": "external-service-port",
|
||||
"protocol": "PROTOCOL_HTTP2"
|
||||
}
|
||||
]
|
||||
}
|
||||
}
|
|
@ -0,0 +1,42 @@
|
|||
{
|
||||
"id": {
|
||||
"type": {
|
||||
"group": "catalog",
|
||||
"group_version": "v1alpha1",
|
||||
"kind": "Service"
|
||||
},
|
||||
"tenancy": {
|
||||
"partition": "default",
|
||||
"namespace": "default",
|
||||
"peer_name": "local"
|
||||
},
|
||||
"name": "grpc-api"
|
||||
},
|
||||
"data": {
|
||||
"@type": "hashicorp.consul.catalog.v1alpha1.Service",
|
||||
"workloads": {
|
||||
"names": [
|
||||
"api-1",
|
||||
"api-2",
|
||||
"api-3",
|
||||
"api-4",
|
||||
"api-5",
|
||||
"api-6",
|
||||
"api-7",
|
||||
"api-8",
|
||||
"api-9",
|
||||
"api-20"
|
||||
]
|
||||
},
|
||||
"ports": [
|
||||
{
|
||||
"target_port": "grpc",
|
||||
"protocol": "PROTOCOL_GRPC"
|
||||
},
|
||||
{
|
||||
"target_port": "mesh",
|
||||
"protocol": "PROTOCOL_MESH"
|
||||
}
|
||||
]
|
||||
}
|
||||
}
|
|
@ -0,0 +1,29 @@
|
|||
{
|
||||
"id": {
|
||||
"type": {
|
||||
"group": "catalog",
|
||||
"group_version": "v1alpha1",
|
||||
"kind": "Service"
|
||||
},
|
||||
"tenancy": {
|
||||
"partition": "default",
|
||||
"namespace": "default",
|
||||
"peer_name": "local"
|
||||
},
|
||||
"name": "http-api"
|
||||
},
|
||||
"data": {
|
||||
"@type": "hashicorp.consul.catalog.v1alpha1.Service",
|
||||
"workloads": {
|
||||
"prefixes": [
|
||||
"api-1"
|
||||
]
|
||||
},
|
||||
"ports": [
|
||||
{
|
||||
"target_port": "http",
|
||||
"protocol": "PROTOCOL_HTTP"
|
||||
}
|
||||
]
|
||||
}
|
||||
}
|
|
@ -0,0 +1,33 @@
|
|||
{
|
||||
"id": {
|
||||
"type": {
|
||||
"group": "catalog",
|
||||
"group_version": "v1alpha1",
|
||||
"kind": "HealthStatus"
|
||||
},
|
||||
"tenancy": {
|
||||
"partition": "default",
|
||||
"namespace": "default",
|
||||
"peer_name": "local"
|
||||
},
|
||||
"name": "node-1-health"
|
||||
},
|
||||
"owner": {
|
||||
"type": {
|
||||
"group": "catalog",
|
||||
"group_version": "v1alpha1",
|
||||
"kind": "Node"
|
||||
},
|
||||
"tenancy": {
|
||||
"partition": "default",
|
||||
"namespace": "default",
|
||||
"peer_name": "local"
|
||||
},
|
||||
"name": "node-1"
|
||||
},
|
||||
"data": {
|
||||
"@type": "hashicorp.consul.catalog.v1alpha1.HealthStatus",
|
||||
"type": "synthetic",
|
||||
"status": "HEALTH_PASSING"
|
||||
}
|
||||
}
|
|
@ -0,0 +1,27 @@
|
|||
{
|
||||
"id": {
|
||||
"type": {
|
||||
"group": "catalog",
|
||||
"group_version": "v1alpha1",
|
||||
"kind": "Node"
|
||||
},
|
||||
"tenancy": {
|
||||
"partition": "default",
|
||||
"namespace": "default",
|
||||
"peer_name": "local"
|
||||
},
|
||||
"name": "node-1"
|
||||
},
|
||||
"data": {
|
||||
"@type": "hashicorp.consul.catalog.v1alpha1.Node",
|
||||
"addresses": [
|
||||
{
|
||||
"host": "198.18.1.1",
|
||||
"external": true
|
||||
},
|
||||
{
|
||||
"host": "172.16.0.1"
|
||||
}
|
||||
]
|
||||
}
|
||||
}
|
|
@ -0,0 +1,33 @@
|
|||
{
|
||||
"id": {
|
||||
"type": {
|
||||
"group": "catalog",
|
||||
"group_version": "v1alpha1",
|
||||
"kind": "HealthStatus"
|
||||
},
|
||||
"tenancy": {
|
||||
"partition": "default",
|
||||
"namespace": "default",
|
||||
"peer_name": "local"
|
||||
},
|
||||
"name": "node-2-health"
|
||||
},
|
||||
"owner": {
|
||||
"type": {
|
||||
"group": "catalog",
|
||||
"group_version": "v1alpha1",
|
||||
"kind": "Node"
|
||||
},
|
||||
"tenancy": {
|
||||
"partition": "default",
|
||||
"namespace": "default",
|
||||
"peer_name": "local"
|
||||
},
|
||||
"name": "node-2"
|
||||
},
|
||||
"data": {
|
||||
"@type": "hashicorp.consul.catalog.v1alpha1.HealthStatus",
|
||||
"type": "synthetic",
|
||||
"status": "HEALTH_WARNING"
|
||||
}
|
||||
}
|
|
@ -0,0 +1,27 @@
|
|||
{
|
||||
"id": {
|
||||
"type": {
|
||||
"group": "catalog",
|
||||
"group_version": "v1alpha1",
|
||||
"kind": "Node"
|
||||
},
|
||||
"tenancy": {
|
||||
"partition": "default",
|
||||
"namespace": "default",
|
||||
"peer_name": "local"
|
||||
},
|
||||
"name": "node-2"
|
||||
},
|
||||
"data": {
|
||||
"@type": "hashicorp.consul.catalog.v1alpha1.Node",
|
||||
"addresses": [
|
||||
{
|
||||
"host": "198.18.1.2",
|
||||
"external": true
|
||||
},
|
||||
{
|
||||
"host": "172.16.0.2"
|
||||
}
|
||||
]
|
||||
}
|
||||
}
|
|
@ -0,0 +1,33 @@
|
|||
{
|
||||
"id": {
|
||||
"type": {
|
||||
"group": "catalog",
|
||||
"group_version": "v1alpha1",
|
||||
"kind": "HealthStatus"
|
||||
},
|
||||
"tenancy": {
|
||||
"partition": "default",
|
||||
"namespace": "default",
|
||||
"peer_name": "local"
|
||||
},
|
||||
"name": "node-3-health"
|
||||
},
|
||||
"owner": {
|
||||
"type": {
|
||||
"group": "catalog",
|
||||
"group_version": "v1alpha1",
|
||||
"kind": "Node"
|
||||
},
|
||||
"tenancy": {
|
||||
"partition": "default",
|
||||
"namespace": "default",
|
||||
"peer_name": "local"
|
||||
},
|
||||
"name": "node-3"
|
||||
},
|
||||
"data": {
|
||||
"@type": "hashicorp.consul.catalog.v1alpha1.HealthStatus",
|
||||
"type": "synthetic",
|
||||
"status": "HEALTH_CRITICAL"
|
||||
}
|
||||
}
|
|
@ -0,0 +1,27 @@
|
|||
{
|
||||
"id": {
|
||||
"type": {
|
||||
"group": "catalog",
|
||||
"group_version": "v1alpha1",
|
||||
"kind": "Node"
|
||||
},
|
||||
"tenancy": {
|
||||
"partition": "default",
|
||||
"namespace": "default",
|
||||
"peer_name": "local"
|
||||
},
|
||||
"name": "node-3"
|
||||
},
|
||||
"data": {
|
||||
"@type": "hashicorp.consul.catalog.v1alpha1.Node",
|
||||
"addresses": [
|
||||
{
|
||||
"host": "198.18.1.3",
|
||||
"external": true
|
||||
},
|
||||
{
|
||||
"host": "172.16.0.3"
|
||||
}
|
||||
]
|
||||
}
|
||||
}
|
|
@ -0,0 +1,33 @@
|
|||
{
|
||||
"id": {
|
||||
"type": {
|
||||
"group": "catalog",
|
||||
"group_version": "v1alpha1",
|
||||
"kind": "HealthStatus"
|
||||
},
|
||||
"tenancy": {
|
||||
"partition": "default",
|
||||
"namespace": "default",
|
||||
"peer_name": "local"
|
||||
},
|
||||
"name": "node-4-health"
|
||||
},
|
||||
"owner": {
|
||||
"type": {
|
||||
"group": "catalog",
|
||||
"group_version": "v1alpha1",
|
||||
"kind": "Node"
|
||||
},
|
||||
"tenancy": {
|
||||
"partition": "default",
|
||||
"namespace": "default",
|
||||
"peer_name": "local"
|
||||
},
|
||||
"name": "node-4"
|
||||
},
|
||||
"data": {
|
||||
"@type": "hashicorp.consul.catalog.v1alpha1.HealthStatus",
|
||||
"type": "synthetic",
|
||||
"status": "HEALTH_MAINTENANCE"
|
||||
}
|
||||
}
|
|
@ -0,0 +1,27 @@
|
|||
{
|
||||
"id": {
|
||||
"type": {
|
||||
"group": "catalog",
|
||||
"group_version": "v1alpha1",
|
||||
"kind": "Node"
|
||||
},
|
||||
"tenancy": {
|
||||
"partition": "default",
|
||||
"namespace": "default",
|
||||
"peer_name": "local"
|
||||
},
|
||||
"name": "node-4"
|
||||
},
|
||||
"data": {
|
||||
"@type": "hashicorp.consul.catalog.v1alpha1.Node",
|
||||
"addresses": [
|
||||
{
|
||||
"host": "198.18.1.4",
|
||||
"external": true
|
||||
},
|
||||
{
|
||||
"host": "172.16.0.4"
|
||||
}
|
||||
]
|
||||
}
|
||||
}
|
39
internal/catalog/catalogtest/run_test.go
Normal file
39
internal/catalog/catalogtest/run_test.go
Normal file
|
@ -0,0 +1,39 @@
|
|||
package catalogtest
|
||||
|
||||
import (
|
||||
"testing"
|
||||
|
||||
svctest "github.com/hashicorp/consul/agent/grpc-external/services/resource/testing"
|
||||
"github.com/hashicorp/consul/internal/catalog"
|
||||
"github.com/hashicorp/consul/internal/catalog/internal/controllers"
|
||||
"github.com/hashicorp/consul/internal/controller"
|
||||
"github.com/hashicorp/consul/internal/resource/reaper"
|
||||
"github.com/hashicorp/consul/proto-public/pbresource"
|
||||
"github.com/hashicorp/consul/sdk/testutil"
|
||||
)
|
||||
|
||||
func runInMemResourceServiceAndControllers(t *testing.T, deps controllers.Dependencies) pbresource.ResourceServiceClient {
|
||||
t.Helper()
|
||||
|
||||
ctx := testutil.TestContext(t)
|
||||
|
||||
// Create the in-mem resource service
|
||||
client := svctest.RunResourceService(t, catalog.RegisterTypes)
|
||||
|
||||
// Setup/Run the controller manager
|
||||
mgr := controller.NewManager(client, testutil.Logger(t))
|
||||
catalog.RegisterControllers(mgr, deps)
|
||||
|
||||
// We also depend on the reaper to take care of cleaning up owned health statuses and
|
||||
// service endpoints so we must enable that controller as well
|
||||
reaper.RegisterControllers(mgr)
|
||||
mgr.SetRaftLeader(true)
|
||||
go mgr.Run(ctx)
|
||||
|
||||
return client
|
||||
}
|
||||
|
||||
func TestControllers_Integration(t *testing.T) {
|
||||
client := runInMemResourceServiceAndControllers(t, catalog.DefaultControllerDependencies())
|
||||
RunCatalogV1Alpha1IntegrationTest(t, client)
|
||||
}
|
707
internal/catalog/catalogtest/test_integration_v1alpha1.go
Normal file
707
internal/catalog/catalogtest/test_integration_v1alpha1.go
Normal file
|
@ -0,0 +1,707 @@
|
|||
package catalogtest
|
||||
|
||||
import (
|
||||
"embed"
|
||||
"fmt"
|
||||
"testing"
|
||||
|
||||
"github.com/hashicorp/consul/internal/catalog"
|
||||
"github.com/hashicorp/consul/internal/catalog/internal/controllers/endpoints"
|
||||
"github.com/hashicorp/consul/internal/catalog/internal/controllers/nodehealth"
|
||||
"github.com/hashicorp/consul/internal/catalog/internal/controllers/workloadhealth"
|
||||
"github.com/hashicorp/consul/internal/catalog/internal/types"
|
||||
rtest "github.com/hashicorp/consul/internal/resource/resourcetest"
|
||||
pbcatalog "github.com/hashicorp/consul/proto-public/pbcatalog/v1alpha1"
|
||||
"github.com/hashicorp/consul/proto-public/pbresource"
|
||||
"github.com/hashicorp/consul/proto/private/prototest"
|
||||
"github.com/hashicorp/consul/sdk/testutil"
|
||||
"github.com/stretchr/testify/require"
|
||||
)
|
||||
|
||||
var (
|
||||
//go:embed integration_test_data
|
||||
testData embed.FS
|
||||
)
|
||||
|
||||
// RunCatalogV1Alpha1IntegrationTest will push up a bunch of catalog related data and then
|
||||
// verify that all the expected reconciliations happened correctly. This test is
|
||||
// intended to exercise a large swathe of behavior of the overall catalog package.
|
||||
// Besides just controller reconciliation behavior, the intent is also to verify
|
||||
// that integrations with the resource service are also working (i.e. the various
|
||||
// validation, mutation and ACL hooks get invoked and are working properly)
|
||||
//
|
||||
// This test specifically is not doing any sort of lifecycle related tests to ensure
|
||||
// that modification to values results in re-reconciliation as expected. Instead there
|
||||
// is another RunCatalogIntegrationTestLifeCycle function that can be used for those
|
||||
// purposes. The two are distinct so that the data being published and the assertions
|
||||
// made against the system can be reused in upgrade tests.
|
||||
func RunCatalogV1Alpha1IntegrationTest(t *testing.T, client pbresource.ResourceServiceClient) {
|
||||
t.Helper()
|
||||
|
||||
PublishCatalogV1Alpha1IntegrationTestData(t, client)
|
||||
VerifyCatalogV1Alpha1IntegrationTestResults(t, client)
|
||||
}
|
||||
|
||||
// PublishCatalogV1Alpha1IntegrationTestData will perform a whole bunch of resource writes
|
||||
// for Service, ServiceEndpoints, Workload, Node and HealthStatus objects
|
||||
func PublishCatalogV1Alpha1IntegrationTestData(t *testing.T, client pbresource.ResourceServiceClient) {
|
||||
t.Helper()
|
||||
|
||||
c := rtest.NewClient(client)
|
||||
|
||||
resources := rtest.ParseResourcesFromFilesystem(t, testData, "integration_test_data/v1alpha1")
|
||||
c.PublishResources(t, resources)
|
||||
}
|
||||
|
||||
func VerifyCatalogV1Alpha1IntegrationTestResults(t *testing.T, client pbresource.ResourceServiceClient) {
|
||||
t.Helper()
|
||||
|
||||
c := rtest.NewClient(client)
|
||||
|
||||
testutil.RunStep(t, "resources-exist", func(t *testing.T) {
|
||||
c.RequireResourceExists(t, rtest.Resource(catalog.ServiceV1Alpha1Type, "api").ID())
|
||||
c.RequireResourceExists(t, rtest.Resource(catalog.ServiceV1Alpha1Type, "http-api").ID())
|
||||
c.RequireResourceExists(t, rtest.Resource(catalog.ServiceV1Alpha1Type, "grpc-api").ID())
|
||||
c.RequireResourceExists(t, rtest.Resource(catalog.ServiceV1Alpha1Type, "foo").ID())
|
||||
|
||||
for i := 1; i < 5; i++ {
|
||||
nodeId := rtest.Resource(catalog.NodeV1Alpha1Type, fmt.Sprintf("node-%d", i)).ID()
|
||||
c.RequireResourceExists(t, nodeId)
|
||||
|
||||
res := c.RequireResourceExists(t, rtest.Resource(catalog.HealthStatusV1Alpha1Type, fmt.Sprintf("node-%d-health", i)).ID())
|
||||
rtest.RequireOwner(t, res, nodeId, true)
|
||||
}
|
||||
|
||||
for i := 1; i < 21; i++ {
|
||||
workloadId := rtest.Resource(catalog.WorkloadV1Alpha1Type, fmt.Sprintf("api-%d", i)).ID()
|
||||
c.RequireResourceExists(t, workloadId)
|
||||
|
||||
res := c.RequireResourceExists(t, rtest.Resource(catalog.HealthStatusV1Alpha1Type, fmt.Sprintf("api-%d-health", i)).ID())
|
||||
rtest.RequireOwner(t, res, workloadId, true)
|
||||
}
|
||||
})
|
||||
|
||||
testutil.RunStep(t, "node-health-reconciliation", func(t *testing.T) {
|
||||
c.WaitForStatusCondition(t, rtest.Resource(catalog.NodeV1Alpha1Type, "node-1").ID(), nodehealth.StatusKey, nodehealth.ConditionPassing)
|
||||
c.WaitForStatusCondition(t, rtest.Resource(catalog.NodeV1Alpha1Type, "node-2").ID(), nodehealth.StatusKey, nodehealth.ConditionWarning)
|
||||
c.WaitForStatusCondition(t, rtest.Resource(catalog.NodeV1Alpha1Type, "node-3").ID(), nodehealth.StatusKey, nodehealth.ConditionCritical)
|
||||
c.WaitForStatusCondition(t, rtest.Resource(catalog.NodeV1Alpha1Type, "node-4").ID(), nodehealth.StatusKey, nodehealth.ConditionMaintenance)
|
||||
})
|
||||
|
||||
testutil.RunStep(t, "workload-health-reconciliation", func(t *testing.T) {
|
||||
c.WaitForStatusCondition(t, rtest.Resource(catalog.WorkloadV1Alpha1Type, "api-1").ID(), workloadhealth.StatusKey, workloadhealth.ConditionNodeAndWorkloadPassing)
|
||||
c.WaitForStatusCondition(t, rtest.Resource(catalog.WorkloadV1Alpha1Type, "api-2").ID(), workloadhealth.StatusKey, workloadhealth.ConditionWorkloadWarning)
|
||||
c.WaitForStatusCondition(t, rtest.Resource(catalog.WorkloadV1Alpha1Type, "api-3").ID(), workloadhealth.StatusKey, workloadhealth.ConditionWorkloadCritical)
|
||||
c.WaitForStatusCondition(t, rtest.Resource(catalog.WorkloadV1Alpha1Type, "api-4").ID(), workloadhealth.StatusKey, workloadhealth.ConditionWorkloadMaintenance)
|
||||
c.WaitForStatusCondition(t, rtest.Resource(catalog.WorkloadV1Alpha1Type, "api-5").ID(), workloadhealth.StatusKey, workloadhealth.ConditionNodeWarning)
|
||||
c.WaitForStatusCondition(t, rtest.Resource(catalog.WorkloadV1Alpha1Type, "api-6").ID(), workloadhealth.StatusKey, workloadhealth.ConditionNodeAndWorkloadWarning)
|
||||
c.WaitForStatusCondition(t, rtest.Resource(catalog.WorkloadV1Alpha1Type, "api-7").ID(), workloadhealth.StatusKey, workloadhealth.ConditionNodeAndWorkloadCritical)
|
||||
c.WaitForStatusCondition(t, rtest.Resource(catalog.WorkloadV1Alpha1Type, "api-8").ID(), workloadhealth.StatusKey, workloadhealth.ConditionNodeAndWorkloadMaintenance)
|
||||
c.WaitForStatusCondition(t, rtest.Resource(catalog.WorkloadV1Alpha1Type, "api-9").ID(), workloadhealth.StatusKey, workloadhealth.ConditionNodeCritical)
|
||||
c.WaitForStatusCondition(t, rtest.Resource(catalog.WorkloadV1Alpha1Type, "api-10").ID(), workloadhealth.StatusKey, workloadhealth.ConditionNodeAndWorkloadCritical)
|
||||
c.WaitForStatusCondition(t, rtest.Resource(catalog.WorkloadV1Alpha1Type, "api-11").ID(), workloadhealth.StatusKey, workloadhealth.ConditionNodeAndWorkloadCritical)
|
||||
c.WaitForStatusCondition(t, rtest.Resource(catalog.WorkloadV1Alpha1Type, "api-12").ID(), workloadhealth.StatusKey, workloadhealth.ConditionNodeAndWorkloadMaintenance)
|
||||
c.WaitForStatusCondition(t, rtest.Resource(catalog.WorkloadV1Alpha1Type, "api-13").ID(), workloadhealth.StatusKey, workloadhealth.ConditionNodeMaintenance)
|
||||
c.WaitForStatusCondition(t, rtest.Resource(catalog.WorkloadV1Alpha1Type, "api-14").ID(), workloadhealth.StatusKey, workloadhealth.ConditionNodeAndWorkloadMaintenance)
|
||||
c.WaitForStatusCondition(t, rtest.Resource(catalog.WorkloadV1Alpha1Type, "api-15").ID(), workloadhealth.StatusKey, workloadhealth.ConditionNodeAndWorkloadMaintenance)
|
||||
c.WaitForStatusCondition(t, rtest.Resource(catalog.WorkloadV1Alpha1Type, "api-16").ID(), workloadhealth.StatusKey, workloadhealth.ConditionNodeAndWorkloadMaintenance)
|
||||
c.WaitForStatusCondition(t, rtest.Resource(catalog.WorkloadV1Alpha1Type, "api-17").ID(), workloadhealth.StatusKey, workloadhealth.ConditionWorkloadPassing)
|
||||
c.WaitForStatusCondition(t, rtest.Resource(catalog.WorkloadV1Alpha1Type, "api-18").ID(), workloadhealth.StatusKey, workloadhealth.ConditionWorkloadWarning)
|
||||
c.WaitForStatusCondition(t, rtest.Resource(catalog.WorkloadV1Alpha1Type, "api-19").ID(), workloadhealth.StatusKey, workloadhealth.ConditionWorkloadCritical)
|
||||
c.WaitForStatusCondition(t, rtest.Resource(catalog.WorkloadV1Alpha1Type, "api-20").ID(), workloadhealth.StatusKey, workloadhealth.ConditionWorkloadMaintenance)
|
||||
})
|
||||
|
||||
testutil.RunStep(t, "service-reconciliation", func(t *testing.T) {
|
||||
c.WaitForStatusCondition(t, rtest.Resource(catalog.ServiceV1Alpha1Type, "foo").ID(), endpoints.StatusKey, endpoints.ConditionUnmanaged)
|
||||
c.WaitForStatusCondition(t, rtest.Resource(catalog.ServiceV1Alpha1Type, "api").ID(), endpoints.StatusKey, endpoints.ConditionManaged)
|
||||
c.WaitForStatusCondition(t, rtest.Resource(catalog.ServiceV1Alpha1Type, "http-api").ID(), endpoints.StatusKey, endpoints.ConditionManaged)
|
||||
c.WaitForStatusCondition(t, rtest.Resource(catalog.ServiceV1Alpha1Type, "grpc-api").ID(), endpoints.StatusKey, endpoints.ConditionManaged)
|
||||
})
|
||||
|
||||
testutil.RunStep(t, "service-endpoints-generation", func(t *testing.T) {
|
||||
verifyServiceEndpoints(t, c, rtest.Resource(catalog.ServiceEndpointsV1Alpha1Type, "foo").ID(), expectedFooServiceEndpoints())
|
||||
verifyServiceEndpoints(t, c, rtest.Resource(catalog.ServiceEndpointsV1Alpha1Type, "api").ID(), expectedApiServiceEndpoints(t, c))
|
||||
verifyServiceEndpoints(t, c, rtest.Resource(catalog.ServiceEndpointsV1Alpha1Type, "http-api").ID(), expectedHTTPApiServiceEndpoints(t, c))
|
||||
verifyServiceEndpoints(t, c, rtest.Resource(catalog.ServiceEndpointsV1Alpha1Type, "grpc-api").ID(), expectedGRPCApiServiceEndpoints(t, c))
|
||||
})
|
||||
}
|
||||
|
||||
func expectedFooServiceEndpoints() *pbcatalog.ServiceEndpoints {
|
||||
return &pbcatalog.ServiceEndpoints{
|
||||
Endpoints: []*pbcatalog.Endpoint{
|
||||
{
|
||||
Addresses: []*pbcatalog.WorkloadAddress{
|
||||
{Host: "198.18.0.1"},
|
||||
},
|
||||
Ports: map[string]*pbcatalog.WorkloadPort{
|
||||
"external-service-port": {
|
||||
Port: 9876,
|
||||
Protocol: pbcatalog.Protocol_PROTOCOL_HTTP2,
|
||||
},
|
||||
},
|
||||
HealthStatus: pbcatalog.Health_HEALTH_PASSING,
|
||||
},
|
||||
},
|
||||
}
|
||||
}
|
||||
|
||||
func expectedApiServiceEndpoints(t *testing.T, c *rtest.Client) *pbcatalog.ServiceEndpoints {
|
||||
return &pbcatalog.ServiceEndpoints{
|
||||
Endpoints: []*pbcatalog.Endpoint{
|
||||
// api-1
|
||||
{
|
||||
TargetRef: c.ResolveResourceID(t, rtest.Resource(types.WorkloadV1Alpha1Type, "api-1").ID()),
|
||||
Addresses: []*pbcatalog.WorkloadAddress{
|
||||
{Host: "172.16.1.1", Ports: []string{"grpc", "http", "mesh"}},
|
||||
{Host: "198.18.2.1", External: true, Ports: []string{"mesh"}},
|
||||
},
|
||||
Ports: map[string]*pbcatalog.WorkloadPort{
|
||||
"http": {Port: 8080, Protocol: pbcatalog.Protocol_PROTOCOL_HTTP},
|
||||
"grpc": {Port: 9090, Protocol: pbcatalog.Protocol_PROTOCOL_GRPC},
|
||||
"mesh": {Port: 10000, Protocol: pbcatalog.Protocol_PROTOCOL_MESH},
|
||||
},
|
||||
HealthStatus: pbcatalog.Health_HEALTH_PASSING,
|
||||
},
|
||||
// api-2
|
||||
{
|
||||
TargetRef: c.ResolveResourceID(t, rtest.Resource(types.WorkloadV1Alpha1Type, "api-2").ID()),
|
||||
Addresses: []*pbcatalog.WorkloadAddress{
|
||||
{Host: "172.16.1.2", Ports: []string{"grpc", "http", "mesh"}},
|
||||
{Host: "198.18.2.2", External: true, Ports: []string{"mesh"}},
|
||||
},
|
||||
Ports: map[string]*pbcatalog.WorkloadPort{
|
||||
"http": {Port: 8080, Protocol: pbcatalog.Protocol_PROTOCOL_HTTP},
|
||||
"grpc": {Port: 9090, Protocol: pbcatalog.Protocol_PROTOCOL_GRPC},
|
||||
"mesh": {Port: 10000, Protocol: pbcatalog.Protocol_PROTOCOL_MESH},
|
||||
},
|
||||
HealthStatus: pbcatalog.Health_HEALTH_WARNING,
|
||||
},
|
||||
// api-3
|
||||
{
|
||||
TargetRef: c.ResolveResourceID(t, rtest.Resource(types.WorkloadV1Alpha1Type, "api-3").ID()),
|
||||
Addresses: []*pbcatalog.WorkloadAddress{
|
||||
{Host: "172.16.1.3", Ports: []string{"grpc", "http", "mesh"}},
|
||||
{Host: "198.18.2.3", External: true, Ports: []string{"mesh"}},
|
||||
},
|
||||
Ports: map[string]*pbcatalog.WorkloadPort{
|
||||
"http": {Port: 8080, Protocol: pbcatalog.Protocol_PROTOCOL_HTTP},
|
||||
"grpc": {Port: 9090, Protocol: pbcatalog.Protocol_PROTOCOL_GRPC},
|
||||
"mesh": {Port: 10000, Protocol: pbcatalog.Protocol_PROTOCOL_MESH},
|
||||
},
|
||||
HealthStatus: pbcatalog.Health_HEALTH_CRITICAL,
|
||||
},
|
||||
// api-4
|
||||
{
|
||||
TargetRef: c.ResolveResourceID(t, rtest.Resource(types.WorkloadV1Alpha1Type, "api-4").ID()),
|
||||
Addresses: []*pbcatalog.WorkloadAddress{
|
||||
{Host: "172.16.1.4", Ports: []string{"grpc", "http", "mesh"}},
|
||||
{Host: "198.18.2.4", External: true, Ports: []string{"mesh"}},
|
||||
},
|
||||
Ports: map[string]*pbcatalog.WorkloadPort{
|
||||
"http": {Port: 8080, Protocol: pbcatalog.Protocol_PROTOCOL_HTTP},
|
||||
"grpc": {Port: 9090, Protocol: pbcatalog.Protocol_PROTOCOL_GRPC},
|
||||
"mesh": {Port: 10000, Protocol: pbcatalog.Protocol_PROTOCOL_MESH},
|
||||
},
|
||||
HealthStatus: pbcatalog.Health_HEALTH_MAINTENANCE,
|
||||
},
|
||||
// api-5
|
||||
{
|
||||
TargetRef: c.ResolveResourceID(t, rtest.Resource(types.WorkloadV1Alpha1Type, "api-5").ID()),
|
||||
Addresses: []*pbcatalog.WorkloadAddress{
|
||||
{Host: "172.16.1.5", Ports: []string{"grpc", "http", "mesh"}},
|
||||
{Host: "198.18.2.5", External: true, Ports: []string{"mesh"}},
|
||||
},
|
||||
Ports: map[string]*pbcatalog.WorkloadPort{
|
||||
"http": {Port: 8080, Protocol: pbcatalog.Protocol_PROTOCOL_HTTP},
|
||||
"grpc": {Port: 9090, Protocol: pbcatalog.Protocol_PROTOCOL_GRPC},
|
||||
"mesh": {Port: 10000, Protocol: pbcatalog.Protocol_PROTOCOL_MESH},
|
||||
},
|
||||
HealthStatus: pbcatalog.Health_HEALTH_WARNING,
|
||||
},
|
||||
// api-6
|
||||
{
|
||||
TargetRef: c.ResolveResourceID(t, rtest.Resource(types.WorkloadV1Alpha1Type, "api-6").ID()),
|
||||
Addresses: []*pbcatalog.WorkloadAddress{
|
||||
{Host: "172.16.1.6", Ports: []string{"grpc", "http", "mesh"}},
|
||||
{Host: "198.18.2.6", External: true, Ports: []string{"mesh"}},
|
||||
},
|
||||
Ports: map[string]*pbcatalog.WorkloadPort{
|
||||
"http": {Port: 8080, Protocol: pbcatalog.Protocol_PROTOCOL_HTTP},
|
||||
"grpc": {Port: 9090, Protocol: pbcatalog.Protocol_PROTOCOL_GRPC},
|
||||
"mesh": {Port: 10000, Protocol: pbcatalog.Protocol_PROTOCOL_MESH},
|
||||
},
|
||||
HealthStatus: pbcatalog.Health_HEALTH_WARNING,
|
||||
},
|
||||
// api-7
|
||||
{
|
||||
TargetRef: c.ResolveResourceID(t, rtest.Resource(types.WorkloadV1Alpha1Type, "api-7").ID()),
|
||||
Addresses: []*pbcatalog.WorkloadAddress{
|
||||
{Host: "172.16.1.7", Ports: []string{"grpc", "http", "mesh"}},
|
||||
{Host: "198.18.2.7", External: true, Ports: []string{"mesh"}},
|
||||
},
|
||||
Ports: map[string]*pbcatalog.WorkloadPort{
|
||||
"http": {Port: 8080, Protocol: pbcatalog.Protocol_PROTOCOL_HTTP},
|
||||
"grpc": {Port: 9090, Protocol: pbcatalog.Protocol_PROTOCOL_GRPC},
|
||||
"mesh": {Port: 10000, Protocol: pbcatalog.Protocol_PROTOCOL_MESH},
|
||||
},
|
||||
HealthStatus: pbcatalog.Health_HEALTH_CRITICAL,
|
||||
},
|
||||
// api-8
|
||||
{
|
||||
TargetRef: c.ResolveResourceID(t, rtest.Resource(types.WorkloadV1Alpha1Type, "api-8").ID()),
|
||||
Addresses: []*pbcatalog.WorkloadAddress{
|
||||
{Host: "172.16.1.8", Ports: []string{"grpc", "http", "mesh"}},
|
||||
{Host: "198.18.2.8", External: true, Ports: []string{"mesh"}},
|
||||
},
|
||||
Ports: map[string]*pbcatalog.WorkloadPort{
|
||||
"http": {Port: 8080, Protocol: pbcatalog.Protocol_PROTOCOL_HTTP},
|
||||
"grpc": {Port: 9090, Protocol: pbcatalog.Protocol_PROTOCOL_GRPC},
|
||||
"mesh": {Port: 10000, Protocol: pbcatalog.Protocol_PROTOCOL_MESH},
|
||||
},
|
||||
HealthStatus: pbcatalog.Health_HEALTH_MAINTENANCE,
|
||||
},
|
||||
// api-9
|
||||
{
|
||||
TargetRef: c.ResolveResourceID(t, rtest.Resource(types.WorkloadV1Alpha1Type, "api-9").ID()),
|
||||
Addresses: []*pbcatalog.WorkloadAddress{
|
||||
{Host: "172.16.1.9", Ports: []string{"grpc", "http", "mesh"}},
|
||||
{Host: "198.18.2.9", External: true, Ports: []string{"mesh"}},
|
||||
},
|
||||
Ports: map[string]*pbcatalog.WorkloadPort{
|
||||
"http": {Port: 8080, Protocol: pbcatalog.Protocol_PROTOCOL_HTTP},
|
||||
"grpc": {Port: 9090, Protocol: pbcatalog.Protocol_PROTOCOL_GRPC},
|
||||
"mesh": {Port: 10000, Protocol: pbcatalog.Protocol_PROTOCOL_MESH},
|
||||
},
|
||||
HealthStatus: pbcatalog.Health_HEALTH_CRITICAL,
|
||||
},
|
||||
// api-10
|
||||
{
|
||||
TargetRef: c.ResolveResourceID(t, rtest.Resource(types.WorkloadV1Alpha1Type, "api-10").ID()),
|
||||
Addresses: []*pbcatalog.WorkloadAddress{
|
||||
{Host: "172.16.1.10", Ports: []string{"grpc", "http", "mesh"}},
|
||||
{Host: "198.18.2.10", External: true, Ports: []string{"mesh"}},
|
||||
},
|
||||
Ports: map[string]*pbcatalog.WorkloadPort{
|
||||
"http": {Port: 8080, Protocol: pbcatalog.Protocol_PROTOCOL_HTTP},
|
||||
"grpc": {Port: 9090, Protocol: pbcatalog.Protocol_PROTOCOL_GRPC},
|
||||
"mesh": {Port: 10000, Protocol: pbcatalog.Protocol_PROTOCOL_MESH},
|
||||
},
|
||||
HealthStatus: pbcatalog.Health_HEALTH_CRITICAL,
|
||||
},
|
||||
// api-11
|
||||
{
|
||||
TargetRef: c.ResolveResourceID(t, rtest.Resource(types.WorkloadV1Alpha1Type, "api-11").ID()),
|
||||
Addresses: []*pbcatalog.WorkloadAddress{
|
||||
{Host: "172.16.1.11", Ports: []string{"grpc", "http", "mesh"}},
|
||||
{Host: "198.18.2.11", External: true, Ports: []string{"mesh"}},
|
||||
},
|
||||
Ports: map[string]*pbcatalog.WorkloadPort{
|
||||
"http": {Port: 8080, Protocol: pbcatalog.Protocol_PROTOCOL_HTTP},
|
||||
"grpc": {Port: 9090, Protocol: pbcatalog.Protocol_PROTOCOL_GRPC},
|
||||
"mesh": {Port: 10000, Protocol: pbcatalog.Protocol_PROTOCOL_MESH},
|
||||
},
|
||||
HealthStatus: pbcatalog.Health_HEALTH_CRITICAL,
|
||||
},
|
||||
// api-12
|
||||
{
|
||||
TargetRef: c.ResolveResourceID(t, rtest.Resource(types.WorkloadV1Alpha1Type, "api-12").ID()),
|
||||
Addresses: []*pbcatalog.WorkloadAddress{
|
||||
{Host: "172.16.1.12", Ports: []string{"grpc", "http", "mesh"}},
|
||||
{Host: "198.18.2.12", External: true, Ports: []string{"mesh"}},
|
||||
},
|
||||
Ports: map[string]*pbcatalog.WorkloadPort{
|
||||
"http": {Port: 8080, Protocol: pbcatalog.Protocol_PROTOCOL_HTTP},
|
||||
"grpc": {Port: 9090, Protocol: pbcatalog.Protocol_PROTOCOL_GRPC},
|
||||
"mesh": {Port: 10000, Protocol: pbcatalog.Protocol_PROTOCOL_MESH},
|
||||
},
|
||||
HealthStatus: pbcatalog.Health_HEALTH_MAINTENANCE,
|
||||
},
|
||||
// api-13
|
||||
{
|
||||
TargetRef: c.ResolveResourceID(t, rtest.Resource(types.WorkloadV1Alpha1Type, "api-13").ID()),
|
||||
Addresses: []*pbcatalog.WorkloadAddress{
|
||||
{Host: "172.16.1.13", Ports: []string{"grpc", "http", "mesh"}},
|
||||
{Host: "198.18.2.13", External: true, Ports: []string{"mesh"}},
|
||||
},
|
||||
Ports: map[string]*pbcatalog.WorkloadPort{
|
||||
"http": {Port: 8080, Protocol: pbcatalog.Protocol_PROTOCOL_HTTP},
|
||||
"grpc": {Port: 9090, Protocol: pbcatalog.Protocol_PROTOCOL_GRPC},
|
||||
"mesh": {Port: 10000, Protocol: pbcatalog.Protocol_PROTOCOL_MESH},
|
||||
},
|
||||
HealthStatus: pbcatalog.Health_HEALTH_MAINTENANCE,
|
||||
},
|
||||
// api-14
|
||||
{
|
||||
TargetRef: c.ResolveResourceID(t, rtest.Resource(types.WorkloadV1Alpha1Type, "api-14").ID()),
|
||||
Addresses: []*pbcatalog.WorkloadAddress{
|
||||
{Host: "172.16.1.14", Ports: []string{"grpc", "http", "mesh"}},
|
||||
{Host: "198.18.2.14", External: true, Ports: []string{"mesh"}},
|
||||
},
|
||||
Ports: map[string]*pbcatalog.WorkloadPort{
|
||||
"http": {Port: 8080, Protocol: pbcatalog.Protocol_PROTOCOL_HTTP},
|
||||
"grpc": {Port: 9090, Protocol: pbcatalog.Protocol_PROTOCOL_GRPC},
|
||||
"mesh": {Port: 10000, Protocol: pbcatalog.Protocol_PROTOCOL_MESH},
|
||||
},
|
||||
HealthStatus: pbcatalog.Health_HEALTH_MAINTENANCE,
|
||||
},
|
||||
// api-15
|
||||
{
|
||||
TargetRef: c.ResolveResourceID(t, rtest.Resource(types.WorkloadV1Alpha1Type, "api-15").ID()),
|
||||
Addresses: []*pbcatalog.WorkloadAddress{
|
||||
{Host: "172.16.1.15", Ports: []string{"grpc", "http", "mesh"}},
|
||||
{Host: "198.18.2.15", External: true, Ports: []string{"mesh"}},
|
||||
},
|
||||
Ports: map[string]*pbcatalog.WorkloadPort{
|
||||
"http": {Port: 8080, Protocol: pbcatalog.Protocol_PROTOCOL_HTTP},
|
||||
"grpc": {Port: 9090, Protocol: pbcatalog.Protocol_PROTOCOL_GRPC},
|
||||
"mesh": {Port: 10000, Protocol: pbcatalog.Protocol_PROTOCOL_MESH},
|
||||
},
|
||||
HealthStatus: pbcatalog.Health_HEALTH_MAINTENANCE,
|
||||
},
|
||||
// api-16
|
||||
{
|
||||
TargetRef: c.ResolveResourceID(t, rtest.Resource(types.WorkloadV1Alpha1Type, "api-16").ID()),
|
||||
Addresses: []*pbcatalog.WorkloadAddress{
|
||||
{Host: "172.16.1.16", Ports: []string{"grpc", "http", "mesh"}},
|
||||
{Host: "198.18.2.16", External: true, Ports: []string{"mesh"}},
|
||||
},
|
||||
Ports: map[string]*pbcatalog.WorkloadPort{
|
||||
"http": {Port: 8080, Protocol: pbcatalog.Protocol_PROTOCOL_HTTP},
|
||||
"grpc": {Port: 9090, Protocol: pbcatalog.Protocol_PROTOCOL_GRPC},
|
||||
"mesh": {Port: 10000, Protocol: pbcatalog.Protocol_PROTOCOL_MESH},
|
||||
},
|
||||
HealthStatus: pbcatalog.Health_HEALTH_MAINTENANCE,
|
||||
},
|
||||
// api-17
|
||||
{
|
||||
TargetRef: c.ResolveResourceID(t, rtest.Resource(types.WorkloadV1Alpha1Type, "api-17").ID()),
|
||||
Addresses: []*pbcatalog.WorkloadAddress{
|
||||
{Host: "172.16.1.17", Ports: []string{"grpc", "http", "mesh"}},
|
||||
{Host: "198.18.2.17", External: true, Ports: []string{"mesh"}},
|
||||
},
|
||||
Ports: map[string]*pbcatalog.WorkloadPort{
|
||||
"http": {Port: 8080, Protocol: pbcatalog.Protocol_PROTOCOL_HTTP},
|
||||
"grpc": {Port: 9090, Protocol: pbcatalog.Protocol_PROTOCOL_GRPC},
|
||||
"mesh": {Port: 10000, Protocol: pbcatalog.Protocol_PROTOCOL_MESH},
|
||||
},
|
||||
HealthStatus: pbcatalog.Health_HEALTH_PASSING,
|
||||
},
|
||||
// api-18
|
||||
{
|
||||
TargetRef: c.ResolveResourceID(t, rtest.Resource(types.WorkloadV1Alpha1Type, "api-18").ID()),
|
||||
Addresses: []*pbcatalog.WorkloadAddress{
|
||||
{Host: "172.16.1.18", Ports: []string{"grpc", "http", "mesh"}},
|
||||
{Host: "198.18.2.18", External: true, Ports: []string{"mesh"}},
|
||||
},
|
||||
Ports: map[string]*pbcatalog.WorkloadPort{
|
||||
"http": {Port: 8080, Protocol: pbcatalog.Protocol_PROTOCOL_HTTP},
|
||||
"grpc": {Port: 9090, Protocol: pbcatalog.Protocol_PROTOCOL_GRPC},
|
||||
"mesh": {Port: 10000, Protocol: pbcatalog.Protocol_PROTOCOL_MESH},
|
||||
},
|
||||
HealthStatus: pbcatalog.Health_HEALTH_WARNING,
|
||||
},
|
||||
// api-19
|
||||
{
|
||||
TargetRef: c.ResolveResourceID(t, rtest.Resource(types.WorkloadV1Alpha1Type, "api-19").ID()),
|
||||
Addresses: []*pbcatalog.WorkloadAddress{
|
||||
{Host: "172.16.1.19", Ports: []string{"grpc", "http", "mesh"}},
|
||||
{Host: "198.18.2.19", External: true, Ports: []string{"mesh"}},
|
||||
},
|
||||
Ports: map[string]*pbcatalog.WorkloadPort{
|
||||
"http": {Port: 8080, Protocol: pbcatalog.Protocol_PROTOCOL_HTTP},
|
||||
"grpc": {Port: 9090, Protocol: pbcatalog.Protocol_PROTOCOL_GRPC},
|
||||
"mesh": {Port: 10000, Protocol: pbcatalog.Protocol_PROTOCOL_MESH},
|
||||
},
|
||||
HealthStatus: pbcatalog.Health_HEALTH_CRITICAL,
|
||||
},
|
||||
// api-20
|
||||
{
|
||||
TargetRef: c.ResolveResourceID(t, rtest.Resource(types.WorkloadV1Alpha1Type, "api-20").ID()),
|
||||
Addresses: []*pbcatalog.WorkloadAddress{
|
||||
{Host: "172.16.1.20", Ports: []string{"grpc", "http", "mesh"}},
|
||||
{Host: "198.18.2.20", External: true, Ports: []string{"mesh"}},
|
||||
},
|
||||
Ports: map[string]*pbcatalog.WorkloadPort{
|
||||
"http": {Port: 8080, Protocol: pbcatalog.Protocol_PROTOCOL_HTTP},
|
||||
"grpc": {Port: 9090, Protocol: pbcatalog.Protocol_PROTOCOL_GRPC},
|
||||
"mesh": {Port: 10000, Protocol: pbcatalog.Protocol_PROTOCOL_MESH},
|
||||
},
|
||||
HealthStatus: pbcatalog.Health_HEALTH_MAINTENANCE,
|
||||
},
|
||||
},
|
||||
}
|
||||
}
|
||||
|
||||
func expectedHTTPApiServiceEndpoints(t *testing.T, c *rtest.Client) *pbcatalog.ServiceEndpoints {
|
||||
return &pbcatalog.ServiceEndpoints{
|
||||
Endpoints: []*pbcatalog.Endpoint{
|
||||
// api-1
|
||||
{
|
||||
TargetRef: c.ResolveResourceID(t, rtest.Resource(types.WorkloadV1Alpha1Type, "api-1").ID()),
|
||||
Addresses: []*pbcatalog.WorkloadAddress{
|
||||
{Host: "172.16.1.1", Ports: []string{"http"}},
|
||||
},
|
||||
Ports: map[string]*pbcatalog.WorkloadPort{
|
||||
"http": {Port: 8080, Protocol: pbcatalog.Protocol_PROTOCOL_HTTP},
|
||||
},
|
||||
HealthStatus: pbcatalog.Health_HEALTH_PASSING,
|
||||
},
|
||||
// api-10
|
||||
{
|
||||
TargetRef: c.ResolveResourceID(t, rtest.Resource(types.WorkloadV1Alpha1Type, "api-10").ID()),
|
||||
Addresses: []*pbcatalog.WorkloadAddress{
|
||||
{Host: "172.16.1.10", Ports: []string{"http"}},
|
||||
},
|
||||
Ports: map[string]*pbcatalog.WorkloadPort{
|
||||
"http": {Port: 8080, Protocol: pbcatalog.Protocol_PROTOCOL_HTTP},
|
||||
},
|
||||
HealthStatus: pbcatalog.Health_HEALTH_CRITICAL,
|
||||
},
|
||||
// api-11
|
||||
{
|
||||
TargetRef: c.ResolveResourceID(t, rtest.Resource(types.WorkloadV1Alpha1Type, "api-11").ID()),
|
||||
Addresses: []*pbcatalog.WorkloadAddress{
|
||||
{Host: "172.16.1.11", Ports: []string{"http"}},
|
||||
},
|
||||
Ports: map[string]*pbcatalog.WorkloadPort{
|
||||
"http": {Port: 8080, Protocol: pbcatalog.Protocol_PROTOCOL_HTTP},
|
||||
},
|
||||
HealthStatus: pbcatalog.Health_HEALTH_CRITICAL,
|
||||
},
|
||||
// api-12
|
||||
{
|
||||
TargetRef: c.ResolveResourceID(t, rtest.Resource(types.WorkloadV1Alpha1Type, "api-12").ID()),
|
||||
Addresses: []*pbcatalog.WorkloadAddress{
|
||||
{Host: "172.16.1.12", Ports: []string{"http"}},
|
||||
},
|
||||
Ports: map[string]*pbcatalog.WorkloadPort{
|
||||
"http": {Port: 8080, Protocol: pbcatalog.Protocol_PROTOCOL_HTTP},
|
||||
},
|
||||
HealthStatus: pbcatalog.Health_HEALTH_MAINTENANCE,
|
||||
},
|
||||
// api-13
|
||||
{
|
||||
TargetRef: c.ResolveResourceID(t, rtest.Resource(types.WorkloadV1Alpha1Type, "api-13").ID()),
|
||||
Addresses: []*pbcatalog.WorkloadAddress{
|
||||
{Host: "172.16.1.13", Ports: []string{"http"}},
|
||||
},
|
||||
Ports: map[string]*pbcatalog.WorkloadPort{
|
||||
"http": {Port: 8080, Protocol: pbcatalog.Protocol_PROTOCOL_HTTP},
|
||||
},
|
||||
HealthStatus: pbcatalog.Health_HEALTH_MAINTENANCE,
|
||||
},
|
||||
// api-14
|
||||
{
|
||||
TargetRef: c.ResolveResourceID(t, rtest.Resource(types.WorkloadV1Alpha1Type, "api-14").ID()),
|
||||
Addresses: []*pbcatalog.WorkloadAddress{
|
||||
{Host: "172.16.1.14", Ports: []string{"http"}},
|
||||
},
|
||||
Ports: map[string]*pbcatalog.WorkloadPort{
|
||||
"http": {Port: 8080, Protocol: pbcatalog.Protocol_PROTOCOL_HTTP},
|
||||
},
|
||||
HealthStatus: pbcatalog.Health_HEALTH_MAINTENANCE,
|
||||
},
|
||||
// api-15
|
||||
{
|
||||
TargetRef: c.ResolveResourceID(t, rtest.Resource(types.WorkloadV1Alpha1Type, "api-15").ID()),
|
||||
Addresses: []*pbcatalog.WorkloadAddress{
|
||||
{Host: "172.16.1.15", Ports: []string{"http"}},
|
||||
},
|
||||
Ports: map[string]*pbcatalog.WorkloadPort{
|
||||
"http": {Port: 8080, Protocol: pbcatalog.Protocol_PROTOCOL_HTTP},
|
||||
},
|
||||
HealthStatus: pbcatalog.Health_HEALTH_MAINTENANCE,
|
||||
},
|
||||
// api-16
|
||||
{
|
||||
TargetRef: c.ResolveResourceID(t, rtest.Resource(types.WorkloadV1Alpha1Type, "api-16").ID()),
|
||||
Addresses: []*pbcatalog.WorkloadAddress{
|
||||
{Host: "172.16.1.16", Ports: []string{"http"}},
|
||||
},
|
||||
Ports: map[string]*pbcatalog.WorkloadPort{
|
||||
"http": {Port: 8080, Protocol: pbcatalog.Protocol_PROTOCOL_HTTP},
|
||||
},
|
||||
HealthStatus: pbcatalog.Health_HEALTH_MAINTENANCE,
|
||||
},
|
||||
// api-17
|
||||
{
|
||||
TargetRef: c.ResolveResourceID(t, rtest.Resource(types.WorkloadV1Alpha1Type, "api-17").ID()),
|
||||
Addresses: []*pbcatalog.WorkloadAddress{
|
||||
{Host: "172.16.1.17", Ports: []string{"http"}},
|
||||
},
|
||||
Ports: map[string]*pbcatalog.WorkloadPort{
|
||||
"http": {Port: 8080, Protocol: pbcatalog.Protocol_PROTOCOL_HTTP},
|
||||
},
|
||||
HealthStatus: pbcatalog.Health_HEALTH_PASSING,
|
||||
},
|
||||
// api-18
|
||||
{
|
||||
TargetRef: c.ResolveResourceID(t, rtest.Resource(types.WorkloadV1Alpha1Type, "api-18").ID()),
|
||||
Addresses: []*pbcatalog.WorkloadAddress{
|
||||
{Host: "172.16.1.18", Ports: []string{"http"}},
|
||||
},
|
||||
Ports: map[string]*pbcatalog.WorkloadPort{
|
||||
"http": {Port: 8080, Protocol: pbcatalog.Protocol_PROTOCOL_HTTP},
|
||||
},
|
||||
HealthStatus: pbcatalog.Health_HEALTH_WARNING,
|
||||
},
|
||||
// api-19
|
||||
{
|
||||
TargetRef: c.ResolveResourceID(t, rtest.Resource(types.WorkloadV1Alpha1Type, "api-19").ID()),
|
||||
Addresses: []*pbcatalog.WorkloadAddress{
|
||||
{Host: "172.16.1.19", Ports: []string{"http"}},
|
||||
},
|
||||
Ports: map[string]*pbcatalog.WorkloadPort{
|
||||
"http": {Port: 8080, Protocol: pbcatalog.Protocol_PROTOCOL_HTTP},
|
||||
},
|
||||
HealthStatus: pbcatalog.Health_HEALTH_CRITICAL,
|
||||
},
|
||||
},
|
||||
}
|
||||
}
|
||||
|
||||
func expectedGRPCApiServiceEndpoints(t *testing.T, c *rtest.Client) *pbcatalog.ServiceEndpoints {
|
||||
return &pbcatalog.ServiceEndpoints{
|
||||
Endpoints: []*pbcatalog.Endpoint{
|
||||
// api-1
|
||||
{
|
||||
TargetRef: c.ResolveResourceID(t, rtest.Resource(types.WorkloadV1Alpha1Type, "api-1").ID()),
|
||||
Addresses: []*pbcatalog.WorkloadAddress{
|
||||
{Host: "172.16.1.1", Ports: []string{"grpc", "mesh"}},
|
||||
{Host: "198.18.2.1", External: true, Ports: []string{"mesh"}},
|
||||
},
|
||||
Ports: map[string]*pbcatalog.WorkloadPort{
|
||||
"grpc": {Port: 9090, Protocol: pbcatalog.Protocol_PROTOCOL_GRPC},
|
||||
"mesh": {Port: 10000, Protocol: pbcatalog.Protocol_PROTOCOL_MESH},
|
||||
},
|
||||
HealthStatus: pbcatalog.Health_HEALTH_PASSING,
|
||||
},
|
||||
// api-2
|
||||
{
|
||||
TargetRef: c.ResolveResourceID(t, rtest.Resource(types.WorkloadV1Alpha1Type, "api-2").ID()),
|
||||
Addresses: []*pbcatalog.WorkloadAddress{
|
||||
{Host: "172.16.1.2", Ports: []string{"grpc", "mesh"}},
|
||||
{Host: "198.18.2.2", External: true, Ports: []string{"mesh"}},
|
||||
},
|
||||
Ports: map[string]*pbcatalog.WorkloadPort{
|
||||
"grpc": {Port: 9090, Protocol: pbcatalog.Protocol_PROTOCOL_GRPC},
|
||||
"mesh": {Port: 10000, Protocol: pbcatalog.Protocol_PROTOCOL_MESH},
|
||||
},
|
||||
HealthStatus: pbcatalog.Health_HEALTH_WARNING,
|
||||
},
|
||||
// api-3
|
||||
{
|
||||
TargetRef: c.ResolveResourceID(t, rtest.Resource(types.WorkloadV1Alpha1Type, "api-3").ID()),
|
||||
Addresses: []*pbcatalog.WorkloadAddress{
|
||||
{Host: "172.16.1.3", Ports: []string{"grpc", "mesh"}},
|
||||
{Host: "198.18.2.3", External: true, Ports: []string{"mesh"}},
|
||||
},
|
||||
Ports: map[string]*pbcatalog.WorkloadPort{
|
||||
"grpc": {Port: 9090, Protocol: pbcatalog.Protocol_PROTOCOL_GRPC},
|
||||
"mesh": {Port: 10000, Protocol: pbcatalog.Protocol_PROTOCOL_MESH},
|
||||
},
|
||||
HealthStatus: pbcatalog.Health_HEALTH_CRITICAL,
|
||||
},
|
||||
// api-4
|
||||
{
|
||||
TargetRef: c.ResolveResourceID(t, rtest.Resource(types.WorkloadV1Alpha1Type, "api-4").ID()),
|
||||
Addresses: []*pbcatalog.WorkloadAddress{
|
||||
{Host: "172.16.1.4", Ports: []string{"grpc", "mesh"}},
|
||||
{Host: "198.18.2.4", External: true, Ports: []string{"mesh"}},
|
||||
},
|
||||
Ports: map[string]*pbcatalog.WorkloadPort{
|
||||
"grpc": {Port: 9090, Protocol: pbcatalog.Protocol_PROTOCOL_GRPC},
|
||||
"mesh": {Port: 10000, Protocol: pbcatalog.Protocol_PROTOCOL_MESH},
|
||||
},
|
||||
HealthStatus: pbcatalog.Health_HEALTH_MAINTENANCE,
|
||||
},
|
||||
// api-5
|
||||
{
|
||||
TargetRef: c.ResolveResourceID(t, rtest.Resource(types.WorkloadV1Alpha1Type, "api-5").ID()),
|
||||
Addresses: []*pbcatalog.WorkloadAddress{
|
||||
{Host: "172.16.1.5", Ports: []string{"grpc", "mesh"}},
|
||||
{Host: "198.18.2.5", External: true, Ports: []string{"mesh"}},
|
||||
},
|
||||
Ports: map[string]*pbcatalog.WorkloadPort{
|
||||
"grpc": {Port: 9090, Protocol: pbcatalog.Protocol_PROTOCOL_GRPC},
|
||||
"mesh": {Port: 10000, Protocol: pbcatalog.Protocol_PROTOCOL_MESH},
|
||||
},
|
||||
HealthStatus: pbcatalog.Health_HEALTH_WARNING,
|
||||
},
|
||||
// api-6
|
||||
{
|
||||
TargetRef: c.ResolveResourceID(t, rtest.Resource(types.WorkloadV1Alpha1Type, "api-6").ID()),
|
||||
Addresses: []*pbcatalog.WorkloadAddress{
|
||||
{Host: "172.16.1.6", Ports: []string{"grpc", "mesh"}},
|
||||
{Host: "198.18.2.6", External: true, Ports: []string{"mesh"}},
|
||||
},
|
||||
Ports: map[string]*pbcatalog.WorkloadPort{
|
||||
"grpc": {Port: 9090, Protocol: pbcatalog.Protocol_PROTOCOL_GRPC},
|
||||
"mesh": {Port: 10000, Protocol: pbcatalog.Protocol_PROTOCOL_MESH},
|
||||
},
|
||||
HealthStatus: pbcatalog.Health_HEALTH_WARNING,
|
||||
},
|
||||
// api-7
|
||||
{
|
||||
TargetRef: c.ResolveResourceID(t, rtest.Resource(types.WorkloadV1Alpha1Type, "api-7").ID()),
|
||||
Addresses: []*pbcatalog.WorkloadAddress{
|
||||
{Host: "172.16.1.7", Ports: []string{"grpc", "mesh"}},
|
||||
{Host: "198.18.2.7", External: true, Ports: []string{"mesh"}},
|
||||
},
|
||||
Ports: map[string]*pbcatalog.WorkloadPort{
|
||||
"grpc": {Port: 9090, Protocol: pbcatalog.Protocol_PROTOCOL_GRPC},
|
||||
"mesh": {Port: 10000, Protocol: pbcatalog.Protocol_PROTOCOL_MESH},
|
||||
},
|
||||
HealthStatus: pbcatalog.Health_HEALTH_CRITICAL,
|
||||
},
|
||||
// api-8
|
||||
{
|
||||
TargetRef: c.ResolveResourceID(t, rtest.Resource(types.WorkloadV1Alpha1Type, "api-8").ID()),
|
||||
Addresses: []*pbcatalog.WorkloadAddress{
|
||||
{Host: "172.16.1.8", Ports: []string{"grpc", "mesh"}},
|
||||
{Host: "198.18.2.8", External: true, Ports: []string{"mesh"}},
|
||||
},
|
||||
Ports: map[string]*pbcatalog.WorkloadPort{
|
||||
"grpc": {Port: 9090, Protocol: pbcatalog.Protocol_PROTOCOL_GRPC},
|
||||
"mesh": {Port: 10000, Protocol: pbcatalog.Protocol_PROTOCOL_MESH},
|
||||
},
|
||||
HealthStatus: pbcatalog.Health_HEALTH_MAINTENANCE,
|
||||
},
|
||||
// api-9
|
||||
{
|
||||
TargetRef: c.ResolveResourceID(t, rtest.Resource(types.WorkloadV1Alpha1Type, "api-9").ID()),
|
||||
Addresses: []*pbcatalog.WorkloadAddress{
|
||||
{Host: "172.16.1.9", Ports: []string{"grpc", "mesh"}},
|
||||
{Host: "198.18.2.9", External: true, Ports: []string{"mesh"}},
|
||||
},
|
||||
Ports: map[string]*pbcatalog.WorkloadPort{
|
||||
"grpc": {Port: 9090, Protocol: pbcatalog.Protocol_PROTOCOL_GRPC},
|
||||
"mesh": {Port: 10000, Protocol: pbcatalog.Protocol_PROTOCOL_MESH},
|
||||
},
|
||||
HealthStatus: pbcatalog.Health_HEALTH_CRITICAL,
|
||||
},
|
||||
// api-20
|
||||
{
|
||||
TargetRef: c.ResolveResourceID(t, rtest.Resource(types.WorkloadV1Alpha1Type, "api-20").ID()),
|
||||
Addresses: []*pbcatalog.WorkloadAddress{
|
||||
{Host: "172.16.1.20", Ports: []string{"grpc", "mesh"}},
|
||||
{Host: "198.18.2.20", External: true, Ports: []string{"mesh"}},
|
||||
},
|
||||
Ports: map[string]*pbcatalog.WorkloadPort{
|
||||
"grpc": {Port: 9090, Protocol: pbcatalog.Protocol_PROTOCOL_GRPC},
|
||||
"mesh": {Port: 10000, Protocol: pbcatalog.Protocol_PROTOCOL_MESH},
|
||||
},
|
||||
HealthStatus: pbcatalog.Health_HEALTH_MAINTENANCE,
|
||||
},
|
||||
},
|
||||
}
|
||||
}
|
||||
|
||||
func verifyServiceEndpoints(t *testing.T, c *rtest.Client, id *pbresource.ID, expected *pbcatalog.ServiceEndpoints) {
|
||||
c.WaitForResourceState(t, id, func(t rtest.T, res *pbresource.Resource) {
|
||||
var actual pbcatalog.ServiceEndpoints
|
||||
err := res.Data.UnmarshalTo(&actual)
|
||||
require.NoError(t, err)
|
||||
prototest.AssertElementsMatch(t, expected.Endpoints, actual.Endpoints)
|
||||
})
|
||||
}
|
|
@ -6,6 +6,7 @@ package catalog
|
|||
import (
|
||||
"github.com/hashicorp/consul/internal/catalog/internal/controllers"
|
||||
"github.com/hashicorp/consul/internal/catalog/internal/mappers/nodemapper"
|
||||
"github.com/hashicorp/consul/internal/catalog/internal/mappers/selectiontracker"
|
||||
"github.com/hashicorp/consul/internal/catalog/internal/types"
|
||||
"github.com/hashicorp/consul/internal/controller"
|
||||
"github.com/hashicorp/consul/internal/resource"
|
||||
|
@ -52,6 +53,7 @@ type ControllerDependencies = controllers.Dependencies
|
|||
func DefaultControllerDependencies() ControllerDependencies {
|
||||
return ControllerDependencies{
|
||||
WorkloadHealthNodeMapper: nodemapper.New(),
|
||||
EndpointsWorkloadMapper: selectiontracker.New(),
|
||||
}
|
||||
}
|
||||
|
||||
|
|
384
internal/catalog/internal/controllers/endpoints/controller.go
Normal file
384
internal/catalog/internal/controllers/endpoints/controller.go
Normal file
|
@ -0,0 +1,384 @@
|
|||
// Copyright (c) HashiCorp, Inc.
|
||||
// SPDX-License-Identifier: MPL-2.0
|
||||
|
||||
package endpoints
|
||||
|
||||
import (
|
||||
"context"
|
||||
"sort"
|
||||
|
||||
"github.com/hashicorp/consul/internal/catalog/internal/controllers/workloadhealth"
|
||||
"github.com/hashicorp/consul/internal/catalog/internal/types"
|
||||
"github.com/hashicorp/consul/internal/controller"
|
||||
"github.com/hashicorp/consul/internal/resource"
|
||||
pbcatalog "github.com/hashicorp/consul/proto-public/pbcatalog/v1alpha1"
|
||||
"github.com/hashicorp/consul/proto-public/pbresource"
|
||||
"google.golang.org/protobuf/proto"
|
||||
"google.golang.org/protobuf/types/known/anypb"
|
||||
)
|
||||
|
||||
const (
|
||||
endpointsMetaManagedBy = "managed-by-controller"
|
||||
)
|
||||
|
||||
// The WorkloadMapper interface is used to provide an implementation around being able
|
||||
// to map a watch even for a Workload resource and translate it to reconciliation requests
|
||||
type WorkloadMapper interface {
|
||||
// MapWorkload conforms to the controller.DependencyMapper signature. Given a Workload
|
||||
// resource it should report the resource IDs that have selected the workload.
|
||||
MapWorkload(context.Context, controller.Runtime, *pbresource.Resource) ([]controller.Request, error)
|
||||
|
||||
// TrackIDForSelector should be used to associate the specified WorkloadSelector with
|
||||
// the given resource ID. Future calls to MapWorkload
|
||||
TrackIDForSelector(*pbresource.ID, *pbcatalog.WorkloadSelector)
|
||||
|
||||
// UntrackID should be used to inform the tracker to forget about the specified ID
|
||||
UntrackID(*pbresource.ID)
|
||||
}
|
||||
|
||||
// ServiceEndpointsController creates a controller to perform automatic endpoint management for
|
||||
// services.
|
||||
func ServiceEndpointsController(workloadMap WorkloadMapper) controller.Controller {
|
||||
if workloadMap == nil {
|
||||
panic("No WorkloadMapper was provided to the ServiceEndpointsController constructor")
|
||||
}
|
||||
|
||||
return controller.ForType(types.ServiceEndpointsType).
|
||||
WithWatch(types.ServiceType, controller.ReplaceType(types.ServiceEndpointsType)).
|
||||
WithWatch(types.WorkloadType, workloadMap.MapWorkload).
|
||||
WithReconciler(newServiceEndpointsReconciler(workloadMap))
|
||||
}
|
||||
|
||||
type serviceEndpointsReconciler struct {
|
||||
workloadMap WorkloadMapper
|
||||
}
|
||||
|
||||
func newServiceEndpointsReconciler(workloadMap WorkloadMapper) *serviceEndpointsReconciler {
|
||||
return &serviceEndpointsReconciler{
|
||||
workloadMap: workloadMap,
|
||||
}
|
||||
}
|
||||
|
||||
// Reconcile will reconcile one ServiceEndpoints resource in response to some event.
|
||||
func (r *serviceEndpointsReconciler) Reconcile(ctx context.Context, rt controller.Runtime, req controller.Request) error {
|
||||
// The runtime is passed by value so replacing it here for the remainder of this
|
||||
// reconciliation request processing will not affect future invocations.
|
||||
rt.Logger = rt.Logger.With("resource-id", req.ID, "controller", StatusKey)
|
||||
|
||||
rt.Logger.Trace("reconciling service endpoints")
|
||||
|
||||
endpointsID := req.ID
|
||||
serviceID := &pbresource.ID{
|
||||
Type: types.ServiceType,
|
||||
Tenancy: endpointsID.Tenancy,
|
||||
Name: endpointsID.Name,
|
||||
}
|
||||
|
||||
// First we read and unmarshal the service
|
||||
|
||||
serviceData, err := getServiceData(ctx, rt, serviceID)
|
||||
if err != nil {
|
||||
rt.Logger.Error("error retrieving corresponding Service", "error", err)
|
||||
return err
|
||||
}
|
||||
|
||||
// Check if the service exists. If it doesn't we can avoid a bunch of other work.
|
||||
if serviceData == nil {
|
||||
rt.Logger.Trace("service has been deleted")
|
||||
|
||||
// The service was deleted so we need to update the WorkloadMapper to tell it to
|
||||
// stop tracking this service
|
||||
r.workloadMap.UntrackID(req.ID)
|
||||
|
||||
// Note that because we configured ServiceEndpoints to be owned by the service,
|
||||
// the service endpoints object should eventually be automatically deleted.
|
||||
// There is no reason to attempt deletion here.
|
||||
return nil
|
||||
}
|
||||
|
||||
// Now read and unmarshal the endpoints. We don't need this data just yet but all
|
||||
// code paths from this point on will need this regardless of branching so we pull
|
||||
// it now.
|
||||
endpointsData, err := getEndpointsData(ctx, rt, endpointsID)
|
||||
if err != nil {
|
||||
rt.Logger.Error("error retrieving existing endpoints", "error", err)
|
||||
return err
|
||||
}
|
||||
|
||||
var status *pbresource.Condition
|
||||
|
||||
if serviceUnderManagement(serviceData.service) {
|
||||
rt.Logger.Trace("service is enabled for automatic endpoint management")
|
||||
// This service should have its endpoints automatically managed
|
||||
status = ConditionManaged
|
||||
|
||||
// Inform the WorkloadMapper to track this service and its selectors. So
|
||||
// future workload updates that would be matched by the services selectors
|
||||
// cause this service to be rereconciled.
|
||||
r.workloadMap.TrackIDForSelector(req.ID, serviceData.service.GetWorkloads())
|
||||
|
||||
// Now read and umarshal all workloads selected by the service. It is imperative
|
||||
// that this happens after we notify the selection tracker to be tracking that
|
||||
// selection criteria. If the order were reversed we could potentially miss
|
||||
// workload creations that should be selected if they happen after gathering
|
||||
// the workloads but before tracking the selector. Tracking first ensures that
|
||||
// any event that happens after that would get mapped to an event for these
|
||||
// endpoints.
|
||||
workloadData, err := getWorkloadData(ctx, rt, serviceData)
|
||||
if err != nil {
|
||||
rt.Logger.Trace("error retrieving selected workloads", "error", err)
|
||||
return err
|
||||
}
|
||||
|
||||
// Calculate the latest endpoints from the already gathered workloads
|
||||
latestEndpoints := workloadsToEndpoints(serviceData.service, workloadData)
|
||||
|
||||
// Before writing the endpoints actually check to see if they are changed
|
||||
if endpointsData == nil || !proto.Equal(endpointsData.endpoints, latestEndpoints) {
|
||||
rt.Logger.Trace("endpoints have changed")
|
||||
|
||||
// First encode the endpoints data as an Any type.
|
||||
endpointData, err := anypb.New(latestEndpoints)
|
||||
if err != nil {
|
||||
rt.Logger.Error("error marshalling latest endpoints", "error", err)
|
||||
return err
|
||||
}
|
||||
|
||||
// Now perform the write. The endpoints resource should be owned by the service
|
||||
// so that it will automatically be deleted upon service deletion. We are using
|
||||
// a special metadata entry to track that this controller is responsible for
|
||||
// the management of this resource.
|
||||
_, err = rt.Client.Write(ctx, &pbresource.WriteRequest{
|
||||
Resource: &pbresource.Resource{
|
||||
Id: req.ID,
|
||||
Owner: serviceData.resource.Id,
|
||||
Metadata: map[string]string{
|
||||
endpointsMetaManagedBy: StatusKey,
|
||||
},
|
||||
Data: endpointData,
|
||||
},
|
||||
})
|
||||
if err != nil {
|
||||
rt.Logger.Error("error writing generated endpoints", "error", err)
|
||||
return err
|
||||
} else {
|
||||
rt.Logger.Trace("updated endpoints were successfully written")
|
||||
}
|
||||
}
|
||||
} else {
|
||||
rt.Logger.Trace("endpoints are not being automatically managed")
|
||||
// This service is not having its endpoints automatically managed
|
||||
status = ConditionUnmanaged
|
||||
|
||||
// Inform the WorkloadMapper that it no longer needs to track this service
|
||||
// as it is no longer under endpoint management
|
||||
r.workloadMap.UntrackID(req.ID)
|
||||
|
||||
// Delete the managed ServiceEndpoints if necessary if the metadata would
|
||||
// indicate that they were previously managed by this controller
|
||||
if endpointsData != nil && endpointsData.resource.Metadata[endpointsMetaManagedBy] == StatusKey {
|
||||
rt.Logger.Trace("removing previous managed endpoints")
|
||||
|
||||
// This performs a CAS deletion to protect against the case where the user
|
||||
// has overwritten the endpoints since we fetched them.
|
||||
_, err := rt.Client.Delete(ctx, &pbresource.DeleteRequest{
|
||||
Id: endpointsData.resource.Id,
|
||||
Version: endpointsData.resource.Version,
|
||||
})
|
||||
|
||||
// Potentially we could look for CAS failures by checking if the gRPC
|
||||
// status code is Aborted. However its an edge case and there could
|
||||
// possibly be other reasons why the gRPC status code would be aborted
|
||||
// besides CAS version mismatches. The simplest thing to do is to just
|
||||
// propagate the error and retry reconciliation later.
|
||||
if err != nil {
|
||||
rt.Logger.Error("error deleting previously managed endpoints", "error", err)
|
||||
return err
|
||||
}
|
||||
}
|
||||
}
|
||||
|
||||
// Update the Service status if necessary. Mainly we want to inform the user
|
||||
// whether we are automatically managing the endpoints to set expectations
|
||||
// for that object existing or not.
|
||||
newStatus := &pbresource.Status{
|
||||
ObservedGeneration: serviceData.resource.Generation,
|
||||
Conditions: []*pbresource.Condition{
|
||||
status,
|
||||
},
|
||||
}
|
||||
// If the status is unchanged then we should return and avoid the unnecessary write
|
||||
if resource.EqualStatus(serviceData.resource.Status[StatusKey], newStatus, false) {
|
||||
return nil
|
||||
}
|
||||
|
||||
_, err = rt.Client.WriteStatus(ctx, &pbresource.WriteStatusRequest{
|
||||
Id: serviceData.resource.Id,
|
||||
Key: StatusKey,
|
||||
Status: newStatus,
|
||||
})
|
||||
|
||||
if err != nil {
|
||||
rt.Logger.Error("error updating the service's status", "error", err, "service", serviceID)
|
||||
}
|
||||
return err
|
||||
}
|
||||
|
||||
// determineWorkloadHealth will find the workload-health controller's status
|
||||
// within the resource status and parse the workloads health out of it. If
|
||||
// the workload-health controller has yet to reconcile the workload health
|
||||
// or the status isn't in the expected form then this function will return
|
||||
// HEALTH_CRITICAL.
|
||||
func determineWorkloadHealth(workload *pbresource.Resource) pbcatalog.Health {
|
||||
status, found := workload.Status[workloadhealth.StatusKey]
|
||||
if !found {
|
||||
return pbcatalog.Health_HEALTH_CRITICAL
|
||||
}
|
||||
|
||||
for _, condition := range status.Conditions {
|
||||
if condition.Type == workloadhealth.StatusConditionHealthy {
|
||||
raw, found := pbcatalog.Health_value[condition.Reason]
|
||||
if found {
|
||||
return pbcatalog.Health(raw)
|
||||
}
|
||||
return pbcatalog.Health_HEALTH_CRITICAL
|
||||
}
|
||||
}
|
||||
return pbcatalog.Health_HEALTH_CRITICAL
|
||||
}
|
||||
|
||||
// serviceUnderManagement detects whether this service should have its
|
||||
// endpoints automatically managed by the controller
|
||||
func serviceUnderManagement(svc *pbcatalog.Service) bool {
|
||||
sel := svc.GetWorkloads()
|
||||
if sel == nil {
|
||||
// The selector wasn't present at all. Therefore this service is not under
|
||||
// automatic endpoint management.
|
||||
return false
|
||||
}
|
||||
|
||||
if len(sel.Names) < 1 && len(sel.Prefixes) < 1 {
|
||||
// The selector was set in the request but the list of workload names
|
||||
// and prefixes were both empty. Therefore this service is not under
|
||||
// automatic endpoint management
|
||||
return false
|
||||
}
|
||||
|
||||
// Some workload selection criteria exists, so this service is considered
|
||||
// under automatic endpoint management.
|
||||
return true
|
||||
}
|
||||
|
||||
// workloadsToEndpoints will translate the Workload resources into a ServiceEndpoints resource
|
||||
func workloadsToEndpoints(svc *pbcatalog.Service, workloads []*workloadData) *pbcatalog.ServiceEndpoints {
|
||||
var endpoints []*pbcatalog.Endpoint
|
||||
|
||||
for _, workload := range workloads {
|
||||
endpoint := workloadToEndpoint(svc, workload)
|
||||
if endpoint != nil {
|
||||
endpoints = append(endpoints, endpoint)
|
||||
}
|
||||
}
|
||||
|
||||
return &pbcatalog.ServiceEndpoints{
|
||||
Endpoints: endpoints,
|
||||
}
|
||||
}
|
||||
|
||||
// workloadToEndpoint will convert a workload resource into a singular Endpoint to be
|
||||
// put within a ServiceEndpoints resource.
|
||||
//
|
||||
// The conversion process involves parsing the workloads health and filtering its
|
||||
// addresses and ports down to just what the service wants to consume.
|
||||
//
|
||||
// Determining the workloads health requires the workload-health controller to already
|
||||
// have reconciled the workloads health and stored it within the resources Status field.
|
||||
// Any unreconciled workload health will be represented in the ServiceEndpoints with
|
||||
// the ANY health status.
|
||||
func workloadToEndpoint(svc *pbcatalog.Service, data *workloadData) *pbcatalog.Endpoint {
|
||||
health := determineWorkloadHealth(data.resource)
|
||||
|
||||
endpointPorts := make(map[string]*pbcatalog.WorkloadPort)
|
||||
|
||||
// Create the endpoints filtered ports map. Only workload ports specified in
|
||||
// one of the services ports are included. Ports with a protocol mismatch
|
||||
// between the service and workload will be excluded as well.
|
||||
for _, svcPort := range svc.Ports {
|
||||
workloadPort, found := data.workload.Ports[svcPort.TargetPort]
|
||||
if !found {
|
||||
// this workload doesn't have this port so ignore it
|
||||
continue
|
||||
}
|
||||
|
||||
if workloadPort.Protocol != svcPort.Protocol {
|
||||
// workload port mismatch - ignore it
|
||||
continue
|
||||
}
|
||||
|
||||
endpointPorts[svcPort.TargetPort] = workloadPort
|
||||
}
|
||||
|
||||
var workloadAddrs []*pbcatalog.WorkloadAddress
|
||||
|
||||
// Now we filter down the addresses and their corresponding port usage to just
|
||||
// what the service needs to consume. If the address isn't being used to serve
|
||||
// any of the services target ports, it will be entirely excluded from the
|
||||
// address list. If some but not all of its ports are served, then the list
|
||||
// of ports will be reduced to just the intersection of the service ports
|
||||
// and the workload addresses ports
|
||||
for _, addr := range data.workload.Addresses {
|
||||
var ports []string
|
||||
|
||||
if len(addr.Ports) > 0 {
|
||||
// The workload address has defined ports, filter these as necessary
|
||||
|
||||
for _, portName := range addr.Ports {
|
||||
// check if the workload port has been selected by the service
|
||||
_, found := endpointPorts[portName]
|
||||
if !found {
|
||||
// this port isn't selected by the service so drop this port
|
||||
continue
|
||||
}
|
||||
|
||||
ports = append(ports, portName)
|
||||
}
|
||||
} else {
|
||||
// The workload address doesn't specify ports. This lack of port specification
|
||||
// means that all ports are exposed on the interface so here we create a list
|
||||
// of all the port names exposed by the service.
|
||||
for portName := range endpointPorts {
|
||||
ports = append(ports, portName)
|
||||
}
|
||||
}
|
||||
|
||||
// sort the ports to keep them stable and prevent unnecessary rewrites when the endpoints
|
||||
// get diffed
|
||||
sort.Slice(ports, func(i, j int) bool {
|
||||
return ports[i] < ports[j]
|
||||
})
|
||||
|
||||
// Only record this workload address if one or more of its ports were consumed
|
||||
// by the service.
|
||||
if len(ports) > 0 {
|
||||
workloadAddrs = append(workloadAddrs, &pbcatalog.WorkloadAddress{
|
||||
Host: addr.Host,
|
||||
External: addr.External,
|
||||
Ports: ports,
|
||||
})
|
||||
}
|
||||
}
|
||||
|
||||
// If all the workload addresses were filtered out then we should completely ignore
|
||||
// the workload. While the name matched nothing else did so it isn't useable as
|
||||
// far as the service is concerned.
|
||||
if len(workloadAddrs) < 1 {
|
||||
return nil
|
||||
}
|
||||
|
||||
return &pbcatalog.Endpoint{
|
||||
TargetRef: data.resource.Id,
|
||||
HealthStatus: health,
|
||||
Addresses: workloadAddrs,
|
||||
Ports: endpointPorts,
|
||||
}
|
||||
}
|
|
@ -0,0 +1,709 @@
|
|||
package endpoints
|
||||
|
||||
import (
|
||||
"context"
|
||||
"testing"
|
||||
|
||||
svctest "github.com/hashicorp/consul/agent/grpc-external/services/resource/testing"
|
||||
"github.com/hashicorp/consul/internal/catalog/internal/controllers/workloadhealth"
|
||||
"github.com/hashicorp/consul/internal/catalog/internal/mappers/selectiontracker"
|
||||
"github.com/hashicorp/consul/internal/catalog/internal/types"
|
||||
"github.com/hashicorp/consul/internal/controller"
|
||||
rtest "github.com/hashicorp/consul/internal/resource/resourcetest"
|
||||
pbcatalog "github.com/hashicorp/consul/proto-public/pbcatalog/v1alpha1"
|
||||
"github.com/hashicorp/consul/proto-public/pbresource"
|
||||
"github.com/hashicorp/consul/proto/private/prototest"
|
||||
"github.com/hashicorp/consul/sdk/testutil"
|
||||
"github.com/hashicorp/consul/sdk/testutil/retry"
|
||||
"github.com/stretchr/testify/require"
|
||||
"github.com/stretchr/testify/suite"
|
||||
)
|
||||
|
||||
var (
|
||||
badId = rtest.Resource(&pbresource.Type{Group: "not", Kind: "found", GroupVersion: "vfake"}, "foo").ID()
|
||||
)
|
||||
|
||||
func TestWorkloadsToEndpoints(t *testing.T) {
|
||||
// This test's purpose is to ensure that converting multiple workloads to endpoints
|
||||
// happens as expected. It is not concerned with the data in each endpoint but rather
|
||||
// the removal of unconvertable workloads (nil endpoints returned by workloadToEndpoint).
|
||||
|
||||
// The workload to endpoint conversion only cares about the service ports
|
||||
service := &pbcatalog.Service{
|
||||
Ports: []*pbcatalog.ServicePort{
|
||||
{TargetPort: "http2", Protocol: pbcatalog.Protocol_PROTOCOL_HTTP2},
|
||||
},
|
||||
}
|
||||
|
||||
workloadAddresses := []*pbcatalog.WorkloadAddress{
|
||||
{Host: "127.0.0.1"},
|
||||
}
|
||||
|
||||
// This workload is port-matched with the service and should show up as an
|
||||
// endpoint in the final set.
|
||||
workloadData1 := &pbcatalog.Workload{
|
||||
Addresses: workloadAddresses,
|
||||
Ports: map[string]*pbcatalog.WorkloadPort{
|
||||
"http2": {Port: 8080, Protocol: pbcatalog.Protocol_PROTOCOL_HTTP2},
|
||||
},
|
||||
}
|
||||
|
||||
// This workload is NOT port-matched with the service and should be omitted.
|
||||
workloadData2 := &pbcatalog.Workload{
|
||||
Addresses: workloadAddresses,
|
||||
Ports: map[string]*pbcatalog.WorkloadPort{
|
||||
"http": {Port: 8080, Protocol: pbcatalog.Protocol_PROTOCOL_HTTP},
|
||||
},
|
||||
}
|
||||
|
||||
// Build out the workloads.
|
||||
workloads := []*workloadData{
|
||||
{
|
||||
// this workload should result in an endpoints
|
||||
resource: rtest.Resource(types.WorkloadType, "foo").
|
||||
WithData(t, workloadData1).
|
||||
Build(),
|
||||
workload: workloadData1,
|
||||
},
|
||||
{
|
||||
// this workload should be filtered out
|
||||
resource: rtest.Resource(types.WorkloadType, "bar").
|
||||
WithData(t, workloadData2).
|
||||
Build(),
|
||||
workload: workloadData2,
|
||||
},
|
||||
}
|
||||
|
||||
endpoints := workloadsToEndpoints(service, workloads)
|
||||
require.Len(t, endpoints.Endpoints, 1)
|
||||
prototest.AssertDeepEqual(t, workloads[0].resource.Id, endpoints.Endpoints[0].TargetRef)
|
||||
}
|
||||
|
||||
func TestWorkloadToEndpoint(t *testing.T) {
|
||||
// This test handles ensuring that the bulk of the functionality of
|
||||
// the workloadToEndpoint function works correctly.
|
||||
//
|
||||
// * WorkloadPorts that are not selected by one service port are ignored
|
||||
// and not present in the resulting Endpoint
|
||||
// * WorkloadPorts that have a protocol mismatch with the service port
|
||||
// are ignored and not present in the resulting Endpoint
|
||||
// * WorkloadAddresses with 0 non-ignored ports are omitted from the
|
||||
// resulting Endpoint.
|
||||
// * Specifying no ports for a WorkloadAddress will use all the non-ignored
|
||||
// ports. These are explicitly set but that is intended to be an
|
||||
// implementation detail at this point.
|
||||
|
||||
service := &pbcatalog.Service{
|
||||
Ports: []*pbcatalog.ServicePort{
|
||||
// the workload will not have this port so it should be ignored
|
||||
{TargetPort: "not-found", Protocol: pbcatalog.Protocol_PROTOCOL_GRPC},
|
||||
{TargetPort: "http", Protocol: pbcatalog.Protocol_PROTOCOL_HTTP},
|
||||
// the workload will have a different protocol for this port and so it
|
||||
// will be ignored.
|
||||
{TargetPort: "grpc", Protocol: pbcatalog.Protocol_PROTOCOL_GRPC},
|
||||
},
|
||||
}
|
||||
|
||||
workload := &pbcatalog.Workload{
|
||||
Addresses: []*pbcatalog.WorkloadAddress{
|
||||
// this address will be in the endpoint with all the ports that are
|
||||
// not filtered out - so just http
|
||||
{Host: "127.0.0.1"},
|
||||
// this address will be in the endpoint but with a filtered ports list
|
||||
{Host: "198.18.1.1", Ports: []string{"http", "grpc"}},
|
||||
// this address should not show up in the endpoint because the port it
|
||||
// uses is filtered out
|
||||
{Host: "198.8.0.1", Ports: []string{"grpc"}},
|
||||
},
|
||||
Ports: map[string]*pbcatalog.WorkloadPort{
|
||||
"http": {Port: 8080, Protocol: pbcatalog.Protocol_PROTOCOL_HTTP},
|
||||
// the protocol is wrong here so it will not show up in the endpoints.
|
||||
"grpc": {Port: 9090, Protocol: pbcatalog.Protocol_PROTOCOL_HTTP2},
|
||||
},
|
||||
}
|
||||
|
||||
data := &workloadData{
|
||||
resource: rtest.Resource(types.WorkloadType, "foo").
|
||||
WithData(t, workload).
|
||||
Build(),
|
||||
workload: workload,
|
||||
}
|
||||
|
||||
expected := &pbcatalog.Endpoint{
|
||||
TargetRef: data.resource.Id,
|
||||
Addresses: []*pbcatalog.WorkloadAddress{
|
||||
{Host: "127.0.0.1", Ports: []string{"http"}},
|
||||
{Host: "198.18.1.1", Ports: []string{"http"}},
|
||||
},
|
||||
Ports: map[string]*pbcatalog.WorkloadPort{
|
||||
"http": workload.Ports["http"],
|
||||
},
|
||||
// The health is critical because we are not setting the workload's
|
||||
// health status. The tests for determineWorkloadHealth will ensure
|
||||
// that we can properly determine the health status and the overall
|
||||
// controller tests will prove that the integration works as expected.
|
||||
HealthStatus: pbcatalog.Health_HEALTH_CRITICAL,
|
||||
}
|
||||
|
||||
prototest.AssertDeepEqual(t, expected, workloadToEndpoint(service, data))
|
||||
}
|
||||
|
||||
func TestWorkloadToEndpoint_AllAddressesFiltered(t *testing.T) {
|
||||
// This test checks the specific case where the workload has no
|
||||
// address/port combinations that remain unfiltered. In this
|
||||
// case we want to ensure nil is returned instead of an Endpoint
|
||||
// with no addresses.
|
||||
|
||||
service := &pbcatalog.Service{
|
||||
Ports: []*pbcatalog.ServicePort{
|
||||
{TargetPort: "not-found", Protocol: pbcatalog.Protocol_PROTOCOL_HTTP},
|
||||
},
|
||||
}
|
||||
|
||||
workload := &pbcatalog.Workload{
|
||||
Addresses: []*pbcatalog.WorkloadAddress{
|
||||
{Host: "127.0.0.1"},
|
||||
},
|
||||
Ports: map[string]*pbcatalog.WorkloadPort{
|
||||
"http": {Port: 8080, Protocol: pbcatalog.Protocol_PROTOCOL_HTTP},
|
||||
},
|
||||
}
|
||||
|
||||
data := &workloadData{
|
||||
resource: rtest.Resource(types.WorkloadType, "foo").
|
||||
WithData(t, workload).
|
||||
Build(),
|
||||
workload: workload,
|
||||
}
|
||||
|
||||
require.Nil(t, workloadToEndpoint(service, data))
|
||||
}
|
||||
|
||||
func TestServiceUnderManagement(t *testing.T) {
|
||||
// This test ensures that we can properly detect when a service
|
||||
// should have endpoints generated for it vs when those endpoints
|
||||
// are not being automatically managed.
|
||||
|
||||
type testCase struct {
|
||||
svc *pbcatalog.Service
|
||||
managed bool
|
||||
}
|
||||
|
||||
cases := map[string]testCase{
|
||||
"nil": {
|
||||
svc: nil,
|
||||
managed: false,
|
||||
},
|
||||
"nil-selector": {
|
||||
svc: &pbcatalog.Service{Workloads: nil},
|
||||
managed: false,
|
||||
},
|
||||
"empty-selector": {
|
||||
svc: &pbcatalog.Service{Workloads: &pbcatalog.WorkloadSelector{}},
|
||||
managed: false,
|
||||
},
|
||||
"exact-match": {
|
||||
svc: &pbcatalog.Service{Workloads: &pbcatalog.WorkloadSelector{
|
||||
Names: []string{"foo"},
|
||||
}},
|
||||
managed: true,
|
||||
},
|
||||
"prefix-match": {
|
||||
svc: &pbcatalog.Service{Workloads: &pbcatalog.WorkloadSelector{
|
||||
Prefixes: []string{"foo"},
|
||||
}},
|
||||
managed: true,
|
||||
},
|
||||
"multiple": {
|
||||
svc: &pbcatalog.Service{Workloads: &pbcatalog.WorkloadSelector{
|
||||
Names: []string{"foo"},
|
||||
Prefixes: []string{"api-"},
|
||||
}},
|
||||
managed: true,
|
||||
},
|
||||
}
|
||||
|
||||
for name, tcase := range cases {
|
||||
t.Run(name, func(t *testing.T) {
|
||||
require.Equal(t, tcase.managed, serviceUnderManagement(tcase.svc))
|
||||
})
|
||||
}
|
||||
}
|
||||
|
||||
func TestDetermineWorkloadHealth(t *testing.T) {
|
||||
// This test ensures that parsing workload health out of the
|
||||
// resource status works as expected.
|
||||
|
||||
type testCase struct {
|
||||
res *pbresource.Resource
|
||||
expected pbcatalog.Health
|
||||
}
|
||||
|
||||
cases := map[string]testCase{
|
||||
"no-status": {
|
||||
res: rtest.Resource(types.WorkloadType, "foo").Build(),
|
||||
expected: pbcatalog.Health_HEALTH_CRITICAL,
|
||||
},
|
||||
"condition-not-found": {
|
||||
res: rtest.Resource(types.WorkloadType, "foo").
|
||||
WithStatus(workloadhealth.StatusKey, &pbresource.Status{
|
||||
Conditions: []*pbresource.Condition{
|
||||
{
|
||||
Type: "other",
|
||||
State: pbresource.Condition_STATE_TRUE,
|
||||
Reason: "NOT_RELEVANT",
|
||||
},
|
||||
},
|
||||
}).
|
||||
Build(),
|
||||
expected: pbcatalog.Health_HEALTH_CRITICAL,
|
||||
},
|
||||
"invalid-reason": {
|
||||
res: rtest.Resource(types.WorkloadType, "foo").
|
||||
WithStatus(workloadhealth.StatusKey, &pbresource.Status{
|
||||
Conditions: []*pbresource.Condition{
|
||||
{
|
||||
Type: workloadhealth.StatusConditionHealthy,
|
||||
State: pbresource.Condition_STATE_TRUE,
|
||||
Reason: "INVALID_HEALTH_STATUS_REASON",
|
||||
},
|
||||
},
|
||||
}).
|
||||
Build(),
|
||||
expected: pbcatalog.Health_HEALTH_CRITICAL,
|
||||
},
|
||||
"passing": {
|
||||
res: rtest.Resource(types.WorkloadType, "foo").
|
||||
WithStatus(workloadhealth.StatusKey, &pbresource.Status{
|
||||
Conditions: []*pbresource.Condition{
|
||||
{
|
||||
Type: workloadhealth.StatusConditionHealthy,
|
||||
State: pbresource.Condition_STATE_TRUE,
|
||||
Reason: pbcatalog.Health_HEALTH_PASSING.String(),
|
||||
},
|
||||
},
|
||||
}).
|
||||
Build(),
|
||||
expected: pbcatalog.Health_HEALTH_PASSING,
|
||||
},
|
||||
"warning": {
|
||||
res: rtest.Resource(types.WorkloadType, "foo").
|
||||
WithStatus(workloadhealth.StatusKey, &pbresource.Status{
|
||||
Conditions: []*pbresource.Condition{
|
||||
{
|
||||
Type: workloadhealth.StatusConditionHealthy,
|
||||
State: pbresource.Condition_STATE_TRUE,
|
||||
Reason: pbcatalog.Health_HEALTH_WARNING.String(),
|
||||
},
|
||||
},
|
||||
}).
|
||||
Build(),
|
||||
expected: pbcatalog.Health_HEALTH_WARNING,
|
||||
},
|
||||
"critical": {
|
||||
res: rtest.Resource(types.WorkloadType, "foo").
|
||||
WithStatus(workloadhealth.StatusKey, &pbresource.Status{
|
||||
Conditions: []*pbresource.Condition{
|
||||
{
|
||||
Type: workloadhealth.StatusConditionHealthy,
|
||||
State: pbresource.Condition_STATE_TRUE,
|
||||
Reason: pbcatalog.Health_HEALTH_CRITICAL.String(),
|
||||
},
|
||||
},
|
||||
}).
|
||||
Build(),
|
||||
expected: pbcatalog.Health_HEALTH_CRITICAL,
|
||||
},
|
||||
"maintenance": {
|
||||
res: rtest.Resource(types.WorkloadType, "foo").
|
||||
WithStatus(workloadhealth.StatusKey, &pbresource.Status{
|
||||
Conditions: []*pbresource.Condition{
|
||||
{
|
||||
Type: workloadhealth.StatusConditionHealthy,
|
||||
State: pbresource.Condition_STATE_TRUE,
|
||||
Reason: pbcatalog.Health_HEALTH_MAINTENANCE.String(),
|
||||
},
|
||||
},
|
||||
}).
|
||||
Build(),
|
||||
expected: pbcatalog.Health_HEALTH_MAINTENANCE,
|
||||
},
|
||||
}
|
||||
|
||||
for name, tcase := range cases {
|
||||
t.Run(name, func(t *testing.T) {
|
||||
require.Equal(t, tcase.expected, determineWorkloadHealth(tcase.res))
|
||||
})
|
||||
}
|
||||
}
|
||||
|
||||
type controllerSuite struct {
|
||||
suite.Suite
|
||||
|
||||
ctx context.Context
|
||||
client *rtest.Client
|
||||
rt controller.Runtime
|
||||
|
||||
tracker *selectiontracker.WorkloadSelectionTracker
|
||||
reconciler *serviceEndpointsReconciler
|
||||
}
|
||||
|
||||
func (suite *controllerSuite) SetupTest() {
|
||||
suite.ctx = testutil.TestContext(suite.T())
|
||||
client := svctest.RunResourceService(suite.T(), types.Register)
|
||||
suite.rt = controller.Runtime{
|
||||
Client: client,
|
||||
Logger: testutil.Logger(suite.T()),
|
||||
}
|
||||
suite.client = rtest.NewClient(client)
|
||||
suite.tracker = selectiontracker.New()
|
||||
suite.reconciler = newServiceEndpointsReconciler(suite.tracker)
|
||||
}
|
||||
|
||||
func (suite *controllerSuite) requireTracking(workload *pbresource.Resource, ids ...*pbresource.ID) {
|
||||
reqs, err := suite.tracker.MapWorkload(suite.ctx, suite.rt, workload)
|
||||
require.NoError(suite.T(), err)
|
||||
require.Len(suite.T(), reqs, len(ids))
|
||||
for _, id := range ids {
|
||||
prototest.AssertContainsElement(suite.T(), reqs, controller.Request{ID: id})
|
||||
}
|
||||
}
|
||||
|
||||
func (suite *controllerSuite) requireEndpoints(resource *pbresource.Resource, expected ...*pbcatalog.Endpoint) {
|
||||
var svcEndpoints pbcatalog.ServiceEndpoints
|
||||
require.NoError(suite.T(), resource.Data.UnmarshalTo(&svcEndpoints))
|
||||
require.Len(suite.T(), svcEndpoints.Endpoints, len(expected))
|
||||
prototest.AssertElementsMatch(suite.T(), expected, svcEndpoints.Endpoints)
|
||||
}
|
||||
|
||||
func (suite *controllerSuite) TestReconcile_ServiceNotFound() {
|
||||
// This test's purpose is to ensure that when we are reconciling
|
||||
// endpoints for a service that no longer exists, we stop
|
||||
// tracking the endpoints resource ID in the selection tracker.
|
||||
|
||||
// generate a workload resource to use for checking if it maps
|
||||
// to a service endpoints object
|
||||
workload := rtest.Resource(types.WorkloadType, "foo").Build()
|
||||
|
||||
// ensure that the tracker knows about the service prior to
|
||||
// calling reconcile so that we can ensure it removes tracking
|
||||
id := rtest.Resource(types.ServiceEndpointsType, "not-found").ID()
|
||||
suite.tracker.TrackIDForSelector(id, &pbcatalog.WorkloadSelector{Prefixes: []string{""}})
|
||||
|
||||
// verify that mapping the workload to service endpoints returns a
|
||||
// non-empty list prior to reconciliation which should remove the
|
||||
// tracking.
|
||||
suite.requireTracking(workload, id)
|
||||
|
||||
// Because the endpoints don't exist, this reconcile call should
|
||||
// cause tracking of the endpoints to be removed
|
||||
err := suite.reconciler.Reconcile(suite.ctx, suite.rt, controller.Request{ID: id})
|
||||
require.NoError(suite.T(), err)
|
||||
|
||||
// Now ensure that the tracking was removed
|
||||
suite.requireTracking(workload)
|
||||
}
|
||||
|
||||
func (suite *controllerSuite) TestReconcile_NoSelector_NoEndpoints() {
|
||||
// This test's purpose is to ensure that the service's status is
|
||||
// updated to record that its endpoints are not being automatically
|
||||
// managed. Additionally, with no endpoints pre-existing it will
|
||||
// not attempt to delete them.
|
||||
|
||||
service := rtest.Resource(types.ServiceType, "test").
|
||||
WithData(suite.T(), &pbcatalog.Service{
|
||||
Ports: []*pbcatalog.ServicePort{
|
||||
{TargetPort: "http", Protocol: pbcatalog.Protocol_PROTOCOL_HTTP},
|
||||
},
|
||||
}).
|
||||
Write(suite.T(), suite.client)
|
||||
|
||||
endpointsID := rtest.Resource(types.ServiceEndpointsType, "test").ID()
|
||||
|
||||
err := suite.reconciler.Reconcile(suite.ctx, suite.rt, controller.Request{ID: endpointsID})
|
||||
require.NoError(suite.T(), err)
|
||||
|
||||
suite.client.RequireStatusCondition(suite.T(), service.Id, StatusKey, ConditionUnmanaged)
|
||||
}
|
||||
|
||||
func (suite *controllerSuite) TestReconcile_NoSelector_ManagedEndpoints() {
|
||||
// This test's purpose is to ensure that when moving from managed endpoints
|
||||
// to unmanaged endpoints for a service, any already generated managed endpoints
|
||||
// get deleted.
|
||||
|
||||
service := rtest.Resource(types.ServiceType, "test").
|
||||
WithData(suite.T(), &pbcatalog.Service{
|
||||
Ports: []*pbcatalog.ServicePort{
|
||||
{TargetPort: "http", Protocol: pbcatalog.Protocol_PROTOCOL_HTTP},
|
||||
},
|
||||
}).
|
||||
Write(suite.T(), suite.client)
|
||||
|
||||
endpoints := rtest.Resource(types.ServiceEndpointsType, "test").
|
||||
WithData(suite.T(), &pbcatalog.ServiceEndpoints{}).
|
||||
// this marks these endpoints as under management
|
||||
WithMeta(endpointsMetaManagedBy, StatusKey).
|
||||
Write(suite.T(), suite.client)
|
||||
|
||||
err := suite.reconciler.Reconcile(suite.ctx, suite.rt, controller.Request{ID: endpoints.Id})
|
||||
require.NoError(suite.T(), err)
|
||||
// the status should indicate the services endpoints are not being managed
|
||||
suite.client.RequireStatusCondition(suite.T(), service.Id, StatusKey, ConditionUnmanaged)
|
||||
// endpoints under management should be deleted
|
||||
suite.client.RequireResourceNotFound(suite.T(), endpoints.Id)
|
||||
}
|
||||
|
||||
func (suite *controllerSuite) TestReconcile_NoSelector_UnmanagedEndpoints() {
|
||||
// This test's purpose is to ensure that when re-reconciling a service that
|
||||
// doesn't have its endpoints managed, that we do not delete any unmanaged
|
||||
// ServiceEndpoints resource that the user would have manually written.
|
||||
|
||||
service := rtest.Resource(types.ServiceType, "test").
|
||||
WithData(suite.T(), &pbcatalog.Service{
|
||||
Ports: []*pbcatalog.ServicePort{
|
||||
{TargetPort: "http", Protocol: pbcatalog.Protocol_PROTOCOL_HTTP},
|
||||
},
|
||||
}).
|
||||
Write(suite.T(), suite.client)
|
||||
|
||||
endpoints := rtest.Resource(types.ServiceEndpointsType, "test").
|
||||
WithData(suite.T(), &pbcatalog.ServiceEndpoints{}).
|
||||
Write(suite.T(), suite.client)
|
||||
|
||||
err := suite.reconciler.Reconcile(suite.ctx, suite.rt, controller.Request{ID: endpoints.Id})
|
||||
require.NoError(suite.T(), err)
|
||||
// the status should indicate the services endpoints are not being managed
|
||||
suite.client.RequireStatusCondition(suite.T(), service.Id, StatusKey, ConditionUnmanaged)
|
||||
// unmanaged endpoints should not be deleted when the service is unmanaged
|
||||
suite.client.RequireResourceExists(suite.T(), endpoints.Id)
|
||||
}
|
||||
|
||||
func (suite *controllerSuite) TestReconcile_Managed_NoPreviousEndpoints() {
|
||||
// This test's purpose is to ensure the managed endpoint generation occurs
|
||||
// as expected when there are no pre-existing endpoints.
|
||||
|
||||
service := rtest.Resource(types.ServiceType, "test").
|
||||
WithData(suite.T(), &pbcatalog.Service{
|
||||
Workloads: &pbcatalog.WorkloadSelector{
|
||||
Prefixes: []string{""},
|
||||
},
|
||||
Ports: []*pbcatalog.ServicePort{
|
||||
{TargetPort: "http", Protocol: pbcatalog.Protocol_PROTOCOL_HTTP},
|
||||
},
|
||||
}).
|
||||
Write(suite.T(), suite.client)
|
||||
|
||||
endpointsID := rtest.Resource(types.ServiceEndpointsType, "test").ID()
|
||||
|
||||
rtest.Resource(types.WorkloadType, "test-workload").
|
||||
WithData(suite.T(), &pbcatalog.Workload{
|
||||
Addresses: []*pbcatalog.WorkloadAddress{{Host: "127.0.0.1"}},
|
||||
Ports: map[string]*pbcatalog.WorkloadPort{
|
||||
"http": {Port: 8080, Protocol: pbcatalog.Protocol_PROTOCOL_HTTP},
|
||||
},
|
||||
}).
|
||||
Write(suite.T(), suite.client)
|
||||
|
||||
err := suite.reconciler.Reconcile(suite.ctx, suite.rt, controller.Request{ID: endpointsID})
|
||||
require.NoError(suite.T(), err)
|
||||
|
||||
// Verify that the services status has been set to indicate endpoints are automatically managed.
|
||||
suite.client.RequireStatusCondition(suite.T(), service.Id, StatusKey, ConditionManaged)
|
||||
|
||||
// The service endpoints metadata should include our tag to indcate it was generated by this controller
|
||||
res := suite.client.RequireResourceMeta(suite.T(), endpointsID, endpointsMetaManagedBy, StatusKey)
|
||||
|
||||
var endpoints pbcatalog.ServiceEndpoints
|
||||
err = res.Data.UnmarshalTo(&endpoints)
|
||||
require.NoError(suite.T(), err)
|
||||
require.Len(suite.T(), endpoints.Endpoints, 1)
|
||||
// We are not going to retest that the workloads to endpoints conversion process
|
||||
// The length check should be sufficient to prove the endpoints are being
|
||||
// converted. The unit tests for the workloadsToEndpoints functions prove that
|
||||
// the process works correctly in all cases.
|
||||
}
|
||||
|
||||
func (suite *controllerSuite) TestReconcile_Managed_ExistingEndpoints() {
|
||||
// This test's purpose is to ensure that when the current set of endpoints
|
||||
// differs from any prior set of endpoints that the resource gets rewritten.
|
||||
|
||||
service := rtest.Resource(types.ServiceType, "test").
|
||||
WithData(suite.T(), &pbcatalog.Service{
|
||||
Workloads: &pbcatalog.WorkloadSelector{
|
||||
Prefixes: []string{""},
|
||||
},
|
||||
Ports: []*pbcatalog.ServicePort{
|
||||
{TargetPort: "http", Protocol: pbcatalog.Protocol_PROTOCOL_HTTP},
|
||||
},
|
||||
}).
|
||||
Write(suite.T(), suite.client)
|
||||
|
||||
endpoints := rtest.Resource(types.ServiceEndpointsType, "test").
|
||||
WithData(suite.T(), &pbcatalog.ServiceEndpoints{}).
|
||||
WithOwner(service.Id).
|
||||
Write(suite.T(), suite.client)
|
||||
|
||||
rtest.Resource(types.WorkloadType, "test-workload").
|
||||
WithData(suite.T(), &pbcatalog.Workload{
|
||||
Addresses: []*pbcatalog.WorkloadAddress{{Host: "127.0.0.1"}},
|
||||
Ports: map[string]*pbcatalog.WorkloadPort{
|
||||
"http": {Port: 8080, Protocol: pbcatalog.Protocol_PROTOCOL_HTTP},
|
||||
},
|
||||
}).
|
||||
Write(suite.T(), suite.client)
|
||||
|
||||
err := suite.reconciler.Reconcile(suite.ctx, suite.rt, controller.Request{ID: endpoints.Id})
|
||||
require.NoError(suite.T(), err)
|
||||
|
||||
suite.client.RequireStatusCondition(suite.T(), service.Id, StatusKey, ConditionManaged)
|
||||
res := suite.client.RequireResourceMeta(suite.T(), endpoints.Id, endpointsMetaManagedBy, StatusKey)
|
||||
|
||||
var newEndpoints pbcatalog.ServiceEndpoints
|
||||
err = res.Data.UnmarshalTo(&newEndpoints)
|
||||
require.NoError(suite.T(), err)
|
||||
require.Len(suite.T(), newEndpoints.Endpoints, 1)
|
||||
}
|
||||
|
||||
func (suite *controllerSuite) TestController() {
|
||||
// This test's purpose is to exercise the controller in a halfway realistic way.
|
||||
// Generally we are trying to go through the whole lifecycle of creating services,
|
||||
// adding workloads, modifying workload health and modifying the service selection
|
||||
// criteria. This isn't a full integration test as that would require also
|
||||
// executing the workload health controller. Instead workload health status is
|
||||
// synthesized as necessary.
|
||||
|
||||
// Run the controller manager
|
||||
mgr := controller.NewManager(suite.client, suite.rt.Logger)
|
||||
mgr.Register(ServiceEndpointsController(suite.tracker))
|
||||
mgr.SetRaftLeader(true)
|
||||
go mgr.Run(suite.ctx)
|
||||
|
||||
// Add a service - there are no workloads so an empty endpoints
|
||||
// object should be created.
|
||||
service := rtest.Resource(types.ServiceType, "api").
|
||||
WithData(suite.T(), &pbcatalog.Service{
|
||||
Workloads: &pbcatalog.WorkloadSelector{
|
||||
Prefixes: []string{"api-"},
|
||||
},
|
||||
Ports: []*pbcatalog.ServicePort{
|
||||
{TargetPort: "http", Protocol: pbcatalog.Protocol_PROTOCOL_HTTP},
|
||||
},
|
||||
}).
|
||||
Write(suite.T(), suite.client)
|
||||
|
||||
// Wait for the controller to record that the endpoints are being managed
|
||||
res := suite.client.WaitForReconciliation(suite.T(), service.Id, StatusKey)
|
||||
// Check that the services status was updated accordingly
|
||||
rtest.RequireStatusCondition(suite.T(), res, StatusKey, ConditionManaged)
|
||||
|
||||
// Check that the endpoints resource exists and contains 0 endpoints
|
||||
endpointsID := rtest.Resource(types.ServiceEndpointsType, "api").ID()
|
||||
endpoints := suite.client.RequireResourceExists(suite.T(), endpointsID)
|
||||
suite.requireEndpoints(endpoints)
|
||||
|
||||
// Now add a workload that would be selected by the service. Leave
|
||||
// the workload in a state where its health has not been reconciled
|
||||
workload := rtest.Resource(types.WorkloadType, "api-1").
|
||||
WithData(suite.T(), &pbcatalog.Workload{
|
||||
Addresses: []*pbcatalog.WorkloadAddress{{Host: "127.0.0.1"}},
|
||||
Ports: map[string]*pbcatalog.WorkloadPort{
|
||||
"http": {Port: 8080, Protocol: pbcatalog.Protocol_PROTOCOL_HTTP},
|
||||
},
|
||||
Identity: "api",
|
||||
}).
|
||||
Write(suite.T(), suite.client)
|
||||
|
||||
// Wait for the endpoints to be regenerated
|
||||
endpoints = suite.client.WaitForNewVersion(suite.T(), endpointsID, endpoints.Version)
|
||||
|
||||
// Verify that the generated endpoints now contain the workload
|
||||
suite.requireEndpoints(endpoints, &pbcatalog.Endpoint{
|
||||
TargetRef: workload.Id,
|
||||
Addresses: []*pbcatalog.WorkloadAddress{
|
||||
{Host: "127.0.0.1", Ports: []string{"http"}},
|
||||
},
|
||||
Ports: map[string]*pbcatalog.WorkloadPort{
|
||||
"http": {Port: 8080, Protocol: pbcatalog.Protocol_PROTOCOL_HTTP},
|
||||
},
|
||||
HealthStatus: pbcatalog.Health_HEALTH_CRITICAL,
|
||||
})
|
||||
|
||||
// Update the health status of the workload
|
||||
suite.client.WriteStatus(suite.ctx, &pbresource.WriteStatusRequest{
|
||||
Id: workload.Id,
|
||||
Key: workloadhealth.StatusKey,
|
||||
Status: &pbresource.Status{
|
||||
ObservedGeneration: workload.Generation,
|
||||
Conditions: []*pbresource.Condition{
|
||||
{
|
||||
Type: workloadhealth.StatusConditionHealthy,
|
||||
State: pbresource.Condition_STATE_TRUE,
|
||||
Reason: "HEALTH_PASSING",
|
||||
},
|
||||
},
|
||||
},
|
||||
})
|
||||
|
||||
// Wait for the endpoints to be regenerated
|
||||
endpoints = suite.client.WaitForNewVersion(suite.T(), endpointsID, endpoints.Version)
|
||||
|
||||
// ensure the endpoint was put into the passing state
|
||||
suite.requireEndpoints(endpoints, &pbcatalog.Endpoint{
|
||||
TargetRef: workload.Id,
|
||||
Addresses: []*pbcatalog.WorkloadAddress{
|
||||
{Host: "127.0.0.1", Ports: []string{"http"}},
|
||||
},
|
||||
Ports: map[string]*pbcatalog.WorkloadPort{
|
||||
"http": {Port: 8080, Protocol: pbcatalog.Protocol_PROTOCOL_HTTP},
|
||||
},
|
||||
HealthStatus: pbcatalog.Health_HEALTH_PASSING,
|
||||
})
|
||||
|
||||
// rewrite the service to add more selection criteria. This should trigger
|
||||
// reconciliation but shouldn't result in updating the endpoints because
|
||||
// the actual list of currently selected workloads has not changed
|
||||
rtest.Resource(types.ServiceType, "api").
|
||||
WithData(suite.T(), &pbcatalog.Service{
|
||||
Workloads: &pbcatalog.WorkloadSelector{
|
||||
Prefixes: []string{"api-"},
|
||||
Names: []string{"doesnt-matter"},
|
||||
},
|
||||
Ports: []*pbcatalog.ServicePort{
|
||||
{TargetPort: "http", Protocol: pbcatalog.Protocol_PROTOCOL_HTTP},
|
||||
},
|
||||
}).
|
||||
Write(suite.T(), suite.client)
|
||||
|
||||
// Wait for the service status' observed generation to get bumped
|
||||
service = suite.client.WaitForReconciliation(suite.T(), service.Id, StatusKey)
|
||||
|
||||
// Verify that the endpoints were not regenerated
|
||||
suite.client.RequireVersionUnchanged(suite.T(), endpointsID, endpoints.Version)
|
||||
|
||||
// Delete the endpoints. The controller should bring these back momentarily
|
||||
suite.client.Delete(suite.ctx, &pbresource.DeleteRequest{Id: endpointsID})
|
||||
|
||||
// Wait for controller to recreate the endpoints
|
||||
retry.Run(suite.T(), func(r *retry.R) {
|
||||
suite.client.RequireResourceExists(r, endpointsID)
|
||||
})
|
||||
|
||||
// Move the service to having unmanaged endpoints
|
||||
rtest.Resource(types.ServiceType, "api").
|
||||
WithData(suite.T(), &pbcatalog.Service{
|
||||
Ports: []*pbcatalog.ServicePort{
|
||||
{TargetPort: "http", Protocol: pbcatalog.Protocol_PROTOCOL_HTTP},
|
||||
},
|
||||
}).
|
||||
Write(suite.T(), suite.client)
|
||||
|
||||
res = suite.client.WaitForReconciliation(suite.T(), service.Id, StatusKey)
|
||||
rtest.RequireStatusCondition(suite.T(), res, StatusKey, ConditionUnmanaged)
|
||||
|
||||
// Verify that the endpoints were deleted
|
||||
suite.client.RequireResourceNotFound(suite.T(), endpointsID)
|
||||
}
|
||||
|
||||
func TestController(t *testing.T) {
|
||||
suite.Run(t, new(controllerSuite))
|
||||
}
|
|
@ -0,0 +1,177 @@
|
|||
package endpoints
|
||||
|
||||
import (
|
||||
"context"
|
||||
"sort"
|
||||
|
||||
"github.com/hashicorp/consul/internal/catalog/internal/types"
|
||||
"github.com/hashicorp/consul/internal/controller"
|
||||
"github.com/hashicorp/consul/internal/resource"
|
||||
pbcatalog "github.com/hashicorp/consul/proto-public/pbcatalog/v1alpha1"
|
||||
"github.com/hashicorp/consul/proto-public/pbresource"
|
||||
"google.golang.org/grpc/codes"
|
||||
"google.golang.org/grpc/status"
|
||||
)
|
||||
|
||||
type serviceData struct {
|
||||
resource *pbresource.Resource
|
||||
service *pbcatalog.Service
|
||||
}
|
||||
|
||||
type endpointsData struct {
|
||||
resource *pbresource.Resource
|
||||
endpoints *pbcatalog.ServiceEndpoints
|
||||
}
|
||||
|
||||
type workloadData struct {
|
||||
resource *pbresource.Resource
|
||||
workload *pbcatalog.Workload
|
||||
}
|
||||
|
||||
// getServiceData will read the service with the given ID and unmarshal the
|
||||
// Data field. The return value is a struct that contains the retrieved
|
||||
// resource as well as the unmsashalled form. If the resource doesn't
|
||||
// exist, nil will be returned. Any other error either with retrieving
|
||||
// the resource or unmarshalling it will cause the error to be returned
|
||||
// to the caller
|
||||
func getServiceData(ctx context.Context, rt controller.Runtime, id *pbresource.ID) (*serviceData, error) {
|
||||
rsp, err := rt.Client.Read(ctx, &pbresource.ReadRequest{Id: id})
|
||||
switch {
|
||||
case status.Code(err) == codes.NotFound:
|
||||
return nil, nil
|
||||
case err != nil:
|
||||
return nil, err
|
||||
}
|
||||
|
||||
var service pbcatalog.Service
|
||||
err = rsp.Resource.Data.UnmarshalTo(&service)
|
||||
if err != nil {
|
||||
return nil, resource.NewErrDataParse(&service, err)
|
||||
}
|
||||
|
||||
return &serviceData{resource: rsp.Resource, service: &service}, nil
|
||||
}
|
||||
|
||||
// getEndpointsData will read the endpoints with the given ID and unmarshal the
|
||||
// Data field. The return value is a struct that contains the retrieved
|
||||
// resource as well as the unmsashalled form. If the resource doesn't
|
||||
// exist, nil will be returned. Any other error either with retrieving
|
||||
// the resource or unmarshalling it will cause the error to be returned
|
||||
// to the caller
|
||||
func getEndpointsData(ctx context.Context, rt controller.Runtime, id *pbresource.ID) (*endpointsData, error) {
|
||||
rsp, err := rt.Client.Read(ctx, &pbresource.ReadRequest{Id: id})
|
||||
switch {
|
||||
case status.Code(err) == codes.NotFound:
|
||||
return nil, nil
|
||||
case err != nil:
|
||||
return nil, err
|
||||
}
|
||||
|
||||
var endpoints pbcatalog.ServiceEndpoints
|
||||
err = rsp.Resource.Data.UnmarshalTo(&endpoints)
|
||||
if err != nil {
|
||||
return nil, resource.NewErrDataParse(&endpoints, err)
|
||||
}
|
||||
|
||||
return &endpointsData{resource: rsp.Resource, endpoints: &endpoints}, nil
|
||||
}
|
||||
|
||||
// getWorkloadData will retrieve all workloads for the given services selector
|
||||
// and unmarhshal them, returning a slic of objects hold both the resource and
|
||||
// unmarshaled forms. Unmarshalling errors, or other resource service errors
|
||||
// will be returned to the caller.
|
||||
func getWorkloadData(ctx context.Context, rt controller.Runtime, svc *serviceData) ([]*workloadData, error) {
|
||||
workloadResources, err := gatherWorkloadsForService(ctx, rt, svc)
|
||||
if err != nil {
|
||||
return nil, err
|
||||
}
|
||||
|
||||
var results []*workloadData
|
||||
for _, res := range workloadResources {
|
||||
var workload pbcatalog.Workload
|
||||
err = res.Data.UnmarshalTo(&workload)
|
||||
if err != nil {
|
||||
return nil, resource.NewErrDataParse(&workload, err)
|
||||
}
|
||||
|
||||
results = append(results, &workloadData{resource: res, workload: &workload})
|
||||
}
|
||||
|
||||
return results, nil
|
||||
}
|
||||
|
||||
// gatherWorkloadsForService will retrieve all the unique workloads for a given selector.
|
||||
// NotFound errors for workloads selected by Name will be ignored. Any other
|
||||
// resource service errors will be returned to the caller. Prior to returning
|
||||
// the slice of resources, they will be sorted by name. The consistent ordering
|
||||
// will allow callers to diff two versions of the data to determine if anything
|
||||
// has changed but it also will make testing a little easier.
|
||||
func gatherWorkloadsForService(ctx context.Context, rt controller.Runtime, svc *serviceData) ([]*pbresource.Resource, error) {
|
||||
var workloads []*pbresource.Resource
|
||||
|
||||
sel := svc.service.GetWorkloads()
|
||||
|
||||
// this map will track all the gathered workloads by name, this is mainly to deduplicate workloads if they
|
||||
// are specified multiple times throughout the list of selection criteria
|
||||
workloadNames := make(map[string]struct{})
|
||||
|
||||
// First gather all the prefix matched workloads. We could do this second but by doing
|
||||
// it first its possible we can avoid some resource service calls to read individual
|
||||
// workloads selected by name if they are also matched by a prefix.
|
||||
for _, prefix := range sel.GetPrefixes() {
|
||||
rsp, err := rt.Client.List(ctx, &pbresource.ListRequest{
|
||||
Type: types.WorkloadType,
|
||||
Tenancy: svc.resource.Id.Tenancy,
|
||||
NamePrefix: prefix,
|
||||
})
|
||||
if err != nil {
|
||||
return nil, err
|
||||
}
|
||||
|
||||
// append all workloads in the list response to our list of all selected workloads
|
||||
for _, workload := range rsp.Resources {
|
||||
// ignore duplicate workloads
|
||||
if _, found := workloadNames[workload.Id.Name]; !found {
|
||||
workloads = append(workloads, workload)
|
||||
workloadNames[workload.Id.Name] = struct{}{}
|
||||
}
|
||||
}
|
||||
}
|
||||
|
||||
// Now gather the exact match selections
|
||||
for _, name := range sel.GetNames() {
|
||||
// ignore names we have already fetched
|
||||
if _, found := workloadNames[name]; found {
|
||||
continue
|
||||
}
|
||||
|
||||
workloadID := &pbresource.ID{
|
||||
Type: types.WorkloadType,
|
||||
Tenancy: svc.resource.Id.Tenancy,
|
||||
Name: name,
|
||||
}
|
||||
|
||||
rsp, err := rt.Client.Read(ctx, &pbresource.ReadRequest{Id: workloadID})
|
||||
switch {
|
||||
case status.Code(err) == codes.NotFound:
|
||||
// Ignore not found errors as services may select workloads that do not
|
||||
// yet exist. This is not considered an error state or mis-configuration
|
||||
// as the user could be getting ready to add the workloads.
|
||||
continue
|
||||
case err != nil:
|
||||
return nil, err
|
||||
}
|
||||
|
||||
workloads = append(workloads, rsp.Resource)
|
||||
workloadNames[rsp.Resource.Id.Name] = struct{}{}
|
||||
}
|
||||
|
||||
// Sorting ensures deterministic output. This will help for testing but
|
||||
// the real reason to do this is so we will be able to diff the set of
|
||||
// workloads endpoints to determine if we need to update them.
|
||||
sort.Slice(workloads, func(i, j int) bool {
|
||||
return workloads[i].Id.Name < workloads[j].Id.Name
|
||||
})
|
||||
|
||||
return workloads, nil
|
||||
}
|
|
@ -0,0 +1,260 @@
|
|||
package endpoints
|
||||
|
||||
import (
|
||||
"context"
|
||||
"testing"
|
||||
|
||||
svctest "github.com/hashicorp/consul/agent/grpc-external/services/resource/testing"
|
||||
"github.com/hashicorp/consul/internal/catalog/internal/types"
|
||||
"github.com/hashicorp/consul/internal/controller"
|
||||
"github.com/hashicorp/consul/internal/resource"
|
||||
rtest "github.com/hashicorp/consul/internal/resource/resourcetest"
|
||||
pbcatalog "github.com/hashicorp/consul/proto-public/pbcatalog/v1alpha1"
|
||||
"github.com/hashicorp/consul/proto-public/pbresource"
|
||||
"github.com/hashicorp/consul/proto/private/prototest"
|
||||
"github.com/hashicorp/consul/sdk/testutil"
|
||||
"github.com/stretchr/testify/require"
|
||||
"github.com/stretchr/testify/suite"
|
||||
"google.golang.org/grpc/codes"
|
||||
"google.golang.org/grpc/status"
|
||||
)
|
||||
|
||||
type reconciliationDataSuite struct {
|
||||
suite.Suite
|
||||
|
||||
ctx context.Context
|
||||
client pbresource.ResourceServiceClient
|
||||
rt controller.Runtime
|
||||
|
||||
apiServiceData *pbcatalog.Service
|
||||
apiService *pbresource.Resource
|
||||
apiEndpoints *pbresource.Resource
|
||||
api1Workload *pbresource.Resource
|
||||
api2Workload *pbresource.Resource
|
||||
api123Workload *pbresource.Resource
|
||||
web1Workload *pbresource.Resource
|
||||
web2Workload *pbresource.Resource
|
||||
}
|
||||
|
||||
func (suite *reconciliationDataSuite) SetupTest() {
|
||||
suite.ctx = testutil.TestContext(suite.T())
|
||||
suite.client = svctest.RunResourceService(suite.T(), types.Register)
|
||||
suite.rt = controller.Runtime{
|
||||
Client: suite.client,
|
||||
Logger: testutil.Logger(suite.T()),
|
||||
}
|
||||
|
||||
suite.apiServiceData = &pbcatalog.Service{
|
||||
Workloads: &pbcatalog.WorkloadSelector{
|
||||
// This services selectors are specially crafted to exercise both the
|
||||
// dedeuplication and sorting behaviors of gatherWorkloadsForService
|
||||
Prefixes: []string{"api-"},
|
||||
Names: []string{"api-1", "web-2", "web-1", "api-1", "not-found"},
|
||||
},
|
||||
Ports: []*pbcatalog.ServicePort{
|
||||
{
|
||||
TargetPort: "http",
|
||||
Protocol: pbcatalog.Protocol_PROTOCOL_HTTP,
|
||||
},
|
||||
},
|
||||
}
|
||||
|
||||
suite.apiService = rtest.Resource(types.ServiceType, "api").
|
||||
WithData(suite.T(), suite.apiServiceData).
|
||||
Write(suite.T(), suite.client)
|
||||
|
||||
suite.api1Workload = rtest.Resource(types.WorkloadType, "api-1").
|
||||
WithData(suite.T(), &pbcatalog.Workload{
|
||||
Addresses: []*pbcatalog.WorkloadAddress{
|
||||
{Host: "127.0.0.1"},
|
||||
},
|
||||
Ports: map[string]*pbcatalog.WorkloadPort{
|
||||
"http": {Port: 8080, Protocol: pbcatalog.Protocol_PROTOCOL_HTTP},
|
||||
},
|
||||
Identity: "api",
|
||||
}).
|
||||
Write(suite.T(), suite.client)
|
||||
|
||||
suite.api2Workload = rtest.Resource(types.WorkloadType, "api-2").
|
||||
WithData(suite.T(), &pbcatalog.Workload{
|
||||
Addresses: []*pbcatalog.WorkloadAddress{
|
||||
{Host: "127.0.0.1"},
|
||||
},
|
||||
Ports: map[string]*pbcatalog.WorkloadPort{
|
||||
"http": {Port: 8080, Protocol: pbcatalog.Protocol_PROTOCOL_HTTP},
|
||||
},
|
||||
Identity: "api",
|
||||
}).
|
||||
Write(suite.T(), suite.client)
|
||||
|
||||
suite.api123Workload = rtest.Resource(types.WorkloadType, "api-123").
|
||||
WithData(suite.T(), &pbcatalog.Workload{
|
||||
Addresses: []*pbcatalog.WorkloadAddress{
|
||||
{Host: "127.0.0.1"},
|
||||
},
|
||||
Ports: map[string]*pbcatalog.WorkloadPort{
|
||||
"http": {Port: 8080, Protocol: pbcatalog.Protocol_PROTOCOL_HTTP},
|
||||
},
|
||||
Identity: "api",
|
||||
}).
|
||||
Write(suite.T(), suite.client)
|
||||
|
||||
suite.web1Workload = rtest.Resource(types.WorkloadType, "web-1").
|
||||
WithData(suite.T(), &pbcatalog.Workload{
|
||||
Addresses: []*pbcatalog.WorkloadAddress{
|
||||
{Host: "127.0.0.1"},
|
||||
},
|
||||
Ports: map[string]*pbcatalog.WorkloadPort{
|
||||
"http": {Port: 8080, Protocol: pbcatalog.Protocol_PROTOCOL_HTTP},
|
||||
},
|
||||
Identity: "web",
|
||||
}).
|
||||
Write(suite.T(), suite.client)
|
||||
|
||||
suite.web2Workload = rtest.Resource(types.WorkloadType, "web-2").
|
||||
WithData(suite.T(), &pbcatalog.Workload{
|
||||
Addresses: []*pbcatalog.WorkloadAddress{
|
||||
{Host: "127.0.0.1"},
|
||||
},
|
||||
Ports: map[string]*pbcatalog.WorkloadPort{
|
||||
"http": {Port: 8080, Protocol: pbcatalog.Protocol_PROTOCOL_HTTP},
|
||||
},
|
||||
Identity: "web",
|
||||
}).
|
||||
Write(suite.T(), suite.client)
|
||||
|
||||
suite.apiEndpoints = rtest.Resource(types.ServiceEndpointsType, "api").
|
||||
WithData(suite.T(), &pbcatalog.ServiceEndpoints{
|
||||
Endpoints: []*pbcatalog.Endpoint{
|
||||
{
|
||||
TargetRef: rtest.Resource(types.WorkloadType, "api-1").ID(),
|
||||
Addresses: []*pbcatalog.WorkloadAddress{
|
||||
{
|
||||
Host: "127.0.0.1",
|
||||
Ports: []string{"http"},
|
||||
},
|
||||
},
|
||||
Ports: map[string]*pbcatalog.WorkloadPort{
|
||||
"http": {Port: 8080, Protocol: pbcatalog.Protocol_PROTOCOL_HTTP},
|
||||
},
|
||||
HealthStatus: pbcatalog.Health_HEALTH_PASSING,
|
||||
},
|
||||
},
|
||||
}).
|
||||
Write(suite.T(), suite.client)
|
||||
}
|
||||
|
||||
func (suite *reconciliationDataSuite) TestGetServiceData_NotFound() {
|
||||
// This test's purposes is to ensure that NotFound errors when retrieving
|
||||
// the service data are ignored properly.
|
||||
data, err := getServiceData(suite.ctx, suite.rt, rtest.Resource(types.ServiceType, "not-found").ID())
|
||||
require.NoError(suite.T(), err)
|
||||
require.Nil(suite.T(), data)
|
||||
}
|
||||
|
||||
func (suite *reconciliationDataSuite) TestGetServiceData_ReadError() {
|
||||
// This test's purpose is to ensure that Read errors other than NotFound
|
||||
// are propagated back to the caller. Specifying a resource ID with an
|
||||
// unregistered type is the easiest way to force a resource service error.
|
||||
badType := &pbresource.Type{
|
||||
Group: "not",
|
||||
Kind: "found",
|
||||
GroupVersion: "vfake",
|
||||
}
|
||||
data, err := getServiceData(suite.ctx, suite.rt, rtest.Resource(badType, "foo").ID())
|
||||
require.Error(suite.T(), err)
|
||||
require.Equal(suite.T(), codes.InvalidArgument, status.Code(err))
|
||||
require.Nil(suite.T(), data)
|
||||
}
|
||||
|
||||
func (suite *reconciliationDataSuite) TestGetServiceData_UnmarshalError() {
|
||||
// This test's purpose is to ensure that unmarshlling errors are returned
|
||||
// to the caller. We are using a resource id that points to an endpoints
|
||||
// object instead of a service to ensure that the data will be unmarshallable.
|
||||
data, err := getServiceData(suite.ctx, suite.rt, rtest.Resource(types.ServiceEndpointsType, "api").ID())
|
||||
require.Error(suite.T(), err)
|
||||
var parseErr resource.ErrDataParse
|
||||
require.ErrorAs(suite.T(), err, &parseErr)
|
||||
require.Nil(suite.T(), data)
|
||||
}
|
||||
|
||||
func (suite *reconciliationDataSuite) TestGetServiceData_Ok() {
|
||||
// This test's purpose is to ensure that the happy path for
|
||||
// retrieving a service works as expected.
|
||||
data, err := getServiceData(suite.ctx, suite.rt, suite.apiService.Id)
|
||||
require.NoError(suite.T(), err)
|
||||
require.NotNil(suite.T(), data)
|
||||
require.NotNil(suite.T(), data.resource)
|
||||
prototest.AssertDeepEqual(suite.T(), suite.apiService.Id, data.resource.Id)
|
||||
require.Len(suite.T(), data.service.Ports, 1)
|
||||
}
|
||||
|
||||
func (suite *reconciliationDataSuite) TestGetEndpointsData_NotFound() {
|
||||
// This test's purposes is to ensure that NotFound errors when retrieving
|
||||
// the endpoint data are ignored properly.
|
||||
data, err := getEndpointsData(suite.ctx, suite.rt, rtest.Resource(types.ServiceEndpointsType, "not-found").ID())
|
||||
require.NoError(suite.T(), err)
|
||||
require.Nil(suite.T(), data)
|
||||
}
|
||||
|
||||
func (suite *reconciliationDataSuite) TestGetEndpointsData_ReadError() {
|
||||
// This test's purpose is to ensure that Read errors other than NotFound
|
||||
// are propagated back to the caller. Specifying a resource ID with an
|
||||
// unregistered type is the easiest way to force a resource service error.
|
||||
badType := &pbresource.Type{
|
||||
Group: "not",
|
||||
Kind: "found",
|
||||
GroupVersion: "vfake",
|
||||
}
|
||||
data, err := getEndpointsData(suite.ctx, suite.rt, rtest.Resource(badType, "foo").ID())
|
||||
require.Error(suite.T(), err)
|
||||
require.Equal(suite.T(), codes.InvalidArgument, status.Code(err))
|
||||
require.Nil(suite.T(), data)
|
||||
}
|
||||
|
||||
func (suite *reconciliationDataSuite) TestGetEndpointsData_UnmarshalError() {
|
||||
// This test's purpose is to ensure that unmarshlling errors are returned
|
||||
// to the caller. We are using a resource id that points to a service object
|
||||
// instead of an endpoints object to ensure that the data will be unmarshallable.
|
||||
data, err := getEndpointsData(suite.ctx, suite.rt, rtest.Resource(types.ServiceType, "api").ID())
|
||||
require.Error(suite.T(), err)
|
||||
var parseErr resource.ErrDataParse
|
||||
require.ErrorAs(suite.T(), err, &parseErr)
|
||||
require.Nil(suite.T(), data)
|
||||
}
|
||||
|
||||
func (suite *reconciliationDataSuite) TestGetEndpointsData_Ok() {
|
||||
// This test's purpose is to ensure that the happy path for
|
||||
// retrieving an endpoints object works as expected.
|
||||
data, err := getEndpointsData(suite.ctx, suite.rt, suite.apiEndpoints.Id)
|
||||
require.NoError(suite.T(), err)
|
||||
require.NotNil(suite.T(), data)
|
||||
require.NotNil(suite.T(), data.resource)
|
||||
prototest.AssertDeepEqual(suite.T(), suite.apiEndpoints.Id, data.resource.Id)
|
||||
require.Len(suite.T(), data.endpoints.Endpoints, 1)
|
||||
}
|
||||
|
||||
func (suite *reconciliationDataSuite) TestGetWorkloadData() {
|
||||
// This test's purpose is to ensure that gather workloads for
|
||||
// a service work as expected. The services selector was crafted
|
||||
// to exercise the deduplication behavior as well as the sorting
|
||||
// behavior. The assertions in this test will verify that only
|
||||
// unique workloads are returned and that they are ordered.
|
||||
|
||||
data, err := getWorkloadData(suite.ctx, suite.rt, &serviceData{
|
||||
resource: suite.apiService,
|
||||
service: suite.apiServiceData,
|
||||
})
|
||||
|
||||
require.NoError(suite.T(), err)
|
||||
require.Len(suite.T(), data, 5)
|
||||
prototest.AssertDeepEqual(suite.T(), suite.api1Workload, data[0].resource)
|
||||
prototest.AssertDeepEqual(suite.T(), suite.api123Workload, data[1].resource)
|
||||
prototest.AssertDeepEqual(suite.T(), suite.api2Workload, data[2].resource)
|
||||
prototest.AssertDeepEqual(suite.T(), suite.web1Workload, data[3].resource)
|
||||
prototest.AssertDeepEqual(suite.T(), suite.web2Workload, data[4].resource)
|
||||
}
|
||||
|
||||
func TestReconciliationData(t *testing.T) {
|
||||
suite.Run(t, new(reconciliationDataSuite))
|
||||
}
|
33
internal/catalog/internal/controllers/endpoints/status.go
Normal file
33
internal/catalog/internal/controllers/endpoints/status.go
Normal file
|
@ -0,0 +1,33 @@
|
|||
// Copyright (c) HashiCorp, Inc.
|
||||
// SPDX-License-Identifier: MPL-2.0
|
||||
|
||||
package endpoints
|
||||
|
||||
import "github.com/hashicorp/consul/proto-public/pbresource"
|
||||
|
||||
const (
|
||||
StatusKey = "consul.io/endpoint-manager"
|
||||
StatusConditionEndpointsManaged = "EndpointsManaged"
|
||||
|
||||
StatusReasonSelectorNotFound = "SelectorNotFound"
|
||||
StatusReasonSelectorFound = "SelectorFound"
|
||||
|
||||
SelectorFoundMessage = "A valid workload selector is present within the service."
|
||||
SelectorNotFoundMessage = "Either the workload selector was not present or contained no selection criteria."
|
||||
)
|
||||
|
||||
var (
|
||||
ConditionManaged = &pbresource.Condition{
|
||||
Type: StatusConditionEndpointsManaged,
|
||||
State: pbresource.Condition_STATE_TRUE,
|
||||
Reason: StatusReasonSelectorFound,
|
||||
Message: SelectorFoundMessage,
|
||||
}
|
||||
|
||||
ConditionUnmanaged = &pbresource.Condition{
|
||||
Type: StatusConditionEndpointsManaged,
|
||||
State: pbresource.Condition_STATE_FALSE,
|
||||
Reason: StatusReasonSelectorNotFound,
|
||||
Message: SelectorNotFoundMessage,
|
||||
}
|
||||
)
|
|
@ -50,22 +50,10 @@ func (r *nodeHealthReconciler) Reconcile(ctx context.Context, rt controller.Runt
|
|||
return err
|
||||
}
|
||||
|
||||
message := NodeHealthyMessage
|
||||
statusState := pbresource.Condition_STATE_TRUE
|
||||
if health != pbcatalog.Health_HEALTH_PASSING {
|
||||
statusState = pbresource.Condition_STATE_FALSE
|
||||
message = NodeUnhealthyMessage
|
||||
}
|
||||
|
||||
newStatus := &pbresource.Status{
|
||||
ObservedGeneration: res.Generation,
|
||||
Conditions: []*pbresource.Condition{
|
||||
{
|
||||
Type: StatusConditionHealthy,
|
||||
State: statusState,
|
||||
Reason: health.String(),
|
||||
Message: message,
|
||||
},
|
||||
Conditions[health],
|
||||
},
|
||||
}
|
||||
|
||||
|
|
|
@ -3,6 +3,11 @@
|
|||
|
||||
package nodehealth
|
||||
|
||||
import (
|
||||
pbcatalog "github.com/hashicorp/consul/proto-public/pbcatalog/v1alpha1"
|
||||
"github.com/hashicorp/consul/proto-public/pbresource"
|
||||
)
|
||||
|
||||
const (
|
||||
StatusKey = "consul.io/node-health"
|
||||
StatusConditionHealthy = "healthy"
|
||||
|
@ -10,3 +15,40 @@ const (
|
|||
NodeHealthyMessage = "All node health checks are passing"
|
||||
NodeUnhealthyMessage = "One or more node health checks are not passing"
|
||||
)
|
||||
|
||||
var (
|
||||
ConditionPassing = &pbresource.Condition{
|
||||
Type: StatusConditionHealthy,
|
||||
State: pbresource.Condition_STATE_TRUE,
|
||||
Reason: pbcatalog.Health_HEALTH_PASSING.String(),
|
||||
Message: NodeHealthyMessage,
|
||||
}
|
||||
|
||||
ConditionWarning = &pbresource.Condition{
|
||||
Type: StatusConditionHealthy,
|
||||
State: pbresource.Condition_STATE_FALSE,
|
||||
Reason: pbcatalog.Health_HEALTH_WARNING.String(),
|
||||
Message: NodeUnhealthyMessage,
|
||||
}
|
||||
|
||||
ConditionCritical = &pbresource.Condition{
|
||||
Type: StatusConditionHealthy,
|
||||
State: pbresource.Condition_STATE_FALSE,
|
||||
Reason: pbcatalog.Health_HEALTH_CRITICAL.String(),
|
||||
Message: NodeUnhealthyMessage,
|
||||
}
|
||||
|
||||
ConditionMaintenance = &pbresource.Condition{
|
||||
Type: StatusConditionHealthy,
|
||||
State: pbresource.Condition_STATE_FALSE,
|
||||
Reason: pbcatalog.Health_HEALTH_MAINTENANCE.String(),
|
||||
Message: NodeUnhealthyMessage,
|
||||
}
|
||||
|
||||
Conditions = map[pbcatalog.Health]*pbresource.Condition{
|
||||
pbcatalog.Health_HEALTH_PASSING: ConditionPassing,
|
||||
pbcatalog.Health_HEALTH_WARNING: ConditionWarning,
|
||||
pbcatalog.Health_HEALTH_CRITICAL: ConditionCritical,
|
||||
pbcatalog.Health_HEALTH_MAINTENANCE: ConditionMaintenance,
|
||||
}
|
||||
)
|
||||
|
|
|
@ -4,6 +4,7 @@
|
|||
package controllers
|
||||
|
||||
import (
|
||||
"github.com/hashicorp/consul/internal/catalog/internal/controllers/endpoints"
|
||||
"github.com/hashicorp/consul/internal/catalog/internal/controllers/nodehealth"
|
||||
"github.com/hashicorp/consul/internal/catalog/internal/controllers/workloadhealth"
|
||||
"github.com/hashicorp/consul/internal/controller"
|
||||
|
@ -11,9 +12,11 @@ import (
|
|||
|
||||
type Dependencies struct {
|
||||
WorkloadHealthNodeMapper workloadhealth.NodeMapper
|
||||
EndpointsWorkloadMapper endpoints.WorkloadMapper
|
||||
}
|
||||
|
||||
func Register(mgr *controller.Manager, deps Dependencies) {
|
||||
mgr.Register(nodehealth.NodeHealthController())
|
||||
mgr.Register(workloadhealth.WorkloadHealthController(deps.WorkloadHealthNodeMapper))
|
||||
mgr.Register(endpoints.ServiceEndpointsController(deps.EndpointsWorkloadMapper))
|
||||
}
|
||||
|
|
|
@ -58,7 +58,7 @@ type workloadHealthReconciler struct {
|
|||
}
|
||||
|
||||
func (r *workloadHealthReconciler) Reconcile(ctx context.Context, rt controller.Runtime, req controller.Request) error {
|
||||
// The runtime is passed by value so replacing it here for the remaineder of this
|
||||
// The runtime is passed by value so replacing it here for the remainder of this
|
||||
// reconciliation request processing will not affect future invocations.
|
||||
rt.Logger = rt.Logger.With("resource-id", req.ID, "controller", StatusKey)
|
||||
|
||||
|
@ -90,6 +90,12 @@ func (r *workloadHealthReconciler) Reconcile(ctx context.Context, rt controller.
|
|||
if workload.NodeName != "" {
|
||||
nodeID := r.nodeMap.NodeIDFromWorkload(res, &workload)
|
||||
r.nodeMap.TrackWorkload(res.Id, nodeID)
|
||||
|
||||
// It is important that getting the nodes health happens after tracking the
|
||||
// Workload with the node mapper. If the order were reversed we could
|
||||
// potentially miss events for data that changes after we read the node but
|
||||
// before we configured the node mapper to map subsequent events to this
|
||||
// workload.
|
||||
nodeHealth, err = getNodeHealth(ctx, rt, nodeID)
|
||||
if err != nil {
|
||||
rt.Logger.Error("error looking up node health", "error", err, "node-id", nodeID)
|
||||
|
@ -114,33 +120,15 @@ func (r *workloadHealthReconciler) Reconcile(ctx context.Context, rt controller.
|
|||
health = workloadHealth
|
||||
}
|
||||
|
||||
statusState := pbresource.Condition_STATE_TRUE
|
||||
if health != pbcatalog.Health_HEALTH_PASSING {
|
||||
statusState = pbresource.Condition_STATE_FALSE
|
||||
}
|
||||
|
||||
message := WorkloadHealthyMessage
|
||||
condition := WorkloadConditions[workloadHealth]
|
||||
if workload.NodeName != "" {
|
||||
message = NodeAndWorkloadHealthyMessage
|
||||
}
|
||||
switch {
|
||||
case workloadHealth != pbcatalog.Health_HEALTH_PASSING && nodeHealth != pbcatalog.Health_HEALTH_PASSING:
|
||||
message = NodeAndWorkloadUnhealthyMessage
|
||||
case workloadHealth != pbcatalog.Health_HEALTH_PASSING:
|
||||
message = WorkloadUnhealthyMessage
|
||||
case nodeHealth != pbcatalog.Health_HEALTH_PASSING:
|
||||
message = nodehealth.NodeUnhealthyMessage
|
||||
condition = NodeAndWorkloadConditions[workloadHealth][nodeHealth]
|
||||
}
|
||||
|
||||
newStatus := &pbresource.Status{
|
||||
ObservedGeneration: res.Generation,
|
||||
Conditions: []*pbresource.Condition{
|
||||
{
|
||||
Type: StatusConditionHealthy,
|
||||
State: statusState,
|
||||
Reason: health.String(),
|
||||
Message: message,
|
||||
},
|
||||
condition,
|
||||
},
|
||||
}
|
||||
|
||||
|
|
|
@ -1,5 +1,11 @@
|
|||
package workloadhealth
|
||||
|
||||
import (
|
||||
"github.com/hashicorp/consul/internal/catalog/internal/controllers/nodehealth"
|
||||
pbcatalog "github.com/hashicorp/consul/proto-public/pbcatalog/v1alpha1"
|
||||
"github.com/hashicorp/consul/proto-public/pbresource"
|
||||
)
|
||||
|
||||
const (
|
||||
StatusKey = "consul.io/workload-health"
|
||||
StatusConditionHealthy = "healthy"
|
||||
|
@ -9,3 +15,122 @@ const (
|
|||
NodeAndWorkloadUnhealthyMessage = "One or more workload and node health checks are not passing"
|
||||
WorkloadUnhealthyMessage = "One or more workload health checks are not passing"
|
||||
)
|
||||
|
||||
var (
|
||||
ConditionWorkloadPassing = &pbresource.Condition{
|
||||
Type: StatusConditionHealthy,
|
||||
State: pbresource.Condition_STATE_TRUE,
|
||||
Reason: pbcatalog.Health_HEALTH_PASSING.String(),
|
||||
Message: WorkloadHealthyMessage,
|
||||
}
|
||||
|
||||
ConditionWorkloadWarning = &pbresource.Condition{
|
||||
Type: StatusConditionHealthy,
|
||||
State: pbresource.Condition_STATE_FALSE,
|
||||
Reason: pbcatalog.Health_HEALTH_WARNING.String(),
|
||||
Message: WorkloadUnhealthyMessage,
|
||||
}
|
||||
|
||||
ConditionWorkloadCritical = &pbresource.Condition{
|
||||
Type: StatusConditionHealthy,
|
||||
State: pbresource.Condition_STATE_FALSE,
|
||||
Reason: pbcatalog.Health_HEALTH_CRITICAL.String(),
|
||||
Message: WorkloadUnhealthyMessage,
|
||||
}
|
||||
|
||||
ConditionWorkloadMaintenance = &pbresource.Condition{
|
||||
Type: StatusConditionHealthy,
|
||||
State: pbresource.Condition_STATE_FALSE,
|
||||
Reason: pbcatalog.Health_HEALTH_MAINTENANCE.String(),
|
||||
Message: WorkloadUnhealthyMessage,
|
||||
}
|
||||
|
||||
ConditionNodeAndWorkloadPassing = &pbresource.Condition{
|
||||
Type: StatusConditionHealthy,
|
||||
State: pbresource.Condition_STATE_TRUE,
|
||||
Reason: pbcatalog.Health_HEALTH_PASSING.String(),
|
||||
Message: NodeAndWorkloadHealthyMessage,
|
||||
}
|
||||
|
||||
ConditionNodeAndWorkloadWarning = &pbresource.Condition{
|
||||
Type: StatusConditionHealthy,
|
||||
State: pbresource.Condition_STATE_FALSE,
|
||||
Reason: pbcatalog.Health_HEALTH_WARNING.String(),
|
||||
Message: NodeAndWorkloadUnhealthyMessage,
|
||||
}
|
||||
|
||||
ConditionNodeAndWorkloadCritical = &pbresource.Condition{
|
||||
Type: StatusConditionHealthy,
|
||||
State: pbresource.Condition_STATE_FALSE,
|
||||
Reason: pbcatalog.Health_HEALTH_CRITICAL.String(),
|
||||
Message: NodeAndWorkloadUnhealthyMessage,
|
||||
}
|
||||
|
||||
ConditionNodeAndWorkloadMaintenance = &pbresource.Condition{
|
||||
Type: StatusConditionHealthy,
|
||||
State: pbresource.Condition_STATE_FALSE,
|
||||
Reason: pbcatalog.Health_HEALTH_MAINTENANCE.String(),
|
||||
Message: NodeAndWorkloadUnhealthyMessage,
|
||||
}
|
||||
|
||||
ConditionNodeWarning = &pbresource.Condition{
|
||||
Type: StatusConditionHealthy,
|
||||
State: pbresource.Condition_STATE_FALSE,
|
||||
Reason: pbcatalog.Health_HEALTH_WARNING.String(),
|
||||
Message: nodehealth.NodeUnhealthyMessage,
|
||||
}
|
||||
|
||||
ConditionNodeCritical = &pbresource.Condition{
|
||||
Type: StatusConditionHealthy,
|
||||
State: pbresource.Condition_STATE_FALSE,
|
||||
Reason: pbcatalog.Health_HEALTH_CRITICAL.String(),
|
||||
Message: nodehealth.NodeUnhealthyMessage,
|
||||
}
|
||||
|
||||
ConditionNodeMaintenance = &pbresource.Condition{
|
||||
Type: StatusConditionHealthy,
|
||||
State: pbresource.Condition_STATE_FALSE,
|
||||
Reason: pbcatalog.Health_HEALTH_MAINTENANCE.String(),
|
||||
Message: nodehealth.NodeUnhealthyMessage,
|
||||
}
|
||||
|
||||
// WorkloadConditions is a map of the workloadhealth to the status condition
|
||||
// used to represent that health.
|
||||
WorkloadConditions = map[pbcatalog.Health]*pbresource.Condition{
|
||||
pbcatalog.Health_HEALTH_PASSING: ConditionWorkloadPassing,
|
||||
pbcatalog.Health_HEALTH_WARNING: ConditionWorkloadWarning,
|
||||
pbcatalog.Health_HEALTH_CRITICAL: ConditionWorkloadCritical,
|
||||
pbcatalog.Health_HEALTH_MAINTENANCE: ConditionWorkloadMaintenance,
|
||||
}
|
||||
|
||||
// NodeAndWorkloadConditions is a map whose ultimate values are the status conditions
|
||||
// used to represent the combined health of a workload and its associated node.
|
||||
// The outer map's keys are the workloads health and the inner maps keys are the nodes
|
||||
// health
|
||||
NodeAndWorkloadConditions = map[pbcatalog.Health]map[pbcatalog.Health]*pbresource.Condition{
|
||||
pbcatalog.Health_HEALTH_PASSING: {
|
||||
pbcatalog.Health_HEALTH_PASSING: ConditionNodeAndWorkloadPassing,
|
||||
pbcatalog.Health_HEALTH_WARNING: ConditionNodeWarning,
|
||||
pbcatalog.Health_HEALTH_CRITICAL: ConditionNodeCritical,
|
||||
pbcatalog.Health_HEALTH_MAINTENANCE: ConditionNodeMaintenance,
|
||||
},
|
||||
pbcatalog.Health_HEALTH_WARNING: {
|
||||
pbcatalog.Health_HEALTH_PASSING: ConditionWorkloadWarning,
|
||||
pbcatalog.Health_HEALTH_WARNING: ConditionNodeAndWorkloadWarning,
|
||||
pbcatalog.Health_HEALTH_CRITICAL: ConditionNodeAndWorkloadCritical,
|
||||
pbcatalog.Health_HEALTH_MAINTENANCE: ConditionNodeAndWorkloadMaintenance,
|
||||
},
|
||||
pbcatalog.Health_HEALTH_CRITICAL: {
|
||||
pbcatalog.Health_HEALTH_PASSING: ConditionWorkloadCritical,
|
||||
pbcatalog.Health_HEALTH_WARNING: ConditionNodeAndWorkloadCritical,
|
||||
pbcatalog.Health_HEALTH_CRITICAL: ConditionNodeAndWorkloadCritical,
|
||||
pbcatalog.Health_HEALTH_MAINTENANCE: ConditionNodeAndWorkloadMaintenance,
|
||||
},
|
||||
pbcatalog.Health_HEALTH_MAINTENANCE: {
|
||||
pbcatalog.Health_HEALTH_PASSING: ConditionWorkloadMaintenance,
|
||||
pbcatalog.Health_HEALTH_WARNING: ConditionNodeAndWorkloadMaintenance,
|
||||
pbcatalog.Health_HEALTH_CRITICAL: ConditionNodeAndWorkloadMaintenance,
|
||||
pbcatalog.Health_HEALTH_MAINTENANCE: ConditionNodeAndWorkloadMaintenance,
|
||||
},
|
||||
}
|
||||
)
|
||||
|
|
|
@ -0,0 +1,163 @@
|
|||
// Copyright (c) HashiCorp, Inc.
|
||||
// SPDX-License-Identifier: MPL-2.0
|
||||
|
||||
package selectiontracker
|
||||
|
||||
import (
|
||||
"context"
|
||||
"sync"
|
||||
|
||||
"github.com/hashicorp/consul/internal/controller"
|
||||
"github.com/hashicorp/consul/internal/radix"
|
||||
"github.com/hashicorp/consul/internal/resource"
|
||||
"github.com/hashicorp/consul/lib/stringslice"
|
||||
pbcatalog "github.com/hashicorp/consul/proto-public/pbcatalog/v1alpha1"
|
||||
"github.com/hashicorp/consul/proto-public/pbresource"
|
||||
)
|
||||
|
||||
type WorkloadSelectionTracker struct {
|
||||
lock sync.Mutex
|
||||
prefixes *radix.Tree[[]controller.Request]
|
||||
exact *radix.Tree[[]controller.Request]
|
||||
|
||||
// workloadSelectors contains a map keyed on resource names with values
|
||||
// being the selector that resource is currently associated with. This map
|
||||
// is kept mainly to make tracking removal operations more efficient.
|
||||
// Generally any operation that could take advantage of knowing where
|
||||
// in the trees the resource id is referenced can use this to prevent
|
||||
// needing to search the whole tree.
|
||||
workloadSelectors map[string]*pbcatalog.WorkloadSelector
|
||||
}
|
||||
|
||||
func New() *WorkloadSelectionTracker {
|
||||
return &WorkloadSelectionTracker{
|
||||
prefixes: radix.New[[]controller.Request](),
|
||||
exact: radix.New[[]controller.Request](),
|
||||
workloadSelectors: make(map[string]*pbcatalog.WorkloadSelector),
|
||||
}
|
||||
}
|
||||
|
||||
// MapWorkload will return a slice of controller.Requests with 1 resource for
|
||||
// each resource that selects the specified Workload resource.
|
||||
func (t *WorkloadSelectionTracker) MapWorkload(_ context.Context, _ controller.Runtime, res *pbresource.Resource) ([]controller.Request, error) {
|
||||
t.lock.Lock()
|
||||
defer t.lock.Unlock()
|
||||
|
||||
var reqs []controller.Request
|
||||
|
||||
// gather the list of all resources that select the specified workload using a prefix match
|
||||
t.prefixes.WalkPath(res.Id.Name, func(path string, requests []controller.Request) bool {
|
||||
reqs = append(reqs, requests...)
|
||||
return false
|
||||
})
|
||||
|
||||
// gather the list of all resources that select the specified workload using an exact match
|
||||
exactReqs, _ := t.exact.Get(res.Id.Name)
|
||||
|
||||
// return the combined list of all resources that select the specified workload
|
||||
return append(reqs, exactReqs...), nil
|
||||
}
|
||||
|
||||
// TrackIDForSelector will associate workloads matching the specified workload
|
||||
// selector with the given resource id.
|
||||
func (t *WorkloadSelectionTracker) TrackIDForSelector(id *pbresource.ID, selector *pbcatalog.WorkloadSelector) {
|
||||
t.lock.Lock()
|
||||
defer t.lock.Unlock()
|
||||
|
||||
if previousSelector, found := t.workloadSelectors[id.Name]; found {
|
||||
if stringslice.Equal(previousSelector.Names, selector.Names) &&
|
||||
stringslice.Equal(previousSelector.Prefixes, selector.Prefixes) {
|
||||
// the selector is unchanged so do nothing
|
||||
return
|
||||
}
|
||||
|
||||
// Potentially we could detect differences and do more minimal work. However
|
||||
// users are not expected to alter workload selectors often and therefore
|
||||
// not optimizing this further is probably fine. Therefore we are going
|
||||
// to wipe all tracking of the id and reinsert things.
|
||||
t.untrackID(id)
|
||||
}
|
||||
|
||||
// loop over all the exact matching rules and associate those workload names
|
||||
// with the given resource id
|
||||
for _, name := range selector.GetNames() {
|
||||
// lookup any resource id associations for the given workload name
|
||||
leaf, _ := t.exact.Get(name)
|
||||
|
||||
// append the ID to the existing request list
|
||||
t.exact.Insert(name, append(leaf, controller.Request{ID: id}))
|
||||
}
|
||||
|
||||
// loop over all the prefix matching rules and associate those prefixes
|
||||
// with the given resource id.
|
||||
for _, prefix := range selector.GetPrefixes() {
|
||||
// lookup any resource id associations for the given workload name prefix
|
||||
leaf, _ := t.prefixes.Get(prefix)
|
||||
|
||||
// append the new resource ID to the existing request list
|
||||
t.prefixes.Insert(prefix, append(leaf, controller.Request{ID: id}))
|
||||
}
|
||||
|
||||
t.workloadSelectors[id.Name] = selector
|
||||
}
|
||||
|
||||
// UntrackID causes the tracker to stop tracking the given resource ID
|
||||
func (t *WorkloadSelectionTracker) UntrackID(id *pbresource.ID) {
|
||||
t.lock.Lock()
|
||||
defer t.lock.Unlock()
|
||||
t.untrackID(id)
|
||||
}
|
||||
|
||||
// untrackID should be called to stop tracking a resource ID.
|
||||
// This method assumes the lock is already held. Besides modifying
|
||||
// the prefix & name trees to not reference this ID, it will also
|
||||
// delete any corresponding entry within the workloadSelectors map
|
||||
func (t *WorkloadSelectionTracker) untrackID(id *pbresource.ID) {
|
||||
selector, found := t.workloadSelectors[id.Name]
|
||||
if !found {
|
||||
return
|
||||
}
|
||||
|
||||
removeIDFromTreeAtPaths(t.exact, id, selector.Names)
|
||||
removeIDFromTreeAtPaths(t.prefixes, id, selector.Prefixes)
|
||||
|
||||
// If we don't do this deletion then reinsertion of the id for
|
||||
// tracking in the future could prevent selection criteria from
|
||||
// being properly inserted into the radix trees.
|
||||
delete(t.workloadSelectors, id.Name)
|
||||
}
|
||||
|
||||
// removeIDFromTree will remove the given resource ID from all leaf nodes in the radix tree.
|
||||
func removeIDFromTreeAtPaths(t *radix.Tree[[]controller.Request], id *pbresource.ID, paths []string) {
|
||||
for _, path := range paths {
|
||||
requests, _ := t.Get(path)
|
||||
|
||||
foundIdx := -1
|
||||
for idx, req := range requests {
|
||||
if resource.EqualID(req.ID, id) {
|
||||
foundIdx = idx
|
||||
break
|
||||
}
|
||||
}
|
||||
|
||||
if foundIdx != -1 {
|
||||
l := len(requests)
|
||||
|
||||
if l == 1 {
|
||||
requests = nil
|
||||
} else if foundIdx == l-1 {
|
||||
requests = requests[:foundIdx]
|
||||
} else if foundIdx == 0 {
|
||||
requests = requests[1:]
|
||||
} else {
|
||||
requests = append(requests[:foundIdx], requests[foundIdx+1:]...)
|
||||
}
|
||||
|
||||
if len(requests) > 1 {
|
||||
t.Insert(path, requests)
|
||||
} else {
|
||||
t.Delete(path)
|
||||
}
|
||||
}
|
||||
}
|
||||
}
|
|
@ -0,0 +1,275 @@
|
|||
package selectiontracker
|
||||
|
||||
import (
|
||||
"context"
|
||||
"testing"
|
||||
|
||||
"github.com/hashicorp/consul/internal/catalog/internal/types"
|
||||
"github.com/hashicorp/consul/internal/controller"
|
||||
"github.com/hashicorp/consul/internal/radix"
|
||||
rtest "github.com/hashicorp/consul/internal/resource/resourcetest"
|
||||
pbcatalog "github.com/hashicorp/consul/proto-public/pbcatalog/v1alpha1"
|
||||
"github.com/hashicorp/consul/proto-public/pbresource"
|
||||
"github.com/hashicorp/consul/proto/private/prototest"
|
||||
"github.com/stretchr/testify/require"
|
||||
"github.com/stretchr/testify/suite"
|
||||
)
|
||||
|
||||
var (
|
||||
workloadData = &pbcatalog.Workload{
|
||||
Addresses: []*pbcatalog.WorkloadAddress{
|
||||
{
|
||||
Host: "198.18.0.1",
|
||||
},
|
||||
},
|
||||
Ports: map[string]*pbcatalog.WorkloadPort{
|
||||
"http": {
|
||||
Port: 8080,
|
||||
Protocol: pbcatalog.Protocol_PROTOCOL_HTTP,
|
||||
},
|
||||
},
|
||||
}
|
||||
)
|
||||
|
||||
func TestRemoveIDFromTreeAtPaths(t *testing.T) {
|
||||
tree := radix.New[[]controller.Request]()
|
||||
|
||||
toRemove := rtest.Resource(types.ServiceEndpointsType, "blah").ID()
|
||||
other1 := rtest.Resource(types.ServiceEndpointsType, "other1").ID()
|
||||
other2 := rtest.Resource(types.ServiceEndpointsType, "other1").ID()
|
||||
|
||||
// we are trying to create a tree such that removal of the toRemove id causes a
|
||||
// few things to happen.
|
||||
//
|
||||
// * All the slice modification conditions are executed
|
||||
// - removal from beginning of the list
|
||||
// - removal from the end of the list
|
||||
// - removal of only element in the list
|
||||
// - removal from middle of the list
|
||||
// * Paths without matching ids are ignored
|
||||
|
||||
notMatching := []controller.Request{
|
||||
{ID: other1},
|
||||
{ID: other2},
|
||||
}
|
||||
|
||||
matchAtBeginning := []controller.Request{
|
||||
{ID: toRemove},
|
||||
{ID: other1},
|
||||
{ID: other2},
|
||||
}
|
||||
|
||||
matchAtEnd := []controller.Request{
|
||||
{ID: other1},
|
||||
{ID: other2},
|
||||
{ID: toRemove},
|
||||
}
|
||||
|
||||
matchInMiddle := []controller.Request{
|
||||
{ID: other1},
|
||||
{ID: toRemove},
|
||||
{ID: other2},
|
||||
}
|
||||
|
||||
matchOnly := []controller.Request{
|
||||
{ID: toRemove},
|
||||
}
|
||||
|
||||
tree.Insert("no-match", notMatching)
|
||||
tree.Insert("match-beginning", matchAtBeginning)
|
||||
tree.Insert("match-end", matchAtEnd)
|
||||
tree.Insert("match-middle", matchInMiddle)
|
||||
tree.Insert("match-only", matchOnly)
|
||||
|
||||
removeIDFromTreeAtPaths(tree, toRemove, []string{
|
||||
"no-match",
|
||||
"match-beginning",
|
||||
"match-end",
|
||||
"match-middle",
|
||||
"match-only",
|
||||
})
|
||||
|
||||
reqs, found := tree.Get("no-match")
|
||||
require.True(t, found)
|
||||
require.Equal(t, notMatching, reqs)
|
||||
|
||||
reqs, found = tree.Get("match-beginning")
|
||||
require.True(t, found)
|
||||
require.Equal(t, notMatching, reqs)
|
||||
|
||||
reqs, found = tree.Get("match-end")
|
||||
require.True(t, found)
|
||||
require.Equal(t, notMatching, reqs)
|
||||
|
||||
reqs, found = tree.Get("match-middle")
|
||||
require.True(t, found)
|
||||
require.Equal(t, notMatching, reqs)
|
||||
|
||||
// The last tracked request should cause removal from the tree
|
||||
_, found = tree.Get("match-only")
|
||||
require.False(t, found)
|
||||
}
|
||||
|
||||
type selectionTrackerSuite struct {
|
||||
suite.Suite
|
||||
|
||||
rt controller.Runtime
|
||||
tracker *WorkloadSelectionTracker
|
||||
|
||||
workloadAPI1 *pbresource.Resource
|
||||
workloadWeb1 *pbresource.Resource
|
||||
endpointsFoo *pbresource.ID
|
||||
endpointsBar *pbresource.ID
|
||||
}
|
||||
|
||||
func (suite *selectionTrackerSuite) SetupTest() {
|
||||
suite.tracker = New()
|
||||
|
||||
suite.workloadAPI1 = rtest.Resource(types.WorkloadType, "api-1").WithData(suite.T(), workloadData).Build()
|
||||
suite.workloadWeb1 = rtest.Resource(types.WorkloadType, "web-1").WithData(suite.T(), workloadData).Build()
|
||||
suite.endpointsFoo = rtest.Resource(types.ServiceEndpointsType, "foo").ID()
|
||||
suite.endpointsBar = rtest.Resource(types.ServiceEndpointsType, "bar").ID()
|
||||
}
|
||||
|
||||
func (suite *selectionTrackerSuite) requireMappedIDs(workload *pbresource.Resource, ids ...*pbresource.ID) {
|
||||
suite.T().Helper()
|
||||
|
||||
reqs, err := suite.tracker.MapWorkload(context.Background(), suite.rt, workload)
|
||||
require.NoError(suite.T(), err)
|
||||
require.Len(suite.T(), reqs, len(ids))
|
||||
for _, id := range ids {
|
||||
prototest.AssertContainsElement(suite.T(), reqs, controller.Request{ID: id})
|
||||
}
|
||||
}
|
||||
|
||||
func (suite *selectionTrackerSuite) TestMapWorkload_Empty() {
|
||||
// If we aren't tracking anything than the default mapping behavior
|
||||
// should be to return an empty list of requests.
|
||||
suite.requireMappedIDs(suite.workloadAPI1)
|
||||
}
|
||||
|
||||
func (suite *selectionTrackerSuite) TestUntrackID_Empty() {
|
||||
// this test has no assertions but mainly is here to prove that things
|
||||
// dont explode if this is attempted.
|
||||
suite.tracker.UntrackID(suite.endpointsFoo)
|
||||
}
|
||||
|
||||
func (suite *selectionTrackerSuite) TestTrackAndMap_SingleResource_MultipleWorkloadMappings() {
|
||||
// This test aims to prove that tracking a resources workload selector and
|
||||
// then mapping a workload back to that resource works as expected when the
|
||||
// result set is a single resource. This test will ensure that both prefix
|
||||
// and exact match criteria are handle correctly and that one resource
|
||||
// can be mapped from multiple distinct workloads.
|
||||
|
||||
// associate the foo endpoints with some workloads
|
||||
suite.tracker.TrackIDForSelector(suite.endpointsFoo, &pbcatalog.WorkloadSelector{
|
||||
Names: []string{"bar", "api", "web-1"},
|
||||
Prefixes: []string{"api-"},
|
||||
})
|
||||
|
||||
// Ensure that mappings tracked by prefix work.
|
||||
suite.requireMappedIDs(suite.workloadAPI1, suite.endpointsFoo)
|
||||
|
||||
// Ensure that mappings tracked by exact match work.
|
||||
suite.requireMappedIDs(suite.workloadWeb1, suite.endpointsFoo)
|
||||
}
|
||||
|
||||
func (suite *selectionTrackerSuite) TestTrackAndMap_MultiResource_SingleWorkloadMapping() {
|
||||
// This test aims to prove that multiple resources selecting of a workload
|
||||
// will result in multiple requests when mapping that workload.
|
||||
|
||||
// associate the foo endpoints with some workloads
|
||||
suite.tracker.TrackIDForSelector(suite.endpointsFoo, &pbcatalog.WorkloadSelector{
|
||||
Prefixes: []string{"api-"},
|
||||
})
|
||||
|
||||
// associate the bar endpoints with some workloads
|
||||
suite.tracker.TrackIDForSelector(suite.endpointsBar, &pbcatalog.WorkloadSelector{
|
||||
Names: []string{"api-1"},
|
||||
})
|
||||
|
||||
// now the mapping should return both endpoints resource ids
|
||||
suite.requireMappedIDs(suite.workloadAPI1, suite.endpointsFoo, suite.endpointsBar)
|
||||
}
|
||||
|
||||
func (suite *selectionTrackerSuite) TestDuplicateTracking() {
|
||||
// This test aims to prove that tracking some ID multiple times doesn't
|
||||
// result in multiple requests for the same ID
|
||||
|
||||
// associate the foo endpoints with some workloads 3 times without changing
|
||||
// the selection criteria. The second two times should be no-ops
|
||||
suite.tracker.TrackIDForSelector(suite.endpointsFoo, &pbcatalog.WorkloadSelector{
|
||||
Prefixes: []string{"api-"},
|
||||
})
|
||||
|
||||
suite.tracker.TrackIDForSelector(suite.endpointsFoo, &pbcatalog.WorkloadSelector{
|
||||
Prefixes: []string{"api-"},
|
||||
})
|
||||
|
||||
suite.tracker.TrackIDForSelector(suite.endpointsFoo, &pbcatalog.WorkloadSelector{
|
||||
Prefixes: []string{"api-"},
|
||||
})
|
||||
|
||||
// regardless of the number of times tracked we should only see a single request
|
||||
suite.requireMappedIDs(suite.workloadAPI1, suite.endpointsFoo)
|
||||
}
|
||||
|
||||
func (suite *selectionTrackerSuite) TestModifyTracking() {
|
||||
// This test aims to prove that modifying selection criteria for a resource
|
||||
// works as expected. Adding new criteria results in all being tracked.
|
||||
// Removal of some criteria does't result in removal of all etc. More or
|
||||
// less we want to ensure that updating selection criteria leaves the
|
||||
// tracker in a consistent/expected state.
|
||||
|
||||
// track the web-1 workload
|
||||
suite.tracker.TrackIDForSelector(suite.endpointsFoo, &pbcatalog.WorkloadSelector{
|
||||
Names: []string{"web-1"},
|
||||
})
|
||||
|
||||
// ensure that api-1 isn't mapped but web-1 is
|
||||
suite.requireMappedIDs(suite.workloadAPI1)
|
||||
suite.requireMappedIDs(suite.workloadWeb1, suite.endpointsFoo)
|
||||
|
||||
// now also track the api- prefix
|
||||
suite.tracker.TrackIDForSelector(suite.endpointsFoo, &pbcatalog.WorkloadSelector{
|
||||
Names: []string{"web-1"},
|
||||
Prefixes: []string{"api-"},
|
||||
})
|
||||
|
||||
// ensure that both workloads are mapped appropriately
|
||||
suite.requireMappedIDs(suite.workloadAPI1, suite.endpointsFoo)
|
||||
suite.requireMappedIDs(suite.workloadWeb1, suite.endpointsFoo)
|
||||
|
||||
// now remove the web tracking
|
||||
suite.tracker.TrackIDForSelector(suite.endpointsFoo, &pbcatalog.WorkloadSelector{
|
||||
Prefixes: []string{"api-"},
|
||||
})
|
||||
|
||||
// ensure that only api-1 is mapped
|
||||
suite.requireMappedIDs(suite.workloadAPI1, suite.endpointsFoo)
|
||||
suite.requireMappedIDs(suite.workloadWeb1)
|
||||
}
|
||||
|
||||
func (suite *selectionTrackerSuite) TestRemove() {
|
||||
// This test aims to prove that removal of a resource from tracking
|
||||
// actually prevents subsequent mapping calls from returning the
|
||||
// workload.
|
||||
|
||||
// track the web-1 workload
|
||||
suite.tracker.TrackIDForSelector(suite.endpointsFoo, &pbcatalog.WorkloadSelector{
|
||||
Names: []string{"web-1"},
|
||||
})
|
||||
|
||||
// ensure that api-1 isn't mapped but web-1 is
|
||||
suite.requireMappedIDs(suite.workloadWeb1, suite.endpointsFoo)
|
||||
|
||||
// untrack the resource
|
||||
suite.tracker.UntrackID(suite.endpointsFoo)
|
||||
|
||||
// ensure that we no longer map the previous workload to the resource
|
||||
suite.requireMappedIDs(suite.workloadWeb1)
|
||||
}
|
||||
|
||||
func TestWorkloadSelectionSuite(t *testing.T) {
|
||||
suite.Run(t, new(selectionTrackerSuite))
|
||||
}
|
|
@ -60,3 +60,12 @@ type errTooMuchMesh struct {
|
|||
func (err errTooMuchMesh) Error() string {
|
||||
return fmt.Sprintf("protocol \"mesh\" was specified in more than 1 port: %+v", err.Ports)
|
||||
}
|
||||
|
||||
type errInvalidEndpointsOwnerName struct {
|
||||
Name string
|
||||
OwnerName string
|
||||
}
|
||||
|
||||
func (err errInvalidEndpointsOwnerName) Error() string {
|
||||
return fmt.Sprintf("invalid owner name %q. ServiceEndpoints objects must be owned by a Service with the same name: %q", err.OwnerName, err.Name)
|
||||
}
|
||||
|
|
|
@ -52,6 +52,10 @@ func TestErrorStrings(t *testing.T) {
|
|||
"errTooMuchMesh": errTooMuchMesh{
|
||||
Ports: []string{"http", "grpc"},
|
||||
},
|
||||
"errInvalidEndpointsOwnerName": errInvalidEndpointsOwnerName{
|
||||
Name: "foo",
|
||||
OwnerName: "bar",
|
||||
},
|
||||
"errNotDNSLabel": errNotDNSLabel,
|
||||
"errNotIPAddress": errNotIPAddress,
|
||||
"errUnixSocketMultiport": errUnixSocketMultiport,
|
||||
|
|
|
@ -73,7 +73,7 @@ func ValidateHealthStatus(res *pbresource.Resource) error {
|
|||
Wrapped: resource.ErrMissing,
|
||||
})
|
||||
} else if !resource.EqualType(res.Owner.Type, WorkloadType) && !resource.EqualType(res.Owner.Type, NodeType) {
|
||||
err = multierror.Append(err, resource.ErrOwnerInvalid{ResourceType: res.Id.Type, OwnerType: res.Owner.Type})
|
||||
err = multierror.Append(err, resource.ErrOwnerTypeInvalid{ResourceType: res.Id.Type, OwnerType: res.Owner.Type})
|
||||
}
|
||||
|
||||
return err
|
||||
|
|
|
@ -203,11 +203,11 @@ func TestValidateHealthStatus_InvalidOwner(t *testing.T) {
|
|||
res := createHealthStatusResource(t, data, tcase.owner)
|
||||
err := ValidateHealthStatus(res)
|
||||
require.Error(t, err)
|
||||
expected := resource.ErrOwnerInvalid{
|
||||
expected := resource.ErrOwnerTypeInvalid{
|
||||
ResourceType: HealthStatusType,
|
||||
OwnerType: tcase.owner.Type,
|
||||
}
|
||||
var actual resource.ErrOwnerInvalid
|
||||
var actual resource.ErrOwnerTypeInvalid
|
||||
require.ErrorAs(t, err, &actual)
|
||||
require.Equal(t, expected, actual)
|
||||
})
|
||||
|
|
|
@ -71,7 +71,7 @@ func ValidateService(res *pbresource.Resource) error {
|
|||
},
|
||||
},
|
||||
})
|
||||
} else {
|
||||
} else if port.VirtualPort != 0 {
|
||||
usedVirtualPorts[port.VirtualPort] = idx
|
||||
}
|
||||
|
||||
|
|
|
@ -30,10 +30,48 @@ func RegisterServiceEndpoints(r resource.Registry) {
|
|||
r.Register(resource.Registration{
|
||||
Type: ServiceEndpointsV1Alpha1Type,
|
||||
Proto: &pbcatalog.ServiceEndpoints{},
|
||||
Validate: nil,
|
||||
Validate: ValidateServiceEndpoints,
|
||||
Mutate: MutateServiceEndpoints,
|
||||
})
|
||||
}
|
||||
|
||||
func MutateServiceEndpoints(res *pbresource.Resource) error {
|
||||
if res.Owner == nil {
|
||||
res.Owner = &pbresource.ID{
|
||||
Type: ServiceV1Alpha1Type,
|
||||
Tenancy: res.Id.Tenancy,
|
||||
Name: res.Id.Name,
|
||||
}
|
||||
}
|
||||
|
||||
var err error
|
||||
if !resource.EqualType(res.Owner.Type, ServiceV1Alpha1Type) {
|
||||
err = multierror.Append(err, resource.ErrOwnerTypeInvalid{
|
||||
ResourceType: ServiceEndpointsV1Alpha1Type,
|
||||
OwnerType: res.Owner.Type,
|
||||
})
|
||||
}
|
||||
|
||||
if !resource.EqualTenancy(res.Owner.Tenancy, res.Id.Tenancy) {
|
||||
err = multierror.Append(err, resource.ErrOwnerTenantInvalid{
|
||||
ResourceTenancy: res.Id.Tenancy,
|
||||
OwnerTenancy: res.Owner.Tenancy,
|
||||
})
|
||||
}
|
||||
|
||||
if res.Owner.Name != res.Id.Name {
|
||||
err = multierror.Append(err, resource.ErrInvalidField{
|
||||
Name: "name",
|
||||
Wrapped: errInvalidEndpointsOwnerName{
|
||||
Name: res.Id.Name,
|
||||
OwnerName: res.Owner.Name,
|
||||
},
|
||||
})
|
||||
}
|
||||
|
||||
return err
|
||||
}
|
||||
|
||||
func ValidateServiceEndpoints(res *pbresource.Resource) error {
|
||||
var svcEndpoints pbcatalog.ServiceEndpoints
|
||||
|
||||
|
|
|
@ -45,6 +45,18 @@ func TestValidateService_Ok(t *testing.T) {
|
|||
VirtualPort: 42,
|
||||
Protocol: pbcatalog.Protocol_PROTOCOL_HTTP,
|
||||
},
|
||||
{
|
||||
TargetPort: "other",
|
||||
// leaving VirtualPort unset to verify that seeing
|
||||
// a zero virtual port multiple times is fine.
|
||||
Protocol: pbcatalog.Protocol_PROTOCOL_HTTP2,
|
||||
},
|
||||
{
|
||||
TargetPort: "other2",
|
||||
// leaving VirtualPort unset to verify that seeing
|
||||
// a zero virtual port multiple times is fine.
|
||||
Protocol: pbcatalog.Protocol_PROTOCOL_GRPC,
|
||||
},
|
||||
},
|
||||
VirtualIps: []string{"198.18.0.1"},
|
||||
}
|
||||
|
|
1
internal/catalog/internal/types/testdata/errInvalidEndpointsOwnerName.golden
vendored
Normal file
1
internal/catalog/internal/types/testdata/errInvalidEndpointsOwnerName.golden
vendored
Normal file
|
@ -0,0 +1 @@
|
|||
invalid owner name "bar". ServiceEndpoints objects must be owned by a Service with the same name: "foo"
|
|
@ -152,46 +152,6 @@ type Reconciler interface {
|
|||
Reconcile(ctx context.Context, rt Runtime, req Request) error
|
||||
}
|
||||
|
||||
// DependencyMapper is called when a dependency watched via WithWatch is changed
|
||||
// to determine which of the controller's managed resources need to be reconciled.
|
||||
type DependencyMapper func(
|
||||
ctx context.Context,
|
||||
rt Runtime,
|
||||
res *pbresource.Resource,
|
||||
) ([]Request, error)
|
||||
|
||||
// MapOwner implements a DependencyMapper that returns the updated resource's owner.
|
||||
func MapOwner(_ context.Context, _ Runtime, res *pbresource.Resource) ([]Request, error) {
|
||||
var reqs []Request
|
||||
if res.Owner != nil {
|
||||
reqs = append(reqs, Request{ID: res.Owner})
|
||||
}
|
||||
return reqs, nil
|
||||
}
|
||||
|
||||
func MapOwnerFiltered(filter *pbresource.Type) DependencyMapper {
|
||||
return func(_ context.Context, _ Runtime, res *pbresource.Resource) ([]Request, error) {
|
||||
if res.Owner == nil {
|
||||
return nil, nil
|
||||
}
|
||||
|
||||
ownerType := res.Owner.GetType()
|
||||
if ownerType.Group != filter.Group {
|
||||
return nil, nil
|
||||
}
|
||||
|
||||
if ownerType.GroupVersion != filter.GroupVersion {
|
||||
return nil, nil
|
||||
}
|
||||
|
||||
if ownerType.Kind != filter.Kind {
|
||||
return nil, nil
|
||||
}
|
||||
|
||||
return []Request{{ID: res.Owner}}, nil
|
||||
}
|
||||
}
|
||||
|
||||
// Placement determines where and how many replicas of the controller will run.
|
||||
type Placement int
|
||||
|
||||
|
|
|
@ -9,7 +9,6 @@ import (
|
|||
"testing"
|
||||
"time"
|
||||
|
||||
"github.com/google/go-cmp/cmp/cmpopts"
|
||||
"github.com/stretchr/testify/require"
|
||||
|
||||
svctest "github.com/hashicorp/consul/agent/grpc-external/services/resource/testing"
|
||||
|
@ -267,76 +266,3 @@ func testContext(t *testing.T) context.Context {
|
|||
|
||||
return ctx
|
||||
}
|
||||
|
||||
func resourceID(group string, version string, kind string, name string) *pbresource.ID {
|
||||
return &pbresource.ID{
|
||||
Type: &pbresource.Type{
|
||||
Group: group,
|
||||
GroupVersion: version,
|
||||
Kind: kind,
|
||||
},
|
||||
Tenancy: &pbresource.Tenancy{
|
||||
Partition: "default",
|
||||
Namespace: "default",
|
||||
PeerName: "local",
|
||||
},
|
||||
Name: name,
|
||||
}
|
||||
}
|
||||
|
||||
func TestMapOwnerFiltered(t *testing.T) {
|
||||
mapper := controller.MapOwnerFiltered(&pbresource.Type{
|
||||
Group: "foo",
|
||||
GroupVersion: "v1",
|
||||
Kind: "bar",
|
||||
})
|
||||
|
||||
type testCase struct {
|
||||
owner *pbresource.ID
|
||||
matches bool
|
||||
}
|
||||
|
||||
cases := map[string]testCase{
|
||||
"nil-owner": {
|
||||
owner: nil,
|
||||
matches: false,
|
||||
},
|
||||
"group-mismatch": {
|
||||
owner: resourceID("other", "v1", "bar", "irrelevant"),
|
||||
matches: false,
|
||||
},
|
||||
"group-version-mismatch": {
|
||||
owner: resourceID("foo", "v2", "bar", "irrelevant"),
|
||||
matches: false,
|
||||
},
|
||||
"kind-mismatch": {
|
||||
owner: resourceID("foo", "v1", "baz", "irrelevant"),
|
||||
matches: false,
|
||||
},
|
||||
"match": {
|
||||
owner: resourceID("foo", "v1", "bar", "irrelevant"),
|
||||
matches: true,
|
||||
},
|
||||
}
|
||||
|
||||
for name, tcase := range cases {
|
||||
t.Run(name, func(t *testing.T) {
|
||||
// the runtime is not used by the mapper so its fine to pass an empty struct
|
||||
req, err := mapper(context.Background(), controller.Runtime{}, &pbresource.Resource{
|
||||
Id: resourceID("foo", "v1", "other", "x"),
|
||||
Owner: tcase.owner,
|
||||
})
|
||||
|
||||
// The mapper has no error paths at present
|
||||
require.NoError(t, err)
|
||||
|
||||
if tcase.matches {
|
||||
require.NotNil(t, req)
|
||||
require.Len(t, req, 1)
|
||||
prototest.AssertDeepEqual(t, req[0].ID, tcase.owner, cmpopts.EquateEmpty())
|
||||
} else {
|
||||
require.Nil(t, req)
|
||||
}
|
||||
})
|
||||
}
|
||||
}
|
||||
|
|
58
internal/controller/dependency_mappers.go
Normal file
58
internal/controller/dependency_mappers.go
Normal file
|
@ -0,0 +1,58 @@
|
|||
package controller
|
||||
|
||||
import (
|
||||
"context"
|
||||
|
||||
"github.com/hashicorp/consul/internal/resource"
|
||||
"github.com/hashicorp/consul/proto-public/pbresource"
|
||||
)
|
||||
|
||||
// DependencyMapper is called when a dependency watched via WithWatch is changed
|
||||
// to determine which of the controller's managed resources need to be reconciled.
|
||||
type DependencyMapper func(
|
||||
ctx context.Context,
|
||||
rt Runtime,
|
||||
res *pbresource.Resource,
|
||||
) ([]Request, error)
|
||||
|
||||
// MapOwner implements a DependencyMapper that returns the updated resource's owner.
|
||||
func MapOwner(_ context.Context, _ Runtime, res *pbresource.Resource) ([]Request, error) {
|
||||
var reqs []Request
|
||||
if res.Owner != nil {
|
||||
reqs = append(reqs, Request{ID: res.Owner})
|
||||
}
|
||||
return reqs, nil
|
||||
}
|
||||
|
||||
// MapOwnerFiltered creates a DependencyMapper that returns owner IDs as Requests
|
||||
// if the type of the owner ID matches the given filter type.
|
||||
func MapOwnerFiltered(filter *pbresource.Type) DependencyMapper {
|
||||
return func(_ context.Context, _ Runtime, res *pbresource.Resource) ([]Request, error) {
|
||||
if res.Owner == nil {
|
||||
return nil, nil
|
||||
}
|
||||
|
||||
if !resource.EqualType(res.Owner.GetType(), filter) {
|
||||
return nil, nil
|
||||
}
|
||||
|
||||
return []Request{{ID: res.Owner}}, nil
|
||||
}
|
||||
}
|
||||
|
||||
// ReplaceType creates a DependencyMapper that returns request IDs with the same
|
||||
// name and tenancy as the original resource but with the type replaced with
|
||||
// the type specified as this functions parameter.
|
||||
func ReplaceType(desiredType *pbresource.Type) DependencyMapper {
|
||||
return func(_ context.Context, _ Runtime, res *pbresource.Resource) ([]Request, error) {
|
||||
return []Request{
|
||||
{
|
||||
ID: &pbresource.ID{
|
||||
Type: desiredType,
|
||||
Tenancy: res.Id.Tenancy,
|
||||
Name: res.Id.Name,
|
||||
},
|
||||
},
|
||||
}, nil
|
||||
}
|
||||
}
|
137
internal/controller/dependency_mappers_test.go
Normal file
137
internal/controller/dependency_mappers_test.go
Normal file
|
@ -0,0 +1,137 @@
|
|||
package controller
|
||||
|
||||
import (
|
||||
"context"
|
||||
"testing"
|
||||
|
||||
"github.com/google/go-cmp/cmp/cmpopts"
|
||||
"github.com/hashicorp/consul/proto-public/pbresource"
|
||||
"github.com/hashicorp/consul/proto/private/prototest"
|
||||
"github.com/stretchr/testify/require"
|
||||
)
|
||||
|
||||
func resourceID(group string, version string, kind string, name string) *pbresource.ID {
|
||||
return &pbresource.ID{
|
||||
Type: &pbresource.Type{
|
||||
Group: group,
|
||||
GroupVersion: version,
|
||||
Kind: kind,
|
||||
},
|
||||
Tenancy: &pbresource.Tenancy{
|
||||
Partition: "default",
|
||||
Namespace: "default",
|
||||
PeerName: "local",
|
||||
},
|
||||
Name: name,
|
||||
}
|
||||
}
|
||||
|
||||
func TestMapOwner(t *testing.T) {
|
||||
owner := resourceID("foo", "v99", "bar", "object")
|
||||
|
||||
res := &pbresource.Resource{
|
||||
Id: resourceID("something", "v1", "else", "x"),
|
||||
Owner: owner,
|
||||
}
|
||||
|
||||
reqs, err := MapOwner(context.Background(), Runtime{}, res)
|
||||
require.NoError(t, err)
|
||||
require.Len(t, reqs, 1)
|
||||
prototest.AssertDeepEqual(t, owner, reqs[0].ID)
|
||||
}
|
||||
|
||||
func TestMapOwnerFiltered(t *testing.T) {
|
||||
mapper := MapOwnerFiltered(&pbresource.Type{
|
||||
Group: "foo",
|
||||
GroupVersion: "v1",
|
||||
Kind: "bar",
|
||||
})
|
||||
|
||||
type testCase struct {
|
||||
owner *pbresource.ID
|
||||
matches bool
|
||||
}
|
||||
|
||||
cases := map[string]testCase{
|
||||
"nil-owner": {
|
||||
owner: nil,
|
||||
matches: false,
|
||||
},
|
||||
"group-mismatch": {
|
||||
owner: resourceID("other", "v1", "bar", "irrelevant"),
|
||||
matches: false,
|
||||
},
|
||||
"group-version-mismatch": {
|
||||
owner: resourceID("foo", "v2", "bar", "irrelevant"),
|
||||
matches: false,
|
||||
},
|
||||
"kind-mismatch": {
|
||||
owner: resourceID("foo", "v1", "baz", "irrelevant"),
|
||||
matches: false,
|
||||
},
|
||||
"match": {
|
||||
owner: resourceID("foo", "v1", "bar", "irrelevant"),
|
||||
matches: true,
|
||||
},
|
||||
}
|
||||
|
||||
for name, tcase := range cases {
|
||||
t.Run(name, func(t *testing.T) {
|
||||
// the runtime is not used by the mapper so its fine to pass an empty struct
|
||||
req, err := mapper(context.Background(), Runtime{}, &pbresource.Resource{
|
||||
Id: resourceID("foo", "v1", "other", "x"),
|
||||
Owner: tcase.owner,
|
||||
})
|
||||
|
||||
// The mapper has no error paths at present
|
||||
require.NoError(t, err)
|
||||
|
||||
if tcase.matches {
|
||||
require.NotNil(t, req)
|
||||
require.Len(t, req, 1)
|
||||
prototest.AssertDeepEqual(t, req[0].ID, tcase.owner, cmpopts.EquateEmpty())
|
||||
} else {
|
||||
require.Nil(t, req)
|
||||
}
|
||||
})
|
||||
}
|
||||
}
|
||||
|
||||
func TestReplaceType(t *testing.T) {
|
||||
rtype := &pbresource.Type{
|
||||
Group: "foo",
|
||||
GroupVersion: "v1",
|
||||
Kind: "bar",
|
||||
}
|
||||
|
||||
tenant := &pbresource.Tenancy{
|
||||
Partition: "not",
|
||||
Namespace: "using",
|
||||
PeerName: "the-defaults",
|
||||
}
|
||||
|
||||
in := &pbresource.Resource{
|
||||
Id: &pbresource.ID{
|
||||
Type: &pbresource.Type{
|
||||
Group: "other",
|
||||
GroupVersion: "v2",
|
||||
Kind: "baz",
|
||||
},
|
||||
Tenancy: tenant,
|
||||
Name: "arr-matey",
|
||||
},
|
||||
}
|
||||
|
||||
mapper := ReplaceType(rtype)
|
||||
|
||||
reqs, err := mapper(nil, Runtime{}, in)
|
||||
require.NoError(t, err)
|
||||
require.Len(t, reqs, 1)
|
||||
|
||||
expected := &pbresource.ID{
|
||||
Type: rtype,
|
||||
Tenancy: tenant,
|
||||
Name: "arr-matey",
|
||||
}
|
||||
prototest.AssertDeepEqual(t, expected, reqs[0].ID)
|
||||
}
|
4
internal/radix/doc.go
Normal file
4
internal/radix/doc.go
Normal file
|
@ -0,0 +1,4 @@
|
|||
// This packages contents were originally copied from github.com/armon/go-radix.
|
||||
// After the intial copy all the data structures were made to use Go 1.18 generics
|
||||
// instead of relying on the use of interface{} or the any type.
|
||||
package radix
|
568
internal/radix/radix.go
Normal file
568
internal/radix/radix.go
Normal file
|
@ -0,0 +1,568 @@
|
|||
package radix
|
||||
|
||||
import (
|
||||
"sort"
|
||||
"strings"
|
||||
)
|
||||
|
||||
// WalkFn is used when walking the tree. Takes a
|
||||
// key and value, returning if iteration should
|
||||
// be terminated.
|
||||
type WalkFn[T any] func(s string, v T) bool
|
||||
|
||||
// leafNode is used to represent a value
|
||||
type leafNode[T any] struct {
|
||||
key string
|
||||
val T
|
||||
}
|
||||
|
||||
// edge is used to represent an edge node
|
||||
type edge[T any] struct {
|
||||
label byte
|
||||
node *node[T]
|
||||
}
|
||||
|
||||
type node[T any] struct {
|
||||
// leaf is used to store possible leaf
|
||||
leaf *leafNode[T]
|
||||
|
||||
// prefix is the common prefix we ignore
|
||||
prefix string
|
||||
|
||||
// Edges should be stored in-order for iteration.
|
||||
// We avoid a fully materialized slice to save memory,
|
||||
// since in most cases we expect to be sparse
|
||||
edges edges[T]
|
||||
}
|
||||
|
||||
func (n *node[T]) isLeaf() bool {
|
||||
return n.leaf != nil
|
||||
}
|
||||
|
||||
func (n *node[T]) addEdge(e edge[T]) {
|
||||
num := len(n.edges)
|
||||
idx := sort.Search(num, func(i int) bool {
|
||||
return n.edges[i].label >= e.label
|
||||
})
|
||||
|
||||
n.edges = append(n.edges, edge[T]{})
|
||||
copy(n.edges[idx+1:], n.edges[idx:])
|
||||
n.edges[idx] = e
|
||||
}
|
||||
|
||||
func (n *node[T]) updateEdge(label byte, node *node[T]) {
|
||||
num := len(n.edges)
|
||||
idx := sort.Search(num, func(i int) bool {
|
||||
return n.edges[i].label >= label
|
||||
})
|
||||
if idx < num && n.edges[idx].label == label {
|
||||
n.edges[idx].node = node
|
||||
return
|
||||
}
|
||||
panic("replacing missing edge")
|
||||
}
|
||||
|
||||
func (n *node[T]) getEdge(label byte) *node[T] {
|
||||
num := len(n.edges)
|
||||
idx := sort.Search(num, func(i int) bool {
|
||||
return n.edges[i].label >= label
|
||||
})
|
||||
if idx < num && n.edges[idx].label == label {
|
||||
return n.edges[idx].node
|
||||
}
|
||||
return nil
|
||||
}
|
||||
|
||||
func (n *node[T]) delEdge(label byte) {
|
||||
num := len(n.edges)
|
||||
idx := sort.Search(num, func(i int) bool {
|
||||
return n.edges[i].label >= label
|
||||
})
|
||||
if idx < num && n.edges[idx].label == label {
|
||||
copy(n.edges[idx:], n.edges[idx+1:])
|
||||
n.edges[len(n.edges)-1] = edge[T]{}
|
||||
n.edges = n.edges[:len(n.edges)-1]
|
||||
}
|
||||
}
|
||||
|
||||
type edges[T any] []edge[T]
|
||||
|
||||
func (e edges[T]) Len() int {
|
||||
return len(e)
|
||||
}
|
||||
|
||||
func (e edges[T]) Less(i, j int) bool {
|
||||
return e[i].label < e[j].label
|
||||
}
|
||||
|
||||
func (e edges[T]) Swap(i, j int) {
|
||||
e[i], e[j] = e[j], e[i]
|
||||
}
|
||||
|
||||
func (e edges[T]) Sort() {
|
||||
sort.Sort(e)
|
||||
}
|
||||
|
||||
// Tree implements a radix tree. This can be treated as a
|
||||
// Dictionary abstract data type. The main advantage over
|
||||
// a standard hash map is prefix-based lookups and
|
||||
// ordered iteration,
|
||||
type Tree[T any] struct {
|
||||
root *node[T]
|
||||
size int
|
||||
}
|
||||
|
||||
// New returns an empty Tree
|
||||
func New[T any]() *Tree[T] {
|
||||
return NewFromMap[T](nil)
|
||||
}
|
||||
|
||||
// NewFromMap returns a new tree containing the keys
|
||||
// from an existing map
|
||||
func NewFromMap[T any](m map[string]T) *Tree[T] {
|
||||
t := &Tree[T]{root: &node[T]{}}
|
||||
for k, v := range m {
|
||||
t.Insert(k, v)
|
||||
}
|
||||
return t
|
||||
}
|
||||
|
||||
// Len is used to return the number of elements in the tree
|
||||
func (t *Tree[T]) Len() int {
|
||||
return t.size
|
||||
}
|
||||
|
||||
// longestPrefix finds the length of the shared prefix
|
||||
// of two strings
|
||||
func longestPrefix(k1, k2 string) int {
|
||||
max := len(k1)
|
||||
if l := len(k2); l < max {
|
||||
max = l
|
||||
}
|
||||
var i int
|
||||
for i = 0; i < max; i++ {
|
||||
if k1[i] != k2[i] {
|
||||
break
|
||||
}
|
||||
}
|
||||
return i
|
||||
}
|
||||
|
||||
// Insert is used to add a newentry or update
|
||||
// an existing entry. Returns true if an existing record is updated.
|
||||
func (t *Tree[T]) Insert(s string, v T) (T, bool) {
|
||||
var zeroVal T
|
||||
var parent *node[T]
|
||||
n := t.root
|
||||
search := s
|
||||
for {
|
||||
// Handle key exhaution
|
||||
if len(search) == 0 {
|
||||
if n.isLeaf() {
|
||||
old := n.leaf.val
|
||||
n.leaf.val = v
|
||||
return old, true
|
||||
}
|
||||
|
||||
n.leaf = &leafNode[T]{
|
||||
key: s,
|
||||
val: v,
|
||||
}
|
||||
t.size++
|
||||
return zeroVal, false
|
||||
}
|
||||
|
||||
// Look for the edge
|
||||
parent = n
|
||||
n = n.getEdge(search[0])
|
||||
|
||||
// No edge, create one
|
||||
if n == nil {
|
||||
e := edge[T]{
|
||||
label: search[0],
|
||||
node: &node[T]{
|
||||
leaf: &leafNode[T]{
|
||||
key: s,
|
||||
val: v,
|
||||
},
|
||||
prefix: search,
|
||||
},
|
||||
}
|
||||
parent.addEdge(e)
|
||||
t.size++
|
||||
return zeroVal, false
|
||||
}
|
||||
|
||||
// Determine longest prefix of the search key on match
|
||||
commonPrefix := longestPrefix(search, n.prefix)
|
||||
if commonPrefix == len(n.prefix) {
|
||||
search = search[commonPrefix:]
|
||||
continue
|
||||
}
|
||||
|
||||
// Split the node
|
||||
t.size++
|
||||
child := &node[T]{
|
||||
prefix: search[:commonPrefix],
|
||||
}
|
||||
parent.updateEdge(search[0], child)
|
||||
|
||||
// Restore the existing node
|
||||
child.addEdge(edge[T]{
|
||||
label: n.prefix[commonPrefix],
|
||||
node: n,
|
||||
})
|
||||
n.prefix = n.prefix[commonPrefix:]
|
||||
|
||||
// Create a new leaf node
|
||||
leaf := &leafNode[T]{
|
||||
key: s,
|
||||
val: v,
|
||||
}
|
||||
|
||||
// If the new key is a subset, add to this node
|
||||
search = search[commonPrefix:]
|
||||
if len(search) == 0 {
|
||||
child.leaf = leaf
|
||||
return zeroVal, false
|
||||
}
|
||||
|
||||
// Create a new edge for the node
|
||||
child.addEdge(edge[T]{
|
||||
label: search[0],
|
||||
node: &node[T]{
|
||||
leaf: leaf,
|
||||
prefix: search,
|
||||
},
|
||||
})
|
||||
return zeroVal, false
|
||||
}
|
||||
}
|
||||
|
||||
// Delete is used to delete a key, returning the previous
|
||||
// value and if it was deleted
|
||||
func (t *Tree[T]) Delete(s string) (T, bool) {
|
||||
var zeroVal T
|
||||
|
||||
var parent *node[T]
|
||||
var label byte
|
||||
n := t.root
|
||||
search := s
|
||||
for {
|
||||
// Check for key exhaution
|
||||
if len(search) == 0 {
|
||||
if !n.isLeaf() {
|
||||
break
|
||||
}
|
||||
goto DELETE
|
||||
}
|
||||
|
||||
// Look for an edge
|
||||
parent = n
|
||||
label = search[0]
|
||||
n = n.getEdge(label)
|
||||
if n == nil {
|
||||
break
|
||||
}
|
||||
|
||||
// Consume the search prefix
|
||||
if strings.HasPrefix(search, n.prefix) {
|
||||
search = search[len(n.prefix):]
|
||||
} else {
|
||||
break
|
||||
}
|
||||
}
|
||||
return zeroVal, false
|
||||
|
||||
DELETE:
|
||||
// Delete the leaf
|
||||
leaf := n.leaf
|
||||
n.leaf = nil
|
||||
t.size--
|
||||
|
||||
// Check if we should delete this node from the parent
|
||||
if parent != nil && len(n.edges) == 0 {
|
||||
parent.delEdge(label)
|
||||
}
|
||||
|
||||
// Check if we should merge this node
|
||||
if n != t.root && len(n.edges) == 1 {
|
||||
n.mergeChild()
|
||||
}
|
||||
|
||||
// Check if we should merge the parent's other child
|
||||
if parent != nil && parent != t.root && len(parent.edges) == 1 && !parent.isLeaf() {
|
||||
parent.mergeChild()
|
||||
}
|
||||
|
||||
return leaf.val, true
|
||||
}
|
||||
|
||||
// DeletePrefix is used to delete the subtree under a prefix
|
||||
// Returns how many nodes were deleted
|
||||
// Use this to delete large subtrees efficiently
|
||||
func (t *Tree[T]) DeletePrefix(s string) int {
|
||||
return t.deletePrefix(nil, t.root, s)
|
||||
}
|
||||
|
||||
// delete does a recursive deletion
|
||||
func (t *Tree[T]) deletePrefix(parent, n *node[T], prefix string) int {
|
||||
// Check for key exhaustion
|
||||
if len(prefix) == 0 {
|
||||
// Remove the leaf node
|
||||
subTreeSize := 0
|
||||
//recursively walk from all edges of the node to be deleted
|
||||
recursiveWalk(n, func(s string, v T) bool {
|
||||
subTreeSize++
|
||||
return false
|
||||
})
|
||||
if n.isLeaf() {
|
||||
n.leaf = nil
|
||||
}
|
||||
n.edges = nil // deletes the entire subtree
|
||||
|
||||
// Check if we should merge the parent's other child
|
||||
if parent != nil && parent != t.root && len(parent.edges) == 1 && !parent.isLeaf() {
|
||||
parent.mergeChild()
|
||||
}
|
||||
t.size -= subTreeSize
|
||||
return subTreeSize
|
||||
}
|
||||
|
||||
// Look for an edge
|
||||
label := prefix[0]
|
||||
child := n.getEdge(label)
|
||||
if child == nil || (!strings.HasPrefix(child.prefix, prefix) && !strings.HasPrefix(prefix, child.prefix)) {
|
||||
return 0
|
||||
}
|
||||
|
||||
// Consume the search prefix
|
||||
if len(child.prefix) > len(prefix) {
|
||||
prefix = prefix[len(prefix):]
|
||||
} else {
|
||||
prefix = prefix[len(child.prefix):]
|
||||
}
|
||||
return t.deletePrefix(n, child, prefix)
|
||||
}
|
||||
|
||||
func (n *node[T]) mergeChild() {
|
||||
e := n.edges[0]
|
||||
child := e.node
|
||||
n.prefix = n.prefix + child.prefix
|
||||
n.leaf = child.leaf
|
||||
n.edges = child.edges
|
||||
}
|
||||
|
||||
// Get is used to lookup a specific key, returning
|
||||
// the value and if it was found
|
||||
func (t *Tree[T]) Get(s string) (T, bool) {
|
||||
var zeroVal T
|
||||
n := t.root
|
||||
search := s
|
||||
for {
|
||||
// Check for key exhaution
|
||||
if len(search) == 0 {
|
||||
if n.isLeaf() {
|
||||
return n.leaf.val, true
|
||||
}
|
||||
break
|
||||
}
|
||||
|
||||
// Look for an edge
|
||||
n = n.getEdge(search[0])
|
||||
if n == nil {
|
||||
break
|
||||
}
|
||||
|
||||
// Consume the search prefix
|
||||
if strings.HasPrefix(search, n.prefix) {
|
||||
search = search[len(n.prefix):]
|
||||
} else {
|
||||
break
|
||||
}
|
||||
}
|
||||
return zeroVal, false
|
||||
}
|
||||
|
||||
// LongestPrefix is like Get, but instead of an
|
||||
// exact match, it will return the longest prefix match.
|
||||
func (t *Tree[T]) LongestPrefix(s string) (string, T, bool) {
|
||||
var zeroVal T
|
||||
var last *leafNode[T]
|
||||
n := t.root
|
||||
search := s
|
||||
for {
|
||||
// Look for a leaf node
|
||||
if n.isLeaf() {
|
||||
last = n.leaf
|
||||
}
|
||||
|
||||
// Check for key exhaution
|
||||
if len(search) == 0 {
|
||||
break
|
||||
}
|
||||
|
||||
// Look for an edge
|
||||
n = n.getEdge(search[0])
|
||||
if n == nil {
|
||||
break
|
||||
}
|
||||
|
||||
// Consume the search prefix
|
||||
if strings.HasPrefix(search, n.prefix) {
|
||||
search = search[len(n.prefix):]
|
||||
} else {
|
||||
break
|
||||
}
|
||||
}
|
||||
if last != nil {
|
||||
return last.key, last.val, true
|
||||
}
|
||||
return "", zeroVal, false
|
||||
}
|
||||
|
||||
// Minimum is used to return the minimum value in the tree
|
||||
func (t *Tree[T]) Minimum() (string, T, bool) {
|
||||
var zeroVal T
|
||||
n := t.root
|
||||
for {
|
||||
if n.isLeaf() {
|
||||
return n.leaf.key, n.leaf.val, true
|
||||
}
|
||||
if len(n.edges) > 0 {
|
||||
n = n.edges[0].node
|
||||
} else {
|
||||
break
|
||||
}
|
||||
}
|
||||
return "", zeroVal, false
|
||||
}
|
||||
|
||||
// Maximum is used to return the maximum value in the tree
|
||||
func (t *Tree[T]) Maximum() (string, T, bool) {
|
||||
var zeroVal T
|
||||
n := t.root
|
||||
for {
|
||||
if num := len(n.edges); num > 0 {
|
||||
n = n.edges[num-1].node
|
||||
continue
|
||||
}
|
||||
if n.isLeaf() {
|
||||
return n.leaf.key, n.leaf.val, true
|
||||
}
|
||||
break
|
||||
}
|
||||
return "", zeroVal, false
|
||||
}
|
||||
|
||||
// Walk is used to walk the tree
|
||||
func (t *Tree[T]) Walk(fn WalkFn[T]) {
|
||||
recursiveWalk(t.root, fn)
|
||||
}
|
||||
|
||||
// WalkPrefix is used to walk the tree under a prefix
|
||||
func (t *Tree[T]) WalkPrefix(prefix string, fn WalkFn[T]) {
|
||||
n := t.root
|
||||
search := prefix
|
||||
for {
|
||||
// Check for key exhaustion
|
||||
if len(search) == 0 {
|
||||
recursiveWalk(n, fn)
|
||||
return
|
||||
}
|
||||
|
||||
// Look for an edge
|
||||
n = n.getEdge(search[0])
|
||||
if n == nil {
|
||||
return
|
||||
}
|
||||
|
||||
// Consume the search prefix
|
||||
if strings.HasPrefix(search, n.prefix) {
|
||||
search = search[len(n.prefix):]
|
||||
continue
|
||||
}
|
||||
if strings.HasPrefix(n.prefix, search) {
|
||||
// Child may be under our search prefix
|
||||
recursiveWalk(n, fn)
|
||||
}
|
||||
return
|
||||
}
|
||||
}
|
||||
|
||||
// WalkPath is used to walk the tree, but only visiting nodes
|
||||
// from the root down to a given leaf. Where WalkPrefix walks
|
||||
// all the entries *under* the given prefix, this walks the
|
||||
// entries *above* the given prefix.
|
||||
func (t *Tree[T]) WalkPath(path string, fn WalkFn[T]) {
|
||||
n := t.root
|
||||
search := path
|
||||
for {
|
||||
// Visit the leaf values if any
|
||||
if n.leaf != nil && fn(n.leaf.key, n.leaf.val) {
|
||||
return
|
||||
}
|
||||
|
||||
// Check for key exhaution
|
||||
if len(search) == 0 {
|
||||
return
|
||||
}
|
||||
|
||||
// Look for an edge
|
||||
n = n.getEdge(search[0])
|
||||
if n == nil {
|
||||
return
|
||||
}
|
||||
|
||||
// Consume the search prefix
|
||||
if strings.HasPrefix(search, n.prefix) {
|
||||
search = search[len(n.prefix):]
|
||||
} else {
|
||||
break
|
||||
}
|
||||
}
|
||||
}
|
||||
|
||||
// recursiveWalk is used to do a pre-order walk of a node
|
||||
// recursively. Returns true if the walk should be aborted
|
||||
func recursiveWalk[T any](n *node[T], fn WalkFn[T]) bool {
|
||||
// Visit the leaf values if any
|
||||
if n.leaf != nil && fn(n.leaf.key, n.leaf.val) {
|
||||
return true
|
||||
}
|
||||
|
||||
// Recurse on the children
|
||||
i := 0
|
||||
k := len(n.edges) // keeps track of number of edges in previous iteration
|
||||
for i < k {
|
||||
e := n.edges[i]
|
||||
if recursiveWalk(e.node, fn) {
|
||||
return true
|
||||
}
|
||||
// It is a possibility that the WalkFn modified the node we are
|
||||
// iterating on. If there are no more edges, mergeChild happened,
|
||||
// so the last edge became the current node n, on which we'll
|
||||
// iterate one last time.
|
||||
if len(n.edges) == 0 {
|
||||
return recursiveWalk(n, fn)
|
||||
}
|
||||
// If there are now less edges than in the previous iteration,
|
||||
// then do not increment the loop index, since the current index
|
||||
// points to a new edge. Otherwise, get to the next index.
|
||||
if len(n.edges) >= k {
|
||||
i++
|
||||
}
|
||||
k = len(n.edges)
|
||||
}
|
||||
return false
|
||||
}
|
||||
|
||||
// ToMap is used to walk the tree and convert it into a map
|
||||
func (t *Tree[T]) ToMap() map[string]T {
|
||||
out := make(map[string]T, t.size)
|
||||
t.Walk(func(k string, v T) bool {
|
||||
out[k] = v
|
||||
return false
|
||||
})
|
||||
return out
|
||||
}
|
408
internal/radix/radix_test.go
Normal file
408
internal/radix/radix_test.go
Normal file
|
@ -0,0 +1,408 @@
|
|||
package radix
|
||||
|
||||
import (
|
||||
crand "crypto/rand"
|
||||
"fmt"
|
||||
"reflect"
|
||||
"sort"
|
||||
"strconv"
|
||||
"testing"
|
||||
)
|
||||
|
||||
func TestRadix(t *testing.T) {
|
||||
var min, max string
|
||||
inp := make(map[string]interface{})
|
||||
for i := 0; i < 1000; i++ {
|
||||
gen := generateUUID()
|
||||
inp[gen] = i
|
||||
if gen < min || i == 0 {
|
||||
min = gen
|
||||
}
|
||||
if gen > max || i == 0 {
|
||||
max = gen
|
||||
}
|
||||
}
|
||||
|
||||
r := NewFromMap(inp)
|
||||
if r.Len() != len(inp) {
|
||||
t.Fatalf("bad length: %v %v", r.Len(), len(inp))
|
||||
}
|
||||
|
||||
r.Walk(func(k string, v interface{}) bool {
|
||||
println(k)
|
||||
return false
|
||||
})
|
||||
|
||||
for k, v := range inp {
|
||||
out, ok := r.Get(k)
|
||||
if !ok {
|
||||
t.Fatalf("missing key: %v", k)
|
||||
}
|
||||
if out != v {
|
||||
t.Fatalf("value mis-match: %v %v", out, v)
|
||||
}
|
||||
}
|
||||
|
||||
// Check min and max
|
||||
outMin, _, _ := r.Minimum()
|
||||
if outMin != min {
|
||||
t.Fatalf("bad minimum: %v %v", outMin, min)
|
||||
}
|
||||
outMax, _, _ := r.Maximum()
|
||||
if outMax != max {
|
||||
t.Fatalf("bad maximum: %v %v", outMax, max)
|
||||
}
|
||||
|
||||
for k, v := range inp {
|
||||
out, ok := r.Delete(k)
|
||||
if !ok {
|
||||
t.Fatalf("missing key: %v", k)
|
||||
}
|
||||
if out != v {
|
||||
t.Fatalf("value mis-match: %v %v", out, v)
|
||||
}
|
||||
}
|
||||
if r.Len() != 0 {
|
||||
t.Fatalf("bad length: %v", r.Len())
|
||||
}
|
||||
}
|
||||
|
||||
func TestRoot(t *testing.T) {
|
||||
r := New[bool]()
|
||||
_, ok := r.Delete("")
|
||||
if ok {
|
||||
t.Fatalf("bad")
|
||||
}
|
||||
_, ok = r.Insert("", true)
|
||||
if ok {
|
||||
t.Fatalf("bad")
|
||||
}
|
||||
val, ok := r.Get("")
|
||||
if !ok || val != true {
|
||||
t.Fatalf("bad: %v", val)
|
||||
}
|
||||
val, ok = r.Delete("")
|
||||
if !ok || val != true {
|
||||
t.Fatalf("bad: %v", val)
|
||||
}
|
||||
}
|
||||
|
||||
func TestDelete(t *testing.T) {
|
||||
|
||||
r := New[bool]()
|
||||
|
||||
s := []string{"", "A", "AB"}
|
||||
|
||||
for _, ss := range s {
|
||||
r.Insert(ss, true)
|
||||
}
|
||||
|
||||
for _, ss := range s {
|
||||
_, ok := r.Delete(ss)
|
||||
if !ok {
|
||||
t.Fatalf("bad %q", ss)
|
||||
}
|
||||
}
|
||||
}
|
||||
|
||||
func TestDeletePrefix(t *testing.T) {
|
||||
type exp struct {
|
||||
inp []string
|
||||
prefix string
|
||||
out []string
|
||||
numDeleted int
|
||||
}
|
||||
|
||||
cases := []exp{
|
||||
{[]string{"", "A", "AB", "ABC", "R", "S"}, "A", []string{"", "R", "S"}, 3},
|
||||
{[]string{"", "A", "AB", "ABC", "R", "S"}, "ABC", []string{"", "A", "AB", "R", "S"}, 1},
|
||||
{[]string{"", "A", "AB", "ABC", "R", "S"}, "", []string{}, 6},
|
||||
{[]string{"", "A", "AB", "ABC", "R", "S"}, "S", []string{"", "A", "AB", "ABC", "R"}, 1},
|
||||
{[]string{"", "A", "AB", "ABC", "R", "S"}, "SS", []string{"", "A", "AB", "ABC", "R", "S"}, 0},
|
||||
}
|
||||
|
||||
for _, test := range cases {
|
||||
r := New[bool]()
|
||||
for _, ss := range test.inp {
|
||||
r.Insert(ss, true)
|
||||
}
|
||||
|
||||
deleted := r.DeletePrefix(test.prefix)
|
||||
if deleted != test.numDeleted {
|
||||
t.Fatalf("Bad delete, expected %v to be deleted but got %v", test.numDeleted, deleted)
|
||||
}
|
||||
|
||||
out := []string{}
|
||||
fn := func(s string, v bool) bool {
|
||||
out = append(out, s)
|
||||
return false
|
||||
}
|
||||
r.Walk(fn)
|
||||
|
||||
if !reflect.DeepEqual(out, test.out) {
|
||||
t.Fatalf("mis-match: %v %v", out, test.out)
|
||||
}
|
||||
}
|
||||
}
|
||||
|
||||
func TestLongestPrefix(t *testing.T) {
|
||||
r := New[interface{}]()
|
||||
|
||||
keys := []string{
|
||||
"",
|
||||
"foo",
|
||||
"foobar",
|
||||
"foobarbaz",
|
||||
"foobarbazzip",
|
||||
"foozip",
|
||||
}
|
||||
for _, k := range keys {
|
||||
r.Insert(k, nil)
|
||||
}
|
||||
if r.Len() != len(keys) {
|
||||
t.Fatalf("bad len: %v %v", r.Len(), len(keys))
|
||||
}
|
||||
|
||||
type exp struct {
|
||||
inp string
|
||||
out string
|
||||
}
|
||||
cases := []exp{
|
||||
{"a", ""},
|
||||
{"abc", ""},
|
||||
{"fo", ""},
|
||||
{"foo", "foo"},
|
||||
{"foob", "foo"},
|
||||
{"foobar", "foobar"},
|
||||
{"foobarba", "foobar"},
|
||||
{"foobarbaz", "foobarbaz"},
|
||||
{"foobarbazzi", "foobarbaz"},
|
||||
{"foobarbazzip", "foobarbazzip"},
|
||||
{"foozi", "foo"},
|
||||
{"foozip", "foozip"},
|
||||
{"foozipzap", "foozip"},
|
||||
}
|
||||
for _, test := range cases {
|
||||
m, _, ok := r.LongestPrefix(test.inp)
|
||||
if !ok {
|
||||
t.Fatalf("no match: %v", test)
|
||||
}
|
||||
if m != test.out {
|
||||
t.Fatalf("mis-match: %v %v", m, test)
|
||||
}
|
||||
}
|
||||
}
|
||||
|
||||
func TestWalkPrefix(t *testing.T) {
|
||||
r := New[interface{}]()
|
||||
|
||||
keys := []string{
|
||||
"foobar",
|
||||
"foo/bar/baz",
|
||||
"foo/baz/bar",
|
||||
"foo/zip/zap",
|
||||
"zipzap",
|
||||
}
|
||||
for _, k := range keys {
|
||||
r.Insert(k, nil)
|
||||
}
|
||||
if r.Len() != len(keys) {
|
||||
t.Fatalf("bad len: %v %v", r.Len(), len(keys))
|
||||
}
|
||||
|
||||
type exp struct {
|
||||
inp string
|
||||
out []string
|
||||
}
|
||||
cases := []exp{
|
||||
{
|
||||
"f",
|
||||
[]string{"foobar", "foo/bar/baz", "foo/baz/bar", "foo/zip/zap"},
|
||||
},
|
||||
{
|
||||
"foo",
|
||||
[]string{"foobar", "foo/bar/baz", "foo/baz/bar", "foo/zip/zap"},
|
||||
},
|
||||
{
|
||||
"foob",
|
||||
[]string{"foobar"},
|
||||
},
|
||||
{
|
||||
"foo/",
|
||||
[]string{"foo/bar/baz", "foo/baz/bar", "foo/zip/zap"},
|
||||
},
|
||||
{
|
||||
"foo/b",
|
||||
[]string{"foo/bar/baz", "foo/baz/bar"},
|
||||
},
|
||||
{
|
||||
"foo/ba",
|
||||
[]string{"foo/bar/baz", "foo/baz/bar"},
|
||||
},
|
||||
{
|
||||
"foo/bar",
|
||||
[]string{"foo/bar/baz"},
|
||||
},
|
||||
{
|
||||
"foo/bar/baz",
|
||||
[]string{"foo/bar/baz"},
|
||||
},
|
||||
{
|
||||
"foo/bar/bazoo",
|
||||
[]string{},
|
||||
},
|
||||
{
|
||||
"z",
|
||||
[]string{"zipzap"},
|
||||
},
|
||||
}
|
||||
|
||||
for _, test := range cases {
|
||||
out := []string{}
|
||||
fn := func(s string, v interface{}) bool {
|
||||
out = append(out, s)
|
||||
return false
|
||||
}
|
||||
r.WalkPrefix(test.inp, fn)
|
||||
sort.Strings(out)
|
||||
sort.Strings(test.out)
|
||||
if !reflect.DeepEqual(out, test.out) {
|
||||
t.Fatalf("mis-match: %v %v", out, test.out)
|
||||
}
|
||||
}
|
||||
}
|
||||
|
||||
func TestWalkPath(t *testing.T) {
|
||||
r := New[interface{}]()
|
||||
|
||||
keys := []string{
|
||||
"foo",
|
||||
"foo/bar",
|
||||
"foo/bar/baz",
|
||||
"foo/baz/bar",
|
||||
"foo/zip/zap",
|
||||
"zipzap",
|
||||
}
|
||||
for _, k := range keys {
|
||||
r.Insert(k, nil)
|
||||
}
|
||||
if r.Len() != len(keys) {
|
||||
t.Fatalf("bad len: %v %v", r.Len(), len(keys))
|
||||
}
|
||||
|
||||
type exp struct {
|
||||
inp string
|
||||
out []string
|
||||
}
|
||||
cases := []exp{
|
||||
{
|
||||
"f",
|
||||
[]string{},
|
||||
},
|
||||
{
|
||||
"foo",
|
||||
[]string{"foo"},
|
||||
},
|
||||
{
|
||||
"foo/",
|
||||
[]string{"foo"},
|
||||
},
|
||||
{
|
||||
"foo/ba",
|
||||
[]string{"foo"},
|
||||
},
|
||||
{
|
||||
"foo/bar",
|
||||
[]string{"foo", "foo/bar"},
|
||||
},
|
||||
{
|
||||
"foo/bar/baz",
|
||||
[]string{"foo", "foo/bar", "foo/bar/baz"},
|
||||
},
|
||||
{
|
||||
"foo/bar/bazoo",
|
||||
[]string{"foo", "foo/bar", "foo/bar/baz"},
|
||||
},
|
||||
{
|
||||
"z",
|
||||
[]string{},
|
||||
},
|
||||
}
|
||||
|
||||
for _, test := range cases {
|
||||
out := []string{}
|
||||
fn := func(s string, v interface{}) bool {
|
||||
out = append(out, s)
|
||||
return false
|
||||
}
|
||||
r.WalkPath(test.inp, fn)
|
||||
sort.Strings(out)
|
||||
sort.Strings(test.out)
|
||||
if !reflect.DeepEqual(out, test.out) {
|
||||
t.Fatalf("mis-match: %v %v", out, test.out)
|
||||
}
|
||||
}
|
||||
}
|
||||
|
||||
func TestWalkDelete(t *testing.T) {
|
||||
r := New[interface{}]()
|
||||
r.Insert("init0/0", nil)
|
||||
r.Insert("init0/1", nil)
|
||||
r.Insert("init0/2", nil)
|
||||
r.Insert("init0/3", nil)
|
||||
r.Insert("init1/0", nil)
|
||||
r.Insert("init1/1", nil)
|
||||
r.Insert("init1/2", nil)
|
||||
r.Insert("init1/3", nil)
|
||||
r.Insert("init2", nil)
|
||||
|
||||
deleteFn := func(s string, v interface{}) bool {
|
||||
r.Delete(s)
|
||||
return false
|
||||
}
|
||||
|
||||
r.WalkPrefix("init1", deleteFn)
|
||||
|
||||
for _, s := range []string{"init0/0", "init0/1", "init0/2", "init0/3", "init2"} {
|
||||
if _, ok := r.Get(s); !ok {
|
||||
t.Fatalf("expecting to still find %q", s)
|
||||
}
|
||||
}
|
||||
if n := r.Len(); n != 5 {
|
||||
t.Fatalf("expected to find exactly 5 nodes, instead found %d: %v", n, r.ToMap())
|
||||
}
|
||||
|
||||
r.Walk(deleteFn)
|
||||
if n := r.Len(); n != 0 {
|
||||
t.Fatalf("expected to find exactly 0 nodes, instead found %d: %v", n, r.ToMap())
|
||||
}
|
||||
}
|
||||
|
||||
// generateUUID is used to generate a random UUID
|
||||
func generateUUID() string {
|
||||
buf := make([]byte, 16)
|
||||
if _, err := crand.Read(buf); err != nil {
|
||||
panic(fmt.Errorf("failed to read random bytes: %v", err))
|
||||
}
|
||||
|
||||
return fmt.Sprintf("%08x-%04x-%04x-%04x-%12x",
|
||||
buf[0:4],
|
||||
buf[4:6],
|
||||
buf[6:8],
|
||||
buf[8:10],
|
||||
buf[10:16])
|
||||
}
|
||||
|
||||
func BenchmarkInsert(b *testing.B) {
|
||||
r := New[bool]()
|
||||
for i := 0; i < 10000; i++ {
|
||||
r.Insert(fmt.Sprintf("init%d", i), true)
|
||||
}
|
||||
b.ResetTimer()
|
||||
for n := 0; n < b.N; n++ {
|
||||
_, updated := r.Insert(strconv.Itoa(n), true)
|
||||
if updated {
|
||||
b.Fatal("bad")
|
||||
}
|
||||
}
|
||||
}
|
|
@ -92,12 +92,12 @@ func (err ErrInvalidMapKey) Unwrap() error {
|
|||
return err.Wrapped
|
||||
}
|
||||
|
||||
type ErrOwnerInvalid struct {
|
||||
type ErrOwnerTypeInvalid struct {
|
||||
ResourceType *pbresource.Type
|
||||
OwnerType *pbresource.Type
|
||||
}
|
||||
|
||||
func (err ErrOwnerInvalid) Error() string {
|
||||
func (err ErrOwnerTypeInvalid) Error() string {
|
||||
return fmt.Sprintf(
|
||||
"resources of type %s cannot be owned by resources with type %s",
|
||||
ToGVK(err.ResourceType),
|
||||
|
@ -105,6 +105,20 @@ func (err ErrOwnerInvalid) Error() string {
|
|||
)
|
||||
}
|
||||
|
||||
type ErrOwnerTenantInvalid struct {
|
||||
ResourceType *pbresource.Type
|
||||
ResourceTenancy *pbresource.Tenancy
|
||||
OwnerTenancy *pbresource.Tenancy
|
||||
}
|
||||
|
||||
func (err ErrOwnerTenantInvalid) Error() string {
|
||||
return fmt.Sprintf(
|
||||
"resource in partition %s, namespace %s and peer %s cannot be owned by a resource in partition %s, namespace %s and peer %s",
|
||||
err.ResourceTenancy.Partition, err.ResourceTenancy.Namespace, err.ResourceTenancy.PeerName,
|
||||
err.OwnerTenancy.Partition, err.OwnerTenancy.Namespace, err.OwnerTenancy.PeerName,
|
||||
)
|
||||
}
|
||||
|
||||
type ErrInvalidReferenceType struct {
|
||||
AllowedType *pbresource.Type
|
||||
}
|
||||
|
|
|
@ -63,7 +63,7 @@ func TestErrorStrings(t *testing.T) {
|
|||
Key: "http",
|
||||
Wrapped: fakeWrappedErr,
|
||||
},
|
||||
"ErrOwnerInvalid": ErrOwnerInvalid{
|
||||
"ErrOwnerInvalid": ErrOwnerTypeInvalid{
|
||||
ResourceType: &pbresource.Type{Group: "foo", GroupVersion: "v1", Kind: "bar"},
|
||||
OwnerType: &pbresource.Type{Group: "other", GroupVersion: "v2", Kind: "something"},
|
||||
},
|
||||
|
|
|
@ -2,7 +2,6 @@ package resourcetest
|
|||
|
||||
import (
|
||||
"context"
|
||||
"testing"
|
||||
|
||||
"github.com/hashicorp/consul/proto-public/pbresource"
|
||||
"github.com/oklog/ulid/v2"
|
||||
|
@ -38,13 +37,24 @@ func Resource(rtype *pbresource.Type, name string) *resourceBuilder {
|
|||
}
|
||||
}
|
||||
|
||||
func (b *resourceBuilder) WithData(t *testing.T, data protoreflect.ProtoMessage) *resourceBuilder {
|
||||
func (b *resourceBuilder) WithData(t T, data protoreflect.ProtoMessage) *resourceBuilder {
|
||||
t.Helper()
|
||||
|
||||
anyData, err := anypb.New(data)
|
||||
require.NoError(t, err)
|
||||
b.resource.Data = anyData
|
||||
return b
|
||||
}
|
||||
|
||||
func (b *resourceBuilder) WithMeta(key string, value string) *resourceBuilder {
|
||||
if b.resource.Metadata == nil {
|
||||
b.resource.Metadata = make(map[string]string)
|
||||
}
|
||||
|
||||
b.resource.Metadata[key] = value
|
||||
return b
|
||||
}
|
||||
|
||||
func (b *resourceBuilder) WithOwner(id *pbresource.ID) *resourceBuilder {
|
||||
b.resource.Owner = id
|
||||
return b
|
||||
|
@ -91,7 +101,13 @@ func (b *resourceBuilder) Build() *pbresource.Resource {
|
|||
return res
|
||||
}
|
||||
|
||||
func (b *resourceBuilder) Write(t *testing.T, client pbresource.ResourceServiceClient) *pbresource.Resource {
|
||||
func (b *resourceBuilder) ID() *pbresource.ID {
|
||||
return b.resource.Id
|
||||
}
|
||||
|
||||
func (b *resourceBuilder) Write(t T, client pbresource.ResourceServiceClient) *pbresource.Resource {
|
||||
t.Helper()
|
||||
|
||||
res := b.resource
|
||||
|
||||
rsp, err := client.Write(context.Background(), &pbresource.WriteRequest{
|
||||
|
@ -101,7 +117,9 @@ func (b *resourceBuilder) Write(t *testing.T, client pbresource.ResourceServiceC
|
|||
require.NoError(t, err)
|
||||
|
||||
if !b.dontCleanup {
|
||||
t.Cleanup(func() {
|
||||
cleaner, ok := t.(CleanupT)
|
||||
require.True(t, ok, "T does not implement a Cleanup method and cannot be used with automatic resource cleanup")
|
||||
cleaner.Cleanup(func() {
|
||||
_, err := client.Delete(context.Background(), &pbresource.DeleteRequest{
|
||||
Id: rsp.Resource.Id,
|
||||
})
|
||||
|
|
218
internal/resource/resourcetest/client.go
Normal file
218
internal/resource/resourcetest/client.go
Normal file
|
@ -0,0 +1,218 @@
|
|||
package resourcetest
|
||||
|
||||
import (
|
||||
"context"
|
||||
"math/rand"
|
||||
"time"
|
||||
|
||||
"github.com/hashicorp/consul/internal/resource"
|
||||
"github.com/hashicorp/consul/proto-public/pbresource"
|
||||
"github.com/hashicorp/consul/sdk/testutil/retry"
|
||||
"github.com/stretchr/testify/require"
|
||||
"golang.org/x/exp/slices"
|
||||
"google.golang.org/grpc/codes"
|
||||
"google.golang.org/grpc/status"
|
||||
)
|
||||
|
||||
type Client struct {
|
||||
pbresource.ResourceServiceClient
|
||||
|
||||
timeout time.Duration
|
||||
wait time.Duration
|
||||
}
|
||||
|
||||
func NewClient(client pbresource.ResourceServiceClient) *Client {
|
||||
return &Client{
|
||||
ResourceServiceClient: client,
|
||||
timeout: 7 * time.Second,
|
||||
wait: 25 * time.Millisecond,
|
||||
}
|
||||
}
|
||||
|
||||
func (client *Client) SetRetryerConfig(timeout time.Duration, wait time.Duration) {
|
||||
client.timeout = timeout
|
||||
client.wait = wait
|
||||
}
|
||||
|
||||
func (client *Client) retry(t T, fn func(r *retry.R)) {
|
||||
retryer := &retry.Timer{Timeout: client.timeout, Wait: client.wait}
|
||||
retry.RunWith(retryer, t, fn)
|
||||
}
|
||||
|
||||
func (client *Client) PublishResources(t T, resources []*pbresource.Resource) {
|
||||
// Randomize the order of insertion. Generally insertion order shouldn't matter as the
|
||||
// controllers should eventually converge on the desired state. The exception to this
|
||||
// is that you cannot insert resources with owner refs before the resource they are
|
||||
// owned by or insert a resource into a non-default tenant before that tenant exists.
|
||||
rand.Shuffle(len(resources), func(i, j int) {
|
||||
temp := resources[i]
|
||||
resources[i] = resources[j]
|
||||
resources[j] = temp
|
||||
})
|
||||
|
||||
// This slice will be used to track the resources actually published each round. When
|
||||
// a resource with an owner ID is encountered we will not attempt to write but defer it
|
||||
// to the next round of publishing
|
||||
var written []*pbresource.ID
|
||||
|
||||
for len(resources) > 0 {
|
||||
var left []*pbresource.Resource
|
||||
published := 0
|
||||
for _, res := range resources {
|
||||
|
||||
// check that any owner references would be satisfied
|
||||
if res.Owner != nil {
|
||||
found := slices.ContainsFunc(written, func(id *pbresource.ID) bool {
|
||||
return resource.EqualID(res.Owner, id)
|
||||
})
|
||||
|
||||
// the owner hasn't yet been published then we cannot publish this resource
|
||||
if !found {
|
||||
left = append(left, res)
|
||||
continue
|
||||
}
|
||||
}
|
||||
|
||||
t.Logf("Writing resource %s with type %s", res.Id.Name, resource.ToGVK(res.Id.Type))
|
||||
_, err := client.Write(context.Background(), &pbresource.WriteRequest{
|
||||
Resource: res,
|
||||
})
|
||||
require.NoError(t, err)
|
||||
|
||||
// track the number o
|
||||
published += 1
|
||||
written = append(written, res.Id)
|
||||
}
|
||||
|
||||
// the next round only has this subset of resources to attempt writing
|
||||
resources = left
|
||||
|
||||
// if we didn't publish any resources this round then nothing would
|
||||
// enable us to do so by iterating again so we break to prevent infinite
|
||||
// loooping.
|
||||
if published == 0 {
|
||||
break
|
||||
}
|
||||
}
|
||||
|
||||
require.Empty(t, resources, "Could not publish all resources - some resources have invalid owner references")
|
||||
}
|
||||
|
||||
func (client *Client) RequireResourceNotFound(t T, id *pbresource.ID) {
|
||||
t.Helper()
|
||||
|
||||
rsp, err := client.Read(context.Background(), &pbresource.ReadRequest{Id: id})
|
||||
require.Error(t, err)
|
||||
require.Equal(t, codes.NotFound, status.Code(err))
|
||||
require.Nil(t, rsp)
|
||||
}
|
||||
|
||||
func (client *Client) RequireResourceExists(t T, id *pbresource.ID) *pbresource.Resource {
|
||||
t.Helper()
|
||||
|
||||
rsp, err := client.Read(context.Background(), &pbresource.ReadRequest{Id: id})
|
||||
require.NoError(t, err, "error reading %s with type %s", id.Name, resource.ToGVK(id.Type))
|
||||
require.NotNil(t, rsp)
|
||||
return rsp.Resource
|
||||
}
|
||||
|
||||
func (client *Client) RequireVersionUnchanged(t T, id *pbresource.ID, version string) *pbresource.Resource {
|
||||
t.Helper()
|
||||
|
||||
res := client.RequireResourceExists(t, id)
|
||||
RequireVersionUnchanged(t, res, version)
|
||||
return res
|
||||
}
|
||||
|
||||
func (client *Client) RequireVersionChanged(t T, id *pbresource.ID, version string) *pbresource.Resource {
|
||||
t.Helper()
|
||||
|
||||
res := client.RequireResourceExists(t, id)
|
||||
RequireVersionChanged(t, res, version)
|
||||
return res
|
||||
}
|
||||
|
||||
func (client *Client) RequireStatusCondition(t T, id *pbresource.ID, statusKey string, condition *pbresource.Condition) *pbresource.Resource {
|
||||
t.Helper()
|
||||
|
||||
res := client.RequireResourceExists(t, id)
|
||||
RequireStatusCondition(t, res, statusKey, condition)
|
||||
return res
|
||||
}
|
||||
|
||||
func (client *Client) RequireStatusConditionForCurrentGen(t T, id *pbresource.ID, statusKey string, condition *pbresource.Condition) *pbresource.Resource {
|
||||
t.Helper()
|
||||
|
||||
res := client.RequireResourceExists(t, id)
|
||||
RequireStatusConditionForCurrentGen(t, res, statusKey, condition)
|
||||
return res
|
||||
}
|
||||
|
||||
func (client *Client) RequireResourceMeta(t T, id *pbresource.ID, key string, value string) *pbresource.Resource {
|
||||
t.Helper()
|
||||
|
||||
res := client.RequireResourceExists(t, id)
|
||||
RequireResourceMeta(t, res, key, value)
|
||||
return res
|
||||
}
|
||||
|
||||
func (client *Client) RequireReconciledCurrentGen(t T, id *pbresource.ID, statusKey string) *pbresource.Resource {
|
||||
t.Helper()
|
||||
|
||||
res := client.RequireResourceExists(t, id)
|
||||
RequireReconciledCurrentGen(t, res, statusKey)
|
||||
return res
|
||||
}
|
||||
|
||||
func (client *Client) WaitForReconciliation(t T, id *pbresource.ID, statusKey string) *pbresource.Resource {
|
||||
t.Helper()
|
||||
|
||||
var res *pbresource.Resource
|
||||
|
||||
client.retry(t, func(r *retry.R) {
|
||||
res = client.RequireReconciledCurrentGen(r, id, statusKey)
|
||||
})
|
||||
|
||||
return res
|
||||
}
|
||||
|
||||
func (client *Client) WaitForStatusCondition(t T, id *pbresource.ID, statusKey string, condition *pbresource.Condition) *pbresource.Resource {
|
||||
t.Helper()
|
||||
|
||||
var res *pbresource.Resource
|
||||
client.retry(t, func(r *retry.R) {
|
||||
res = client.RequireStatusConditionForCurrentGen(t, id, statusKey, condition)
|
||||
})
|
||||
|
||||
return res
|
||||
}
|
||||
|
||||
func (client *Client) WaitForNewVersion(t T, id *pbresource.ID, version string) *pbresource.Resource {
|
||||
t.Helper()
|
||||
|
||||
var res *pbresource.Resource
|
||||
client.retry(t, func(r *retry.R) {
|
||||
res = client.RequireVersionChanged(r, id, version)
|
||||
})
|
||||
return res
|
||||
}
|
||||
|
||||
func (client *Client) WaitForResourceState(t T, id *pbresource.ID, verify func(T, *pbresource.Resource)) *pbresource.Resource {
|
||||
t.Helper()
|
||||
|
||||
var res *pbresource.Resource
|
||||
client.retry(t, func(r *retry.R) {
|
||||
res = client.RequireResourceExists(r, id)
|
||||
verify(r, res)
|
||||
})
|
||||
|
||||
return res
|
||||
}
|
||||
|
||||
// ResolveResourceID will read the specified resource and returns its full ID.
|
||||
// This is mainly useful to get the ID with the Uid filled out.
|
||||
func (client *Client) ResolveResourceID(t T, id *pbresource.ID) *pbresource.ID {
|
||||
t.Helper()
|
||||
|
||||
return client.RequireResourceExists(t, id).Id
|
||||
}
|
40
internal/resource/resourcetest/fs.go
Normal file
40
internal/resource/resourcetest/fs.go
Normal file
|
@ -0,0 +1,40 @@
|
|||
package resourcetest
|
||||
|
||||
import (
|
||||
"fmt"
|
||||
"io/fs"
|
||||
|
||||
"github.com/hashicorp/consul/proto-public/pbresource"
|
||||
"github.com/stretchr/testify/require"
|
||||
"google.golang.org/protobuf/encoding/protojson"
|
||||
)
|
||||
|
||||
// ParseResourcesFromFilesystem will walk the filesystem at the given path
|
||||
// and parse all files as protobuf/JSON resources.
|
||||
func ParseResourcesFromFilesystem(t T, files fs.FS, path string) []*pbresource.Resource {
|
||||
t.Helper()
|
||||
|
||||
var resources []*pbresource.Resource
|
||||
err := fs.WalkDir(files, path, func(fpath string, dent fs.DirEntry, _ error) error {
|
||||
if dent.IsDir() {
|
||||
return nil
|
||||
}
|
||||
|
||||
data, err := fs.ReadFile(files, fpath)
|
||||
if err != nil {
|
||||
return err
|
||||
}
|
||||
|
||||
var res pbresource.Resource
|
||||
err = protojson.Unmarshal(data, &res)
|
||||
if err != nil {
|
||||
return fmt.Errorf("error decoding data from %s: %w", fpath, err)
|
||||
}
|
||||
|
||||
resources = append(resources, &res)
|
||||
return nil
|
||||
})
|
||||
|
||||
require.NoError(t, err)
|
||||
return resources
|
||||
}
|
62
internal/resource/resourcetest/require.go
Normal file
62
internal/resource/resourcetest/require.go
Normal file
|
@ -0,0 +1,62 @@
|
|||
package resourcetest
|
||||
|
||||
import (
|
||||
"github.com/google/go-cmp/cmp"
|
||||
"github.com/hashicorp/consul/proto-public/pbresource"
|
||||
"github.com/hashicorp/consul/proto/private/prototest"
|
||||
"github.com/stretchr/testify/require"
|
||||
"google.golang.org/protobuf/testing/protocmp"
|
||||
)
|
||||
|
||||
func RequireVersionUnchanged(t T, res *pbresource.Resource, version string) {
|
||||
t.Helper()
|
||||
require.Equal(t, version, res.Version)
|
||||
}
|
||||
|
||||
func RequireVersionChanged(t T, res *pbresource.Resource, version string) {
|
||||
t.Helper()
|
||||
require.NotEqual(t, version, res.Version)
|
||||
}
|
||||
|
||||
func RequireStatusCondition(t T, res *pbresource.Resource, statusKey string, condition *pbresource.Condition) {
|
||||
t.Helper()
|
||||
require.NotNil(t, res.Status)
|
||||
status, found := res.Status[statusKey]
|
||||
require.True(t, found)
|
||||
prototest.AssertContainsElement(t, status.Conditions, condition)
|
||||
}
|
||||
|
||||
func RequireStatusConditionForCurrentGen(t T, res *pbresource.Resource, statusKey string, condition *pbresource.Condition) {
|
||||
t.Helper()
|
||||
require.NotNil(t, res.Status)
|
||||
status, found := res.Status[statusKey]
|
||||
require.True(t, found)
|
||||
require.Equal(t, res.Generation, status.ObservedGeneration)
|
||||
prototest.AssertContainsElement(t, status.Conditions, condition)
|
||||
}
|
||||
|
||||
func RequireResourceMeta(t T, res *pbresource.Resource, key string, value string) {
|
||||
t.Helper()
|
||||
require.NotNil(t, res.Metadata)
|
||||
require.Contains(t, res.Metadata, key)
|
||||
require.Equal(t, value, res.Metadata[key])
|
||||
}
|
||||
|
||||
func RequireReconciledCurrentGen(t T, res *pbresource.Resource, statusKey string) {
|
||||
t.Helper()
|
||||
require.NotNil(t, res.Status)
|
||||
status, found := res.Status[statusKey]
|
||||
require.True(t, found)
|
||||
require.Equal(t, res.Generation, status.ObservedGeneration)
|
||||
}
|
||||
|
||||
func RequireOwner(t T, res *pbresource.Resource, owner *pbresource.ID, ignoreUid bool) {
|
||||
t.Helper()
|
||||
|
||||
var opts []cmp.Option
|
||||
if ignoreUid {
|
||||
opts = append(opts, protocmp.IgnoreFields(owner, "uid"))
|
||||
}
|
||||
|
||||
prototest.AssertDeepEqual(t, res.Owner, owner, opts...)
|
||||
}
|
17
internal/resource/resourcetest/testing.go
Normal file
17
internal/resource/resourcetest/testing.go
Normal file
|
@ -0,0 +1,17 @@
|
|||
package resourcetest
|
||||
|
||||
// T represents the subset of testing.T methods that will be used
|
||||
// by the various functionality in this package
|
||||
type T interface {
|
||||
Helper()
|
||||
Log(args ...interface{})
|
||||
Logf(format string, args ...interface{})
|
||||
Errorf(format string, args ...interface{})
|
||||
Fatalf(format string, args ...interface{})
|
||||
FailNow()
|
||||
}
|
||||
|
||||
type CleanupT interface {
|
||||
T
|
||||
Cleanup(func())
|
||||
}
|
|
@ -59,6 +59,10 @@ func (r *R) Logf(format string, args ...interface{}) {
|
|||
r.log(fmt.Sprintf(format, args...))
|
||||
}
|
||||
|
||||
func (r *R) Log(args ...interface{}) {
|
||||
r.log(fmt.Sprintln(args...))
|
||||
}
|
||||
|
||||
func (r *R) Helper() {}
|
||||
|
||||
// runFailed is a sentinel value to indicate that the func itself
|
||||
|
|
Loading…
Reference in a new issue