open-vault/enos/README.md
Jaymala ca18e2fffe
[QT-19] Enable Enos replication scenario (#17748)
* Add initial replication scenario config

Signed-off-by: Jaymala Sinha <jaymala@hashicorp.com>

* Add support for replication with different backend and seal types

Signed-off-by: Jaymala Sinha <jaymala@hashicorp.com>

* Update Consul versions

Signed-off-by: Jaymala Sinha <jaymala@hashicorp.com>

* Additional config for replicaiton scenario

Signed-off-by: Jaymala Sinha <jaymala@hashicorp.com>

* Update replication scenario modules

Signed-off-by: Jaymala Sinha <jaymala@hashicorp.com>

* Refactor replication modules

Signed-off-by: Jaymala Sinha <jaymala@hashicorp.com>

* Add more steps for replication

Signed-off-by: Jaymala Sinha <jaymala@hashicorp.com>

* Work in progress with unsealing followers on secondary cluster

Signed-off-by: Jaymala Sinha <jaymala@hashicorp.com>

* Add more replication scenario steps

Signed-off-by: Jaymala Sinha <jaymala@hashicorp.com>

* More updates

Signed-off-by: Jaymala Sinha <jaymala@hashicorp.com>

* Working shamir scenario

Signed-off-by: Jaymala Sinha <jaymala@hashicorp.com>

* Update to unify get Vault IP module

Signed-off-by: Jaymala Sinha <jaymala@hashicorp.com>

* Remove duplicate module

Signed-off-by: Jaymala Sinha <jaymala@hashicorp.com>

* Fix race condition for secondary followers unseal

Signed-off-by: Jaymala Sinha <jaymala@hashicorp.com>

* Use consistent naming for module directories

Signed-off-by: Jaymala Sinha <jaymala@hashicorp.com>

* Update replication scenario with latest test matrix

Signed-off-by: Jaymala Sinha <jaymala@hashicorp.com>

* Verify replication with awskms

Signed-off-by: Jaymala Sinha <jaymala@hashicorp.com>

* Add write and retrive data support for all scenarios

* Update all scenarios to verify write and read kv data

Signed-off-by: Jaymala Sinha <jaymala@hashicorp.com>

* Fix write and read data modules

Signed-off-by: Jaymala Sinha <jaymala@hashicorp.com>

* Add comments explaining the module run

Signed-off-by: Jaymala Sinha <jaymala@hashicorp.com>

* Address review feedback and update consul version

Signed-off-by: Jaymala Sinha <jaymala@hashicorp.com>

* Address more review feedback

Signed-off-by: Jaymala Sinha <jaymala@hashicorp.com>

* Remove vault debug logging

Signed-off-by: Jaymala Sinha <jaymala@hashicorp.com>

* Exclude ent.fips1402 and ent.hsm.fips1402 packages from Enos test matrix

Signed-off-by: Jaymala Sinha <jaymala@hashicorp.com>

* Add verification for replication connection status

* Currently this verification fails on Consul due to VAULT-12332

Signed-off-by: Jaymala Sinha <jaymala@hashicorp.com>

* Add replication scenario to Enos README

Signed-off-by: Jaymala Sinha <jaymala@hashicorp.com>

* Update README as per review suggesstions

Signed-off-by: Jaymala Sinha <jaymala@hashicorp.com>

* [QT-452] Add recovery keys to scenario outputs

Signed-off-by: Jaymala Sinha <jaymala@hashicorp.com>

* Fix replication output var

Signed-off-by: Jaymala Sinha <jaymala@hashicorp.com>

* Fix autopilot scenario deps and add retry for read data

Signed-off-by: Jaymala Sinha <jaymala@hashicorp.com>

Signed-off-by: Jaymala Sinha <jaymala@hashicorp.com>
2023-01-13 11:43:26 -05:00

12 KiB

Enos

Enos is an quality testing framework that allows composing and executing quality requirement scenarios as code. For Vault, it is currently used to perform infrastructure integration testing using the artifacts that are created as part of the build workflow. While intended to be executed via Github Actions using the results of the build workflow, scenarios are also executable from a developer machine that has the requisite dependencies and configuration.

Refer to the Enos documentation for further information regarding installation, execution or composing Enos scenarios.

When to use Enos

Determining whether to use vault.NewTestCluster() or Enos for testing a feature or scenario is ultimately up to the author. Sometimes one, the other, or both might be appropriate depending on the requirements. Generally, vault.NewTestCluster() is going to give you faster feedback and execution time, whereas Enos is going to give you a real-world execution and validation of the requirement. Consider the following cases as examples of when one might opt for an Enos scenario:

  • The feature require third-party integrations. Whether that be networked dependencies like a real Consul backend, a real KMS key to test awskms auto-unseal, auto-join discovery using AWS tags, or Cloud hardware KMS's.
  • The feature might behave differently under multiple configuration variants and therefore should be tested with both combinations, e.g. auto-unseal and manual shamir unseal or replication in HA mode with integrated storage or Consul storage.
  • The scenario requires coordination between multiple targets. For example, consider the complex lifecycle event of migrating the seal type or storage, or manually triggering a raft disaster scenario by partitioning the network between the leader and follower nodes. Or perhaps an auto-pilot upgrade between a stable version of Vault and our candidate version.
  • The scenario has specific deployment strategy requirements. For example, if we want to add a regression test for an issue that only arises when the software is deployed in a certain manner.
  • The scenario needs to use actual build artifacts that will be promoted through the pipeline.

Requirements

  • AWS access. HashiCorp Vault developers should use Doormat.
  • Terraform >= 1.2
  • Enos >= v0.0.10. You can install it from a release channel.
  • Access to the QTI org in Terraform Cloud. HashiCorp Vault developers can access a shared token in 1Password or request their own in #team-quality on Slack.
  • An SSH keypair in the AWS region you wish to run the scenario. You can use Doormat to log in to the AWS console to create or upload an existing keypair.
  • A Vault install bundle downloaded from releases.hashicorp.com or Artifactory when using the builder:crt variants. When using the builder:local variants Enos will build a Vault bundle from the current branch for you.

Scenario Variables

In CI, each scenario is executed via Github Actions and has been configured using environment variable inputs that follow the ENOS_VAR_varname pattern.

For local execution you can specify all the required variables using environment variables, or you can update enos.vars.hcl with values and uncomment the lines.

Variables that are required:

  • aws_ssh_keypair_name
  • aws_ssh_private_key_path
  • tfc_api_token
  • vault_bundle_path
  • vault_license_path (only required for non-OSS editions)

See enos.vars.hcl or enos-variables.hcl for further descriptions of the variables.

Executing Scenarios

From the enos directory:

# List all available scenarios
enos scenario list
# Run the smoke or upgrade scenario with an artifact that is built locally. Make sure
# the local machine has been configured as detailed in the requirements
# section. This will execute the scenario and clean up any resources if successful.
enos scenario run smoke builder:local
enos scenario run upgrade builder:local
# To run the same scenario variants that are run in CI, refer to the scenarios listed
# in .github/workflows/enos-run.yml under `jobs.enos.strategy.matrix.include`,
# adding `builder:local` to run locally.
enos scenario run smoke backend:consul consul_version:1.12.3 distro:ubuntu seal:awskms builder:local arch:amd64 edition:oss
# Launch an individual scenario but leave infrastructure up after execution
enos scenario launch smoke builder:local
# Check an individual scenario for validity. This is useful during scenario
# authoring and debugging.
enos scenario validate smoke builder:local
# If you've run the tests and desire to see the outputs, such as the URL or
# credentials, you can run the output command to see them. Please note that
# after "run" or destroy there will be no "outputs" as the infrastructure
# will have been destroyed and state cleared.
enos scenario output smoke builder:local
# Explicitly destroy all existing infrastructure
enos scenario destroy smoke builder:local

Refer to the Enos documentation for further information regarding installation, execution or composing scenarios.

Scenarios

There are current two scenarios: smoke and upgrade. Both begin by building Vault as specified by the selected builder variant (see Variants section below for more information).

Smoke

The smoke scenario creates a Vault cluster using the version from the current branch (either in CI or locally), with the backend specified by the backend variant (raft or consul). Next, it unseals with the appropriate method (awskms or shamir) and performs different verifications depending on the backend and seal type.

Upgrade

The upgrade scenario creates a Vault cluster using the version specified in vault_upgrade_initial_release, with the backend specified by the backend variant (raft or consul). Next, it upgrades the Vault binary that is determined by the builder variant. After the upgrade, it verifies that cluster is at the desired version, along with additional verifications.

Autopilot

The autopilot scenario creates a Vault cluster using the version specified in vault_upgrade_initial_release. It writes test data to the Vault cluster. Next, it creates additional nodes with the candidate version of Vault as determined by the vault_product_version variable set. The module uses AWS auto-join to handle discovery and unseals with auto-unseal or Shamir depending on the seal variant. After the new nodes have joined and been unsealed, it verifies reading stored data on the new nodes. Autopilot upgrade verification checks the upgrade status is "await-server-removal" and the target version is set to the version of upgraded nodes. This test also verifies the undo_logs status for Vault versions 1.13.x

Replication

The replication scenario creates two 3-node Vault clusters and runs following verification steps:

  1. Writes data on the primary cluster
  2. Enables performance replication
  3. Verifies reading stored data from secondary cluster
  4. Verifies initial replication status between both clusters
  5. Replaces the leader node and one standby node on the primary Vault cluster
  6. Verifies updated replication status between both clusters

This scenario verifies the performance replication status on both clusters to have their connection_status as "connected" and that the secondary cluster has known_primaries cluster addresses updated to the active nodes IP addresses of the primary Vault cluster. This scenario currently works around issues VAULT-12311 and VAULT-12309. The scenario fails when the primary storage backend is Consul due to issue VAULT-12332

Variants

Both scenarios support a matrix of variants. In order to achieve broad coverage while keeping test run time reasonable, the variants executed by the enos-run Github Actions are tailored to maximize variant distribution per scenario.

builder:crt

This variant is designed for use in Github Actions. The enos-run.yml workflow downloads the artifact built by the build.yml workflow, unzips it, and sets the vault_bundle_path to the zip file and the vault_local_binary_path to the binary.

builder:local

This variant is for running the Enos scenario locally. It builds the Vault bundle from the current branch, placing the bundle at the vault_bundle_path and the unzipped Vault binary at the vault_local_binary_path.

CI Bootstrap

In order to execute any of the scenarios in this repository, it is first necessary to bootstrap the CI AWS account with the required permissions, service quotas and supporting AWS resources. There are two Terraform modules which are used for this purpose, service-user-iam for the account permissions, and service quotas and bootstrap for the supporting resources.

Supported Regions - enos scenarios are supported in the following regions: "us-east-1", "us-east-2", "us-west-1", "us-west-2"

Bootstrap Process

These steps should be followed to bootstrap this repo for enos scenario execution:

Set up CI service user IAM role and Service Quotas

The service user that is used when executing enos scenarios from any GitHub Action workflow must have a properly configured IAM role granting the access required to create resources in AWS. Additionally, service quotas need to be adjusted to ensure that normal use of the ci account does not cause any service quotas to be exceeded. The service-user-iam module contains the IAM Policy and Role for that grants this access as well as the service quota increase requests to adjust the service quotas. This module should be updated whenever a new AWS resource type is required for a scenario or a service quota limit needs to be increased. Since this is persistent and cannot be created and destroyed each time a scenario is run, the Terraform state will be managed by Terraform Cloud. Here are the steps to configure the GitHub Actions service user:

Pre-requisites

  • Access to the hashicorp-qti organization in Terraform Cloud.
  • Full access to the CI AWS account is required.

Notes:

  • For help with access to Terraform Cloud and the CI Account, contact the QT team on Slack (#team-quality) for an invite. After receiving an invite to Terraform Cloud, a personal access token can be created by clicking User Settings --> Tokens --> Create an API token.
  • Access to the AWS account can be done via Doormat, at: https://doormat.hashicorp.services/.
    • For the vault repo the account is: vault_ci and for the vault-enterprise repo, the account is: vault-enterprise_ci.
    • Access can be requested by clicking: Cloud Access --> AWS --> Request Account Access.
  1. Create the Terraform Cloud Workspace - The name of the workspace to be created depends on the repository for which it is being created, but the pattern is: <repository>-ci-service-user-iam, e.g. vault-ci-service-user-iam. It is important that the execution mode for the workspace be set to local. For help on setting up the workspace, contact the QT team on Slack (#team-quality)

  2. Execute the Terraform module

> cd ./enos/ci/service-user-iam
> export TF_WORKSPACE=<repo name>-ci-service-user-iam
> export TF_TOKEN_app_terraform_io=<Terraform Cloud Token>
> export TF_VAR_repository=<repository name>
> terraform init
> terraform plan
> terraform apply -auto-approve

Bootstrap the CI resources

Bootstrapping of the resources in the CI account is accomplished via the GitHub Actions workflow: enos-bootstrap-ci. Before this workflow can be run a workspace must be created as follows:

  1. Create the Terraform Cloud Workspace - The name workspace to be created depends on the repository for which it is being created, but the pattern is: <repository>-ci-bootstrap, e.g. vault-ci-bootstrap. It is important that the execution mode for the workspace be set to local. For help on setting up the workspace, contact the QT team on Slack (#team-quality).

Once the workspace has been created, changes to the bootstrap module will automatically be applied via the GitHub PR workflow. Each time a PR is created for changes to files within that module the module will be planned via the workflow described above. If the plan is ok and the PR is merged, the module will automatically be applied via the same workflow.