2019-06-28 18:09:14 +00:00
|
|
|
|
---
|
2020-01-18 00:18:09 +00:00
|
|
|
|
layout: docs
|
2020-09-21 19:55:36 +00:00
|
|
|
|
page_title: Integrated Storage - Storage Backends - Configuration
|
2019-06-28 18:09:14 +00:00
|
|
|
|
description: |-
|
|
|
|
|
|
2020-09-21 19:55:36 +00:00
|
|
|
|
The Integrated Storage (Raft) backend is used to persist Vault's data. Unlike all the other
|
2020-01-18 00:18:09 +00:00
|
|
|
|
storage backends, this backend does not operate from a single source for the
|
|
|
|
|
data. Instead all the nodes in a Vault cluster will have a replicated copy of
|
|
|
|
|
the entire data. The data is replicated across the nodes using the Raft
|
|
|
|
|
Consensus Algorithm.
|
2019-06-28 18:09:14 +00:00
|
|
|
|
---
|
|
|
|
|
|
2020-09-21 19:55:36 +00:00
|
|
|
|
# Integrated Storage (Raft) Backend
|
2019-06-28 18:09:14 +00:00
|
|
|
|
|
2020-09-21 19:55:36 +00:00
|
|
|
|
The Integrated Storage backend is used to persist Vault's data. Unlike other storage
|
|
|
|
|
backends, Integrated Storage does not operate from a single source of data. Instead
|
2019-06-28 18:09:14 +00:00
|
|
|
|
all the nodes in a Vault cluster will have a replicated copy of Vault's data.
|
2020-11-05 20:33:50 +00:00
|
|
|
|
Data gets replicated across all the nodes via the [Raft Consensus
|
2019-06-28 18:09:14 +00:00
|
|
|
|
Algorithm][raft].
|
|
|
|
|
|
2020-09-21 19:55:36 +00:00
|
|
|
|
- **High Availability** – the Integrated Storage backend supports high availability.
|
2019-06-28 18:09:14 +00:00
|
|
|
|
|
2020-09-21 19:55:36 +00:00
|
|
|
|
- **HashiCorp Supported** – the Integrated Storage backend is officially supported
|
2019-06-28 18:09:14 +00:00
|
|
|
|
by HashiCorp.
|
|
|
|
|
|
|
|
|
|
```hcl
|
|
|
|
|
storage "raft" {
|
|
|
|
|
path = "/path/to/raft/data"
|
|
|
|
|
node_id = "raft_node_1"
|
|
|
|
|
}
|
|
|
|
|
cluster_addr = "http://127.0.0.1:8201"
|
|
|
|
|
```
|
|
|
|
|
|
2020-09-21 19:55:36 +00:00
|
|
|
|
~> **Note:** When using the Integrated Storage backend, it is required to provide
|
2020-02-14 20:25:53 +00:00
|
|
|
|
`cluster_addr` to indicate the address and port to be used for communication
|
|
|
|
|
between the nodes in the Raft cluster.
|
|
|
|
|
|
2020-09-21 19:55:36 +00:00
|
|
|
|
~> **Note:** When using the Integrated Storage backend, a separate `ha_storage`
|
2020-04-08 20:07:42 +00:00
|
|
|
|
backend cannot be declared.
|
|
|
|
|
|
2020-09-21 19:55:36 +00:00
|
|
|
|
~> **Note:** When using the Integrated Storage backend, it is strongly recommended to
|
2020-06-05 22:02:55 +00:00
|
|
|
|
set `disable_mlock` to `true`, and to disable memory swapping on the system.
|
|
|
|
|
|
2019-06-28 18:09:14 +00:00
|
|
|
|
## `raft` Parameters
|
|
|
|
|
|
|
|
|
|
- `path` `(string: "")` – The file system path where all the Vault data gets
|
|
|
|
|
stored.
|
2019-10-28 16:43:12 +00:00
|
|
|
|
This value can be overridden by setting the `VAULT_RAFT_PATH` environment variable.
|
2019-06-28 18:09:14 +00:00
|
|
|
|
|
|
|
|
|
- `node_id` `(string: "")` - The identifier for the node in the Raft cluster.
|
2019-10-28 16:43:12 +00:00
|
|
|
|
This value can be overridden by setting the `VAULT_RAFT_NODE_ID` environment variable.
|
2019-06-28 18:09:14 +00:00
|
|
|
|
|
2020-03-04 20:58:51 +00:00
|
|
|
|
- `performance_multiplier` `(integer: 0)` - An integer multiplier used by
|
|
|
|
|
servers to scale key Raft timing parameters. Tuning this affects the time it
|
|
|
|
|
takes Vault to detect leader failures and to perform leader elections, at the
|
|
|
|
|
expense of requiring more network and CPU resources for better performance.
|
|
|
|
|
Omitting this value or setting it to 0 uses default timing described below.
|
|
|
|
|
Lower values are used to tighten timing and increase sensitivity while higher
|
2020-04-08 20:07:42 +00:00
|
|
|
|
values relax timings and reduce sensitivity.
|
2020-03-04 20:58:51 +00:00
|
|
|
|
|
|
|
|
|
By default, Vault will use a lower-performance timing that's suitable for
|
|
|
|
|
minimal Vault servers, currently equivalent to setting this to a value of 5
|
|
|
|
|
(this default may be changed in future versions of Vault, depending if the
|
|
|
|
|
target minimum server profile changes). Setting this to a value of 1 will
|
|
|
|
|
configure Raft to its highest-performance mode and is recommended for
|
|
|
|
|
production Vault servers. The maximum allowed value is 10.
|
|
|
|
|
|
|
|
|
|
- `trailing_logs` `(integer: 10000)` - This controls how many log entries are
|
|
|
|
|
left in the log store on disk after a snapshot is made. This should only be
|
|
|
|
|
adjusted when followers cannot catch up to the leader due to a very large
|
|
|
|
|
snapshot size and high write throughput causing log truncation before a
|
|
|
|
|
snapshot can be fully installed. If you need to use this to recover a cluster,
|
|
|
|
|
consider reducing write throughput or the amount of data stored on Vault. The
|
|
|
|
|
default value is 10000 which is suitable for all normal workloads.
|
|
|
|
|
|
|
|
|
|
- `snapshot_threshold` `(integer: 8192)` - This controls the minimum number of raft
|
|
|
|
|
commit entries between snapshots that are saved to disk. This is a low-level
|
|
|
|
|
parameter that should rarely need to be changed. Very busy clusters
|
|
|
|
|
experiencing excessive disk IO may increase this value to reduce disk IO and
|
|
|
|
|
minimize the chances of all servers taking snapshots at the same time.
|
|
|
|
|
Increasing this trades off disk IO for disk space since the log will grow much
|
|
|
|
|
larger and the space in the raft.db file can't be reclaimed till the next
|
|
|
|
|
snapshot. Servers may take longer to recover from crashes or failover if this
|
|
|
|
|
is increased significantly as more logs will need to be replayed.
|
|
|
|
|
|
2020-03-06 00:47:10 +00:00
|
|
|
|
- `retry_join` `(list: [])` - There can be one or more `retry_join` stanzas.
|
|
|
|
|
When the raft cluster is getting bootstrapped, if the connection details of all
|
|
|
|
|
the nodes are known beforehand, then specifying this config stanzas enables the
|
|
|
|
|
nodes to automatically join a raft cluster. All the nodes would mention all
|
|
|
|
|
other nodes that they could join using this config. When one of the nodes is
|
|
|
|
|
initialized, it becomes the leader and all the other nodes will join the
|
|
|
|
|
leader node to form the cluster. When using Shamir seal, the joined nodes will
|
|
|
|
|
still need to be unsealed manually. See the section below that describes the
|
|
|
|
|
parameters accepted by the `retry_join` stanza.
|
|
|
|
|
|
2020-06-01 14:17:24 +00:00
|
|
|
|
- `max_entry_size` `(integer: 1048576)` - This configures the maximum number of
|
|
|
|
|
bytes for a raft entry. It applies to both Put operations and transactions.
|
|
|
|
|
Any put or transaction operation exceeding this configuration value will cause
|
|
|
|
|
the respective operation to fail. Raft has a suggested max size of data in a
|
|
|
|
|
raft log entry. This is based on current architecture, default timing, etc.
|
|
|
|
|
Integrated storage also uses a chunk size that is the threshold used for
|
|
|
|
|
breaking a large value into chunks. By default, the chunk size is the same as
|
|
|
|
|
raft's max size log entry. The default value for this configuration is 1048576
|
|
|
|
|
-- two times the chunking size.
|
|
|
|
|
|
2021-03-24 14:29:10 +00:00
|
|
|
|
- `autopilot_reconcile_interval` `(string: "")` - This is the interval after
|
|
|
|
|
which autopilot will pick up any state changes. State change could mean multiple
|
|
|
|
|
things; for example a newly joined voter node, initially added as non-voter to
|
|
|
|
|
the raft cluster by autopilot has successfully completed the stabilization
|
|
|
|
|
period thereby qualifying for being promoted as a voter, a node that has become
|
|
|
|
|
unhealthy and needs to be shown as such in the state API, a node has been marked
|
|
|
|
|
as dead needing eviction from raft configuration, etc. Defaults to 10s.
|
|
|
|
|
|
2020-03-06 00:47:10 +00:00
|
|
|
|
### `retry_join` stanza
|
|
|
|
|
|
2020-05-07 01:26:08 +00:00
|
|
|
|
- `leader_api_addr` `(string: "")` - Address of a possible leader node.
|
2020-03-06 00:47:10 +00:00
|
|
|
|
|
2020-10-23 20:13:09 +00:00
|
|
|
|
- `auto_join` `(string: "")` - Cloud auto-join configuration.
|
|
|
|
|
|
|
|
|
|
- `auto_join_scheme` `(string: "")` - The optional URI protocol scheme for addresses
|
2020-12-17 21:53:33 +00:00
|
|
|
|
discovered via auto-join.
|
2020-10-23 20:13:09 +00:00
|
|
|
|
|
|
|
|
|
- `auto_join_port` `(uint: "")` - The optional port used for addressed discovered
|
2020-12-17 21:53:33 +00:00
|
|
|
|
via auto-join.
|
2020-10-23 20:13:09 +00:00
|
|
|
|
|
2020-05-07 01:26:08 +00:00
|
|
|
|
- `leader_ca_cert_file` `(string: "")` - File path to the CA cert of the
|
|
|
|
|
possible leader node.
|
2020-03-06 00:47:10 +00:00
|
|
|
|
|
2020-05-07 01:26:08 +00:00
|
|
|
|
- `leader_client_cert_file` `(string: "")` - File path to the client certificate
|
2020-12-17 21:53:33 +00:00
|
|
|
|
for the follower node to establish client authentication with the possible
|
|
|
|
|
leader node.
|
2020-03-06 00:47:10 +00:00
|
|
|
|
|
2020-05-07 01:26:08 +00:00
|
|
|
|
- `leader_client_key_file` `(string: "")` - File path to the client key for the
|
2020-12-17 21:53:33 +00:00
|
|
|
|
follower node to establish client authentication with the possible leader node.
|
2020-05-07 01:26:08 +00:00
|
|
|
|
|
|
|
|
|
- `leader_ca_cert` `(string: "")` - CA cert of the possible leader node.
|
|
|
|
|
|
|
|
|
|
- `leader_client_cert` `(string: "")` - Client certificate for the follower node
|
2020-12-17 21:53:33 +00:00
|
|
|
|
to establish client authentication with the possible leader node.
|
2020-05-07 01:26:08 +00:00
|
|
|
|
|
|
|
|
|
- `leader_client_key` `(string: "")` - Client key for the follower node to
|
2020-12-17 21:53:33 +00:00
|
|
|
|
establish client authentication with the possible leader node.
|
2020-05-07 01:26:08 +00:00
|
|
|
|
|
|
|
|
|
Each `retry_join` block may provide TLS certificates via file paths or as a
|
|
|
|
|
single-line certificate string value with newlines delimited by `\n`, but not a
|
2020-10-23 20:13:09 +00:00
|
|
|
|
combination of both. Each `retry_join` stanza may contain either a `leader_api_addr`
|
|
|
|
|
value or a cloud `auto_join` configuration value, but not both. When an `auto_join`
|
|
|
|
|
value is provided, Vault will automatically attempt to discover and resolve
|
|
|
|
|
potential Raft leader addresses.
|
|
|
|
|
|
|
|
|
|
By default, Vault will attempt to reach discovered peers using HTTPS and port 8200.
|
|
|
|
|
Operators may override these through the `auto_join_scheme` and `auto_join_port`
|
|
|
|
|
fields respectively.
|
2020-03-06 00:47:10 +00:00
|
|
|
|
|
|
|
|
|
Example Configuration:
|
2020-12-17 21:53:33 +00:00
|
|
|
|
|
2020-03-06 00:47:10 +00:00
|
|
|
|
```
|
|
|
|
|
storage "raft" {
|
|
|
|
|
path = "/Users/foo/raft/"
|
|
|
|
|
node_id = "node1"
|
2020-10-23 20:13:09 +00:00
|
|
|
|
|
2020-03-06 00:47:10 +00:00
|
|
|
|
retry_join {
|
|
|
|
|
leader_api_addr = "http://127.0.0.2:8200"
|
2020-06-15 18:36:15 +00:00
|
|
|
|
leader_ca_cert_file = "/path/to/ca1"
|
2020-05-07 01:26:08 +00:00
|
|
|
|
leader_client_cert_file = "/path/to/client/cert1"
|
|
|
|
|
leader_client_key_file = "/path/to/client/key1"
|
2020-03-06 00:47:10 +00:00
|
|
|
|
}
|
|
|
|
|
retry_join {
|
|
|
|
|
leader_api_addr = "http://127.0.0.3:8200"
|
2020-05-07 01:26:08 +00:00
|
|
|
|
leader_ca_cert_file = "/path/to/ca2"
|
|
|
|
|
leader_client_cert_file = "/path/to/client/cert2"
|
|
|
|
|
leader_client_key_file = "/path/to/client/key2"
|
2020-03-06 00:47:10 +00:00
|
|
|
|
}
|
|
|
|
|
retry_join {
|
|
|
|
|
leader_api_addr = "http://127.0.0.4:8200"
|
2020-05-07 01:26:08 +00:00
|
|
|
|
leader_ca_cert_file = "/path/to/ca3"
|
|
|
|
|
leader_client_cert_file = "/path/to/client/cert3"
|
|
|
|
|
leader_client_key_file = "/path/to/client/key3"
|
2020-03-06 00:47:10 +00:00
|
|
|
|
}
|
2020-10-23 20:13:09 +00:00
|
|
|
|
retry_join {
|
|
|
|
|
auto_join = "provider=aws region=eu-west-1 tag_key=vault tag_value=... access_key_id=... secret_access_key=..."
|
|
|
|
|
}
|
2020-03-06 00:47:10 +00:00
|
|
|
|
}
|
|
|
|
|
```
|
|
|
|
|
|
2020-01-18 00:18:09 +00:00
|
|
|
|
[raft]: https://raft.github.io/ 'The Raft Consensus Algorithm'
|