--- layout: "docs" page_title: "Bootstrapping" sidebar_current: "docs-guides-bootstrapping" description: |- Before a Consul cluster can begin to service requests, it is necessary for a server node to be elected leader. For this reason, the first nodes that are started are generally the server nodes. Remember that an agent can run in both client and server mode. Server nodes are responsible for running the consensus protocol, and storing the cluster state. The client nodes are mostly stateless and rely on the server nodes, so they can be started easily. --- # Bootstrapping a Datacenter Before a Consul cluster can begin to service requests, it is necessary for a server node to be elected leader. For this reason, the first nodes that are started are generally the server nodes. Remember that an agent can run in both client and server mode. Server nodes are responsible for running the [consensus protocol](/docs/internals/consensus.html), and storing the cluster state. The client nodes are mostly stateless and rely on the server nodes, so they can be started easily. The recommended way to bootstrap is to use the `-bootstrap-expect` [configuration option](/docs/agent/options.html). This options informs Consul of the expected number of server nodes, and automatically bootstraps when that many servers are available. To prevent inconsistencies and split-brain situations, all servers should specify the same value for `-bootstrap-expect` or specify no value at all. Any server that does not specify a value will not attempt to bootstrap the cluster. There is a [deployment table](/docs/internals/consensus.html#toc_4) that covers various options, but it is recommended to have 3 or 5 total servers per data center. A single server deployment is _**highly**_ discouraged as data loss is inevitable in a failure scenario. Suppose we are starting a 3 server cluster, we can start `Node A`, `Node B` and `Node C` providing the `-bootstrap-expect 3` flag. Once the nodes are started, you should see a message to the effect of: ```text [WARN] raft: EnableSingleNode disabled, and no known peers. Aborting election. ``` This indicates that the nodes are expecting 2 peers, but none are known yet. The servers will not elect themselves leader to prevent a split-brain. We can now join these machines together. Since a join operation is symmetric it does not matter which node initiates it. From any node you can do the following: ```text $ consul join Successfully joined cluster by contacting 3 nodes. ``` Once the join is successful, one of the nodes will output something like: ```text [INFO] consul: adding server foo (Addr: 127.0.0.2:8300) (DC: dc1) [INFO] consul: adding server bar (Addr: 127.0.0.1:8300) (DC: dc1) [INFO] consul: Attempting bootstrap with nodes: [127.0.0.3:8300 127.0.0.2:8300 127.0.0.1:8300] ... [INFO] consul: cluster leadership acquired ``` As a sanity check, the `consul info` command is a useful tool. It can be used to verify `raft.num_peers` is now 2, and you can view the latest log index under `raft.last_log_index`. When running `consul info` on the followers, you should see `raft.last_log_index` converge to the same value as the leader begins replication. That value represents the last log entry that has been stored on disk. Now that the servers are all started and replicating to each other, all the remaining clients can be joined. Clients are much easier, as they can be started and perform a `join` against any existing node. All nodes participate in a gossip protocol to perform basic discovery, so clients will automatically find the servers and register themselves. It should be noted that it is not strictly necessary to start the server nodes before the clients, however most operations will fail until the servers are available. ## Manual Bootstrapping In versions of Consul previous to 0.4, bootstrapping was a more manual process. For a guide on using the `-bootstrap` flag directly, see the [manual bootstrapping guide](/docs/guides/manual-bootstrap.html). This is not recommended, as it is more error prone than automatic bootstrapping.