open-consul/contributing/INTERNALS.md
2021-07-05 18:25:06 -04:00

5.4 KiB

Consul Internals Guide

This guide is intended to help folks who want to understand more about how Consul works from a code perspective, or who are thinking about contributing to Consul. For a high level overview of Consul's design, please see the Consul Architecture Guide as a starting point.

Architecture Overview

Consul is designed around the concept of a Consul Agent. The agent is deployed as a single Go binary and runs on every node in a cluster.

A small subset of agents, usually 3 to 7, run in server mode and participate in the Raft Consensus Protocol. The Consul servers hold a consistent view of the state of the cluster, including the service catalog and the health state of services and nodes as well as other items like Consul's key/value store contents. An agent in server mode is a superset of the client capabilities that follow.

All the remaining agents in a cluster run in client mode. Applications on client nodes use their local agent in client mode to register services and to discover other services or interact with the key/value store. For the latter queries, the agent sends RPC requests internally to one of the Consul servers for the information. None of the key/value data is on any of the client agents, for example, it's always fetched on the fly from a Consul server.

Both client and server mode agents participate in a Gossip Protocol which provides two important mechanisms. First, it allows for agents to learn about all the other agents in the cluster, just by joining initially with a single existing member of the cluster. This allows clients to discover new Consul servers. Second, the gossip protocol provides a distributed failure detector, whereby the agents in the cluster randomly probe each other at regular intervals. Because of this failure detector, Consul can run health checks locally on each agent and just sent edge-triggered updates when the state of a health check changes, confident that if the agent dies altogether then the cluster will detect that. This makes Consul's health checking design very scaleable compared to centralized systems with a central polling type of design.

There are many other aspects of Consul that are well-covered in Consul's Internals Guides.

Source Code Layout

Shared Components

The components in this section are shared between Consul agents in client and server modes.

Directory Contents
command/agent This contains the actual CLI command implementation for the consul agent command, which mostly just invokes an agent object from the agent package.
agent This is where the agent object is defined, and the top level agent package has all of the functionality that's common to both client and server agents. This includes things like service registration, the HTTP and DNS endpoints, watches, and top-level glue for health checks.
agent/config This has all the user-facing configuration processing code, as well as the internal configuration structure that's used by the agent.
agent/checks This has implementations for the different health check types.
agent/ae, agent/local These are used together to power the agent's Anti-Entropy Sync Back process to the Consul servers.
agent/router, agent/pool These are used for routing RPC queries to Consul servers and for connection pooling.
agent/structs This has definitions of all the internal RPC protocol request and response structures.

Other Components

There are several other top-level packages used internally by Consul as well as externally by other applications.

Directory Contents
api This api package provides an official Go API client for Consul, which is also used by Consul's CLI commands to communicate with the local Consul agent.
api/watch This has implementation details for Consul's watches, used both internally to Consul and by the watch CLI command.
website This has the full source code for consul.io. Pull requests can update the source code and Consul's documentation all together.