c4a3f5047b
We recently ran into an issue on a small percentage of nomad-clients where the nomad-client was running successfully, but due to a race condition, could not correctly bind to the docker socket. This caused all of our nomad jobs to be allocated to a single nomad-client instead of being spread evenly across our clients. The only way to discover this was to run `nomad node-status <node>` and count each job allocation per node. This can lead to a fairly long debugging process if there are several nomad-clients. Including the number of allocations for each node in the `node-status` command would save a large amount of debug time. ``` jake@biscuits [12:08:41] [~] -> % nomad node-status ID Datacenter Name Class Drain Status Allocations 2b0aabc5 dc1 biscuits <none> false ready 0 ``` ``` jake@biscuits [12:08:55] [~] -> % nomad node-status ID Datacenter Name Class Drain Status Allocations 2b0aabc5 dc1 biscuits <none> false ready 1 ``` |
||
---|---|---|
api | ||
client | ||
command | ||
demo | ||
dist | ||
Godeps | ||
helper | ||
jobspec | ||
nomad | ||
scheduler | ||
scripts | ||
testutil | ||
vendor | ||
website | ||
.gitattributes | ||
.gitignore | ||
.travis.yml | ||
CHANGELOG.md | ||
commands.go | ||
GNUmakefile | ||
LICENSE | ||
main.go | ||
main_test.go | ||
README.md | ||
Vagrantfile | ||
version.go |
Nomad
- Website: https://www.nomadproject.io
- IRC:
#nomad-tool
on Freenode - Mailing list: Google Groups
Nomad is a cluster manager, designed for both long lived services and short lived batch processing workloads. Developers use a declarative job specification to submit work, and Nomad ensures constraints are satisfied and resource utilization is optimized by efficient task packing. Nomad supports all major operating systems and virtualized, containerized, or standalone applications.
The key features of Nomad are:
-
Docker Support: Jobs can specify tasks which are Docker containers. Nomad will automatically run the containers on clients which have Docker installed, scale up and down based on the number of instances request, and automatically recover from failures.
-
Multi-Datacenter and Multi-Region Aware: Nomad is designed to be a global-scale scheduler. Multiple datacenters can be managed as part of a larger region, and jobs can be scheduled across datacenters if requested. Multiple regions join together and federate jobs making it easy to run jobs anywhere.
-
Operationally Simple: Nomad runs as a single binary that can be either a client or server, and is completely self contained. Nomad does not require any external services for storage or coordination. This means Nomad combines the features of a resource manager and scheduler in a single system.
-
Distributed and Highly-Available: Nomad servers cluster together and perform leader election and state replication to provide high availability in the face of failure. The Nomad scheduling engine is optimized for optimistic concurrency allowing all servers to make scheduling decisions to maximize throughput.
-
HashiCorp Ecosystem: Nomad integrates with the entire HashiCorp ecosystem of tools. Along with all HashiCorp tools, Nomad is designed in the unix philosophy of doing something specific and doing it well. Nomad integrates with tools like Packer, Consul, and Terraform to support building artifacts, service discovery, monitoring and capacity management.
For more information, see the introduction section of the Nomad website.
Getting Started & Documentation
All documentation is available on the Nomad website.
Developing Nomad
If you wish to work on Nomad itself or any of its built-in systems, you will first need Go installed on your machine (version 1.5+ is required).
Developing with Vagrant There is an included Vagrantfile that can help bootstrap the process. The created virtual machine is based off of Ubuntu 14, and installs several of the base libraries that can be used by Nomad.
To use this virtual machine, checkout Nomad and run vagrant up
from the root
of the repository:
$ git clone https://github.com/hashicorp/nomad.git
$ cd nomad
$ vagrant up
The virtual machine will launch, and a provisioning script will install the needed dependencies.
Developing locally
For local dev first make sure Go is properly installed, including setting up a
GOPATH. After setting up Go, clone this
repository into $GOPATH/src/github.com/hashicorp/nomad
. Then you can
download the required build tools such as vet, cover, godep etc by bootstrapping
your environment.
$ make bootstrap
...
Afterwards type make test
. This will run the tests. If this exits with exit status 0,
then everything is working!
$ make test
...
To compile a development version of Nomad, run make dev
. This will put the
Nomad binary in the bin
and $GOPATH/bin
folders:
$ make dev
...
$ bin/nomad
...
To cross-compile Nomad, run make bin
. This will compile Nomad for multiple
platforms and place the resulting binaries into the ./pkg
directory:
$ make bin
...
$ ls ./pkg
...