Go to file
Tim Gross 74486d86fb
scheduler: prevent panic in spread iterator during alloc stop
The spread iterator can panic when processing an evaluation, resulting
in an unrecoverable state in the cluster. Whenever a panicked server
restarts and quorum is restored, the next server to dequeue the
evaluation will panic.

To trigger this state:
* The job must have `max_parallel = 0` and a `canary >= 1`.
* The job must not have a `spread` block.
* The job must have a previous version.
* The previous version must have a `spread` block and at least one
  failed allocation.

In this scenario, the desired changes include `(place 1+) (stop
1+), (ignore n) (canary 1)`. Before the scheduler can place the canary
allocation, it tries to find out which allocations can be
stopped. This passes back through the stack so that we can determine
previous-node penalties, etc. We call `SetJob` on the stack with the
previous version of the job, which will include assessing the `spread`
block (even though the results are unused). The task group spread info
state from that pass through the spread iterator is not reset when we
call `SetJob` again. When the new job version iterates over the
`groupPropertySets`, it will get an empty `spreadAttributeMap`,
resulting in an unexpected nil pointer dereference.

This changeset resets the spread iterator internal state when setting
the job, logging with a bypass around the bug in case we hit similar
cases, and a test that panics the scheduler without the patch.
2022-02-09 19:53:06 -05:00
.changelog scheduler: prevent panic in spread iterator during alloc stop 2022-02-09 19:53:06 -05:00
.circleci Version 1.2.3 2021-12-13 10:12:07 -05:00
.github fix: backport release branch target (#11627) 2021-12-07 09:45:46 -05:00
.tours Make number of scheduler workers reloadable (#11593) 2022-01-06 11:56:13 -05:00
acl api: prevent excessice CPU load on job parse 2022-02-09 19:51:47 -05:00
api cleanup: stop referencing depreceted HeaderMap field 2022-01-12 10:32:54 -06:00
client client: check escaping of alloc dir using symlinks 2022-02-09 19:50:13 -05:00
command api: prevent excessice CPU load on job parse 2022-02-09 19:51:47 -05:00
contributing Version 1.2.3 2021-12-13 10:12:07 -05:00
demo [demo] Kadalu CSI support for Nomad (#11207) 2021-10-06 15:29:15 -04:00
dev docs: swap master for main in Nomad repo 2021-03-08 14:26:31 -05:00
drivers drivers: set world-readable permissions on copied resolv.conf (#11856) 2022-01-14 12:25:23 -05:00
e2e chore: fixup inconsistent method receiver names. (#11704) 2021-12-20 11:44:21 +01:00
helper client: check escaping of alloc dir using symlinks 2022-02-09 19:50:13 -05:00
integrations spelling: registrations 2018-03-11 18:40:53 +00:00
internal/testing/apitests Revert "Return SchedulerConfig instead of SchedulerConfigResponse struct (#10799)" (#11433) 2021-11-02 17:42:52 -04:00
jobspec Parse `job > group > consul` block in HCL1 (#11423) 2021-11-03 13:49:32 -04:00
jobspec2 api: prevent excessice CPU load on job parse 2022-02-09 19:51:47 -05:00
lib chore: fixup inconsistent method receiver names. (#11704) 2021-12-20 11:44:21 +01:00
nomad client: check escaping of alloc dir using symlinks 2022-02-09 19:50:13 -05:00
plugins chore: fixup inconsistent method receiver names. (#11704) 2021-12-20 11:44:21 +01:00
scheduler scheduler: prevent panic in spread iterator during alloc stop 2022-02-09 19:53:06 -05:00
scripts golang security update 1.17.5 2021-12-10 13:50:22 -05:00
terraform terraform: update installed version used to 1.0.11. 2021-11-19 09:33:11 +01:00
testutil cli: refactor operator debug capture (#11466) 2021-11-05 19:43:10 -04:00
tools build: bump go version to 1.17.3 (#11461) 2021-11-05 15:34:24 -04:00
ui ui: fix test (#11870) 2022-01-18 10:36:10 -05:00
version Generate files for 1.2.5 release 2022-01-31 14:54:26 +00:00
website docs: add `nomad.plan.node_rejected` metric (#11860) 2022-01-18 13:47:20 -05:00
.gitattributes Remove invalid gitattributes 2018-02-14 14:47:43 -08:00
.gitignore terraform: update installed version used to 1.0.11. 2021-11-19 09:33:11 +01:00
.golangci.yml chore: fixup inconsistent method receiver names. (#11704) 2021-12-20 11:44:21 +01:00
CHANGELOG.md docs: add 1.2.5 to changelog 2022-01-28 15:08:48 -05:00
GNUmakefile set LAST_RELEASE to 1.2.4 for the 1.2.5 release branch 2022-01-28 14:50:54 -05:00
LICENSE Initial commit 2015-06-01 12:21:00 +02:00
README.md README: Align with Consul README (#9681) 2020-12-18 09:38:34 -08:00
Vagrantfile tools: update virtualbox networking configuration (#11561) 2021-11-24 10:45:58 -05:00
build_linux_arm.go gofmt all the files 2021-10-01 10:14:28 -04:00
go.mod Fix log level parsing from lines that include a timestamp (#11838) 2022-01-13 09:56:35 -05:00
go.sum Fix log level parsing from lines that include a timestamp (#11838) 2022-01-13 09:56:35 -05:00
main.go Raft Debugging Improvements (#11414) 2021-11-04 10:16:12 -04:00
main_test.go Adding initial skeleton 2015-06-01 13:46:21 +02:00

README.md

Nomad Build Status Discuss

HashiCorp Nomad logo

Nomad is a simple and flexible workload orchestrator to deploy and manage containers (docker, podman), non-containerized applications (executable, Java), and virtual machines (qemu) across on-prem and clouds at scale.

Nomad is supported on Linux, Windows, and macOS. A commercial version of Nomad, Nomad Enterprise, is also available.

Nomad provides several key features:

  • Deploy Containers and Legacy Applications: Nomads flexibility as an orchestrator enables an organization to run containers, legacy, and batch applications together on the same infrastructure. Nomad brings core orchestration benefits to legacy applications without needing to containerize via pluggable task drivers.

  • Simple & Reliable: Nomad runs as a single binary and is entirely self contained - combining resource management and scheduling into a single system. Nomad does not require any external services for storage or coordination. Nomad automatically handles application, node, and driver failures. Nomad is distributed and resilient, using leader election and state replication to provide high availability in the event of failures.

  • Device Plugins & GPU Support: Nomad offers built-in support for GPU workloads such as machine learning (ML) and artificial intelligence (AI). Nomad uses device plugins to automatically detect and utilize resources from hardware devices such as GPU, FPGAs, and TPUs.

  • Federation for Multi-Region, Multi-Cloud: Nomad was designed to support infrastructure at a global scale. Nomad supports federation out-of-the-box and can deploy applications across multiple regions and clouds.

  • Proven Scalability: Nomad is optimistically concurrent, which increases throughput and reduces latency for workloads. Nomad has been proven to scale to clusters of 10K+ nodes in real-world production environments.

  • HashiCorp Ecosystem: Nomad integrates seamlessly with Terraform, Consul, Vault for provisioning, service discovery, and secrets management.

Quick Start

Testing

See Learn: Getting Started for instructions on setting up a local Nomad cluster for non-production use.

Optionally, find Terraform manifests for bringing up a development Nomad cluster on a public cloud in the terraform directory.

Production

See Learn: Nomad Reference Architecture for recommended practices and a reference architecture for production deployments.

Documentation

Full, comprehensive documentation is available on the Nomad website: https://www.nomadproject.io/docs

Guides are available on HashiCorp Learn.

Contributing

See the contributing directory for more developer documentation.