Find a file
Tim Gross c67c31e543 csi: ensure that PastClaims are populated with correct mode (#11932)
In the client's `(*csiHook) Postrun()` method, we make an unpublish
RPC that includes a claim in the `CSIVolumeClaimStateUnpublishing`
state and using the mode from the client. But then in the
`(*CSIVolume) Unpublish` RPC handler, we query the volume from the
state store (because we only get an ID from the client). And when we
make the client RPC for the node unpublish step, we use the _current
volume's_ view of the mode. If the volume's mode has been changed
before the old allocations can have their claims released, then we end
up making a CSI RPC that will never succeed.

Why does this code path get the mode from the volume and not the
claim? Because the claim written by the GC job in `(*CoreScheduler)
csiVolumeClaimGC` doesn't have a mode. Instead it just writes a claim
in the unpublishing state to ensure the volumewatcher detects a "past
claim" change and reaps all the claims on the volumes.

Fix this by ensuring that the `CSIVolumeDenormalize` creates past
claims for all nil allocations with a correct access mode set.
2022-01-28 14:43:43 -05:00
.changelog CSI: resolve invalid claim states (#11890) 2022-01-28 14:43:35 -05:00
.circleci Version 1.2.3 2021-12-13 10:12:07 -05:00
.github fix: backport release branch target (#11627) 2021-12-07 09:45:46 -05:00
.tours Make number of scheduler workers reloadable (#11593) 2022-01-06 11:56:13 -05:00
acl
api cleanup: stop referencing depreceted HeaderMap field 2022-01-12 10:32:54 -06:00
client Update IsEmpty to check for pre-1.2.4 fields (#11930) 2022-01-28 14:41:49 -05:00
command Update IsEmpty to check for pre-1.2.4 fields (#11930) 2022-01-28 14:41:49 -05:00
contributing Version 1.2.3 2021-12-13 10:12:07 -05:00
demo
dev
drivers drivers: set world-readable permissions on copied resolv.conf (#11856) 2022-01-14 12:25:23 -05:00
e2e chore: fixup inconsistent method receiver names. (#11704) 2021-12-20 11:44:21 +01:00
helper cli: return error from raft commands if db is open 2021-12-16 11:41:01 -08:00
integrations
internal/testing/apitests Revert "Return SchedulerConfig instead of SchedulerConfigResponse struct (#10799)" (#11433) 2021-11-02 17:42:52 -04:00
jobspec Parse job > group > consul block in HCL1 (#11423) 2021-11-03 13:49:32 -04:00
jobspec2 Expose Consul template configuration parameters (#11606) 2022-01-10 10:19:07 -05:00
lib chore: fixup inconsistent method receiver names. (#11704) 2021-12-20 11:44:21 +01:00
nomad csi: ensure that PastClaims are populated with correct mode (#11932) 2022-01-28 14:43:43 -05:00
plugins chore: fixup inconsistent method receiver names. (#11704) 2021-12-20 11:44:21 +01:00
scheduler CSI: resolve invalid claim states (#11890) 2022-01-28 14:43:35 -05:00
scripts golang security update 1.17.5 2021-12-10 13:50:22 -05:00
terraform terraform: update installed version used to 1.0.11. 2021-11-19 09:33:11 +01:00
testutil cli: refactor operator debug capture (#11466) 2021-11-05 19:43:10 -04:00
tools build: bump go version to 1.17.3 (#11461) 2021-11-05 15:34:24 -04:00
ui ui: fix test (#11870) 2022-01-18 10:36:10 -05:00
version Generate files for 1.2.4 release 2022-01-18 23:43:00 +00:00
website docs: add nomad.plan.node_rejected metric (#11860) 2022-01-18 13:47:20 -05:00
.gitattributes
.gitignore terraform: update installed version used to 1.0.11. 2021-11-19 09:33:11 +01:00
.golangci.yml chore: fixup inconsistent method receiver names. (#11704) 2021-12-20 11:44:21 +01:00
build_linux_arm.go
CHANGELOG.md docs: add 1.2.4 to changelog 2022-01-18 18:31:34 -05:00
GNUmakefile Make number of scheduler workers reloadable (#11593) 2022-01-06 11:56:13 -05:00
go.mod Fix log level parsing from lines that include a timestamp (#11838) 2022-01-13 09:56:35 -05:00
go.sum Fix log level parsing from lines that include a timestamp (#11838) 2022-01-13 09:56:35 -05:00
LICENSE
main.go Raft Debugging Improvements (#11414) 2021-11-04 10:16:12 -04:00
main_test.go
README.md
Vagrantfile tools: update virtualbox networking configuration (#11561) 2021-11-24 10:45:58 -05:00

Nomad Build Status Discuss

HashiCorp Nomad logo

Nomad is a simple and flexible workload orchestrator to deploy and manage containers (docker, podman), non-containerized applications (executable, Java), and virtual machines (qemu) across on-prem and clouds at scale.

Nomad is supported on Linux, Windows, and macOS. A commercial version of Nomad, Nomad Enterprise, is also available.

Nomad provides several key features:

  • Deploy Containers and Legacy Applications: Nomads flexibility as an orchestrator enables an organization to run containers, legacy, and batch applications together on the same infrastructure. Nomad brings core orchestration benefits to legacy applications without needing to containerize via pluggable task drivers.

  • Simple & Reliable: Nomad runs as a single binary and is entirely self contained - combining resource management and scheduling into a single system. Nomad does not require any external services for storage or coordination. Nomad automatically handles application, node, and driver failures. Nomad is distributed and resilient, using leader election and state replication to provide high availability in the event of failures.

  • Device Plugins & GPU Support: Nomad offers built-in support for GPU workloads such as machine learning (ML) and artificial intelligence (AI). Nomad uses device plugins to automatically detect and utilize resources from hardware devices such as GPU, FPGAs, and TPUs.

  • Federation for Multi-Region, Multi-Cloud: Nomad was designed to support infrastructure at a global scale. Nomad supports federation out-of-the-box and can deploy applications across multiple regions and clouds.

  • Proven Scalability: Nomad is optimistically concurrent, which increases throughput and reduces latency for workloads. Nomad has been proven to scale to clusters of 10K+ nodes in real-world production environments.

  • HashiCorp Ecosystem: Nomad integrates seamlessly with Terraform, Consul, Vault for provisioning, service discovery, and secrets management.

Quick Start

Testing

See Learn: Getting Started for instructions on setting up a local Nomad cluster for non-production use.

Optionally, find Terraform manifests for bringing up a development Nomad cluster on a public cloud in the terraform directory.

Production

See Learn: Nomad Reference Architecture for recommended practices and a reference architecture for production deployments.

Documentation

Full, comprehensive documentation is available on the Nomad website: https://www.nomadproject.io/docs

Guides are available on HashiCorp Learn.

Contributing

See the contributing directory for more developer documentation.