2017-06-29 00:10:05 +00:00
|
|
|
---
|
|
|
|
layout: "guides"
|
|
|
|
page_title: "Running Apache Spark on Nomad"
|
2019-05-08 21:40:38 +00:00
|
|
|
sidebar_current: "guides-analytical-workloads-spark-intro"
|
2017-06-29 00:10:05 +00:00
|
|
|
description: |-
|
|
|
|
Learn how to run Apache Spark on a Nomad cluster.
|
|
|
|
---
|
|
|
|
|
|
|
|
# Running Apache Spark on Nomad
|
|
|
|
|
2019-05-08 21:40:38 +00:00
|
|
|
Apache Spark is a popular data processing engine/framework that has been
|
|
|
|
architected to use third-party schedulers. The Nomad ecosystem includes a
|
|
|
|
[fork of Apache Spark](https://github.com/hashicorp/nomad-spark) that natively
|
|
|
|
integrates Nomad as a cluster manager and scheduler for Spark. When running on
|
|
|
|
Nomad, the Spark executors that run Spark tasks for your application, and
|
2017-06-29 00:10:05 +00:00
|
|
|
optionally the application driver itself, run as Nomad tasks in a Nomad job.
|
|
|
|
|
|
|
|
## Next Steps
|
|
|
|
|
2019-05-08 21:40:38 +00:00
|
|
|
The links in the sidebar contain detailed information about specific aspects of
|
2017-06-30 00:12:23 +00:00
|
|
|
the integration, beginning with [Getting Started](/guides/spark/pre.html).
|