Merge pull request #2974 from nak3/fix-monitoring-docs
docs: fix tiny errors in monitoring spark app
This commit is contained in:
commit
fb857ceca9
|
@ -83,13 +83,13 @@ job "spark-history-server" {
|
||||||
}
|
}
|
||||||
```
|
```
|
||||||
|
|
||||||
The job file above can also be found [here](https://github.com/hashicorp/nomad/blob/f-terraform-config/terraform/examples/spark/spark-history-server.nomad).
|
The job file above can also be found [here](https://github.com/hashicorp/nomad/blob/master/terraform/examples/spark/spark-history-server-hdfs.nomad).
|
||||||
|
|
||||||
To run the history server, first [deploy HDFS](/guides/spark/hdfs.html) and then
|
To run the history server, first [deploy HDFS](/guides/spark/hdfs.html) and then
|
||||||
create a directory in HDFS to store events:
|
create a directory in HDFS to store events:
|
||||||
|
|
||||||
```shell
|
```shell
|
||||||
$ hdfs dfs -mkdir /spark-events
|
$ hdfs dfs -fs hdfs://hdfs.service.consul:8020 -mkdir /spark-events
|
||||||
```
|
```
|
||||||
|
|
||||||
You can then deploy the history server with:
|
You can then deploy the history server with:
|
||||||
|
@ -101,7 +101,7 @@ $ nomad run spark-history-server-hdfs.nomad
|
||||||
You can get the private IP for the history server with a Consul DNS lookup:
|
You can get the private IP for the history server with a Consul DNS lookup:
|
||||||
|
|
||||||
```shell
|
```shell
|
||||||
$ dig.spark-history.service.consul
|
$ dig spark-history.service.consul
|
||||||
```
|
```
|
||||||
|
|
||||||
Find the public IP that corresponds to the private IP returned by the `dig`
|
Find the public IP that corresponds to the private IP returned by the `dig`
|
||||||
|
|
Loading…
Reference in a new issue