sahara-image-elements/elements/spark
Pino Toscano 2fb2422da8 Make almost all the element scripts as e/u/pipefail
Following the latest dib-lint reporting, make sure almost all the
scripts are enabling -e, -u, and pipefail.  This eases the discovery of
failing commands, and the usage of unset variables.

There are few exceptions with flags not set:

* elements/hadoop-hdp/install.d/40-install-hdp, set -e
  when installing HDP 2.0, it tries to install tez, which does not seem
  to exist in Hortonworks' repositories

* elements/ssh/install.d/33-ssh, set -e
  the version of augtool (part of augeas 0.10) in older Ubuntu versions
  (like Precise, needed by the cloudera plugin) exits with wrong return
  values in autosave mode

* elements/storm/install.d/60-storm, set -e
  It tries to change the permission of /etc/supervisord.conf, which does
  not seem to exist

Change-Id: Ic1314639dfc6a66c48ca87b6820707a2b0cb1dbd
Partial-Bug: #1435306
2015-04-07 12:46:39 +02:00
..
install.d Make almost all the element scripts as e/u/pipefail 2015-04-07 12:46:39 +02:00
post-install.d Make almost all the element scripts as e/u/pipefail 2015-04-07 12:46:39 +02:00
root.d Make almost all the element scripts as e/u/pipefail 2015-04-07 12:46:39 +02:00
README.rst Add/fix headers in README.rst files 2015-04-02 18:15:30 +02:00
element-deps Add a Spark element 2014-02-06 11:48:54 +01:00

README.rst

spark

Installs Spark on Ubuntu. Requires Hadoop CDH 4 (hadoop-cdh element).

It will install a version of Spark known to be compatible with CDH 4 This behaviour can be controlled also by using 'DIB_SPARK_VERSION' or directly with 'SPARK_DOWNLOAD_URL'

If you set 'SPARK_CUSTOM_DISTRO' to 1, you can point the 'SPARK_DOWNLOAD_URL' variable to a custom Spark distribution created with the make-distribution.sh script included in Spark.