An Ansible role for installing Apache Spark.
spark_version
- Spark version.spark_install_java
- flag toggling the JDK installation using the builtin azavea.java role dependency (default:yes
)spark_cloudera_distribution
- Cloudera distribution version (default:cdh5.4
)spark_symlinks_enabled
(defaultyes
) - ifyes
deploy 2 symlinks (<spark_home>/conf -> /etc/spark/conf ; <spark_home> ->spark_usr_dir
)spark_shims_enabled
(defaultyes
) - ifyes
deploy the shims (like/usr/bin/spark-shell
,/usr/bin/spark-submit
)spark_env_extras
- An optional dictionary with key and value attributes to add tospark-env.sh
(e.g.MESOS_NATIVE_LIBRARY: "/usr/local/lib/libmesos.so"
)spark_defaults_extras
- An optional dictionary with key and value attributes to add tospark-defaults.conf
(e.g."spark.eventLog.enabled": true
)spark_user_groups
- an optional list of (OS)groups the spark user should belong tospark_user_shell
- the spark user's default shell (default:/bin/false
)spark_log4j_loggers
- A list of dictionaries configuring the spark log4j loggers (default: logger config from spark/conf/log4j.properties.template)
See the examples directory.