Skip to content

An Ansible role for installing Apache Spark.

License

Notifications You must be signed in to change notification settings

lhoss/ansible-spark

 
 

Folders and files

NameName
Last commit message
Last commit date

Latest commit

 

History

59 Commits
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 

Repository files navigation

ansible-spark

An Ansible role for installing Apache Spark.

Role Variables

  • spark_version - Spark version.
  • spark_install_java - flag toggling the JDK installation using the builtin azavea.java role dependency (default: yes)
  • spark_cloudera_distribution - Cloudera distribution version (default: cdh5.4)
  • spark_symlinks_enabled (default yes) - if yes deploy 2 symlinks (<spark_home>/conf -> /etc/spark/conf ; <spark_home> -> spark_usr_dir )
  • spark_shims_enabled (default yes) - if yes deploy the shims (like /usr/bin/spark-shell, /usr/bin/spark-submit)
  • spark_env_extras - An optional dictionary with key and value attributes to add to spark-env.sh (e.g. MESOS_NATIVE_LIBRARY: "/usr/local/lib/libmesos.so")
  • spark_defaults_extras - An optional dictionary with key and value attributes to add to spark-defaults.conf (e.g. "spark.eventLog.enabled": true)
  • spark_user_groups - an optional list of (OS)groups the spark user should belong to
  • spark_user_shell - the spark user's default shell (default: /bin/false)
  • spark_log4j_loggers - A list of dictionaries configuring the spark log4j loggers (default: logger config from spark/conf/log4j.properties.template)

Example Playbook

See the examples directory.

About

An Ansible role for installing Apache Spark.

Resources

License

Stars

Watchers

Forks

Packages

No packages published

Languages

  • Shell 100.0%