


In standalone mode, Spark will have a master node (which is the cluster manager) and worker nodes. If you will not be managing Spark using the Mesos or YARN cluster managers, you'll be running Spark in what is called standalone mode. export SPARK_LOG_DIR=/var/log/sparĮxport SPARK_PID_DIR=$/run Configure Spark Nodes to Join the Cluster $ sudo chown spark:root /usr/local/spark/conf/spark-*Įdit the Spark Environment file spark-env.sh. $ sudo cp /usr/local/spark/conf/ /usr/local/spark/conf/nf $ sudo -u spark mkdir $SPARK_HOME/run Create the Spark Configuration FilesĬreate the Spark Configuration files by copying the templates $ sudo cp /usr/local/spark/conf/spark-env.sh.template /usr/local/spark/conf/spark-env.sh $ sudo chown -R spark:root /usr/local/sparkĬreate LOG and PID directories: $ sudo mkdir /var/log/spark $ source /etc/environmentĬreate a Spark user and make it the owner of the SPARK_HOME directory: $ sudo adduser spark -system -home /usr/local/spark/ -disabled-password
#How to install spark ubuntu update
$ sudo mv spark-2.0.2-bin-hadoop2.7/ /usr/local/spark/ Update System Variables $ sudo nano /etc/environmentĪdd an environment variable called SPARK_HOME: export SPARK_HOME=/usr/local/sparkĪt the end of the PATH variable, add $SPARK_HOME/bin: PATH=":/usr/local/spark/bin" $ sudo apt-get install sbt Install Spark 2.0ĭownload Spark 2.0 from this link and unpack the TAR file: $ wget $ scala -version Install SBT 0.13 $ echo "deb /" | sudo tee -a /etc/apt//sbt.list Install Scala 2.11.8 $ wget $ sudo dpkg -i scala-2.11.8.deb If you don't have Java installed, follow this tutorial to get Java 8 installed. Java HotSpot(TM) 64-Bit Server VM (build 25.91-b14, mixed mode) Java(TM) SE Runtime Environment (build 1.8.0_91-b14) Install Scala 2.11Įnsure that you have Java installed. Note: The following steps should be performed on all the nodes in the cluster unless otherwise noted. For more information on installing and using Cassandra, visit this site. This guide assumes you have a Cassandra 3.x cluster that is already up and running.
#How to install spark ubuntu software
If you have any of these software packages installed and configured already, you can skip that step. In this guide, we will be installing Scala 2.11, Spark 2.0 as a service, and the DataStax spark-cassandra-connector library on the client program.
