- Python 3
- Recommended: Anaconda
- Get the latest release from the Apache Spark downloads page and unpack it.
- That's it - check that it works:
cd spark-2.4.0-bin-hadoop2.7/
bin/pyspark
You should see output like this:
Python 3.6.3 |Anaconda custom (64-bit)| (default, Oct 6 2017, 12:04:38)
[GCC 4.2.1 Compatible Clang 4.0.1 (tags/RELEASE_401/final)] on darwin
Type "help", "copyright", "credits" or "license" for more information.
2019-01-09 10:56:19 WARN NativeCodeLoader:62 - Unable to load native-hadoop library for your platform... using builtin-java classes where applicable
Setting default log level to "WARN".
To adjust logging level use sc.setLogLevel(newLevel). For SparkR, use setLogLevel(newLevel).
2019-01-09 10:56:21 WARN Utils:66 - Service 'SparkUI' could not bind on port 4040. Attempting port 4041.
Welcome to
____ __
/ __/__ ___ _____/ /__
_\ \/ _ \/ _ `/ __/ '_/
/__ / .__/\_,_/_/ /_/\_\ version 2.4.0
/_/
Using Python version 3.6.3 (default, Oct 6 2017 12:04:38)
SparkSession available as 'spark'.
>>>
- Clone the git repository on your laptop:
git clone https://github.com/EPCCed/prace-spark-for-data-scientists.git
- We're setting up an environment variable for the Spark installation directory (replace [INSTALLATION_PATH] below with the path of your installation):
export SPARK_HOME=[INSTALLATION_PATH]/spark-2.4.0-bin-hadoop2.7/
- Configure the environment for PySpark to use Jupyter notebooks:
export PYSPARK_DRIVER_PYTHON=jupyter
export PYSPARK_DRIVER_PYTHON_OPTS="notebook"
- Start PySpark with a Jupyter notebook server:
cd prace-spark-for-data-scientists
$SPARK_HOME/bin/pyspark