This guide was created for the students of all the Big Data courses at Politecnico di Torino. This guide must be used only in case you want to run your application without using the distributed cluster. On the distributed cluster, everything is already configured and ready to run Spark application. Please follow this guide only in case you want to execute applications on your own, without the distributed cluster.
It will guide you on the installation of JupyterLab + PySpark.
Three options are available to run your code locally:
Requirement: you need a Google account to be able to use Google Colab. Internet connection is required to run the code.
This is the easiest solution and it does not require any configuration.
Steps:
- Go to the Colab's entry page (link)
- Create a new notebook
- (optional) Verify that no Accellerator (GPU/TPU) is being used
- Runtime -> Change runtime type -> Accelerator: None
- Connect to a runtime by going into the top-right corner of the notebook and clicking
Connect - After connection, Colab should report RAM and Disk usage
- Install PySpark by running a cell with the following code:
!pip install pyspark - Install and configure Java by running a cell with the following code:
import os
!apt-get install openjdk-8-jdk-headless -qq > /dev/null
os.environ["JAVA_HOME"] = "/usr/lib/jvm/java-8-openjdk-amd64"
- Instantiate a SparkSession by running the following code in a cell:
from pyspark.sql import SparkSession
spark = SparkSession.builder.getOrCreate()
sc = spark.sparkContext
You can find an example here.
To upload sample files on Google Colab, you can either:
!wget download_link
To locally install PySpark on your own computer, we suggest you to use Anaconda/Miniconda.
If you do not have Miniconda installed on your system, you can find instructions and installers at this link.
Same can be performed using pip and virtualenv.
Steps (after installation):
- (optional) Create a conda environment by opening a terminal (Anaconda/Miniconda terminal on Windows)
conda create -n pyspark python=3.10
- Activate the newly created environment
conda activate pyspark
- Install JupyterLab, findspark and PySpark
conda install -c conda-forge jupyterlab pyspark findspark
Alternatively, if you do not have conda install on your system, you can use pip and venv.
With pip install pyspark you are able to install pyspark in the activate environment.
To open the JupyterLab web interface, you need to:
- open a terminal (Anaconda/Miniconda terminal on windows)
- go to the directory in which you want to start and save your notebook
cd your_path
- start JupyterLab
conda activate pyspark
jupyter lab
If you prefer to develop your python application in a simple python script, copy these two lines of code at the very beginning of every python script. You can run the application by pressing on the "Run Python file" in top right corner of VSCode.
import findspark
findspark.init()
If you want to use LABINF's computer, you need to install PySpark locally.
- open a terminal (Anaconda terminal on windows)
- install pyspark
pip install pyspark
Then, you can launch JupyterLab by following the Open the JupyterLab web interface section, skipping the conda activate command.
One last additional step must be followed when running your application in LABINF:
the PYSPARK_PYTHON environment variable must be set. You can simply write the following code in the first cell and execute it:
import os
os.environ['PYSPARK_PYTHON'] = 'C:\\tools\\Anaconda3\\python.exe'
You can find an example here.
Always remember to shut down your jupyter server. You can do that by going into File -> Shutdown as shown here.

