- Get ReqlTime data Stream from facebook
- Apply Filters on That data
- Prdouce Filtered Data to kafka
- Consume Data with spark
- Apply Aggregate Function on that data
1- Start Docker Container of Kafka docker-compose --project-directory=".\docker\" up -d
2- get new token from Graph API
- token must have user_likes permission
3- start either producer or consumer
- make sure that you installed dependencies inside producer
npm i, inside pythonSparkpip install -r requirements.txt
3.1- Set Token in env (windows powershell) $env:TOKEN='<Your-Token>'
3.2- Starting Producer node .\producer\producer.js
3.3- start consumer python .\spark\pythonSpark\main.py
