DeepLaugh is a study / analysis that explores the science behind the art of comedy, leveraging the captions and audio to identify triggers of laughter in comedy sets
Following File Downloads YouTube audio and captions and saves them in a folder
- File: Subtitle Scraper.ipynb
Following File Converts the Downloaded Audio into MP3 Format
- File: File: Mp3 Converter.ipynb
Following File Cleans the Downloaded Videos to get rid of excess/unnecessary formatting
File does some basic EDA and creates graphs (mins per video and laughs per video)
- File: Video_Clip_EDA.ipynb
This notebook processes the audio data to form that can be consumed by tree based models. All details of feature engineering can be found here. Audio files used as the raw input can be found here
This notebook walks through the execution of AST transformer to predict laughter points.
This notebook walks through the implementation of the DistillBERT model for text, and combines it with the feature based model developed for audio
Notebook with implementation of training of the whisper-small model to predict laughter. Seen in Future Work section of presentation. Uses files: processed_targets_output.csv and mp3 audio files
- File: laughterAudio.ipynb
This project was done as a part of capstone project for the course: Adavanced Machine Learning (MIS382N) instructed by Prof. Joydeep Ghosh. Contributors are Kedar Godbole, Alex Imhoff, Tara Mary Joseph, Jyoti Kumari, Aman Sharma