A collection of Jupyter notebooks for training purposes, designed to teach the fundamentals of transformer-based NLP models and practical machine learning workflows.
An introduction to how transformer models process text end-to-end. Covers tokenization, inference with pre-trained models, hidden states, attention mechanisms, and sentiment analysis using DistilBERT.
A side-by-side comparison of two training approaches on the MRPC paraphrase detection task:
- Fine-tuning a pre-trained BERT model with the HuggingFace
TrainerAPI. - Full training with a custom PyTorch loop including manual backpropagation, optimizer steps, and learning rate scheduling.
Builds a text classifier on the 20 Newsgroups dataset (18,000 posts across 20 topics) using Google Gemini API embeddings fed into a simple Keras neural network.
This repository is intended solely for educational and training purposes. The notebooks, code, and examples provided here are meant to help learners understand machine learning concepts and should not be used as-is in production environments.