This project implements a basic transformer-based language model from scratch.
- Tokenization
- Embedding Layer
- Positional Encoding
- Self-Attention
- Transformer Block
- Full Language Model
- Install requirements:
pip install -r requirements.txt - Train model:
python train.py - Run predictions:
python predict.py