-
Notifications
You must be signed in to change notification settings - Fork 1
This project is a minimal GPT implementation focused on a question-answering system. The model was trained from scratch on the WikiText-2 dataset without leveraging transfer learning or pretrained models, resulting in limited performance. It serves as an educational experiment to explore the fundamentals of training language models from scratch.
dipanjanpathak/mini-gpt-model
Folders and files
| Name | Name | Last commit message | Last commit date | |
|---|---|---|---|---|
Repository files navigation
About
This project is a minimal GPT implementation focused on a question-answering system. The model was trained from scratch on the WikiText-2 dataset without leveraging transfer learning or pretrained models, resulting in limited performance. It serves as an educational experiment to explore the fundamentals of training language models from scratch.
Stars
Watchers
Forks
Releases
No releases published
Packages 0
No packages published