Skip to content

Experimental small LLM project for learning and testing—outputs are often nonsensical.

Notifications You must be signed in to change notification settings

Amjuks/LLM-Experimental

Repository files navigation

Small LLM Project

This project demonstrates how to build, train, and interact with a small Large Language Model (LLM) using a custom dataset and tokenizer.

Overview

  • Custom Training: Train a language model on your own text data with a custom tokenizer.
  • Model Checkpoints: Save and reuse model checkpoints for evaluation or further training.
  • Web Demo: Interact with the trained model through a simple web interface.
  • Evaluation: Test and explore the model’s outputs using scripts or notebooks.

How It Works

  1. Prepare Data: Process your text data and set up a tokenizer.
  2. Train Model: Run the training script to build your LLM.
  3. Demo & Inference: Use the web app or scripts to interact with the model and see results.

Features

  • End-to-end pipeline: data prep, training, inference, and demo.
  • Customizable for different datasets and tasks.
  • Simple web interface for easy testing.

About

Experimental small LLM project for learning and testing—outputs are often nonsensical.

Topics

Resources

Stars

Watchers

Forks

Releases

No releases published

Packages

No packages published