Experimental small LLM project for learning and testing—outputs are often nonsensical.
-
Updated
Jun 12, 2025 - Python
Experimental small LLM project for learning and testing—outputs are often nonsensical.
A lightweight, end-to-end implementation of Stable Diffusion built from first principles on a single T4 GPU. Features a custom 192-channel U-Net, VAE, and a CLIP encoder, optimized for consumer hardware and trained on approx. 168k images.
A from-scratch PyTorch LLM implementing Sparse Mixture-of-Experts (MoE) with Top-2 gating. Integrates modern Llama-3 components (RMSNorm, SwiGLU, RoPE, GQA) and a custom-coded Byte-Level BPE tokenizer. Pre-trained on a curated corpus of existential & dark philosophical literature.
Add a description, image, and links to the custom-tokenizer topic page so that developers can more easily learn about it.
To associate your repository with the custom-tokenizer topic, visit your repo's landing page and select "manage topics."