Nano Chat

An open-source project implementing the complete LLM pipeline: tokenization, pre-training, fine-tuning for chat, evaluation, and a web UI for interacting with the result.

What It Does

Lets you train your own small language model from scratch for approximately $100 in GPU time. The result won’t compete with frontier models (Claude, GPT, Gemini), but gives you a model under your absolute control — custom training data, custom behavior, no external dependencies.

Why It Matters

Democratizes LLM training to the point where an individual developer can afford to experiment. The practical use cases are niche (custom domain models, research, education about how LLMs work), but the symbolic value is significant: for $100, you can understand and own the entire stack.

See Also