Share
Created: 15 Oct 2025

nanochat: Build Your Own ChatGPT Clone

Minimal, from scratch, full-stack training/inference pipeline in ~8,000 lines of clean code

What is nanochat?

Complete Training Stages

Advanced Training Features

Infrastructure & Deployment

Performance Benchmarks

For as low as ~$100 in cost (~4 hours on an 8XH100 node), you can train a little ChatGPT clone that you can kind of talk to, and which can write stories/poems, answer simple questions.

About ~12 hours surpasses GPT-2 CORE metric.

As you scale up towards ~$1000 (~41.6 hours of training), it becomes more coherent and can solve simple math/code problems and take multiple choice tests.

A depth 30 model trained for 24 hours gets into 40s on MMLU and 70s on ARC-Easy, 20s on GSM8K, etc.

💰 Training Cost Breakdown

🎯 nanochat Goals & Features

Repository Features

  • Capstone project of LLM101n (still being developed)

  • Potential to grow into research harness or benchmark, similar to nanoGPT

  • Cohesive, minimal, readable, hackable, maximally forkable repo

  • Not finished, tuned or optimized - likely quite a bit of low-hanging fruit

  • Generates single markdown report card, summarizing and gamifying the whole thing

  • Not finished, tuned or optimized - likely quite a bit of low-hanging fruit

Chris Prakoso - Augmented Humanity | Practical AI, Data & Analytics

Chris Prakoso

Augmented Humanity | Practical AI, Data & Analytics

Connect with me for cutting-edge AI/ML insights, hands-on LLM tutorials, and the latest in open-source machine learning development