Train your own ChatGPT on Apple Silicon — MLX port of nanochat
-
Updated
Apr 29, 2026 - Python
Train your own ChatGPT on Apple Silicon — MLX port of nanochat
Train Llama 3 models from scratch. Any scale, any personality. By Arianna Method.
A minimal, hackable Vision-Language Model built on Karpathy’s nanochat — add image understanding and multimodal chat for under $200 in compute.
The best GPT that $100-$125 worth of pre-training and finetuning can buy
Ascend NPU fork of nanochat for LLM training with torch_npu/HCCL (experimental)
nanochat's inference engine re-vibed in C++ with GGML.
JAX/Flax NNX port of Karpathy's nanochat, optimized for multi-host Cloud TPU pods.
I built this repo to prove to my granny that I can implement GPT.
Rust-first continuation of nanochat-rs, benchmarked against karpathy/nanochat for faster, reproducible LLM training, evaluation, and inference.
Controlled benchmark of memory mechanisms for transformers, built on nanochat
AI agents running research on single-GPU nanochat training automatically
A real LLM chat app that runs natively on Windows XP / Pentium 4, packaged as a single portable EXE with a custom C+SSE2 inference engine and Win32 GUI.
Run nanochat training efficiently on Huawei Ascend NPUs with minimal code changes, supporting tokenizer, pretraining, and evaluation workflows.
Add a description, image, and links to the nanochat topic page so that developers can more easily learn about it.
To associate your repository with the nanochat topic, visit your repo's landing page and select "manage topics."