Compact and Agent-Native MoE Training System
Production MoE training frameworks deliver peak throughput and broad model coverage, but evolving them for new architectures or system optimizations remains expensive. Meanwhile, the design patterns that make these stacks work for humans, such as plugin systems, registry-based indirection, and heavy compiled extensions, are harder for AI coding agents to navigate.
PithTrain is an MoE training framework designed agent-native from the start: ~10K lines of Python, no implicit indirection, with shipped agent skills for recurring tasks. It delivers production-grade performance, including 4D parallelism, compute-communication overlap, and FP8 training, in a codebase compact enough that an agent (or a human) can read it end-to-end.
NVIDIA Hopper (SM90) or Blackwell (SM100) GPUs are required. CUDA 13.0 and Python >= 3.12 are required. We use uv to manage project dependencies.
git clone https://github.com/mlc-ai/Pith-Train.git && cd Pith-Train
uv venv # skip if you already have a virtual environmentFor users:
uv pip install .For developers:
uv syncPretrain Qwen3-30B-A3B from scratch. Datasets and checkpoints are stored in the workspace folder by default. Other models like DeepSeek-V2-Lite follow the same steps. See examples for available configurations.
1. Prepare the dataset
bash examples/build_tokenized_corpus/launch.sh dclm-qwen3Download and tokenize the DCLM pretraining corpus into mmap-friendly packed sequences. Each model uses its own tokenizer, so switching to a different model requires running this step again.
2. Configure training
Edit examples/pretrain_language_model/qwen3-30b-a3b/script.py to adjust parallelism, batch size, learning rate, and other hyperparameters. The model architecture is defined in the accompanying config.json.
3. Launch training
bash examples/pretrain_language_model/launch.sh qwen3-30b-a3bThe launch script auto-detects GPUs and supports both single-node and multi-node (SLURM) setups. Training resumes from the latest checkpoint automatically, and checkpoints are reshardable across different parallelism.
4. Export checkpoint
bash examples/convert_checkpoint/launch.sh qwen3-30b-a3bConvert a training checkpoint to standard Hugging Face format for evaluation or inference. The same tool also supports importing Hugging Face checkpoints for continued pretraining.
PithTrain is structured in three layers:
- Application — Training loop for pretraining, SFT, and more.
- Engine — The bulk of PithTrain, composed of five modules:
- Model — Protocol interface with implementations for Qwen, DeepSeek, and GPT-OSS architectures.
- Building Blocks — FP8 linear and quantization, ring attention, expert dispatch and deduplication, etc.
- Pipeline Engine — DualPipeV scheduler with 5-stage overlapped forward-backward execution and P2P communication.
- Distributed Training — Pipeline, data, context, and expert parallelism (PP x FSDP x CP x EP).
- Training Infrastructure —
torch.compile, optimizer and LR scheduling, checkpointing, logging, etc.
- Operator — PyTorch (basic ops, NCCL), operator libraries (DeepGEMM, FlashAttention), and Python DSLs (Triton).
PithTrain is developed by contributors from CMU. It is built on top of DeepSeek's DualPipe, which provides the original pipeline parallelism schedule and examples. We thank the CMU Foundation and Language Model (FLAME) Center for providing the compute resources to develop PithTrain. We also acknowledge the support of DGX B200 from NVIDIA.
PithTrain is released under the Apache 2.0 License.