Hugging Face Releases Full SmolLM3 Training Stack and Checkpoints
Hugging Face has open-sourced the full training and evaluation pipeline for SmolLM3, its latest small-scale language model. The release includes more than 100 intermediate checkpoints and supports dual-mode reasoning, multilingual tasks, and long-context inference. This marks a major move toward transparency and reproducibility in compact AI models.
Hugging Face has published the complete training stack for SmolLM3, a 3 billion parameter language model designed for resource-efficient inference. The open repository includes scripts for pretraining using Nanotron, post-training alignment with SFT and Adaptive Preference Optimization, and a full suite of evaluation tools. Also included are over 100 training checkpoints, allowing researchers to audit or resume training at any point in the model’s lifecycle.
This release positions SmolLM3 as a viable alternative in the small-to-medium LLM space. Unlike proprietary models, SmolLM3 is fully open and optimized for multilingual (6 languages) and long-context scenarios, with context lengths up to 128,000 tokens. The transparency of its training pipeline makes it particularly valuable for researchers and engineers building or fine-tuning models for on-device or low-latency environments. Hugging Face’s decision to make every layer of SmolLM3 visible—from data to training to alignment—signals a deep commitment to reproducibility in the open-source AI ecosystem.
Pure Neo Signal:
We love
and you too
If you like what we do, please share it on your social media and feel free to buy us a coffee.