SmolLM3: A Tiny, Multilingual, Long-Context Reasoner

2025-07-09
SmolLM3: A Tiny, Multilingual, Long-Context Reasoner

SmolLM3 is a fully open-source 3B parameter multilingual language model that strikes a compelling balance between efficiency and performance. Outperforming Llama-3.2-3B and Qwen2.5-3B on various benchmarks, it even competes with larger 4B parameter models. Supporting 6 languages and boasting a context length of up to 128k tokens, SmolLM3 features a unique dual-mode reasoning capability (think/no_think). Beyond the model itself, the researchers are releasing the complete engineering blueprint, including architecture details, data mixtures, and training methodology—a valuable resource for anyone building or studying models at this scale.