Alibaba Unveils Qwen2.5-Max: A Massive MoE Language Model
2025-01-28
Alibaba has released Qwen2.5-Max, a large-scale Mixture-of-Experts (MoE) model pre-trained on over 20 trillion tokens and further refined with supervised fine-tuning and reinforcement learning from human feedback. Benchmarks like MMLU-Pro, LiveCodeBench, LiveBench, and Arena-Hard show Qwen2.5-Max outperforming models such as DeepSeek V3. The model is accessible via Qwen Chat and an Alibaba Cloud API. This release represents a significant advancement in scaling large language models and paves the way for future improvements in model intelligence.