Apple Unveils New Generation of Multilingual, Multimodal Foundation Models
2025-07-18

Apple introduced two new multilingual, multimodal foundation language models powering its on-device and server-side intelligence features. A ~3B parameter on-device model, optimized for Apple silicon, and a scalable server model built on a novel Parallel-Track Mixture-of-Experts (PT-MoE) transformer. Both are trained on massive multilingual and multimodal datasets, refined with supervised fine-tuning and reinforcement learning. They support more languages, image understanding, and tool calls, matching or exceeding comparable open-source baselines. A new Swift-centric framework simplifies integration for developers.
AI