China's Moonshot AI has released Kimi K2.5, a groundbreaking open-source multimodal model trained on 15 trillion visual and text tokens, excelling in coding, vision, and agent swarms. The model outperforms proprietary rivals like GPT 5.2 and Gemini 3 Pro in key benchmarks and introduces Kimi Code for developers. Available now via Kimi.com, API, and tools like VSCode, it marks a leap toward AGI in open-source AI.
Highlights
- Native multimodal capabilities for text, images, and videos with 256K context length
- Outperforms GPT 5.2 and Gemini 3 Pro on SWE-Bench coding and VideoMMU benchmarks
- Introduces Kimi Code, an open-source coding agent integrable with VSCode and Cursor
- Agent Swarm mode coordinates up to 100 sub-agents for 4.5x faster complex tasks
- Backed by Alibaba and available on NVIDIA NIM and Hugging Face