Moonshot AI’s Kimi K2.5 Expands What Open-Weight Models Can Do
Moonshot AI, a Beijing-based AI startup known for releasing large open-weight language models aimed at coding and agentic tasks, has released Kimi K2.5, the latest iteration of its open-source Kimi model family.
The company says Kimi K2.5 extends the foundation of the earlier Kimi K2 model with significantly expanded multimodal training and new agentic execution capabilities. The model was pretrained on a staggering 15 trillion mixed visual and text tokens and is described as a native multimodal model, meaning images and text are trained together in a unified architecture, rather than being handled by separate models stitched together after the fact. In a technical blog, Moonshot describes three areas where Kimi K2.5 builds on its predecessor: coding with vision, scaled agent execution, and office productivity tasks in knowledge work.
For software development, Moonshot says Kimi K2.5 has particularly strong capabilities in front-end workflows. The model can turn simple conversational prompts into complete interface code, including interactive layouts and animation effects. Beyond text prompts, the blog highlights K2.5’s ability to reason over images and video as part of the coding process, enabling image- and video-to-code generation as well as visual debugging. In these scenarios, visual inputs help the model interpret user intent, check its outputs against visual references, and refine layouts or interactions, allowing users to guide development with images or video rather than spelling out every design detail in text.
To assess real-world coding performance, Moonshot uses its internal Kimi Code Bench, which spans typical end-to-end engineering tasks including building, debugging, refactoring, testing, and scripting across multiple programming languages. On these internal benchmarks, the company reports “consistent and meaningful improvements” over the previous K2 generation.
(VRVIRUS/Shutterstock)
Another notable change in Kimi K2.5 is the addition of Agent Swarm, which Moonshot describes as a research preview exploring parallel, multi-agent execution. Rather than relying on a single agent to work through tasks step by step, Kimi K2.5 can autonomously spin up and coordinate as many as 100 sub-agents to tackle different parts of a problem at the same time, managing workflows that span up to 1,500 coordinated steps. A trainable “orchestrator” agent is responsible for breaking tasks into parallelizable subtasks and managing their execution, an approach Moonshot says is designed to reduce latency compared to sequential agent pipelines. To prevent the system from defaulting back to single-agent behavior, the orchestrator is trained using a parallel-agent reinforcement learning method that explicitly rewards early parallelism before shifting focus to overall task quality. In internal evaluations, Moonshot reports that Agent Swarm reduced end-to-end execution time by roughly 3 to 4.5x on complex, highly parallelizable workflows.
Beyond coding and agentic execution, Moonshot says Kimi K2.5 is capable of handling more complex knowledge-work workflows. The blog describes the model as able to reason over large, high-density inputs and coordinate multi-step tool use to produce work such as documents, spreadsheets, PDFs, and slide decks. Moonshot evaluates these capabilities using two internal benchmarks focused on professional tasks, measuring both end-to-end output quality and multi-step workflow execution. On those benchmarks, the company reports Kimi K2.5 outperforms its predecessor, K2 Thinking, reflecting stronger performance on structured office tasks.
Kimi K2.5 is available through Moonshot’s Kimi.com web interface, mobile app, API, and its developer-focused Kimi Code environment. On the web and app, users can choose from four modes: Instant, Thinking, Agent, and Agent Swarm. Agent Swarm is currently offered as a beta feature and includes free usage credits for higher-tier paid users. Read more about Kimi K2.5 here.
(Source: NicoElNino/Shutterstock)
While many frontier models remain proprietary and closed, there has been considerable momentum behind open-weight systems, as Kimi’s continued development exemplifies. By releasing model weights and detailed technical documentation, Moonshot AI is enabling independent researchers, startups, and enterprise users to inspect how the system is trained, adapt it to specialized workloads, and deploy it on their own infrastructure rather than relying solely on hosted APIs. This openness matters for cost control, data governance, and scientific reproducibility, particularly for organizations operating under regulatory or sovereignty constraints.
At a geopolitical level, Kimi is also an example of how Chinese AI developers are using open models to build global trust and relevance at a time when access to advanced hardware and proprietary Western models has been more restricted. How widely Kimi K2.5 will be adopted remains to be seen, but the release shows how open-weight models are being extended to handle more complex work and could be rapidly closing the gaps with proprietary systems.
Related

