Moonshot AI

Moonshot AI

+
+
+
+
About

Chinese AI company developing the Kimi series of large language models, including state-of-the-art mixture-of-experts models with long-context capabilities

+
+
+
+
Portfolio Stats
Total Models5
Multimodal1
Benchmarks Run95
Avg Performance73.1%
+
+
+
+
Latest Release
Kimi K2 0905
Released: Sep 5, 2025
+
+
+
+
Release Timeline
Recent model releases by year
2025
5 models
+
+
+
+
Performance Overview
Top models and benchmark performance

Top Performing Models

By avg score

Benchmark Categories

Other
95
68.8%

Model Statistics

Multimodal Ratio
20%
Models with Providers
2

All Models

Complete portfolio of 5 models with advanced filtering

LicenseLinks
#01Moonshot AIKimi K2 Instruct
Kimi K2 is a state-of-the-art mixture-of-experts (MoE) language model with 32 billion activated parameters and 1 trillion total parameters. Trained with the MuonClip optimizer, it achieves exceptional performance across frontier knowledge, reasoning, and coding tasks while being meticulously optimized for agentic capabilities. The instruct variant is post-trained for drop-in, general-purpose chat and agentic experiences without long thinking.
Jul 11, 2025
MIT
71.6%60.0%93.3%--
#02Moonshot AIKimi K2-Instruct-0905
Kimi K2-Instruct-0905 is the latest, most capable version of Kimi K2, achieving state-of-the-art performance in frontier knowledge, math, and coding among non-thinking models. This Mixture-of-Experts model features 32 billion activated parameters and 1 trillion total parameters, meticulously optimized for agentic tasks. Key features include enhanced agentic coding intelligence, extended context length to 256K tokens, and a hybrid architecture trained with MuonClip optimizer on 15.5T tokens. The model achieves 65.8% on SWE-bench Verified (single attempt), 47.3% on SWE-bench Multilingual, and excels at tool use with 70.6% on Tau2-retail. It is a reflex-grade model without long thinking, designed to act and execute complex tasks seamlessly.
Sep 5, 2025
MIT
65.8%60.0%-53.7%-
#03Moonshot AIKimi K2 0905
Kimi K2 0905 is the September update of Kimi K2 0711. It is a large-scale Mixture-of-Experts (MoE) language model developed by Moonshot AI, featuring 1 trillion total parameters with 32 billion active per forward pass. It supports long-context inference up to 256k tokens, extended from the previous 128k. This update improves agentic coding with higher accuracy and better generalization across scaffolds, and enhances frontend coding with more aesthetic and functional outputs for web, 3D, and related tasks. The model is trained with a novel stack incorporating the MuonClip optimizer for stable large-scale MoE training.
Sep 5, 2025
Proprietary
--94.5%--
#04Moonshot AIKimi K2 Base
Kimi K2 base model is a state-of-the-art mixture-of-experts (MoE) language model with 32 billion activated parameters and 1 trillion total parameters. Trained on 15.5 trillion tokens with the MuonClip optimizer, this is the foundation model before instruction tuning. It demonstrates strong performance on knowledge, reasoning, and coding benchmarks while being optimized for agentic capabilities.
Jul 11, 2025
MIT
-----
#05Moonshot AIKimi-k1.5
Kimi 1.5 is a next-generation multimodal large language model developed by Moonshot AI. It incorporates advanced reinforcement learning (RL) and scalable multimodal reasoning, delivering state-of-the-art performance in math, code, vision, and long-context reasoning tasks.
Jan 20, 2025
Proprietary
-----
+
+
+
+
Resources