MoonshotAI: Kimi K2 0905

Kimi-K2-0905

Kimi K2 0905 is the September update to Kimi K2 0711, a large-scale Mixture-of-Experts (MoE) model developed by Moonshot AI. It features 1 trillion total parameters with 32B active per forward pass and extends long-context inference from 128K to 256K tokens.

This release enhances agentic coding with improved accuracy and better generalization across scaffolds, while also boosting frontend development with more refined and functional outputs for web, 3D, and related applications. Optimized for agentic capabilities—spanning advanced tool use, reasoning, and code synthesis—Kimi K2 continues to excel across benchmarks in coding (LiveCodeBench, SWE-bench), reasoning (ZebraLogic, GPQA), and tool use (Tau2, AceBench). Training is powered by a novel stack that incorporates the MuonClip optimizer for stable, large-scale MoE performance.

Conversations

Download TXT
Download PDF
CreatorMoonshot AI
Release DateSeptember, 2025
LicenseModified MIT License
Context Window262,144
Image Input SupportNo
Open Source (Weights)Yes
Parameters1000B, 32.0B active at inference time
Model WeightsClick here

Performance Benchmarks

BenchmarkMetricK2-Instruct-0905K2-Instruct-0711Qwen3-Coder-480B-A35B-InstructGLM-4.5DeepSeek-V3.1Claude-Sonnet-4Claude-Opus-4
SWE-Bench verifiedACC69.2 ± 0.6365.869.6*64.2*66.0*72.7*72.5*
SWE-Bench MultilingualACC55.9 ± 0.7247.354.7*52.754.5*53.3*
Multi-SWE-BenchACC33.5 ± 0.2831.332.731.729.035.7
Terminal-BenchACC44.5 ± 2.0337.537.5*39.9*31.3*36.4*43.2*
SWE-DevACC66.6 ± 0.7261.964.763.253.367.1