Kimi K2 0905

kimi-k2-0905
byMoonShotAI|Created Sep 5, 2025
Chat Completions

Kimi K2 0905 is the September update of Kimi K2 0711, a Mixture-of-Experts (MoE) language model from Moonshot AI with 1 trillion parameters and 32 billion active per pass. The long-context window has been expanded to 256k tokens. This release brings improved agentic coding accuracy and generalization across scaffolds, as well as more aesthetic and functional frontend code for web, 3D, and similar tasks. Kimi K2 remains optimized for advanced tool use, reasoning, and code synthesis, excelling in benchmarks like LiveCodeBench, SWE-bench, ZebraLogic, GPQA, Tau2, and AceBench. Its training uses a novel stack with the MuonClip optimizer for stable large-scale MoE training.

Pricing

Pay-as-you-go rates for this model. More details can be found here.

Input Tokens (1M)

$0.07

Output Tokens (1M)

$1.24

Capabilities

Input Modalities

Text

Output Modalities

Text

Supported Parameters

Available parameters for API requests

Frequency Penalty
Logit Bias
Logprobs
Max Completion Tokens
Parallel Tool Calls
Presence Penalty
Response Format
Stop
Temperature
Tool Choice
Tools
Top P

Usage Analytics

Token usage across the last 30 active days

Uptime

Reliability over the last 7 days

Throughput

Time-To-First-Token (TTFT)

Code Example

Example code for using this model through our API with Python (OpenAI SDK) or cURL. Replace placeholders with your API key and model ID.

Basic request example. Ensure API key permissions. For more details, see our documentation.

from openai import OpenAI

client = OpenAI(
    base_url="https://api.naga.ac/v1",
    api_key="YOUR_API_KEY",
)

resp = client.chat.completions.create(
    model="kimi-k2-0905",
    messages=[
        {{"role": "user", "content": "What's 2+2?"}}
    ],
    temperature=0.2,
)
print(resp.choices[0].message.content)