MiniMax-M1 is a large-scale, open-weight reasoning model with 456B total parameters and 45.9B active per token, leveraging a hybrid Mixture-of-Experts (MoE) architecture and a custom "lightning attention" mechanism. It supports context windows up to 1 million tokens and is optimized for long-context understanding, software engineering, agentic tool use, and mathematical reasoning. The model is trained via a custom reinforcement learning pipeline (CISPO) and demonstrates strong performance on FullStackBench, SWE-bench, MATH, GPQA, and TAU-Bench.
Pricing
Pay-as-you-go rates for this model. More details can be found here.
Input Tokens (1M)
$0.40
Output Tokens (1M)
$0.96
Capabilities
Input Modalities
Text
Output Modalities
Text
Supported Parameters
Available parameters for API requests
Frequency Penalty
Logit Bias
Logprobs
Max Completion Tokens
Presence Penalty
Reasoning Effort
Response Format
Stop
Temperature
Tool Choice
Tools
Top P