Minimax

Token usage over time

Browse models from Minimax

2 models

MiniMax M2

11.3M Tokens

MiniMax-M2 is a compact, efficient language model with 10B active (230B total) parameters, optimized for coding and agentic workflows. It achieves near-frontier reasoning and tool use with low latency and deployment cost. The model excels in code generation, multi-file editing, compile-run-fix cycles, and automated test repair, showing strong results on SWE-Bench and Terminal-Bench. MiniMax-M2 performs well in agentic benchmarks like BrowseComp and GAIA, handling long-term planning, retrieval, and error recovery. With a small activation footprint, it delivers fast inference and high concurrency, making it ideal for developer tools, agents, and applications that demand cost-effective, responsive reasoning.

byMinimax
$0.07/1M input tokens$0.22/1M output tokens

MiniMax M1

659K Tokens

MiniMax-M1 is a large-scale, open-weight reasoning model with 456B total parameters and 45.9B active per token, leveraging a hybrid Mixture-of-Experts (MoE) architecture and a custom "lightning attention" mechanism. It supports context windows up to 1 million tokens and is optimized for long-context understanding, software engineering, agentic tool use, and mathematical reasoning. The model is trained via a custom reinforcement learning pipeline (CISPO) and demonstrates strong performance on FullStackBench, SWE-bench, MATH, GPQA, and TAU-Bench.

byMinimax
$0.40/1M input tokens$0.96/1M output tokens