Minimax
Token usage over time
Browse models from Minimax
MiniMax M2
MiniMax-M2 is a compact, efficient language model with 10B active (230B total) parameters, optimized for coding and agentic workflows. It achieves near-frontier reasoning and tool use with low latency and deployment cost. The model excels in code generation, multi-file editing, compile-run-fix cycles, and automated test repair, showing strong results on SWE-Bench and Terminal-Bench. MiniMax-M2 performs well in agentic benchmarks like BrowseComp and GAIA, handling long-term planning, retrieval, and error recovery. With a small activation footprint, it delivers fast inference and high concurrency, making it ideal for developer tools, agents, and applications that demand cost-effective, responsive reasoning.
MiniMax M1
MiniMax-M1 is a large-scale, open-weight reasoning model with 456B total parameters and 45.9B active per token, leveraging a hybrid Mixture-of-Experts (MoE) architecture and a custom "lightning attention" mechanism. It supports context windows up to 1 million tokens and is optimized for long-context understanding, software engineering, agentic tool use, and mathematical reasoning. The model is trained via a custom reinforcement learning pipeline (CISPO) and demonstrates strong performance on FullStackBench, SWE-bench, MATH, GPQA, and TAU-Bench.