This page collects the public integration surface for the model: supported endpoints, available request parameters, and example calls through the NagaAI API.
Chat CompletionsMiniMax-M1 is a large-scale, open-weight reasoning model with 456B total parameters and 45.9B active per token, leveraging a hybrid Mixture-of-Experts (MoE) architecture and a custom "lightning attention" mechanism. It supports context windows up to 1 million tokens and is optimized for long-context understanding, software engineering, agentic tool use, and mathematical reasoning. The model is trained via a custom reinforcement learning pipeline (CISPO) and demonstrates strong performance on FullStackBench, SWE-bench, MATH, GPQA, and TAU-Bench.