MiniMax M2 benchmark results
Compare MiniMax M2 benchmark results across hosted providers and endpoints. This page summarizes public runs on MMLU, MATH, GSM8K, IFEval, and MuSR, including score, latency, sample coverage, prompts, outputs, and methodology.
Provider Endpoints
MiniMax M2 has 5 public runs across 2 providers. Provider-hosted versions of the same model can differ in quantization, infrastructure, and serving configuration, which affects benchmark results independently of model capability.
How to Compare Endpoints
Use canonical-prompt runs on the same benchmark to compare endpoints fairly. Score differences between providers running the same model family reflect hosting differences rather than model differences. Check the methodology for how runs are defined and what makes them comparable.
MMLU results for MiniMax M2
No public MMLU runs for MiniMax M2 are available yet. Explore all MMLU benchmark results or run this benchmark on your endpoint.
MATH results for MiniMax M2
| Provider | Endpoint | Best Score | Runs |
|---|---|---|---|
| Vertex MaaS | Vertex MaaS / MiniMax M2 | — | 1 run |
GSM8K results for MiniMax M2
| Provider | Endpoint | Best Score | Runs |
|---|---|---|---|
| Vertex MaaS | Vertex MaaS / MiniMax M2 | — | 3 runs |
IFEval results for MiniMax M2
No public IFEval runs for MiniMax M2 are available yet. Explore all IFEval benchmark results or run this benchmark on your endpoint.
MuSR results for MiniMax M2
| Provider | Endpoint | Best Score | Runs |
|---|---|---|---|
| Together AI | Together / MiniMax M2.5 | — | 1 run |
Related
Benchscope is a JavaScript app. If the interactive interface does not load, enable JavaScript or use the links above for the main public sections.