Home/Models/MiniMax/MiniMax M2.1
M

MiniMax M2.1

Input:$0.24/M
Output:$0.96/M
Context:204,800(total input + output per request)
Max Output:131.1K
MiniMax M2.1: Significantly Enhanced Multi-Language Programming, Built for Real-World Complex Tasks
New
Commercial Use
Playground
Overview
Features
Pricing
API

What is MiniMax M2.1

MiniMax M2.1 is the follow-up release to the MiniMax M2 family, published by MiniMax on Dec 23, 2025. It is positioned as an open-source, production-oriented model designed specifically for coding, agentic multi-step workflows (tool use, multi-turn planning), and full-stack app generation (web, mobile, backend). The release emphasizes improved multi-language programming, better mobile / native app capabilities, concise responses, and improved tool/agent generalization.

Main features

  • MoE efficiency: Large total parameter count but only a small active subset per token (architecture designed to trade peak capacity for inference efficiency).
  • Coding-first optimizations: Strong polyglot code comprehension and generation across many languages (Python, TypeScript, Rust, Go, C++, Java, Swift, mobile native languages).
  • Agentic & tool workflows: Designed for tool calls, multi-step plans, and “interleaved thinking” / chained execution of composite instruction constraints.
  • Large context support & long outputs: Targeted at long-context developer workflows and agent trace/history.
  • Low-latency / high-throughput: Practical for interactive coding assistants and scaled agent loops due to the selective-activation design and deployment optimizations.

Technical capabilities and specifications

  • Architecture: Mixture-of-Experts (MoE) design.
  • Parameters: Reported design: ~230 billion total parameters with ~10 billion active parameters used per inference (MoE active subset). This is the same active-parameter efficiency model used in the M2 family.
  • Inference characteristics: Designed for low-latency interactive use, high-throughput batched inference, and agentic loops with frequent tool calls.
  • Streaming / function calling: supports streaming token output and advanced function-call/tool interfaces for structured I/O.

Benchmark performance

MiniMax released comparative benchmark claims and third-party aggregators reported scores at launch; representative published figures include:

Multi-SWE Bench / SWE-Bench (coding/agentic suites): provider and aggregator listings cite 49.4% on Multi-SWE-Bench and 72.5% on SWE-Bench Multilingual for M2.1 (these are scoring aggregates for code-generation and code reasoning tasks).

M2.1 shows comprehensive improvements over M2 in test case generation, code optimization, code review and instruction following and M2.1 outperforms M2 and often matches or surpasses Claude Sonnet 4.5 on several coding subtasks.

MiniMax M2.1

Representative production use cases

  1. IDE code assistant & refactoring: Multi-file refactors, code review suggestions, automated test generation and patch generation across multiple languages.
  2. Agentic “Digital Employee”: Automating repetitive office workflows (searching ticket systems, summarizing documents, interacting with web apps through text-based commands) using tool integration and interleaved thinking.
  3. Multi-language engineering support: Teams that maintain polyglot codebases (Rust, Go, Java, C++, TypeScript) can use M2.1 for cross-language code synthesis and conversions.
  4. Automated code evaluation & test generation: Generating testcases, running code analysis and producing suggested fixes or optimizations as part of CI tooling.
  5. Local/on-prem research and customization: Organizations that require on-prem control can fine-tune or run M2.1 locally using published weights and recommended inference stacks.

How to access and use MiniMax M2.1 API

Step 1: Sign Up for API Key

Log in to cometapi.com. If you are not our user yet, please register first. Sign into your CometAPI console. Get the access credential API key of the interface. Click “Add Token” at the API token in the personal center, get the token key: sk-xxxxx and submit.

Step 2: Send Requests to MiniMax M2.1 API

Select the “minimax-m2.1” endpoint to send the API request and set the request body. The request method and request body are obtained from our website API doc. Our website also provides Apifox test for your convenience. Replace <YOUR_API_KEY> with your actual CometAPI key from your account. Where to call it: Chat-style APIs.

Insert your question or request into the content field—this is what the model will respond to . Process the API response to get the generated answer.

Step 3: Retrieve and Verify Results

Process the API response to get the generated answer. After processing, the API responds with the task status and output data.

FAQ

What is the context window size for MiniMax M2.1?

MiniMax M2.1 supports a 204,800 token context window, enabling handling of large codebases and complex multi-file projects.

Can MiniMax M2.1 build native Android and iOS apps?

Yes, MiniMax M2.1 significantly strengthens native Android (Kotlin) and iOS (Swift/Objective-C) development capabilities, addressing a common weakness in AI coding models.

What programming languages does MiniMax M2.1 support?

MiniMax M2.1 provides systematic enhancement for Rust, Java, Golang, C++, Kotlin, Objective-C, TypeScript, JavaScript, and Python, covering the complete chain from low-level system development to application layer development.

What AI coding tools are compatible with MiniMax M2.1?

MiniMax M2.1 works with Claude Code, Cursor, Cline, Kilo Code, Roo Code, BlackBox AI, and Droid (Factory AI), with support for context management via Skill.md, cursorrule, and agent.md files.

What is the difference between M2.1 and M2.1-lightning?

MiniMax M2.1 outputs at approximately 60 tokens per second with maximum capability, while M2.1-lightning delivers approximately 100 tps for faster, more agile responses at similar quality.

Does MiniMax M2.1 support function calling and tool use?

Yes, MiniMax M2.1 features Interleaved Thinking for systematic problem-solving and supports function calling via the Anthropic-compatible API, enabling Shell, Browser, Python interpreter, and MCP tool orchestration.

How does MiniMax M2.1 compare to Claude Sonnet 4.5 for multilingual coding?

MiniMax M2.1 outperforms Claude Sonnet 4.5 on multilingual coding scenarios and closely approaches Claude Opus 4.5 on SWE-bench Verified, while costing only 8% of Claude's price.

Features for MiniMax M2.1

Explore the key features of MiniMax M2.1, designed to enhance performance and usability. Discover how these capabilities can benefit your projects and improve user experience.

Pricing for MiniMax M2.1

Explore competitive pricing for MiniMax M2.1, designed to fit various budgets and usage needs. Our flexible plans ensure you only pay for what you use, making it easy to scale as your requirements grow. Discover how MiniMax M2.1 can enhance your projects while keeping costs manageable.
Comet Price (USD / M Tokens)Official Price (USD / M Tokens)Discount
Input:$0.24/M
Output:$0.96/M
Input:$0.3/M
Output:$1.2/M
-20%

Sample code and API for MiniMax M2.1

Access comprehensive sample code and API resources for MiniMax M2.1 to streamline your integration process. Our detailed documentation provides step-by-step guidance, helping you leverage the full potential of MiniMax M2.1 in your projects.
Python
JavaScript
Curl
from openai import OpenAI
import os

# Get your CometAPI key from https://api.cometapi.com/console/token
COMETAPI_KEY = os.environ.get("COMETAPI_KEY") or "<YOUR_COMETAPI_KEY>"
BASE_URL = "https://api.cometapi.com/v1"

client = OpenAI(base_url=BASE_URL, api_key=COMETAPI_KEY)

# minimax-m2.1: MiniMax M2.1 model via chat/completions
completion = client.chat.completions.create(
    model="minimax-m2.1",
    messages=[
        {"role": "user", "content": "Hello! Tell me a short joke."}
    ]
)

print(completion.choices[0].message.content)

More Models