Home/Models/MiniMax/MiniMax M2.1
M

MiniMax M2.1

輸入:$0.24/M
輸出:$0.96/M
上下文:204,800(total input + output per request)
最大输出:131.1K
MiniMax M2.1:显著增强的多语言编程能力,专为真实世界的复杂任务而打造
新
商用
Playground
概览
功能亮点
定价
API

What is MiniMax M2.1

MiniMax M2.1 is the follow-up release to the MiniMax M2 family, published by MiniMax on Dec 23, 2025. It is positioned as an open-source, production-oriented model designed specifically for coding, agentic multi-step workflows (tool use, multi-turn planning), and full-stack app generation (web, mobile, backend). The release emphasizes improved multi-language programming, better mobile / native app capabilities, concise responses, and improved tool/agent generalization.

Main features

  • MoE efficiency: Large total parameter count but only a small active subset per token (architecture designed to trade peak capacity for inference efficiency).
  • Coding-first optimizations: Strong polyglot code comprehension and generation across many languages (Python, TypeScript, Rust, Go, C++, Java, Swift, mobile native languages).
  • Agentic & tool workflows: Designed for tool calls, multi-step plans, and “interleaved thinking” / chained execution of composite instruction constraints.
  • Large context support & long outputs: Targeted at long-context developer workflows and agent trace/history.
  • Low-latency / high-throughput: Practical for interactive coding assistants and scaled agent loops due to the selective-activation design and deployment optimizations.

Technical capabilities and specifications

  • Architecture: Mixture-of-Experts (MoE) design.
  • Parameters: Reported design: ~230 billion total parameters with ~10 billion active parameters used per inference (MoE active subset). This is the same active-parameter efficiency model used in the M2 family.
  • Inference characteristics: Designed for low-latency interactive use, high-throughput batched inference, and agentic loops with frequent tool calls.
  • Streaming / function calling: supports streaming token output and advanced function-call/tool interfaces for structured I/O.

Benchmark performance

MiniMax released comparative benchmark claims and third-party aggregators reported scores at launch; representative published figures include:

Multi-SWE Bench / SWE-Bench (coding/agentic suites): provider and aggregator listings cite 49.4% on Multi-SWE-Bench and 72.5% on SWE-Bench Multilingual for M2.1 (these are scoring aggregates for code-generation and code reasoning tasks).

M2.1 shows comprehensive improvements over M2 in test case generation, code optimization, code review and instruction following and M2.1 outperforms M2 and often matches or surpasses Claude Sonnet 4.5 on several coding subtasks.

MiniMax M2.1

Representative production use cases

  1. IDE code assistant & refactoring: Multi-file refactors, code review suggestions, automated test generation and patch generation across multiple languages.
  2. Agentic “Digital Employee”: Automating repetitive office workflows (searching ticket systems, summarizing documents, interacting with web apps through text-based commands) using tool integration and interleaved thinking.
  3. Multi-language engineering support: Teams that maintain polyglot codebases (Rust, Go, Java, C++, TypeScript) can use M2.1 for cross-language code synthesis and conversions.
  4. Automated code evaluation & test generation: Generating testcases, running code analysis and producing suggested fixes or optimizations as part of CI tooling.
  5. Local/on-prem research and customization: Organizations that require on-prem control can fine-tune or run M2.1 locally using published weights and recommended inference stacks.

How to access and use MiniMax M2.1 API

Step 1: Sign Up for API Key

Log in to cometapi.com. If you are not our user yet, please register first. Sign into your CometAPI console. Get the access credential API key of the interface. Click “Add Token” at the API token in the personal center, get the token key: sk-xxxxx and submit.

Step 2: Send Requests to MiniMax M2.1 API

Select the “minimax-m2.1” endpoint to send the API request and set the request body. The request method and request body are obtained from our website API doc. Our website also provides Apifox test for your convenience. Replace <YOUR_API_KEY> with your actual CometAPI key from your account. Where to call it: Chat-style APIs.

Insert your question or request into the content field—this is what the model will respond to . Process the API response to get the generated answer.

Step 3: Retrieve and Verify Results

Process the API response to get the generated answer. After processing, the API responds with the task status and output data.

常见问题

What is the context window size for MiniMax M2.1?

MiniMax M2.1 supports a 204,800 token context window, enabling handling of large codebases and complex multi-file projects.

Can MiniMax M2.1 build native Android and iOS apps?

Yes, MiniMax M2.1 significantly strengthens native Android (Kotlin) and iOS (Swift/Objective-C) development capabilities, addressing a common weakness in AI coding models.

What programming languages does MiniMax M2.1 support?

MiniMax M2.1 provides systematic enhancement for Rust, Java, Golang, C++, Kotlin, Objective-C, TypeScript, JavaScript, and Python, covering the complete chain from low-level system development to application layer development.

What AI coding tools are compatible with MiniMax M2.1?

MiniMax M2.1 works with Claude Code, Cursor, Cline, Kilo Code, Roo Code, BlackBox AI, and Droid (Factory AI), with support for context management via Skill.md, cursorrule, and agent.md files.

What is the difference between M2.1 and M2.1-lightning?

MiniMax M2.1 outputs at approximately 60 tokens per second with maximum capability, while M2.1-lightning delivers approximately 100 tps for faster, more agile responses at similar quality.

Does MiniMax M2.1 support function calling and tool use?

Yes, MiniMax M2.1 features Interleaved Thinking for systematic problem-solving and supports function calling via the Anthropic-compatible API, enabling Shell, Browser, Python interpreter, and MCP tool orchestration.

How does MiniMax M2.1 compare to Claude Sonnet 4.5 for multilingual coding?

MiniMax M2.1 outperforms Claude Sonnet 4.5 on multilingual coding scenarios and closely approaches Claude Opus 4.5 on SWE-bench Verified, while costing only 8% of Claude's price.

MiniMax M2.1 的功能

了解 MiniMax M2.1 的核心能力,帮助提升性能与可用性,并改善整体体验。

MiniMax M2.1 的定价

查看 MiniMax M2.1 的竞争性定价,满足不同预算与使用需求,灵活方案确保随需求扩展。
Comet 价格 (USD / M Tokens)官方定价 (USD / M Tokens)折扣
輸入:$0.24/M
輸出:$0.96/M
輸入:$0.3/M
輸出:$1.2/M
-20%

MiniMax M2.1 的示例代码与 API

获取完整示例代码与 API 资源,简化 MiniMax M2.1 的集成流程,我们提供逐步指导,助你发挥模型潜能。
Python
JavaScript
Curl
from openai import OpenAI
import os

# Get your CometAPI key from https://api.cometapi.com/console/token
COMETAPI_KEY = os.environ.get("COMETAPI_KEY") or "<YOUR_COMETAPI_KEY>"
BASE_URL = "https://api.cometapi.com/v1"

client = OpenAI(base_url=BASE_URL, api_key=COMETAPI_KEY)

# minimax-m2.1: MiniMax M2.1 model via chat/completions
completion = client.chat.completions.create(
    model="minimax-m2.1",
    messages=[
        {"role": "user", "content": "Hello! Tell me a short joke."}
    ]
)

print(completion.choices[0].message.content)

更多模型