模型支持企业博客
500+ AI 模型 API,一次搞定,就在 CometAPI
模型 API
开发者
快速入门文档API 仪表板
资源
AI 模型博客企业更新日志关于
2025 CometAPI。保留所有权利。隐私政策服务条款
Home/Models/OpenAI/GPT-5.1
O

GPT-5.1

输入:$1/M
输出:$8/M
GPT-5.1 是一款通用型指令微调语言模型,专注于跨产品工作流程的文本生成与推理。它支持多轮对话、结构化输出格式,以及面向代码的任务,如起草、重构和解释。典型用例包括聊天助手、检索增强的问答、数据转换,以及在支持的情况下基于工具或 API 的代理式自动化。技术亮点包括以文本为中心的模态、指令遵循、JSON 风格的输出,以及与常见编排框架中的函数调用兼容。
新
商用
Playground
概览
功能亮点
定价
API

GPT 5.1 API is what GPT-5.1 Thinking is the advanced reasoning variant of OpenAI’s GPT-5.1 family, it prioritizes adaptive, higher-quality reasoning while giving developers explicit control over the latency / compute trade-off.

Basic features

  • Adaptive reasoning: the model dynamically adjusts thinking depth per request — faster on routine tasks, more persistent on complex ones. This reduces latency and token use for common queries. explicitly allocates more reasoning time for complex prompts, and is more persistent on multi-step problems; can be slower for hard tasks but gives deeper answers.
  • Reasoning modes: none / low / medium / high (GPT-5.1 defaults to none for low-latency cases; choose higher levels for more demanding tasks). The Responses API exposes a reasoning parameter to control this.
  • Default tone & style: written to be clearer on complex topics (less jargon), more explanatory and “patient.”
  • Context window (tokens / long context) Thinking: much larger — 400K token context for paid tiers.

Key technical details

  • Adaptive compute allocation — training and inference design causes the model to expend fewer reasoning tokens on trivial tasks and proportionally more on difficult tasks. This is not a separate “think engine” but a dynamic allocation within the reasoning pipeline.
  • Reasoning parameter in the Responses API — clients pass a reasoning object (for example reasoning: { "effort": "high" }) to request deeper internal reasoning; setting reasoning: { "effort": "none" } effectively disables the extended internal reasoning pass for lower latency. The Responses API also returns reasoning/token metadata (helpful for cost and debugging). )
  • Tools & parallel tool calls — GPT-5.1 improves parallel tool calling and includes named tools (like apply_patch) that reduce failure modes for programmatic edits; parallelization increases end-to-end throughput for tool-heavy workflows.
  • Prompt cache and persistence — prompt_cache_retention='24h' is supported on Responses and Chat Completions endpoints to retain context across multi-turn sessions (reduces repeated token encoding).

Benchmark performance

Latency / token efficiency examples (vendor-provided): on routine queries, OpenAI reports dramatic reductions in tokens/time (example: an npm listing command that took ~10s / ~250 tokens on GPT-5 now takes ~2s / ~50 tokens on GPT-5.1 in their representative test). Third-party early testers (e.g., asset managers, coding firms) reported 2–3× speedups on many tasks and token-efficiency gains in tool-heavy flows.

OpenAI and early partners published representative benchmark claims and measured improvements:

EvaluationGPT‑5.1 (high)GPT‑5 (high)
SWE-bench Verified (all 500 problems)76.3%72.8%
GPQA Diamond (no tools)88.1%85.7%
AIME 2025 (no tools)94.0%94.6%
FrontierMath (with Python tool)26.7%26.3%
MMMU85.4%84.2%
Tau2-bench Airline67.0%62.6%
Tau2-bench Telecom*95.6%96.7%
Tau2-bench Retail77.9%81.1%
BrowseComp Long Context 128k90.0%90.0%

Limitations & safety considerations

  • Hallucination risk persists. Adaptive reasoning helps on complex problems but does not eliminate hallucinations; higher reasoning_effort improves checks but does not guarantee correctness. Always validate high-stakes outputs.
  • Resource and cost tradeoffs: while GPT-5.1 can be far more token-efficient on simple flows, enabling high reasoning effort or long agentic tool-use can increase token consumption and latency. Use prompt caching to mitigate repeated costs where appropriate.
  • Tool safety: apply_patch and shell tools increase automation power (and risk). Production deployments should gate tool execution (review diffs / commands before execution), use least privilege, and ensure robust CI/CD and operational guardrails.

Comparison with other models

  • vs GPT-5: GPT-5.1 improves adaptive reasoning and instruction adherence; OpenAI reports faster response times on easy tasks and better persistence on hard tasks. GPT-5.1 also adds the none reasoning option and extended prompt caching.
  • vs GPT-4.x / 4.1: GPT-5.1 is designed for more agentic, tool-heavy, and coding tasks; OpenAI and partners report gains on coding benchmarks and multi-step reasoning. For many standard conversational tasks, GPT-5.1 Instant may be comparable to earlier GPT-4.x chat models but with improved steerability and personality presets.
  • vs Anthropic / Claude / other LLMs: ChatGPT 5.1′;s MoA architecture gives it a distinct edge in tasks requiring complex, multi-step reasoning. It scored an unprecedented 98.20 on the HELM benchmark for complex reasoning, compared to Claude 4’s 95.60 and Gemini 2.0 Ultra’s 94.80.

GPT-5.1 的功能

了解 GPT-5.1 的核心能力,帮助提升性能与可用性,并改善整体体验。

GPT-5.1 的定价

查看 GPT-5.1 的竞争性定价,满足不同预算与使用需求,灵活方案确保随需求扩展。
Comet 价格 (USD / M Tokens)官方定价 (USD / M Tokens)折扣
输入:$1/M
输出:$8/M
输入:$1.25/M
输出:$10/M
-20%

GPT-5.1 的示例代码与 API

GPT 5.1 API is what GPT-5.1 Thinking is the advanced reasoning variant of OpenAI’s GPT-5.1 family, it prioritizes adaptive, higher-quality reasoning while giving developers explicit control over the latency / compute trade-off.
POST
/v1/responses
POST
/v1/chat/completions
Python
JavaScript
Curl
from openai import OpenAI
import os

# Get your CometAPI key from https://api.cometapi.com/console/token, and paste it here
COMETAPI_KEY = os.environ.get("COMETAPI_KEY") or "<YOUR_COMETAPI_KEY>"
BASE_URL = "https://api.cometapi.com/v1"

client = OpenAI(base_url=BASE_URL, api_key=COMETAPI_KEY)
response = client.responses.create(
    model="gpt-5.1", input="Tell me a three sentence bedtime story about a unicorn."
)

print(response)

Python Code Example

from openai import OpenAI
import os

# Get your CometAPI key from https://api.cometapi.com/console/token, and paste it here
COMETAPI_KEY = os.environ.get("COMETAPI_KEY") or "<YOUR_COMETAPI_KEY>"
BASE_URL = "https://api.cometapi.com/v1"

client = OpenAI(base_url=BASE_URL, api_key=COMETAPI_KEY)
response = client.responses.create(
    model="gpt-5.1", input="Tell me a three sentence bedtime story about a unicorn."
)

print(response)

JavaScript Code Example

import OpenAI from "openai";

// Get your CometAPI key from https://api.cometapi.com/console/token, and paste it here
const api_key = process.env.COMETAPI_KEY;
const base_url = "https://api.cometapi.com/v1";

const openai = new OpenAI({
  apiKey: api_key,
  baseURL: base_url,
});

const response = await openai.responses.create({
  model: "gpt-5.1",
  input: "Tell me a three sentence bedtime story about a unicorn.",
});

console.log(response);

Curl Code Example

curl https://api.cometapi.com/v1/responses \
  -H "Content-Type: application/json" \
  -H "Authorization: Bearer $COMETAPI_KEY" \
  -d '{
    "model": "gpt-5.1",
    "input": "Tell me a three sentence bedtime story about a unicorn."
  }'

更多模型

A

Claude Opus 4.6

输入:$4/M
输出:$20/M
Claude Opus 4.6 是 Anthropic 的“Opus”级大型语言模型,于 2026 年 2 月发布。其定位为知识工作与研究工作流的主力模型——提升长上下文推理、多步骤规划、工具使用(包括代理型软件工作流),以及计算机使用类任务,如自动生成幻灯片和电子表格。
A

Claude Sonnet 4.6

输入:$2.4/M
输出:$12/M
Claude Sonnet 4.6 是迄今为止我们最强大的 Sonnet 模型。它对模型在编码、计算机使用、长上下文推理、智能体规划、知识工作和设计等方面的能力进行了全面升级。Sonnet 4.6 还在 beta 阶段提供 1M token 上下文窗口。
O

GPT-5.4 nano

输入:$0.16/M
输出:$1/M
GPT-5.4 nano 专为速度和成本最为关键的任务而设计,例如分类、数据提取、排序以及子智能体。
O

GPT-5.4 mini

输入:$0.6/M
输出:$3.6/M
GPT-5.4 mini 将 GPT-5.4 的优势融入到一款更快速、更高效、专为大规模工作负载设计的模型中。
A

Claude Mythos Preview

A

Claude Mythos Preview

即将推出
输入:$60/M
输出:$240/M
Claude Mythos Preview 是我们迄今为止最强大的前沿模型,并显示出 在许多评测基准上的得分相较于我们此前的前沿模型 Claude Opus 4.6 有显著跃升。
X

mimo-v2-pro

输入:$0.8/M
输出:$2.4/M
MiMo-V2-Pro 是 Xiaomi 的旗舰基础模型,拥有超过 1T 的总参数量和 1M 的上下文长度,并针对智能体场景进行了深度优化。它对 OpenClaw 等通用智能体框架具有很强的适配性。在标准 PinchBench 和 ClawBench 基准测试中,它跻身全球第一梯队,感知性能接近 Opus 4.6。MiMo-V2-Pro 旨在作为智能体系统的大脑,协调复杂工作流,推动生产工程任务,并可靠地交付结果。

相关博客

GPT 5.2 Codex 发布:功能、基准测试与访问
Dec 22, 2025

GPT 5.2 Codex 发布:功能、基准测试与访问

OpenAI 发布了 GPT-5.2-Codex,这是一款针对 Codex 优化的 GPT-5.2 版本,专为长时程的智能体式编码任务、大规模重构与迁移、在终端环境中可靠使用工具、改进的 Windows 原生表现,以及更强的网络安全能力而设计。SWE-Bench Pro 和 Terminal-Bench 2.0 等基准测试显示,GPT-5.2-Codex 在智能体式编码模型中处于最先进水平。
目前没有公开、精确的“ChatGPT每日用水量”数据。用水量取决于数据中心的位置与冷却方式、当日负载、时间与季节、电力与水系统等多种因素。

基于学术界对生成式AI的估算:
- 推理阶段(用户问答)大致约为0.5升/20–50次交互,即每次请求约0.01–0.025升。
- 训练阶段的用水是一次性的,规模可达数十万升量级,但不属于“每日”消耗。

因此,若要粗略估算“每日用水”,可用公式:当日总请求量 × 0.01–0.025 升。举例:若当天有1000万次请求,约为10万–25万升。上述数值为研究估算,实际会因模型版本、运营商与部署环境而显著波动。
Dec 6, 2025
chat-gpt

目前没有公开、精确的“ChatGPT每日用水量”数据。用水量取决于数据中心的位置与冷却方式、当日负载、时间与季节、电力与水系统等多种因素。 基于学术界对生成式AI的估算: - 推理阶段(用户问答)大致约为0.5升/20–50次交互,即每次请求约0.01–0.025升。 - 训练阶段的用水是一次性的,规模可达数十万升量级,但不属于“每日”消耗。 因此,若要粗略估算“每日用水”,可用公式:当日总请求量 × 0.01–0.025 升。举例:若当天有1000万次请求,约为10万–25万升。上述数值为研究估算,实际会因模型版本、运营商与部署环境而显著波动。

简短回答:ChatGPT 的全球服务每天可能消耗约 2 million 至 160 million 升水 — 一个非常宽的范围,由……驱动
什么是 GPT-5.1 Pro?专业解读与状态报告
Nov 28, 2025
gpt-5-1
gpt-5-1-instant
gpt-5-1-pro

什么是 GPT-5.1 Pro?专业解读与状态报告

OpenAI 的 GPT-5.1 Pro 是 GPT-5 系列中最新的增量发布:一次生产级的模型更新,对推理能力与延迟/吞吐量进行了优化
Claude Opus 4.5:它怎么样——以及要花多少钱?
Nov 23, 2025
gemini-3-pro
gpt-5-1

Claude Opus 4.5:它怎么样——以及要花多少钱?

社交媒体帖子和调查报道指出,即将推出的 Claude Opus 4.5(常缩写为“Opus 4.5”)——据一些消息来源称其内部代号为 Neptune V6——并且该模型已与外部红队人员共享以进行越狱测试。公开细节仍然零散,因此本文汇总现有报道,解释这次泄露对能力与安全性的暗示,并给出一个有依据的可能定价估算,以及 Opus 4.5 相较于 Google 的 Gemini 3 和 OpenAI 的 GPT-5.1 的潜在表现。
Gemini 3 Pro vs GPT 5.1:哪个更好?全面对比
Nov 18, 2025
gemini-3-pro-preview
gpt-5-1

Gemini 3 Pro vs GPT 5.1:哪个更好?全面对比

OpenAI 的 GPT-5.1 和 Google 的 Gemini 3 Pro 都代表了通用型多模态 AI 持续竞争中渐进但意义重大的进步。