ModelsSupportEnterpriseBlog
500+ AI Model API, All In One API.Just In CometAPI
Models API
Developer
Quick StartDocumentationAPI Dashboard
Resources
AI ModelsBlogEnterpriseChangelogAbout
2025 CometAPI. All right reserved.Privacy PolicyTerms of Service
Home/Models/OpenAI/GPT-5.2 Pro
O

GPT-5.2 Pro

Input:$16.8/M
Output:$134.4/M
Context:400,000
Max Output:128,000
gpt-5.2-pro is the highest-capability, production-oriented member of OpenAI’s GPT-5.2 family, exposed through the Responses API for workloads that demand maximal fidelity, multi-step reasoning, extensive tool use and the largest context/throughput budgets OpenAI offers.
New
Commercial Use
Playground
Overview
Features
Pricing
API

What is GPT-5.2-Pro

GPT-5.2-Pro is the “Pro” tier of OpenAI’s GPT-5.2 family intended for the hardest problems — multi-step reasoning, complex code, large document synthesis, and professional knowledge work. It’s made available in the Responses API to enable multi-turn interactions and advanced API features (tooling, reasoning modes, compaction, etc.). The Pro variant trades throughput and cost for maximum answer quality and stronger safety/consistency in hard domains.

Main features (what gpt-5.2-pro brings to applications)

  • Highest-fidelity reasoning: Pro supports OpenAI’s top reasoning settings (including xhigh) to trade latency and compute for deeper internal reasoning passes and improved chain-of-thought-style solution refinement.
  • Large-context, long-document proficiency: engineered to maintain accuracy across very long contexts (OpenAI benchmarked up through 256k+ tokens for family variants), making the tier suitable for legal/technical document review, enterprise knowledge bases, and long-running agent states.
  • Stronger tool & agent execution: designed to call toolsets reliably (allowed-tools lists, auditing hooks, and richer tool integrations) and to act as a “mega-agent” that can orchestrate multiple subtools and multi-step workflows.
  • Improved factuality & safety mitigations: OpenAI reports notable reductions in hallucination and undesirable responses on internal safety metrics for GPT-5.2 vs prior models, supported by updates in the system card and targeted safety training.

Technical capabilities & specifications (developer-oriented)

  • API endpoint & availability: Responses API is the recommended integration for Pro-level workflows; developers can set reasoning.effort to none|medium|high|xhigh to tune internal compute devoted to reasoning. Pro exposes the highest xhigh fidelity.
  • Reasoning effort levels: none | medium | high | xhigh (Pro and Thinking support xhigh for quality-prioritized runs). This parameter lets you trade cost/latency for quality.
  • Compaction & context management: New compaction features allow the API to manage what the model “remembers” and reduce token usage while preserving relevant context—helpful for long conversations and document workflows.
  • Tooling & custom tools: Models can call custom tools (send raw text to tools while constraining model outputs); stronger tool-calling and agentic patterns in 5.2 reduce the need for elaborate system prompts.

Benchmark performance

Below are the most relevant, reproducible headline numbers for GPT-5.2 Pro (OpenAI’s verified/internal results):

  • GDPval (professional work benchmark): GPT-5.2 Pro — 74.1% (wins/ties) on the GDPval suite — a marked improvement over GPT-5.1. This metric is designed to approximate value in real economic tasks across many occupations.
  • ARC-AGI-1 (general reasoning): GPT-5.2 Pro — 90.5% (Verified); Pro was reported as the first model to cross 90% on this benchmark.
  • Coding & software engineering (SWE-Bench): strong gains in multi-step code reasoning; e.g., SWE-Bench Pro public and SWE-Lancer (IC Diamond) show material improvements over GPT-5.1 — representative family numbers: SWE-Bench Pro public ~55.6% (Thinking; Pro results reported higher on internal runs).
  • Long-context factuality (MRCRv2): GPT-5.2 family shows high retrieval and needle-finding scores across 4k–256k ranges (examples: MRCRv2 8 needles at 16k–32k: 95.3% for GPT-5.2 Thinking; Pro maintained high accuracy at larger windows). These show the family’s resilience to long-context tasks, a Pro selling point.

How gpt-5.2-pro compares with peers and other GPT-5.2 tiers

  • vs GPT-5.2 Thinking / Instant:: gpt-5.2-pro prioritizes fidelity and maximal reasoning (xhigh) over latency/cost. gpt-5.2 (Thinking) sits in the middle for deeper work, and gpt-5.2-chat-latest (Instant) is tuned for low-latency chat. Choose Pro for highest-value, compute-intensive tasks.
  • Versus Google Gemini 3 and other frontier models: GPT-5.2 (family) as OpenAI’s competitive response to Gemini 3. Leaderboards show task-dependent winners — on some graduate-level science and professional benchmarks GPT-5.2 Pro and Gemini 3 are close; in narrow coding or specialized domains outcomes can vary.
  • Versus GPT-5.1 / GPT-5: Pro shows material gains in GDPval, ARC-AGI, coding benchmarks and long-context metrics vs GPT-5.1, and adds new API controls (xhigh reasoning, compaction). OpenAI will keep earlier variants available during transition.

Practical use cases and recommended patterns

High-value use cases where Pro makes sense

  • Complex financial modeling, large spreadsheet synthesis and analysis where accuracy and multi-step reasoning matter (OpenAI reported improved investment banking spreadsheet task scores).
  • Long-document legal or scientific synthesis where the 400k token context preserves entire reports, appendices, and citation chains.
  • High-quality code generation and multi-file refactoring for enterprise codebases (Pro’s higher xhigh reasoning helps with multi-step program transformations).
  • Strategic planning, multi-stage project orchestration, and agentic workflows that use custom tools and require robust tool calling.

When to choose Thinking or Instant instead

  • Choose Instant for fast, lower-cost conversational tasks and editor integrations.
  • Choose Thinking for deeper but latency-sensitive work where cost is constrained but quality still matters.

How to access and use GPT-5.2 pro API

Step 1: Sign Up for API Key

Log in to cometapi.com. If you are not our user yet, please register first. Sign into your CometAPI console. Get the access credential API key of the interface. Click “Add Token” at the API token in the personal center, get the token key: sk-xxxxx and submit.

Step 2: Send Requests to GPT-5.2 pro API

Select the “gpt-5.2-pro” endpoint to send the API request and set the request body. The request method and request body are obtained from our website API doc. Our website also provides Apifox test for your convenience. Replace <YOUR_API_KEY> with your actual CometAPI key from your account. Where to call it: Responses-style APIs.

Insert your question or request into the content field—this is what the model will respond to . Process the API response to get the generated answer.

Step 3: Retrieve and Verify Results

Process the API response to get the generated answer. After processing, the API responds with the task status and output data.

See also Gemini 3 Pro Preview API

FAQ

Why does GPT-5.2 Pro only work with the Responses API?

GPT-5.2 Pro is exclusively available through the Responses API to enable multi-turn model interactions before responding to API requests, supporting advanced workflows like tool chaining and extended reasoning sessions that require persistent state management.

What reasoning effort levels does GPT-5.2 Pro support?

GPT-5.2 Pro supports three reasoning effort levels: medium, high, and xhigh—allowing developers to balance response quality against latency for complex problem-solving tasks.

How does GPT-5.2 Pro handle long-running requests?

Some GPT-5.2 Pro requests may take several minutes to complete due to the model's deep reasoning process. OpenAI recommends using background mode to avoid timeouts on particularly challenging tasks.

What tools can GPT-5.2 Pro access through the Responses API?

GPT-5.2 Pro supports web search, file search, image generation, and MCP (Model Context Protocol), but notably does not support code interpreter or computer use tools.

When should I choose GPT-5.2 Pro over standard GPT-5.2?

Choose GPT-5.2 Pro when your workload demands maximum fidelity, multi-step reasoning, or extensive tool orchestration—it's designed for production scenarios with the largest context and throughput budgets OpenAI offers.

Features for GPT-5.2 Pro

Explore the key features of GPT-5.2 Pro, designed to enhance performance and usability. Discover how these capabilities can benefit your projects and improve user experience.

Pricing for GPT-5.2 Pro

Explore competitive pricing for GPT-5.2 Pro, designed to fit various budgets and usage needs. Our flexible plans ensure you only pay for what you use, making it easy to scale as your requirements grow. Discover how GPT-5.2 Pro can enhance your projects while keeping costs manageable.
Comet Price (USD / M Tokens)Official Price (USD / M Tokens)Discount
Input:$16.8/M
Output:$134.4/M
Input:$21/M
Output:$168/M
-20%

Sample code and API for GPT-5.2 Pro

GPT-5.2-Pro is OpenAI’s highest-quality variant of the GPT-5.2 family designed for the hardest, highest-value knowledge and technical tasks.
POST
/v1/responses
Python
JavaScript
Curl
from openai import OpenAI
import os

# Get your CometAPI key from https://api.cometapi.com/console/token, and paste it here
COMETAPI_KEY = os.environ.get("COMETAPI_KEY") or "<YOUR_COMETAPI_KEY>"
BASE_URL = "https://api.cometapi.com/v1"

client = OpenAI(base_url=BASE_URL, api_key=COMETAPI_KEY)

response = client.responses.create(
    model="gpt-5.2-pro",
    input="How much gold would it take to coat the Statue of Liberty in a 1mm layer?",
    reasoning={"effort": "high"},
)

print(response.output_text)

Python Code Example

from openai import OpenAI
import os

# Get your CometAPI key from https://api.cometapi.com/console/token, and paste it here
COMETAPI_KEY = os.environ.get("COMETAPI_KEY") or "<YOUR_COMETAPI_KEY>"
BASE_URL = "https://api.cometapi.com/v1"

client = OpenAI(base_url=BASE_URL, api_key=COMETAPI_KEY)

response = client.responses.create(
    model="gpt-5.2-pro",
    input="How much gold would it take to coat the Statue of Liberty in a 1mm layer?",
    reasoning={"effort": "high"},
)

print(response.output_text)

JavaScript Code Example

import OpenAI from "openai";

// Get your CometAPI key from https://api.cometapi.com/console/token, and paste it here
const COMETAPI_KEY = process.env.COMETAPI_KEY || "<YOUR_COMETAPI_KEY>";
const BASE_URL = "https://api.cometapi.com/v1";

const client = new OpenAI({
  apiKey: COMETAPI_KEY,
  baseURL: BASE_URL,
});

async function main() {
  const response = await client.responses.create({
    model: "gpt-5.2-pro",
    input: "How much gold would it take to coat the Statue of Liberty in a 1mm layer?",
    reasoning: {
    effort: "high"
  }
  });

  console.log(response.output_text);
}

main();

Curl Code Example

curl https://api.cometapi.com/v1/responses \
     --header "Authorization: Bearer $COMETAPI_KEY" \
     --header "content-type: application/json" \
     --data \
'{
    "model": "gpt-5.2-pro",
    "input": "How much gold would it take to coat the Statue of Liberty in a 1mm layer?",
    "reasoning": {
        "effort": "high"
    }
}'

More Models

A

Claude Opus 4.6

Input:$4/M
Output:$20/M
Claude Opus 4.6 is Anthropic’s “Opus”-class large language model, released February 2026. It is positioned as a workhorse for knowledge-work and research workflows — improving long-context reasoning, multi-step planning, tool use (including agentic software workflows), and computer-use tasks such as automated slide and spreadsheet generation.
A

Claude Sonnet 4.6

Input:$2.4/M
Output:$12/M
Claude Sonnet 4.6 is our most capable Sonnet model yet. It’s a full upgrade of the model’s skills across coding, computer use, long-context reasoning, agent planning, knowledge work, and design. Sonnet 4.6 also features a 1M token context window in beta.
O

GPT-5.4 nano

Input:$0.16/M
Output:$1/M
GPT-5.4 nano is designed for tasks where speed and cost matter most like classification, data extraction, ranking, and sub-agents.
O

GPT-5.4 mini

Input:$0.6/M
Output:$3.6/M
GPT-5.4 mini brings the strengths of GPT-5.4 to a faster, more efficient model designed for high-volume workloads.
A

Claude Mythos Preview

A

Claude Mythos Preview

Coming soon
Input:$60/M
Output:$240/M
Claude Mythos Preview is our most capable frontier model to date, and shows a striking leap in scores on many evaluation benchmarks compared to our previous frontier model, Claude Opus 4.6.
X

mimo-v2-pro

Input:$0.8/M
Output:$2.4/M
MiMo-V2-Pro is Xiaomi's flagship foundation model, featuring over 1T total parameters and a 1M context length, deeply optimized for agentic scenarios. It is highly adaptable to general agent frameworks like OpenClaw. It ranks among the global top tier in the standard PinchBench and ClawBench benchmarks, with perceived performance approaching that of Opus 4.6. MiMo-V2-Pro is designed to serve as the brain of agent systems, orchestrating complex workflows, driving production engineering tasks, and delivering results reliably.

Related Blog

Can ChatGPT make PowerPoints
Mar 26, 2026
chat-gpt

Can ChatGPT make PowerPoints

Over the past two years AI tools have moved from “help me write slide text” to “assemble and export a full .pptx,” and both OpenAI and Microsoft have added features that make one-click or near–one-click PowerPoint creation possible., the question is no longer "Can AI help me work?" but "How much of my work can AI do?" Among the most requested tasks is the creation of slide decks—the ubiquitous currency of business communication. For years, users have dreamt of a simple command: "Hey ChatGPT, make me a presentation." In 2026, that dream is closer than ever to reality, though it comes with nuances that every professional must understand.
Cursor vs Claude Code vs Codex: Which Is Better for Vibe Coding in 2026?
Feb 2, 2026

Cursor vs Claude Code vs Codex: Which Is Better for Vibe Coding in 2026?

In the rapidly evolving world of Vibe Code development, developers are debating which tools enable the most productive, intuitive, and reliable workflows. Today’s comparison puts three leading agents — Cursor, Claude Code, and OpenAI Codex — head-to-head, focusing on the emergent paradigm of “vibe coding,” pricing, features, operations, usage, and real-world performance.
How to run openClaw (Moltbot/ Clawdbot ) on local LLMs without API
Feb 1, 2026
clawdbot
openclaw

How to run openClaw (Moltbot/ Clawdbot ) on local LLMs without API

OpenClaw — the open-source AI assistant that began life as Clawdbot (and briefly Moltbot) — exploded into the spotlight for its persistent memory, deep local access, and broad messaging integrations. That popularity brought questions about how to run the system without depending on cloud APIs. The solution lies in running OpenClaw on local Large Language Models (LLMs) via ollama or other , entirely without an internet connection or subscription fees.
Openclaw(Moltbot /Clawdbot) : Setup Guide+ API hosting Tutorial
Jan 29, 2026
clawdbot

Openclaw(Moltbot /Clawdbot) : Setup Guide+ API hosting Tutorial

Moltbot — previously known to the internet as Clawdbot — erupted onto timelines in late January 2026 as a do-it-yourself, agentic personal assistant that actually executes tasks on your behalf: clearing inboxes, running commands, searching local files, and answering through your chat app of choice (Telegram, WhatsApp, Discord, iMessage, etc.).
How to “Vibe Coding” a Tiny Mobile App as beginner
Jan 28, 2026
vibe-coding

How to “Vibe Coding” a Tiny Mobile App as beginner

Vibe Coding a tiny mobile app as a beginner is all about momentum and feel: start with a very small idea (like a counter or simple to-do app), pick an easy framework or tool (Codex, Cursor etc.), build something that runs as fast as possible without worrying about clean code, then tweak the UI and interactions by feel while testing on your phone. When you get stuck, search or ask AI instead of overthinking, and once it works and feels good, stop and ship it. The goal is flow and confidence, not perfection.