Home/Models/DeepSeek/DeepSeek-Reasoner
D

DeepSeek-Reasoner

Input:$0.44/M
Output:$1.752/M
DeepSeek-Reasoner is DeepSeek’s reasoning-first family of LLMs and API endpoints designed to (1) expose internal chain-of-thought (CoT) reasoning to callers and (2) operate in “thinking” modes tuned for multi-step planning, math, coding and agent/tool use.
Commercial Use
Overview
Features
Pricing
API
Versions

What is DeepSeek-Reasoner?

DeepSeek-Reasoner is the reasoning (or “thinking”) mode/API name for DeepSeek’s reasoning-first models (currently aligned to the DeepSeek-V3.2 family). It is designed to produce an explicit chain of thought (CoT) before emitting a final answer—i.e., the model intentionally generates internal step-by-step reasoning which is exposed (or can be exposed) through the API so callers can inspect or distill it. DeepSeek positions the reasoner variant as the “thinking” counterpart to its non-thinking chat model and markets it for multi-step reasoning, math, coding and agent workflows.

Main features (user-facing)

  • Explicit Chain-of-Thought (CoT) output. API returns a separate reasoning_content field containing the model’s internal stepwise reasoning alongside the final content. This is designed for inspectability and downstream agent logic.
  • “Thinking” vs “Chat” modes. deepseek-reasoner (thinking mode) is distinct from deepseek-chat (non-thinking mode); both were upgraded to the V3.2 generation.
  • Large context windows. DeepSeek exposes very large context lengths . The Reasoner variants are marketed for long-form reasoning and agent memory.
  • JSON output / structured responses. Support for structured JSON outputs useful for programmatic consumption.
  • Agent/agent-builder focus. V3.2 and the Speciale variant are explicitly described as “reasoning-first models built for agents.”

Technical capabilities

  • Inputs: plain text prompts, structured JSON for tool/agent calls, files or long documents (via long context); tokens are standard NLP tokens.
  • Outputs: API returns both reasoning_content (CoT text) and content (final answer). API clients can request only CoT or only final answer by adjusting max_tokens or response parameters. (Practical note: extracting CoT may still be billable as model output.)
  • DeepSeek has iterated via a reasoning-specialized roadmap: base large models (R1 family) followed by focused post-training / reinforcement learning (RLHF-style) and policy-style fine-tuning to improve reasoning depth. The team also uses distillation to compress reasoning capability into smaller, deployable models.
  • The V3.2 series adds agentic post-training for tool-use, hybrid inference (Think / Non-Think), and optimizations for faster “thinking” iterations.
  • Inference efficiency is aided by a sparse attention method (reports call it DeepSeek Sparse Attention — DSA) that focuses compute on relevant segments rather than full dense attention across very long sequences; this reduces cost for very long contexts.

How to access deepseek-reasoner API

Step 1: Sign Up for API Key

Log in to cometapi.com. If you are not our user yet, please register first. Sign into your CometAPI console. Get the access credential API key of the interface. Click “Add Token” at the API token in the personal center, get the token key: sk-xxxxx and submit.

img

Step 2: Send Requests to deepseek-reasoner API

Select the “deepseek-reasoner” endpoint to send the API request and set the request body. The request method and request body are obtained from our website API doc. Our website also provides Apifox test for your convenience. Replace <YOUR_API_KEY> with your actual CometAPI key from your account. base url is Chat format.

Insert your question or request into the content field—this is what the model will respond to . Process the API response to get the generated answer.

Step 3: Retrieve and Verify Results

Process the API response to get the generated answer. After processing, the API responds with the task status and output data.

Features for DeepSeek-Reasoner

Explore the key features of DeepSeek-Reasoner, designed to enhance performance and usability. Discover how these capabilities can benefit your projects and improve user experience.

Pricing for DeepSeek-Reasoner

Explore competitive pricing for DeepSeek-Reasoner, designed to fit various budgets and usage needs. Our flexible plans ensure you only pay for what you use, making it easy to scale as your requirements grow. Discover how DeepSeek-Reasoner can enhance your projects while keeping costs manageable.
Comet Price (USD / M Tokens)Official Price (USD / M Tokens)Discount
Input:$0.44/M
Output:$1.752/M
Input:$0.55/M
Output:$2.19/M
-20%

Sample code and API for DeepSeek-Reasoner

Access comprehensive sample code and API resources for DeepSeek-Reasoner to streamline your integration process. Our detailed documentation provides step-by-step guidance, helping you leverage the full potential of DeepSeek-Reasoner in your projects.
Python
JavaScript
Curl
from openai import OpenAI
import os

# Get your CometAPI key from https://api.cometapi.com/console/token, and paste it here
COMETAPI_KEY = os.environ.get("COMETAPI_KEY") or "<YOUR_COMETAPI_KEY>"
BASE_URL = "https://api.cometapi.com/v1"

client = OpenAI(base_url=BASE_URL, api_key=COMETAPI_KEY)

completion = client.chat.completions.create(
    model="deepseek-r1",
    messages=[
        {"role": "system", "content": "You are a helpful assistant."},
        {"role": "user", "content": "Hello!"},
    ],
)

print(completion.choices[0].message.content)

Versions of DeepSeek-Reasoner

The reason DeepSeek-Reasoner has multiple snapshots may include potential factors such as variations in output after updates requiring older snapshots for consistency, providing developers a transition period for adaptation and migration, and different snapshots corresponding to global or regional endpoints to optimize user experience. For detailed differences between versions, please refer to the official documentation.
version
deepseek-r1t2-chimera
deepseek-r1
deepseek-r1-0528
deepseek-reasoner

More Models