ModelsSupportEnterpriseBlog
500+ AI Model API, All In One API.Just In CometAPI
Models API
Developer
Quick StartDocumentationAPI Dashboard
Resources
AI ModelsBlogEnterpriseChangelogAbout
2025 CometAPI. All right reserved.Privacy PolicyTerms of Service
Home/Models/OpenAI/o1-mini-2024-09-12
O

o1-mini-2024-09-12

Input:$0.88/M
Output:$3.52/M
Commercial Use
Overview
Features
Pricing
API

Technical Specifications of o1-mini-2024-09-12

SpecificationDetails
Model IDo1-mini-2024-09-12
ProviderOpenAI
Model familyo1 series reasoning model
Release snapshotSeptember 12, 2024 snapshot of o1-mini
Primary modalityText input, text output
Core strengthCost-efficient reasoning, especially for STEM, math, and coding tasks
Relative positioningFaster and lower-cost than o1-preview, with strong performance on coding-oriented reasoning workloads
Training approachReinforcement-learning-based reasoning model designed to spend more time thinking before responding
Availability statusSnapshot listed by OpenAI as deprecated

What is o1-mini-2024-09-12?

o1-mini-2024-09-12 is a snapshot of OpenAI’s o1-mini reasoning model, released on September 12, 2024. It belongs to the o1 family, which OpenAI introduced as models that “think before they answer” and are optimized for complex reasoning rather than only fast next-token generation.

Compared with larger o1 variants, o1-mini was positioned as the faster and more economical option for workloads that need strong reasoning without requiring the broadest possible world knowledge. OpenAI specifically highlighted its usefulness for STEM-heavy applications, noting that it performs especially well in math and coding and was designed as a cost-efficient alternative to o1-preview.

In practical terms, this makes o1-mini-2024-09-12 a good fit for developers building applications such as code assistants, technical problem-solving tools, structured analytical workflows, and math-focused copilots. Because this exact snapshot is now marked deprecated in OpenAI’s model documentation, teams using the CometAPI identifier should verify ongoing compatibility and behavior in their own environment.

Main features of o1-mini-2024-09-12

  • Reasoning-first design: OpenAI describes the o1 series as models trained to perform complex reasoning and to spend more time thinking before responding, which can improve performance on multi-step technical tasks.
  • Strong STEM performance: o1-mini was explicitly introduced as excelling at STEM workloads, especially math and coding, making it suitable for engineering and analytical use cases.
  • Lower cost profile: OpenAI stated that o1-mini launched at a significantly lower cost than o1-preview, positioning it as the more budget-friendly reasoning option.
  • Faster response characteristics: The model was presented as faster than o1-preview, which is useful when balancing reasoning quality with latency-sensitive application needs.
  • Good fit for coding applications: OpenAI’s release materials and system documentation repeatedly describe o1-mini as particularly effective for coding-related tasks.
  • Snapshot stability: Using the exact snapshot ID o1-mini-2024-09-12 can help teams target a fixed model version for reproducibility, though OpenAI currently labels this snapshot as deprecated.

How to access and integrate o1-mini-2024-09-12

Step 1: Sign Up for API Key

To access o1-mini-2024-09-12, first create an account on CometAPI and generate an API key from the dashboard. Once you have your key, store it securely as an environment variable so your application can authenticate requests to the API.

Step 2: Send Requests to o1-mini-2024-09-12 API

After getting your API key, send requests to CometAPI’s OpenAI-compatible endpoint while setting the model field to o1-mini-2024-09-12.

curl https://api.cometapi.com/v1/responses \
  -H "Content-Type: application/json" \
  -H "Authorization: Bearer $COMETAPI_API_KEY" \
  -d '{
    "model": "o1-mini-2024-09-12",
    "input": "Write a Python function that solves a quadratic equation and explain the math."
  }'

You can also use the OpenAI SDK format by pointing the client to CometAPI’s base URL and keeping the model name as o1-mini-2024-09-12.

from openai import OpenAI

client = OpenAI(
    api_key="YOUR_COMETAPI_KEY",
    base_url="https://api.cometapi.com/v1"
)

response = client.responses.create(
    model="o1-mini-2024-09-12",
    input="Solve this step by step: If 3x + 5 = 20, what is x?"
)

print(response)

Step 3: Retrieve and Verify Results

Once the API returns a response, parse the output text in your application and validate it for your use case. For reasoning-heavy tasks such as coding, mathematics, or technical analysis, it is a good practice to add automated checks, test cases, or human review to verify that the model’s conclusions are correct before using them in production.

Features for o1-mini-2024-09-12

Explore the key features of o1-mini-2024-09-12, designed to enhance performance and usability. Discover how these capabilities can benefit your projects and improve user experience.

Pricing for o1-mini-2024-09-12

Explore competitive pricing for o1-mini-2024-09-12, designed to fit various budgets and usage needs. Our flexible plans ensure you only pay for what you use, making it easy to scale as your requirements grow. Discover how o1-mini-2024-09-12 can enhance your projects while keeping costs manageable.
Comet Price (USD / M Tokens)Official Price (USD / M Tokens)Discount
Input:$0.88/M
Output:$3.52/M
Input:$1.1/M
Output:$4.4/M
-20%

Sample code and API for o1-mini-2024-09-12

Access comprehensive sample code and API resources for o1-mini-2024-09-12 to streamline your integration process. Our detailed documentation provides step-by-step guidance, helping you leverage the full potential of o1-mini-2024-09-12 in your projects.

More Models

G

Nano Banana 2

Input:$0.4/M
Output:$2.4/M
Core Capabilities Overview: Resolution: Up to 4K (4096×4096), on par with Pro. Reference Image Consistency: Up to 14 reference images (10 objects + 4 characters), maintaining style/character consistency. Extreme Aspect Ratios: New 1:4, 4:1, 1:8, 8:1 ratios added, suitable for long images, posters, and banners. Text Rendering: Advanced text generation, suitable for infographics and marketing poster layouts. Search Enhancement: Integrated Google Search + Image Search. Grounding: Built-in thinking process; complex prompts are reasoned before generation.
A

Claude Opus 4.6

Input:$4/M
Output:$20/M
Claude Opus 4.6 is Anthropic’s “Opus”-class large language model, released February 2026. It is positioned as a workhorse for knowledge-work and research workflows — improving long-context reasoning, multi-step planning, tool use (including agentic software workflows), and computer-use tasks such as automated slide and spreadsheet generation.
A

Claude Sonnet 4.6

Input:$2.4/M
Output:$12/M
Claude Sonnet 4.6 is our most capable Sonnet model yet. It’s a full upgrade of the model’s skills across coding, computer use, long-context reasoning, agent planning, knowledge work, and design. Sonnet 4.6 also features a 1M token context window in beta.
O

GPT-5.4 nano

Input:$0.16/M
Output:$1/M
GPT-5.4 nano is designed for tasks where speed and cost matter most like classification, data extraction, ranking, and sub-agents.
O

GPT-5.4 mini

Input:$0.6/M
Output:$3.6/M
GPT-5.4 mini brings the strengths of GPT-5.4 to a faster, more efficient model designed for high-volume workloads.
A

Claude Mythos Preview

A

Claude Mythos Preview

Coming soon
Input:$60/M
Output:$240/M
Claude Mythos Preview is our most capable frontier model to date, and shows a striking leap in scores on many evaluation benchmarks compared to our previous frontier model, Claude Opus 4.6.