ModelsSupportEnterpriseBlog
500+ AI Model API, All In One API.Just In CometAPI
Models API
Developer
Quick StartDocumentationAPI Dashboard
Resources
AI ModelsBlogEnterpriseChangelogAbout
2025 CometAPI. All right reserved.Privacy PolicyTerms of Service
Home/Models/OpenAI/tts-1-1106
O

tts-1-1106

Input:$12/M
Output:$12/M
Commercial Use
Overview
Features
Pricing
API

Technical Specifications of tts-1-1106

AttributeDetails
Model IDtts-1-1106
Provider familyOpenAI text-to-speech model family
Primary capabilityConverts text input into natural-sounding spoken audio
Typical endpoint/v1/audio/speech
Optimization focusLow-latency, realtime-oriented speech generation
Input modalityText
Output modalityAudio
Supported output formatsmp3, opus, aac, flac, wav, pcm
Voice supportCompatible with OpenAI’s built-in TTS voices; the tts-1 / tts-1-hd family supports a smaller subset including alloy, ash, coral, echo, fable, nova, onyx, sage, and shimmer
Max input length4096 characters per request
Pricing referenceOpenAI lists TTS speech generation pricing at $15.00 per 1M characters for the TTS category
Compliance noteEnd users should be clearly informed when the voice they hear is AI-generated

What is tts-1-1106?

tts-1-1106 is CometAPI’s platform identifier for an OpenAI text-to-speech model in the tts-1 family, designed to transform written text into spoken audio. OpenAI documents tts-1 as a model optimized for speed and realtime use cases, making it suitable for interactive applications that need fast speech generation rather than maximum offline rendering quality.

In practice, this model class is used for scenarios such as narration, voice assistants, accessibility features, conversational interfaces, and automated audio playback. It is accessed through the speech-generation workflow of the Audio API, where developers provide text, select a supported voice, and receive audio in a chosen output format.

Because the exact -1106 suffix appears to be a platform-side identifier rather than the public OpenAI model alias, the safest interpretation is that tts-1-1106 maps to the behavior and integration pattern of OpenAI’s tts-1 generation family. That means developers should expect a fast TTS model focused on responsive synthesis, standard speech endpoint usage, and built-in voice selection.

Main features of tts-1-1106

  • Realtime-oriented speech generation: The underlying tts-1 family is optimized for speed, which makes it well suited for live applications, assistants, and other latency-sensitive audio experiences.
  • Natural-sounding text-to-audio conversion: It converts plain text into lifelike spoken output for playback, narration, and voice-enabled product features.
  • Multiple output formats: Developers can request audio in mp3, opus, aac, flac, wav, or raw pcm, which supports both consumer playback and lower-latency system integration.
  • Built-in voice options: The model family supports a set of preset voices, letting teams choose a delivery style that fits their product tone without training a custom speaker model.
  • Straightforward API integration: The model is designed to work through the standard speech endpoint, reducing implementation complexity for teams already using OpenAI-compatible audio APIs.
  • Language flexibility: OpenAI states its TTS stack generally follows Whisper language support, enabling speech generation across many languages even though voices are primarily optimized for English.
  • Streaming-friendly usage: OpenAI’s speech API supports streamed audio delivery, allowing playback to begin before the full file is finished in suitable implementations.
  • Practical for production apps: With documented rate limits, standardized endpoint behavior, and usage-policy guidance around AI voice disclosure, the model family is suitable for real application deployment.

How to access and integrate tts-1-1106

Step 1: Sign Up for API Key

Sign up on CometAPI and generate your API key from the dashboard. Store the key securely and configure it as an environment variable in your application so your backend can authenticate requests to the tts-1-1106 API.

Step 2: Send Requests to tts-1-1106 API

Send a POST request to the OpenAI-compatible audio speech endpoint through CometAPI, setting model to tts-1-1106 and including the input text plus any supported options such as voice and response_format.

curl --request POST \
  --url https://api.cometapi.com/v1/audio/speech \
  --header "Authorization: Bearer $COMETAPI_API_KEY" \
  --header "Content-Type: application/json" \
  --data '{
    "model": "tts-1-1106",
    "input": "Welcome to CometAPI text to speech.",
    "voice": "alloy",
    "response_format": "mp3"
  }' \
  --output speech.mp3

Step 3: Retrieve and Verify Results

Save the returned audio file or stream the response directly in your application, then verify that the speech content, selected voice, format, and playback quality match your expected output for tts-1-1106.

Features for tts-1-1106

Explore the key features of tts-1-1106, designed to enhance performance and usability. Discover how these capabilities can benefit your projects and improve user experience.

Pricing for tts-1-1106

Explore competitive pricing for tts-1-1106, designed to fit various budgets and usage needs. Our flexible plans ensure you only pay for what you use, making it easy to scale as your requirements grow. Discover how tts-1-1106 can enhance your projects while keeping costs manageable.
Comet Price (USD / M Tokens)Official Price (USD / M Tokens)Discount
Input:$12/M
Output:$12/M
Input:$15/M
Output:$15/M
-20%

Sample code and API for tts-1-1106

Access comprehensive sample code and API resources for tts-1-1106 to streamline your integration process. Our detailed documentation provides step-by-step guidance, helping you leverage the full potential of tts-1-1106 in your projects.

More Models

G

Nano Banana 2

Input:$0.4/M
Output:$2.4/M
Core Capabilities Overview: Resolution: Up to 4K (4096×4096), on par with Pro. Reference Image Consistency: Up to 14 reference images (10 objects + 4 characters), maintaining style/character consistency. Extreme Aspect Ratios: New 1:4, 4:1, 1:8, 8:1 ratios added, suitable for long images, posters, and banners. Text Rendering: Advanced text generation, suitable for infographics and marketing poster layouts. Search Enhancement: Integrated Google Search + Image Search. Grounding: Built-in thinking process; complex prompts are reasoned before generation.
A

Claude Opus 4.6

Input:$4/M
Output:$20/M
Claude Opus 4.6 is Anthropic’s “Opus”-class large language model, released February 2026. It is positioned as a workhorse for knowledge-work and research workflows — improving long-context reasoning, multi-step planning, tool use (including agentic software workflows), and computer-use tasks such as automated slide and spreadsheet generation.
A

Claude Sonnet 4.6

Input:$2.4/M
Output:$12/M
Claude Sonnet 4.6 is our most capable Sonnet model yet. It’s a full upgrade of the model’s skills across coding, computer use, long-context reasoning, agent planning, knowledge work, and design. Sonnet 4.6 also features a 1M token context window in beta.
O

GPT-5.4 nano

Input:$0.16/M
Output:$1/M
GPT-5.4 nano is designed for tasks where speed and cost matter most like classification, data extraction, ranking, and sub-agents.
O

GPT-5.4 mini

Input:$0.6/M
Output:$3.6/M
GPT-5.4 mini brings the strengths of GPT-5.4 to a faster, more efficient model designed for high-volume workloads.
A

Claude Mythos Preview

A

Claude Mythos Preview

Coming soon
Input:$60/M
Output:$240/M
Claude Mythos Preview is our most capable frontier model to date, and shows a striking leap in scores on many evaluation benchmarks compared to our previous frontier model, Claude Opus 4.6.