O

tts-1-hd-1106

Eingabe:$24/M
Ausgabe:$24/M
Kommerzielle Nutzung

Technical Specifications of tts-1-hd-1106

SpecificationDetails
Model IDtts-1-hd-1106
Provider lineageOpenAI text-to-speech model snapshot based on the TTS-1 HD family.
Model typeText-to-speech (TTS): converts text input into spoken audio.
Primary optimizationHigh-quality speech synthesis, with quality prioritized over the lower-latency standard tts-1 variant.
Input modalityText.
Output modalityAudio / spoken speech.
API familyAudio API speech endpoint.
VoicesOpenAI’s text-to-speech guide currently documents 13 built-in voices for the TTS endpoint.
Pricing referenceOpenAI lists tts-1-hd at $30 for input text usage on its model page; snapshot-specific reseller pricing may differ across platforms.
Typical use casesNarration, voiceovers, accessibility audio, spoken assistants, and multilingual audio generation.

What is tts-1-hd-1106?

tts-1-hd-1106 is CometAPI’s platform identifier for an OpenAI high-definition text-to-speech model snapshot associated with the November 2023 TTS release family. It is designed to turn written text into natural-sounding spoken audio while emphasizing audio fidelity and output quality.

In practical terms, this model is suited for applications where polished voice output matters more than the fastest possible response time. Compared with the standard tts-1 family, the HD version is positioned for higher-quality synthesis, making it useful for content narration, professional voice generation, and end-user listening experiences where clarity is important.

Because tts-1-hd-1106 is a snapshot-style identifier, it is best understood as a stable versioned variant of the broader TTS-1 HD model line rather than a separate model family. That interpretation is supported by public model catalogs and community discussion around the 1106 suffix, which aligns with the original November 2023 snapshot naming pattern.

Main features of tts-1-hd-1106

  • High-fidelity speech generation: The model is built for higher-quality text-to-speech output, making it a strong fit for polished audio experiences such as narration, media voiceovers, and premium user-facing audio.
  • Text-to-audio conversion: It accepts written text as input and returns generated speech audio, enabling straightforward speech synthesis pipelines.
  • Snapshot-style versioning: The 1106 suffix indicates a versioned snapshot in the OpenAI naming pattern, which can be useful for teams that want more consistent behavior over time.
  • Voice selection support: OpenAI’s current TTS documentation describes multiple built-in voices available through the speech endpoint, giving developers flexibility in tone and presentation style.
  • Speech endpoint integration: The model belongs to the Audio API speech workflow, so it fits naturally into applications already using OpenAI-compatible audio request patterns.
  • Good fit for accessibility and content narration: Public documentation and ecosystem references consistently position TTS models like this for reading text aloud, generating assistive audio, and creating spoken content from written material.

How to access and integrate tts-1-hd-1106

Step 1: Sign Up for API Key

Sign up on CometAPI and generate your API key from the dashboard. After that, store the key securely in an environment variable such as COMETAPI_API_KEY so your application can authenticate requests safely.

Step 2: Send Requests to tts-1-hd-1106 API

Use CometAPI’s OpenAI-compatible endpoint and specify the model as tts-1-hd-1106.

curl https://api.cometapi.com/v1/audio/speech \
  -H "Content-Type: application/json" \
  -H "Authorization: Bearer $COMETAPI_API_KEY" \
  -d '{
    "model": "tts-1-hd-1106",
    "input": "Hello! This is a sample generated voice response.",
    "voice": "alloy"
  }' --output speech.mp3
from openai import OpenAI

client = OpenAI(
    api_key="<YOUR_COMETAPI_API_KEY>",
    base_url="https://api.cometapi.com/v1"
)

response = client.audio.speech.create(
    model="tts-1-hd-1106",
    voice="alloy",
    input="Hello! This is a sample generated voice response."
)

response.stream_to_file("speech.mp3")

Step 3: Retrieve and Verify Results

Save the returned audio file, then verify that the speech output matches the source text, expected voice, and required quality level. In production, you should also validate file format, playback compatibility, latency, and whether the generated audio meets your narration or accessibility requirements.

Weitere Modelle