模型支持企业博客
500+ AI 模型 API,一次搞定,就在 CometAPI
模型 API
开发者
快速入门文档API 仪表板
资源
AI 模型博客企业更新日志关于
2025 CometAPI。保留所有权利。隐私政策服务条款
Home/Models/OpenAI/Whisper-1
O

Whisper-1

输入:$24/M
输出:$24/M
Speech to text, creating translations
商用
概览
功能亮点
定价
API

Technical Specifications of whisper-1

SpecificationDetails
Model IDwhisper-1
Model typeSpeech-to-text and speech translation
Primary use casesAudio transcription, multilingual speech recognition, speech translation into English
Input modalityAudio
Output modalityText
Supported endpoints/v1/audio/transcriptions, /v1/audio/translations
Streaming supportNot supported for whisper-1
Prompting supportYes, with limited prompt control for formatting, punctuation, and style
Language capabilityMultilingual speech recognition and language identification
Typical integration formatFile upload via multipart form data
Common audio formatsm4a, mp3, mp4, mpeg, mpga, wav, webm
Best fit forConverting spoken content into readable text or English translations

What is whisper-1?

whisper-1 is a speech recognition model available through CometAPI for turning audio into text and creating translations from spoken audio into English. It is designed for developers who need reliable transcription for recorded speech, interviews, meetings, voice notes, subtitles, and multilingual audio workflows.

The model is well suited for applications that need automatic speech recognition across multiple languages. It can transcribe audio in the original language or translate spoken content into English, making it useful for global products, media processing pipelines, support tools, and accessibility solutions.

Because whisper-1 works on uploaded audio files and returns text output, it fits naturally into backend automation, content indexing, caption generation, search enrichment, and analytics pipelines.

Main features of whisper-1

  • Speech-to-text transcription: Converts spoken audio into written text for documents, captions, archives, and application workflows.
  • Speech translation: Creates English text translations from non-English spoken audio, simplifying multilingual content processing.
  • Multilingual recognition: Supports recognition across many languages, making it practical for international and cross-region deployments.
  • Prompt-assisted formatting: Accepts prompts that can help guide punctuation, capitalization, terminology, and transcript style.
  • File-based API workflow: Works well with uploaded audio files, making it easy to integrate into batch jobs, media systems, and backend services.
  • Language identification support: Can be used in workflows where detecting or handling multiple spoken languages is important.
  • Strong fit for content operations: Useful for subtitle generation, searchable transcript creation, customer call logging, interview processing, and voice-note conversion.

How to access and integrate whisper-1

Step 1: Sign Up for API Key

To start using whisper-1, first create an account on CometAPI and generate your API key from the dashboard. After logging in, go to the API management section, create a new key, and store it securely. This key will be required to authenticate every request you send to the whisper-1 API.

Step 2: Send Requests to whisper-1 API

Once you have your API key, you can send requests to the CometAPI endpoint using the whisper-1 model ID. Include your API key in the Authorization header and specify whisper-1 as the target model. For speech workflows, send an audio file to the appropriate transcription or translation endpoint.

curl --request POST \
  --url https://api.cometapi.com/v1/audio/transcriptions \
  --header "Authorization: Bearer YOUR_COMETAPI_KEY" \
  --header "Content-Type: multipart/form-data" \
  --form "model=whisper-1" \
  --form "file=@/path/to/audio.mp3"

For translation workflows, use the translation endpoint with the same model ID:

curl --request POST \
  --url https://api.cometapi.com/v1/audio/translations \
  --header "Authorization: Bearer YOUR_COMETAPI_KEY" \
  --header "Content-Type: multipart/form-data" \
  --form "model=whisper-1" \
  --form "file=@/path/to/audio.mp3"

Step 3: Retrieve and Verify Results

After the request is processed, CometAPI will return the generated text result for your whisper-1 job. Review the response to confirm transcript quality, language handling, punctuation, and completeness. If needed, refine your audio preprocessing or prompting approach and resend the request to improve output consistency for your production use case.

Whisper-1 的功能

了解 Whisper-1 的核心能力,帮助提升性能与可用性,并改善整体体验。

Whisper-1 的定价

查看 Whisper-1 的竞争性定价,满足不同预算与使用需求,灵活方案确保随需求扩展。
Comet 价格 (USD / M Tokens)官方定价 (USD / M Tokens)折扣
输入:$24/M
输出:$24/M
输入:$30/M
输出:$30/M
-20%

Whisper-1 的示例代码与 API

获取完整示例代码与 API 资源,简化 Whisper-1 的集成流程,我们提供逐步指导,助你发挥模型潜能。

更多模型

O

gpt-realtime-1.5

输入:$3.2/M
输出:$12.8/M
用于音频输入与音频输出的最佳语音模型。
O

gpt-audio-1.5

输入:$2/M
输出:$8/M
用于在 Chat Completions 中实现音频输入与音频输出的最佳语音模型。
O

TTS

输入:$12/M
输出:$12/M
OpenAI Text-to-Speech
K

Kling TTS

每次请求:$0.006608
[Speech Synthesis] Newly launched: text-to-broadcast audio online, with preview function ● Can simultaneously generate audio_id, usable with any Keling API.
K

Kling video-to-audio

K

Kling video-to-audio

每次请求:$0.03304
Kling video-to-audio
K

Kling text-to-audio

K

Kling text-to-audio

每次请求:$0.03304
Kling text-to-audio