Hurry! Free Tokens Waiting for You – Register Today!

  • Home
  • Models
    • Grok 4 API
    • Suno v4.5
    • GPT-image-1 API
    • GPT-4.1 API
    • Qwen 3 API
    • Llama 4 API
    • GPT-4o API
    • GPT-4.5 API
    • Claude Opus 4 API
    • Claude Sonnet 4 API
    • DeepSeek R1 API
    • Gemini2.5 pro
    • Runway Gen-3 Alpha API
    • FLUX 1.1 API
    • Kling 1.6 Pro API
    • All Models
  • Enterprise
  • Pricing
  • API Docs
  • Blog
  • Contact
Sign Up
Log in

Video

Runway

Runway/gen4_aleph API

Runway Gen-4 Aleph(model id /gen4_aleph) is Runway’s in-context, video-to-video model that extends the Gen-4 family with powerful video editing, shot continuation and view-synthesis capabilities. In plain terms: Aleph can take an input clip and perform complex edits — add/remove/replace objects, relight, restyle, generate novel camera angles, and even generate the “next shot” in a sequence — driven by text prompts and optional reference images. This release is presented as a major step toward coherent, multi-shot video generation and in-context editing.
Get Free API Key
  • Flexible Solution
  • Constant Updates
import os
from openai import OpenAI

client = OpenAI(
    base_url="https://api.cometapi.com/v1",
    api_key="<YOUR_API_KEY>",    
)

response = client.chat.completions.create(
    model="gen4_aleph",
    messages=[
        {
            "role": "system",
            "content": "You are an AI assistant who knows everything.",
        },
        {
            "role": "user",
            "content": "Tell me, why is the sky blue?"
        },
    ],
)

message = response.choices[0].message.content

print(f"Assistant: {message}")

All AI Models in One API
500+ AI Models

Free For A Limited Time! Register Now 

Get 1M Free Token Instantly!

Runway

Runway/gen4_aleph API

Runway Gen-4 Aleph (commonly referenced as gen4_aleph) is Runway’s in-context video editing + video-to-video generation model built on the Gen-4 family. Aleph was announced by Runway as a major extension to Gen-4 that lets users edit, transform and regenerate parts of an existing video (object removal/insertion, relighting, new camera angles, style/lighting changes) using simple text prompts and optional reference images.

Basic features (what gen4_aleph does)

  • Video-to-video editing: Add, remove or replace objects and characters in a clip.
  • Shot continuation / next-shot generation: Extend a story by generating the next shot that follows the input sequence.
  • Novel view / camera generation: Synthesize different camera angles or reframes from the same scene.
  • Style & lighting control: Change overall aesthetic, color grading and scene lighting.
  • Reference conditioning: Use example images (or reference frames) to keep characters/objects consistent across frames.

Technical details (architecture & inputs)

Model type: multi-task video generation/manipulation built on Runway’s Gen-4 architecture — designed for in-context editing where the input video + prompt jointly determine outputs.

Inputs supported: video URL or upload, text prompts, optional reference images, seeds and framing parameters.

Outputs: edited video clips (various aspect ratios and durations).

Processing characteristics: operates as a video-to-video pipeline (server-side task queue), supports reproducibility via seed values, and exposes optional frame size and duration parameters for control. The CometAPI API schemas and examples show a two-step flow (create generation task → poll/retrieve results).

Benchmark performance

Reported or observable performance notes:

  • Consistency: gen4_aleph preserves character appearance and scene continuity across multiple generated shots better than previous public Gen-series models, enabling multi-shot storytelling.
  • Quality vs. speed: Aleph is offered within the Gen-4 family; certain Gen-4 variants (e.g., “Turbo”) trade cost and speed vs. fidelity. Aleph focuses on in-context editing quality for short clips; reviewers note impressive results but caution that outputs still require postwork for production-grade VFX.

Limitations & known failure modes

  • Clip length / resolution limits: Aleph is optimized for short clips (typical 5s generations); longer continuous video requires Act-Two or other pipelines.
  • Control limits: While Aleph improves continuity, fine-grained control (exact finger positions, micro-facial sync, perfect temporal stability) still needs manual VFX or iterative prompt + mask workflows. Community reports and reviewers highlight occasional artifacts and imperfect occlusion handling.
  • Content & copyright moderation: Runway performs moderation and may reject/terminate tasks for disallowed content. Training data provenance and biases remain community discussion points (previous Gen-3 controversy is a reminder to review legal/ethical guidance).

Representative use cases

  • Post-production & VFX assist: Remove boom mics, replace props, relight scenes to speed up editing.
  • Storyboarding & shot prototyping: Generate next shots or alternate camera angles to explore coverage before shooting.
  • Advertising & social content: Rapid re-styling and object replacement for A/B creatives and quick iterations.
  • Game / asset prototyping: Produce short animated assets and environmental variations for concepting pipelines.

Comparison to other models

Vs. Gen-4 (base): Aleph is an in-context extension of Gen-4 focused on editing existing video rather than only generating novel frames from a still image or text. Gen-4 remains the core generator for new content; Aleph adds powerful edit/transform primitives.

Vs. Gen-4 Turbo: Turbo targets faster, lower-cost iterations for image/video generation; Aleph focuses on multi-task editing workflows (Turbo is about throughput/cost).

Vs OpenAI’s Sora (and similar text→video engines): Aleph is designed primarily as a video-to-video editing engine with strong object-level edits and shot continuity; Sora focuses more on text→video generation from scratch and may produce highly cinematic original footage. Reviewers place Runway as stronger in editing existing footage and creative control, while models like Sora and VEO3 are described as strong for text-to-video generation

How to call gen4_aleph API from CometAPI

runwayml_image_to_video:

Price$0.32000

runwayml_video_to_video:

Price$0.96000

Required Steps

  • Log in to cometapi.com. If you are not our user yet, please register first
  • Get the access credential API key of the interface. Click “Add Token” at the API token in the personal center, get the token key: sk-xxxxx and submit.
  • Get the url of this site: https://api.cometapi.com/

Use Method

  1. Select the “gen4_aleph” endpoint to send the API request and set the request body. The request method and request body are obtained from our website API doc. Our website also provides Apifox test for your convenience.
  2. Replace <YOUR_API_KEY> with your actual CometAPI key from your account.
  3. Insert your question or request into the content field—this is what the model will respond to.
  4. . Process the API response to get the generated answer.

CometAPI provides a fully compatible REST API—for seamless migration. Key details to  API doc:

  • Endpoint: https://api.cometapi.com/runwayml/v1/video_to_video
  • Model Parameter: gen4_aleph
  • Authentication:  Bearer YOUR_CometAPI_API_KEY 
  • Content-Type: application/json .
curl --location --request POST 'https://api.cometapi.com/runwayml/v1/video_to_video' \
--header 'X-Runway-Version: 2024-11-06' \
--header 'Authorization: {{api-key}}' \
--header 'Content-Type: application/json' \
--data-raw '{
"videoUri": "https://filesystem.site/cdn/20250818/c4gCDVPhiBc6TomRTJ7zNg0KwO1PSJ.mp4",
"promptText": "string",
"seed": 4294967295,
"model": "gen4_aleph",
"references": [
{
"type": "image",
"uri": "https://cdn.britannica.com/70/234870-050-D4D024BB/Orange-colored-cat-yawns-displaying-teeth.jpg"
}
],
"ratio": "1280:720",
"contentModeration": {
"publicFigureThreshold": "auto"
}
}'

See Also Runway/Act_two

Start Today

One API
Access 500+ AI Models!

Free For A Limited Time! Register Now
Get 1M Free Token Instantly!

Get Free API Key
API Docs

Related posts

AI Model

Runway/upscale_v1 API

2025-09-14 anna No comments yet

Runway/upscale_v1 is Runway’s production-targeted video upscaler model (model id upscale_v1) designed to increase video resolution by 4× up to 4K (capped at 4096 pixels on a side). It’s available through Runway’s API surface and is also packaged on third-party model hosting marketplaces (e.g., CometAPI).

AI Model

Runway/gen4_image API

2025-09-14 anna No comments yet

Gen-4 Image is Runway’s flagship multimodal image-generation model in the Gen-4 family that supports prompted generation plus visual references (you can “@mention” reference images) to produce highly controllable, stylistically consistent outputs for image and image→video pipelines.

AI Model

Runway/Act_two

2025-09-14 anna No comments yet

Act-Two is Runway’s next-generation AI performance capture and character animation tool: it ingests a short driving performance (a webcam or phone video of someone acting a scene) plus a character reference (image or video) and generates an animated character performance that transfers body, facial expression and hand motion to the character. Act-Two is offered inside Runway’s web product and as a model available through API ecosystem.

500+ AI Model API,All In One API. Just In CometAPI

Models API
  • GPT API
  • Suno API
  • Luma API
  • Sora API
Developer
  • Sign Up
  • API DashBoard
  • Documentation
  • Quick Start
Resources
  • Pricing
  • Enterprise
  • Blog
  • AI Model API Articles
  • Discord Community
Get in touch
  • support@cometapi.com

© CometAPI. All Rights Reserved.  

  • Terms & Service
  • Privacy Policy