Hurry! Free Tokens Waiting for You – Register Today!

  • Home
  • Models
    • Grok 4 API
    • Suno v4.5
    • GPT-image-1 API
    • GPT-4.1 API
    • Qwen 3 API
    • Llama 4 API
    • GPT-4o API
    • GPT-4.5 API
    • Claude Opus 4 API
    • Claude Sonnet 4 API
    • DeepSeek R1 API
    • Gemini2.5 pro
    • Runway Gen-3 Alpha API
    • FLUX 1.1 API
    • Kling 1.6 Pro API
    • All Models
  • Enterprise
  • Pricing
  • API Docs
  • Blog
  • Contact
Sign Up
Log in

Image Generation

Runway

Runway/gen4_image API

Gen-4 Image is Runway’s flagship multimodal image-generation model in the Gen-4 family that supports prompted generation plus visual references (you can “@mention” reference images) to produce highly controllable, stylistically consistent outputs for image and image→video pipelines.
Get Free API Key
  • Flexible Solution
  • Constant Updates
import os
from openai import OpenAI

client = OpenAI(
    base_url="https://api.cometapi.com/v1",
    api_key="<YOUR_API_KEY>",    
)

response = client.chat.completions.create(
    model="Gen-4 Image",
    messages=[
        {
            "role": "system",
            "content": "You are an AI assistant who knows everything.",
        },
        {
            "role": "user",
            "content": "Tell me, why is the sky blue?"
        },
    ],
)

message = response.choices[0].message.content

print(f"Assistant: {message}")

All AI Models in One API
500+ AI Models

Free For A Limited Time! Register Now 

Get 1M Free Token Instantly!

Runway

Runway/gen4_image API

gen4_image is Runway’s flagship multimodal image-generation model in the Gen-4 family that supports prompted generation plus visual references (you can “@mention” reference images) to produce highly controllable, stylistically consistent outputs for image and image→video pipelines.

Introduction — what Gen-4 Image is

gen4_image is Runway’s fourth-generation visual generative model family, engineered to take text prompts + visual references and produce high-fidelity still images or media-ready frames that preserve identity and style across angles and lighting. The model is presented as part of a broader Gen-4 suite (including video variants such as gen4_turbo) and is explicitly designed for creative production — e.g., consistent character rendering, product photography at scale, virtual try-on, and game asset generation.

Key features

  • Reference-based generation (1–3 refs). Use up to three reference images so the model can preserve identity, style, or location while transforming pose, lighting, background, etc.
  • High visual fidelity (production-ready outputs). Outputs target high resolution (1080p options available) with strong detail and stylistic control.
  • Identity & scene consistency. Designed to keep the same character(s) or environment consistent across multiple generations — useful for multi-shot visuals or character-centric assets.
  • Multimodal (text + images) prompts. Combine natural language instructions with reference images to steer composition, mood, clothing, camera angle, etc.
  • Image → image plus text → image workflows. Works as image-to-image (edit/transform) and as text-to-image using references to maintain continuity.
  • Performance tier (Turbo) available. A “Gen-4 Image Turbo” variant trades cost and speed (e.g., ~2.5× faster) while keeping the reference-driven features.
  • Controls & reproducibility. Typical API options include aspect ratio presets, resolution (720p/1080p), seed for reproducibility, and reference tags to point to specific inputs.

Technical details

Input: Text/Image

Outputs: Image

workflow:

  1. User supplies: text prompt + 0–3 reference images (and optional masks, keyframes, camera motion instructions).
  2. Preprocess: references are normalized and encoded; text is tokenized. Identity/style embeddings are extracted and cached for reuse.
  3. Conditioning: text and reference embeddings are fused in the multimodal backbone; optional control signals (pose, depth, mask) are attached.
  4. Sampling / denoising: the decoder runs denoising iterations (diffusion steps) producing an image (or sequence of frames for video).

gen4_image — concrete limits

Temporal / motion edge cases. Reviewers and creators report occasional motion artifacts, odd temporal dynamics (glitches early/late in generated clips), and failures on very complex multi-actor choreography — test with your target scenes.

Compute, cost & queuing. High-quality image→video generation is GPU-heavy; users report queue times and cost/per-render that can be significant for mass production. Plan budget/throughput accordingly.

Creative tradeoffs vs pure artistry models. Gen-4’s strength is consistency; if you need highly stylized, painterly, or “surprising” aesthetic outputs, Midjourney or tuned SDXL checkpoints may produce preferred art directions.

Canonical use cases

  • Pre-production & storyboarding: rapidly create style-consistent character/scene variants from reference photos.
  • Marketing & content generation: fast production of hero images, animated social clips, and campaign assets with consistent brand characters. (Runway lists enterprise examples including live tours and music videos.)
  • Game/asset prototyping & virtual try-on: generate multiple camera angles, outfit variants and environment concepts from a small set of references.

Comparison to other models

  • gen4_image→ best when you need reference / identity consistency (single-character or object kept the same across shots) and when you want image→video and multi-shot pipelines.
  • DALL·E 3 → best for tight prompt-to-image fidelity and a conversational ChatGPT-driven editing flow plus built-in safety/provenance work.
  • SDXL (Stable Diffusion family) → best when you want open models, local/custom fine-tuning, and cost-flexible deployment.
  • Midjourney → best for highly stylized, artistically pleasing renders and strong community-driven presets / “stylize” controls.
  • Runway Gen-4 vs. ByteDance Seedream 4.0 / Google “Nano Banana” type models: recent competitor launches (e.g., Seedream 4.0) emphasize ultra-fast rendering and multi-reference handling aimed at commercial creators; Runway’s advantage is a tightly integrated image→video pipeline and production-oriented controls plus a mature API and SDK ecosystem.

How to call gen4_image API from CometAPI

Price$0.32000

Required Steps

  • Log in to cometapi.com. If you are not our user yet, please register first
  • Get the access credential API key of the interface. Click “Add Token” at the API token in the personal center, get the token key: sk-xxxxx and submit.
  • Get the url of this site: https://api.cometapi.com/

Use Method

  1. Select the “gen4_image” endpoint to send the API request and set the request body. The request method and request body are obtained from our website API doc. Our website also provides Apifox test for your convenience.
  2. Replace <YOUR_API_KEY> with your actual CometAPI key from your account.
  3. Insert your question or request into the content field—this is what the model will respond to.
  4. . Process the API response to get the generated answer.

CometAPI provides a fully compatible REST API—for seamless migration. Key details to  API doc:

  • Endpoint: https://api.cometapi.com/runwayml/v1/text_to_image
  • Model Parameter: gen4_image
  • Authentication:  Bearer YOUR_CometAPI_API_KEY 
  • Content-Type: application/json .
curl --location --request POST 'https://api.cometapi.com/runwayml/v1/text_to_image' \
--header 'X-Runway-Version: 2024-11-06' \
--header 'Authorization: {{api-key}}' \
--header 'Content-Type: application/json' \
--data-raw '{
"promptText": "cat",
"ratio": "1920:1080",
"seed": 4294967295,
"model": "gen4_image",
"referenceImages": [
{
"uri": "https://cdn.britannica.com/70/234870-050-D4D024BB/Orange-colored-cat-yawns-displaying-teeth.jpg",
"tag": "string"
}
],
"contentModeration": {
"publicFigureThreshold": "auto"
}
}'

See also Runway/Act_two

Start Today

One API
Access 500+ AI Models!

Free For A Limited Time! Register Now
Get 1M Free Token Instantly!

Get Free API Key
API Docs

Related posts

AI Model

Runway/upscale_v1 API

2025-09-14 anna No comments yet

Runway/upscale_v1 is Runway’s production-targeted video upscaler model (model id upscale_v1) designed to increase video resolution by 4× up to 4K (capped at 4096 pixels on a side). It’s available through Runway’s API surface and is also packaged on third-party model hosting marketplaces (e.g., CometAPI).

AI Model

Runway/gen4_aleph API

2025-09-14 anna No comments yet

Runway Gen-4 Aleph(model id /gen4_aleph) is Runway’s in-context, video-to-video model that extends the Gen-4 family with powerful video editing, shot continuation and view-synthesis capabilities. In plain terms: Aleph can take an input clip and perform complex edits — add/remove/replace objects, relight, restyle, generate novel camera angles, and even generate the “next shot” in a sequence — driven by text prompts and optional reference images. This release is presented as a major step toward coherent, multi-shot video generation and in-context editing.

AI Model

Runway/Act_two

2025-09-14 anna No comments yet

Act-Two is Runway’s next-generation AI performance capture and character animation tool: it ingests a short driving performance (a webcam or phone video of someone acting a scene) plus a character reference (image or video) and generates an animated character performance that transfers body, facial expression and hand motion to the character. Act-Two is offered inside Runway’s web product and as a model available through API ecosystem.

500+ AI Model API,All In One API. Just In CometAPI

Models API
  • GPT API
  • Suno API
  • Luma API
  • Sora API
Developer
  • Sign Up
  • API DashBoard
  • Documentation
  • Quick Start
Resources
  • Pricing
  • Enterprise
  • Blog
  • AI Model API Articles
  • Discord Community
Get in touch
  • support@cometapi.com

© CometAPI. All Rights Reserved.  

  • Terms & Service
  • Privacy Policy