Midjourney Unveils V1 Video: First AI Video Generation Model

Midjourney has officially unveiled its debut AI video generation tool, the V1 Video Model, marking a significant expansion of its creative capabilities beyond static imagery. The feature is now available to all subscription users and allows them to generate 5-second animated video clips from a single still image, with optional text prompts to guide motion and style.
Key Features of the V1 Video Model
- Image-to-Video Animation: Users can upload their own image or use an existing Midjourney-generated one to produce four 5-second video variants. The animations bring life to static visuals using smooth camera and object movement.
- Extendable Playback Time: Initial clips are 5 seconds long, but users can extend each video by four additional 4-second increments, up to a maximum of 20–21 seconds.
- Motion Modes:
- Low Motion: Subtle effects for mostly static scenes.
- High Motion: Designed for scenes with significant camera pans or character movement.
- Custom Prompts: Motion can be guided by text prompts, allowing users to dictate camera movements, object actions, or transitions.
- Two Animation Styles:
- Auto Mode: Automatically generates animation from the image.
- Manual Mode: Allows prompt editing before animation rendering.
A Vision Beyond Short Clips
Midjourney CEO David Holz stated that this model isn’t just about creating stylish B-roll or ads. It represents a crucial step toward the company’s long-term goal: building a real-time, AI-driven open-world simulation system. This ambitious vision includes:
- A fully interactive 3D space where users can move freely.
- Real-time image rendering and scene response.
- Dynamic characters and environments that respond to user actions.
To achieve this, Midjourney is rolling out its foundational components in phases:
- Image Generation (completed)
- Video Animation (launched)
- 3D Interaction (in development)
- Real-time Rendering Optimization (planned)
Usability and Pricing
- Subscription Access: Priced at $10/month, the V1 model is positioned as a highly accessible creative tool for artists and content creators.
- Rendering Cost: A batch of 20 extended clips (~80 seconds total) consumes about one “fast hour,” costing roughly $4—a lower rate compared to competitors like Google’s Veo 3.
- Video Quality: While not significantly exceeding rivals like Luma Labs’ Dream Machine, Midjourney maintains a high level of visual fidelity. Each frame resembles a digital painting with a cinematic aesthetic.
Limitations
- No Audio Support: Unlike Veo 3 or Dream Machine, Midjourney’s model does not currently generate soundtracks or ambient audio.
- Editing Constraints: No timeline editing, scene transitions, or segment linking are available.
- Rendering Time: Generation speed may lag slightly behind competitors when dealing with complex scenes.
Competitive Landscape
Midjourney enters a crowded field that includes Runway, Luma Labs, Google’s Veo 3, and MiniMax’s Hailuo 02.
Strengths:
- Seamless integration with Midjourney’s image generation workflow.
- Intuitive UI and cost-effective access.
- Ideal for experimental short-form content.
Weaknesses:
- No audio.
- Limited editing and scene duration.
- Fixed resolution at 480p (standard definition).
How to Generate Videos
- Join the Midjourney Discord
Ensure you’re in a channel where the Midjourney Bot is active. If not, add the bot to your server or join the official Midjourney server. - Invoke the Video Command
Use the/imagine
command with your descriptive prompt, appending the--video
parameter. For example:
/imagine “a floating lantern drifting over a serene lake at sunset” --video
- Customize Duration (Optional)
By default, V1 produces a 10‑second clip, but you can adjust length up to 20 seconds with--duration
. E.g.:
/imagine “city skyline at dawn” --video --duration 20s
- Await Generation
The bot processes your request and delivers a video file or link. Processing times may vary based on server load and your subscription tier. - Download or Share
Once generated, click the provided link to view, download, or share your video across platforms .
Compliance and Safety
Midjourney enforces strict guidelines:
- Only images with legal usage rights may be uploaded.
- Generating offensive, pornographic, or inciting content—especially involving real people—is prohibited.
- The system automatically filters violations without charging GPU time.
This launch signifies Midjourney’s serious move into video generation and sets the foundation for its broader ambitions in interactive AI worlds. While it’s not yet a full filmmaking suite, it presents an accessible and visually compelling tool for short-form creativity.
Use MidJourney in CometAPI
CometAPI provides access to over 500 AI models, including open-source and specialized multimodal models for chat, images, code, and more. Its primary strength lies in simplifying the traditionally complex process of AI integration.
CometAPI offer a price far lower than the official price to help you integrate Midjourney API, and you will get $1 in your account after registering and logging in! Welcome to register and experience CometAPI.CometAPI pays as you go.
Important Prerequisite: Before using MidJourney V7, you need to Start building on CometAPI today – sign up here for free access. Please visit docs. Getting started with MidJourney V7 is very simple—just add the --v 7
parameter at the end of your prompt. This simple command tells CometAPI to use the latest V7 model to generate your image.
The latest integration V1 Video Model API will soon appear on CometAPI, so stay tuned!While we finalize V1 Video Model upload, explore our other models on the Models page or try them in the AI Playground.