Runway, a company specializing in video-centric artificial intelligence (AI), announced a new video generation model on Monday. Named Gen-4, this image-to-video generation model builds upon its predecessor, the Gen-3 Alpha AI. The latest version boasts improved consistency in characters, locations, and objects throughout scenes, in addition to enhanced controllable real-world physics. Runway asserts that Gen-4 also features increased adherence to prompts, allowing it to maintain the style, mood, and cinematic qualities of a scene using straightforward commands.
Runway Unveils Gen-4 Image-to-Video Generation Model
The official Runway account shared the announcement on X (formerly Twitter), revealing that Gen-4 is currently being rolled out to the company’s premium tiers and enterprise clients. However, details regarding availability for the free tier remain undisclosed. The announcement further highlighted that “Gen-4 is a significant step forward for fidelity, dynamic motion and controllability in generative media.”
The new model advances beyond Gen-3 Alpha with multiple enhancements aimed at providing a more consistent style and subject across video and image outputs. Runway has also showcased several short films that were entirely created using the Gen-4 video generation capabilities.
In a corresponding blog post, the company elaborated on the new model’s features. According to Runway, with just a single reference image, Gen-4 can produce consistent character representations across varying lighting conditions, locations, and camera angles. Similarly, objects can be incorporated seamlessly into different environments, ensuring uniformity. This functionality aims to assist users in generating videos capable of storytelling or showcasing products while utilizing the same reference image.
Furthermore, by supplying a text description alongside the reference image, users can prompt the AI model to create scenes from diverse perspectives, including close-ups and wide-angle side profiles, even capturing intricate details absent from the original reference. Another claim made by the company suggests that Gen-4 excels in its understanding of real-world physics and motion.
During interactions between subjects and their environments in videos, the model incorporates realistic physics and motion dynamics. Demonstration videos released by the company illustrate this feature, showcasing realistic splashes of water and lifelike movements of bushes.
Despite the advancements presented, Runway did not disclose the dataset utilized to train the AI model for its dynamic and high-fidelity outputs. This lack of information is notable, especially in light of ongoing legal challenges the company faces from artists and competing generative AI firms, who allege that Runway has trained its models using copyrighted material without obtaining permission.