Meta has launched Movie Gen, a powerful AI model designed to produce 1080p videos with synchronised audio. The system can edit videos based on instructions, allowing for personalised content creation using user-supplied images.
With a transformer model containing 30 billion parameters, Movie Gen can generate 16-second videos at 16 frames per second. The model’s advanced techniques improve video motion realism, trained on over 100 million video-text pairs and 1 billion image-text pairs.
Movie Gen outperforms previous models, including Runway Gen3 and OpenAI Sora, particularly in video editing and text-to-video tasks. Benchmarks show its superiority in maintaining video structure and fine details, especially in the TGVE+ test.
Future developments for Movie Gen include improving scene understanding, safeguarding against misuse, and making the system more accessible. Meta envisions applications in social media, film production, and marketing campaigns.