On December 19th, Google unveiled VideoPoet, a video generation model capable of producing videos up to 10 seconds long and automatically generating accompanying music and sound effects based on the video content. VideoPoet extends the video by repeatedly predicting the next frame after the last frame, giving users the impression that the video can be infinitely extended. Unlike other models, VideoPoet utilizes a large language model rather than a diffusion model, integrating multiple functionalities such as text-to-video, video restoration, and video stylization into a single model, offering greater flexibility in use.
Google Releases VideoPoet Video Generation Model, Supporting Up to Ten Seconds of Video and Audio Generation

36氪
This article is from AIbase Daily
Welcome to the [AI Daily] column! This is your daily guide to exploring the world of artificial intelligence. Every day, we present you with hot topics in the AI field, focusing on developers, helping you understand technical trends, and learning about innovative AI product applications.