—
VideoPoet is a state-of-the-art AI video generation model developed by Google Research. Unlike traditional AI tools, VideoPoet is a large language model for video that can generate, edit, and extend videos directly from text, image, or video inputs. It brings together capabilities like video prediction, stylization, inpainting, and audio-to-video synthesis—all within a unified, autoregressive transformer framework.
—
Text-to-Video Generation*: Create short video clips purely from descriptive text prompts
Image-to-Video Animation*: Bring static images to life with smooth, AI-driven motion
Video Inpainting*: Fill in missing or masked regions in existing videos
Stylized Video Generation*: Apply artistic styles or aesthetics to generated motion
Audio-to-Video*: Sync generated visuals to spoken words, music, or sound
Autoregressive Model*: Built on a transformer that predicts video tokens in sequence for high realism
—
—
Currently not publicly available* for commercial use
Demo videos* and research insights available at [Google’s official project page](https://sites.research.google/videopoet/)
> VideoPoet remains in the research phase; there is no public API or product release yet.
—
VideoPoet is one of the first models to unify video generation, extension, and editing in a single transformer-based system. Its ability to blend text, image, audio, and motion into coherent outputs places it at the forefront of generative video research—pushing the boundaries of AI creativity and human–machine collaboration.
—
Overall Score*: 4.7/5 (research status)
—
Find more tools on ThisAIWillDoIt.com.
There are no similar listings