Stable Diffusion for Videos Explained
Last Updated on December 11, 2023 by Editorial Team
Author(s): Louis Bouchard
Originally published on Towards AI.
Deep dive into Stable Video Diffusion
Receive my blogs and more on my AI newsletter and receive free gifts, such as my secrets to success on YouTube!
https://www.youtube.com/embed/TVcE1Ic05lw
What do all recent super-powerful image generation models like DALLE, or Midjourney have in common? Other than their high computing costs, huge training time, and shared hype, they are all based on the same mechanism: diffusion.
Diffusion models are the state-of-the-art results for most image tasks, including text-to-image with DALLE but many other image generation-related tasks too, like image inpainting, style transfer, or image super-resolution.
Then, latent diffusion or the well-known stable diffusion came out, changing everything when it comes to image generation.
But I am not here to talk about old news. We are here to go over the new paper and model released by Stability AI: Stable Video Diffusion. The most recent and open-source video generation model that you can use right now! It takes either images or text and is able to generate cool videos like these automatically. It can even be used to generate multiple views of an object as if it were in 3D.
Iβm Louis from Whatβs AI, and letβs dive into how this new model works!
Video generated by Stable Video Diffusion.
Before getting to videos, letβs do a… Read the full blog for free on Medium.
Join thousands of data leaders on the AI newsletter. Join over 80,000 subscribers and keep up to date with the latest developments in AI. From research to projects and ideas. If you are building an AI startup, an AI-related product, or a service, we invite you to consider becoming aΒ sponsor.
Published via Towards AI