Towards AI Can Help your Team Adopt AI: Corporate Training, Consulting, and Talent Solutions.


Stable Diffusion for Videos Explained
Latest   Machine Learning

Stable Diffusion for Videos Explained

Last Updated on December 11, 2023 by Editorial Team

Author(s): Louis Bouchard

Originally published on Towards AI.

Deep dive into Stable Video Diffusion

Receive my blogs and more on my AI newsletter and receive free gifts, such as my secrets to success on YouTube!

What do all recent super-powerful image generation models like DALLE, or Midjourney have in common? Other than their high computing costs, huge training time, and shared hype, they are all based on the same mechanism: diffusion.

Diffusion models are the state-of-the-art results for most image tasks, including text-to-image with DALLE but many other image generation-related tasks too, like image inpainting, style transfer, or image super-resolution.

Then, latent diffusion or the well-known stable diffusion came out, changing everything when it comes to image generation.

But I am not here to talk about old news. We are here to go over the new paper and model released by Stability AI: Stable Video Diffusion. The most recent and open-source video generation model that you can use right now! It takes either images or text and is able to generate cool videos like these automatically. It can even be used to generate multiple views of an object as if it were in 3D.

I’m Louis from What’s AI, and let’s dive into how this new model works!

Video generated by Stable Video Diffusion.

Before getting to videos, let’s do a… Read the full blog for free on Medium.

Join thousands of data leaders on the AI newsletter. Join over 80,000 subscribers and keep up to date with the latest developments in AI. From research to projects and ideas. If you are building an AI startup, an AI-related product, or a service, we invite you to consider becoming a sponsor.

Published via Towards AI

Feedback ↓