Skip to content

Streaming Video Diffusion: Online Video Editing with Diffusion Models

Notifications You must be signed in to change notification settings

Chenfeng1271/SVDiff

Folders and files

NameName
Last commit message
Last commit date

Latest commit

 

History

5 Commits
 
 
 
 

Repository files navigation

Streaming Video Diffusion: Online Video Editing with Diffusion Models

Arxiv

We present a novel task called online video editing, which is designed to edit streaming frames while maintaining temporal consistency. Unlike existing offline video editing assuming all frames are pre-established and accessible, online video editing is tailored to real-life applications such as live streaming and online chat, requiring (1) fast continual step inference, (2) long-term temporal modeling, and (3) zero-shot video editing capability. To solve these issues, we propose Streaming Video Diffusion (SVDiff), which incorporates the compact spatial-aware temporal recurrence into off-the-shelf Stable Diffusion and is trained with the segment-level scheme on large-scale long videos. This simple yet effective setup allows us to obtain a single model that is capable of executing a broad range of videos and editing each streaming frame with temporal coherence. Our experiments indicate that our model can edit long, high-quality videos with remarkable results, achieving a real-time inference speed of 15.2 FPS at a resolution of 512x512.

Gallery

video.mp4

👇 Citation

Please cite our paper if you find our work useful and use it in your work.

@article{chen2024streaming,
  title={Streaming Video Diffusion: Online Video Editing with Diffusion Models},
  author={Chen, Feng and Yang, Zhen and Zhuang, Bohan and Wu, Qi},
  journal={arXiv preprint arXiv:2405.19726},
  year={2024}
}

About

Streaming Video Diffusion: Online Video Editing with Diffusion Models

Resources

Stars

Watchers

Forks

Releases

No releases published

Packages

No packages published