Skip to content

Vchitect/VideoBooth

Folders and files

NameName
Last commit message
Last commit date

Latest commit

 

History

5 Commits
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 

Repository files navigation

VideoBooth

Paper Project Page Video Visitor

This repository will contain the implementation of the following paper:

VideoBooth: Diffusion-based Video Generation with Image Prompts
Yuming Jiang, Tianxing Wu, Shuai Yang, Chenyang Si, Dahua Lin, Yu Qiao, Chen Change Loy, Ziwei Liu

From MMLab@NTU affliated with S-Lab, Nanyang Technological University and Shanghai AI Laboratory.

Overview

Our VideoBooth generates videos with the subjects specified in the image prompts. overall_structure

TODO

  • Release the training code.
  • Release the training dataset.

Installation

  1. Clone the repository.
git clone https://github.com/Vchitect/VideoBooth.git
cd VideoBooth
  1. Install the environment.
conda env create -f environment.yml
conda activate videobooth
  1. Download pretrained models (Stable Diffusion v1.4, VideoBooth), and put them under the folder ./pretrained_models/.

Inference

Here, we provide one example to perform the inference.

python sample_scripts/sample.py --config sample_scripts/configs/panda.yaml

If you want to use your own image, you need to segment the object first. We use Grounded-SAM to segment the subject from images.

Citation

If you find our repo useful for your research, please consider citing our paper:

@article{jiang2023videobooth,
    author = {Jiang, Yuming and Wu, Tianxing and Yang, Shuai and Si, Chenyang and Lin, Dahua and Qiao, Yu and Loy, Chen Change and Liu, Ziwei},
    title = {VideoBooth: Diffusion-based Video Generation with Image Prompts},
    year = {2023}
}

About

[CVPR2024] VideoBooth: Diffusion-based Video Generation with Image Prompts

Resources

Stars

Watchers

Forks

Releases

No releases published

Packages

No packages published

Languages