目录

EffiVED:Efficient Video Editing via Text-instruction Diffusion Models

Arxiv Link: https://arxiv.org/abs/2403.11568

Origin Videos & Editing Videos Instrctuion
Input image Turn the rabbit into a fox.
Input image make it Van Gogh style
Input image make it a white fox in the desert trail
Input image make it snowy
Input image add a flock of flowers flying.

News

2024.6.5: Release the inference code

To Do List

Release the training dataset and code

Getting Started

This repository is based on I2VGen-XL.

Create Conda Environment (Optional)

It is recommended to install Anaconda.

Windows Installation: https://docs.anaconda.com/anaconda/install/windows/

Linux Installation: https://docs.anaconda.com/anaconda/install/linux/

conda create -n animation python=3.10
conda activate animation

Python Requirements

pip install -r requirements.txt

Running inference

Please download the pretrained model to checkpoints, then modify the test_model with your download model name. You should add your test videos and edited instruction like provided in data/test_list.txt. Then run the following command:

python inference.py --cfg configs/effived_infer.yaml

Training

Obtaining data from image editing datasets.

You can run the following command to generate the video editing pairs:

python scripts/img2seq_augmenter.py

Here we provide a demo to generate the data from MagicBrush. You can download this dataset following this MagicBrush.

Obtaining data from narrow videos.

You can automatically caption the videos using the Video-BLIP2-Preprocessor Script and set the dataset_types and json_path like this:

  - dataset_types: 
      - video_blip
    train_data:
      json_path: 'blip_generated.json'

Then generate the instruction using the code provided in InstructPix2pix and generate the editing videos using CoDeF.

Bibtex

Please cite this paper if you find the code is useful for your research:

@misc{zhang2024effived,
      title={EffiVED:Efficient Video Editing via Text-instruction Diffusion Models}, 
      author={Zhenghao Zhang and Zuozhuo Dai and Long Qin and Weizhi Wang},
      year={2024},
      eprint={2403.11568},
      archivePrefix={arXiv},
      primaryClass={cs.CV}
}

Shoutouts

关于
29.8 MB
邀请码
    Gitlink(确实开源)
  • 加入我们
  • 官网邮箱:gitlink@ccf.org.cn
  • QQ群
  • QQ群
  • 公众号
  • 公众号

版权所有:中国计算机学会技术支持:开源发展技术委员会
京ICP备13000930号-9 京公网安备 11010802032778号