Skip to content
/ TVG Public

code for "TVG: A Training-free Transition Video Generation Method with Diffusion Models"

License

Notifications You must be signed in to change notification settings

SobeyMIL/TVG

Repository files navigation

TVG: A Training-free Transition Video Generation Method with Diffusion Models

Introduction

TVG is a training-free transition video generation method. It can enhance the video generation performance of models under the condition of generating videos given the initial and final frames.

Framework

Transition videos play a crucial role in media production, enhancing the flow and coherence of visual narratives. Traditional methods like morphing often lack artistic appeal and require specialized skills, limiting their effectiveness. Recent advances in diffusion model-based video generation offer new possibilities for creating transitions but face challenges such as poor inter-frame relationship modeling and abrupt content changes. We propose a novel training-free Transition Video Generation (TVG) approach using video-level diffusion models that addresses these limitations without additional training. Our method leverages Gaussian Process Regression (GPR) to model latent representations, ensuring smooth and dynamic transitions between frames. Additionally, we introduce interpolation-based conditional controls and a Frequency-aware Bidirectional Fusion (FBiF) architecture to enhance temporal control and transition reliability. Evaluations of benchmark datasets and custom image pairs demonstrate the effectiveness of our approach in generating high-quality smooth transition videos.

⚙️ Setup

Install Environment via Anaconda (Recommended)

conda create -n TVG python=3.8.5
conda activate TVG
pip install -r requirements.txt

Preparation

Please follow the instructions on DynamiCrafter to download the DynamiCrafter512_interp model and place it in the checkpoints/dynamicrafter_512_interp_v1/model.ckpt path.

Dataset

We used the MorphBench and the TC-Bench-I2V. Please download the MorphBench dataset and place it in EvalData/MorphBench/Animation and EvalData/MorphBench/Metamorphosis. For the TC-Bench-I2V dataset, please follow the official tutorial and save the final frames to the EvalData/TC-Bench/youtube_videos_frames directory. You can refer to the corresponding files in the Prompts for specific paths. The images we collected ourselves are already in the EvalData directory.

💫 Inference

To reproduce the experiments from the paper, please run

bash paper_exp.sh

in the command line.

Results

GIF 1 GIF 2 GIF 3
GIF 4 GIF 5 GIF 6
GIF 7 GIF 8 GIF 9
GIF 10 GIF 11 GIF 12
GIF 13 GIF 14 GIF 15

🖊️ Citation

Please kindly cite our paper if you use our code, data, models or results:

@inproceedings{zhang2024tvg,
        title = {TVG: A Training-free Transition Video Generation Method with Diffusion Models},
        author = {Rui Zhang and Chen Yaosen and Yuegen Liu and  Wei Wang and Xuming Wen and  Hongxia Wang},
        year = {2024},
        booktitle = {arxiv}
}

💞 Acknowledgements

Thanks for the work of DynamiCrafter. Our code is based on the implementation of them.

About

code for "TVG: A Training-free Transition Video Generation Method with Diffusion Models"

Topics

Resources

License

Stars

Watchers

Forks

Releases

No releases published

Packages

No packages published