VideoGenHub is a one-stop library to standardize the inference and evaluation of all the conditional video generation models.
To install from pypi:
pip install videogen-hub
To install from github:
git clone https://github.com/TIGER-AI-Lab/VideoGenHub.git
cd VideoGenHub
cd env_cfg
pip install -r requirements.txt
cd ..
pip install -e .
The requirement of opensora is in env_cfg/opensora.txt
For some models like show one, you need to login through huggingface-cli
.
huggingface-cli login
To reproduce our experiment using benchmark.
For text-to-video generation:
./t2v_inference.sh --<model_name> --<device>
import videogen_hub
model = videogen_hub.load('VideoCrafter2')
video = model.infer_one_video(prompt="A child excitedly swings on a rusty swing set, laughter filling the air.")
# Here video is a torch tensor of shape torch.Size([16, 3, 320, 512])
See Google Colab here: https://colab.research.google.com/drive/145UMsBOe5JLqZ2m0LKqvvqsyRJA1IeaE?usp=sharing
By streamlining research and collaboration, VideoGenHub plays a pivotal role in propelling the field of Video Generation.
We included more than 10 Models in video generation.
Method | Venue | Type |
---|---|---|
LaVie | - | Text-To-Video Generation |
VideoCrafter2 | - | Text-To-Video Generation |
ModelScope | - | Text-To-Video Generation |
StreamingT2V | - | Text-To-Video Generation |
Show 1 | - | Text-To-Video Generation |
OpenSora | - | Text-To-Video Generation |
OpenSora-Plan | - | Text-To-Video Generation |
T2V-Turbo | - | Text-To-Video Generation |
DynamiCrafter2 | - | Image-To-Video Generation |
SEINE | ICLR'24 | Image-To-Video Generation |
Consisti2v | - | Image-To_Video Generation |
I2VGenXL | - | Image-To_Video Generation |
This project is released under the License.
This work is a part of GenAI-Arena work.
Please kindly cite our paper if you use our code, data, models or results:
@misc{jiang2024genai,
title={GenAI Arena: An Open Evaluation Platform for Generative Models},
author={Dongfu Jiang and Max Ku and Tianle Li and Yuansheng Ni and Shizhuo Sun and Rongqi Fan and Wenhu Chen},
year={2024},
eprint={2406.04485},
archivePrefix={arXiv},
primaryClass={cs.AI}
}