VidGen-1M: A Large-Scale Dataset for Text-to-video Generation
Abstract
The quality of video-text pairs fundamentally determines the upper bound of text-to-video models. Currently, the datasets used for training these models suffer from significant shortcomings, including low temporal consistency, poor-quality captions, substandard video quality, and imbalanced data distribution. The prevailing video curation process, which depends on image models for tagging and manual rule-based curation, leads to a high computational load and leaves behind unclean data. As a result, there is a lack of appropriate training datasets for text-to-video models. To address this problem, we present VidGen-1M, a superior training dataset for text-to-video models. Produced through a coarse-to-fine curation strategy, this dataset guarantees high-quality videos and detailed captions with excellent temporal consistency. When used to train the video generation model, this dataset has led to experimental results that surpass those obtained with other models.
Community
The video shows a highway winding through a lush green landscape. The road is surrounded by dense trees and vegetation both sides. The sky is overcast, and the mountains in the distance are partially obscured by clouds. The highway appears to be in good condition, with clear lane markings. There are several vehicles traveling on the road, including cars and trucks. The colors in the video are predominantly green from the trees and grey from the road and sky.
This is an automated message from the Librarian Bot. I found the following papers similar to this paper.
The following papers were recommended by the Semantic Scholar API
- OpenVid-1M: A Large-Scale High-Quality Dataset for Text-to-video Generation (2024)
- MiraData: A Large-Scale Video Dataset with Long Durations and Structured Captions (2024)
- MMTrail: A Multimodal Trailer Video Dataset with Language and Music Descriptions (2024)
- SynthVLM: High-Efficiency and High-Quality Synthetic Data for Vision Language Models (2024)
- VIMI: Grounding Video Generation through Multi-modal Instruction (2024)
Please give a thumbs up to this comment if you found it helpful!
If you want recommendations for any Paper on Hugging Face checkout this Space
You can directly ask Librarian Bot for paper recommendations by tagging it in a comment:
@librarian-bot
recommend
Models citing this paper 1
Datasets citing this paper 1
Spaces citing this paper 0
No Space linking this paper