File size: 3,541 Bytes
7d9d01e
 
 
 
 
 
6d4338e
7d9d01e
 
 
 
 
 
 
 
6d4338e
7d9d01e
 
6d4338e
7d9d01e
 
 
6d4338e
 
 
 
 
7d9d01e
 
 
6d4338e
 
 
7d9d01e
 
 
 
 
 
 
 
 
 
 
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
38
39
40
41
42
43
44
45
# this is .py for store constants 
MODEL_INFO = ['Models', 'Ver.','Abilities']
TASK_INFO = [ 'Resolution', 'FPS', 'Open Source', 'Length', 'Speed', 'Motion', 'Camera', 'Final Sum Score', 'Motion Quality', 'Text-Video Alignment', 'Visual Quality', 'Temporal Consistency']
TASK_INFO_v2 = ['Final Sum Score', 'Motion Quality', 'Text-Video Alignment', 'Visual Quality', 'Temporal Consistency', 'Resolution', 'FPS', 'Open Source', 'Length', 'Speed', 'Motion', 'Camera']

AVG_INFO = ['Final Sum Score', 'Motion Quality', 'Text-Video Alignment', 'Visual Quality', 'Temporal Consistency']
DATA_TITILE_TYPE = ["markdown", "markdown", "number", "number", "number", "number", "number", "number", "number", "number", "number", "number", "number", "number", "number"]
CSV_DIR = "./file/result.csv"

COLUMN_NAMES = MODEL_INFO + TASK_INFO_v2

LEADERBORAD_INTRODUCTION = """# EvalCrafter Leaderboard 🏆

    Welcome to the cutting-edge leaderboard for text-to-video generation, where we meticulously evaluate state-of-the-art generative models using our comprehensive framework, ensuring high-quality results that align with user opinions. Join us in this exciting journey towards excellence! 🛫  

    More methods will be evalcrafted soon, stay tunned ❤️ Join our evaluation by sending an email 📧 ([email protected])! You may also read the [Code](https://github.com/EvalCrafter/EvalCrafter), [Paper](https://arxiv.org/abs/2310.11440), and [Project page](https://evalcrafter.github.io/)  for more detailed information 🤗
    """


TABLE_INTRODUCTION = """In the table below, we summarize each dimension performance of all the models.    """

LEADERBORAD_INFO = """
      The vision and language generative models have been overgrown in recent years. For video generation, various
      open-sourced models and public-available services are released for generating high-visual quality videos. However, 
      these methods often use a few academic metrics, \eg, FVD or IS, to evaluate the performance. We argue that it is 
      hard to judge the large conditional generative models from the simple metrics since these models are often trained on
      very large datasets with multi-aspect abilities. Thus, we propose a new framework and pipeline to exhaustively evaluate
      the performance of the generated videos. To achieve this, we first conduct a new prompt list for text-to-video generation 
      by analyzing the real-world prompt list with the help of the large language model. Then, we evaluate the state-of-the-art video 
      generative models on our carefully designed benchmarks, in terms of visual qualities, content qualities, motion qualities, and 
      text-caption alignment with around 17 objective metrics. To obtain the final leaderboard of the models, we also fit a series of
      coefficients to align the objective metrics to the users' opinions. Based on the proposed opinion alignment method, our final 
      score shows a higher correlation than simply averaging the metrics, showing the effectiveness of the proposed evaluation method.
"""



CITATION_BUTTON_LABEL = "Copy the following snippet to cite these results"
CITATION_BUTTON_TEXT = r"""@inproceedings{Liu2023EvalCrafterBA,
  title={EvalCrafter: Benchmarking and Evaluating Large Video Generation Models},
  author={Yaofang Liu and Xiaodong Cun and Xuebo Liu and Xintao Wang and Yong Zhang and Haoxin Chen and Yang Liu and Tieyong Zeng and Raymond Chan and Ying Shan},
  year={2023},
  url={https://api.semanticscholar.org/CorpusID:264172222}
}"""