Spaces:
Running
Running
update
Browse files- src/about.py +10 -2
src/about.py
CHANGED
|
@@ -21,11 +21,13 @@ NUM_FEWSHOT = 0 # Change with your few shot
|
|
| 21 |
|
| 22 |
|
| 23 |
# Your leaderboard name
|
| 24 |
-
TITLE = """<h1 align="center" id="space-title">
|
| 25 |
|
| 26 |
# What does your leaderboard evaluate?
|
| 27 |
INTRODUCTION_TEXT = """
|
| 28 |
-
|
|
|
|
|
|
|
| 29 |
"""
|
| 30 |
|
| 31 |
# Which evaluations are you running? how can people reproduce what you have?
|
|
@@ -69,4 +71,10 @@ If everything is done, check you can launch the EleutherAIHarness on your model
|
|
| 69 |
|
| 70 |
CITATION_BUTTON_LABEL = "Copy the following snippet to cite these results"
|
| 71 |
CITATION_BUTTON_TEXT = r"""
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 72 |
"""
|
|
|
|
| 21 |
|
| 22 |
|
| 23 |
# Your leaderboard name
|
| 24 |
+
TITLE = """<h1 align="center" id="space-title">TempCompass leaderboard</h1>"""
|
| 25 |
|
| 26 |
# What does your leaderboard evaluate?
|
| 27 |
INTRODUCTION_TEXT = """
|
| 28 |
+
Welcome to the leaderboard of TempCompass! 🏆
|
| 29 |
+
|
| 30 |
+
TempCompass is a benchmark to evaluate the temporal perception ability of Video LLMs. It consists of 410 videos and 7,540 task instructions, covering 11 temporal aspects and 4 task types. Please refer to [our paper](https://arxiv.org/abs/2403.00476) for more details.
|
| 31 |
"""
|
| 32 |
|
| 33 |
# Which evaluations are you running? how can people reproduce what you have?
|
|
|
|
| 71 |
|
| 72 |
CITATION_BUTTON_LABEL = "Copy the following snippet to cite these results"
|
| 73 |
CITATION_BUTTON_TEXT = r"""
|
| 74 |
+
@article{liu2024tempcompass,
|
| 75 |
+
title = {TempCompass: Do Video LLMs Really Understand Videos?},
|
| 76 |
+
author = {Yuanxin Liu and Shicheng Li and Yi Liu and Yuxiang Wang and Shuhuai Ren and Lei Li and Sishuo Chen and Xu Sun and Lu Hou},
|
| 77 |
+
year = {2024},
|
| 78 |
+
journal = {arXiv preprint arXiv: 2403.00476}
|
| 79 |
+
}
|
| 80 |
"""
|