Spaces:
Running
Running
Update constants.py
Browse files- constants.py +4 -3
constants.py
CHANGED
|
@@ -191,15 +191,16 @@ LEADERBORAD_INTRODUCTION = """# V-STaR Leaderboard
|
|
| 191 |
- **Human Alignment:** We conducted extensive experiments and human annotations to validate robustness of V-STaR.
|
| 192 |
- **Valuable Insights:** V-STaR reveals a fundamental weakness in existing Video-LLMs regarding causal spatio-temporal reasoning.
|
| 193 |
|
| 194 |
-
**Join Leaderboard**: Please
|
| 195 |
|
| 196 |
**Credits**: This leaderboard is updated and maintained by the team of [V-STaR Contributors]().
|
| 197 |
"""
|
| 198 |
|
| 199 |
-
SUBMIT_INTRODUCTION = """# Submit on
|
| 200 |
## 🎈
|
| 201 |
⚠️ Please note that you need to obtain the file `results/*.json` by running V-STaR in Github. You may conduct an [Offline Eval]() before submitting.
|
| 202 |
-
|
|
|
|
| 203 |
"""
|
| 204 |
|
| 205 |
TABLE_INTRODUCTION = """
|
|
|
|
| 191 |
- **Human Alignment:** We conducted extensive experiments and human annotations to validate robustness of V-STaR.
|
| 192 |
- **Valuable Insights:** V-STaR reveals a fundamental weakness in existing Video-LLMs regarding causal spatio-temporal reasoning.
|
| 193 |
|
| 194 |
+
**Join Leaderboard**: Please contact us to update your results.
|
| 195 |
|
| 196 |
**Credits**: This leaderboard is updated and maintained by the team of [V-STaR Contributors]().
|
| 197 |
"""
|
| 198 |
|
| 199 |
+
SUBMIT_INTRODUCTION = """# Submit on V-STaR Benchmark Introduction
|
| 200 |
## 🎈
|
| 201 |
⚠️ Please note that you need to obtain the file `results/*.json` by running V-STaR in Github. You may conduct an [Offline Eval]() before submitting.
|
| 202 |
+
|
| 203 |
+
⚠️ Then, please contact us to update your results via [email1](zixu.cheng@qmul.ac.uk) or [email2](hu.jian@qmul.ac.uk).
|
| 204 |
"""
|
| 205 |
|
| 206 |
TABLE_INTRODUCTION = """
|