Datasets:

Modalities:
Text
Formats:
webdataset
Languages:
English
ArXiv:
Libraries:
Datasets
WebDataset
License:
JungleGym commited on
Commit
6a00506
Β·
verified Β·
1 Parent(s): ee1fc5b

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +3 -3
README.md CHANGED
@@ -13,14 +13,14 @@ size_categories:
13
 
14
  # TimeLens-Bench
15
 
16
- πŸ“‘ [**Paper**](TODO) | πŸ’» [**Code**](https://github.com/TencentARC/TimeLens) | 🏠 [**Project Page**](https://timelens-arc-lab.github.io/) | πŸ€— [**Model & Data**](https://huggingface.co/collections/TencentARC/TimeLens)
17
 
18
  ## ✨ Dataset Description
19
 
20
  **TimeLens-Bench** is a comprehensive, high-quality evaluation benchmark for video temporal grounding, proposed in our paper [TimeLens: Rethinking Video Temporal Grounding with Multimodal LLMs](TODO).
21
 
22
- During our annotation process, we identified critical quality issues within existing datasets and performed extensive manual corrections. We observed a **dramatic re-ranking of models** on TimeLens-Bench compared to legacy benchmarks, demonstrating that TimeLens-Bench provides *more reliable* evaluation
23
- (more details in our [paper](TODO) and [project page](https://timelens-arc-lab.github.io/)).
24
  <img src="https://cdn-uploads.huggingface.co/production/uploads/65372e922c6ef949b22c26d9/31s82GO6S5LKlW0-kcIFU.png" alt="performance_comparison_charades-1" width="35%">
25
 
26
  ### Dataset Statistics
 
13
 
14
  # TimeLens-Bench
15
 
16
+ πŸ“‘ [**Paper**](TODO) | πŸ’» [**Code**](https://github.com/TencentARC/TimeLens) | 🏠 [**Project Page**](https://timelens-arc-lab.github.io/) | πŸ€— [**Model & Data**](https://huggingface.co/collections/TencentARC/TimeLens) | πŸ† [**TimeLens-Bench Leaderboard**](https://timelens-arc-lab.github.io/#leaderboard)
17
 
18
  ## ✨ Dataset Description
19
 
20
  **TimeLens-Bench** is a comprehensive, high-quality evaluation benchmark for video temporal grounding, proposed in our paper [TimeLens: Rethinking Video Temporal Grounding with Multimodal LLMs](TODO).
21
 
22
+ During our annotation process, we identified critical quality issues within existing datasets and performed extensive manual corrections. We observed a **dramatic re-ranking of models** on TimeLens-Bench compared to legacy benchmarks, demonstrating that TimeLens-Bench provides **more reliable evaluation** for video temporal grounding.
23
+ (See more details in our [paper](TODO) and [project page](https://timelens-arc-lab.github.io/).)
24
  <img src="https://cdn-uploads.huggingface.co/production/uploads/65372e922c6ef949b22c26d9/31s82GO6S5LKlW0-kcIFU.png" alt="performance_comparison_charades-1" width="35%">
25
 
26
  ### Dataset Statistics