Datasets:

Modalities:
Text
Video
Formats:
json
ArXiv:
Libraries:
Datasets
pandas
License:
CaReBench / README.md
geekifan's picture
Update README.md
03afdda verified
metadata
license: mit
size_categories:
  - 1K<n<10K
configs:
  - config_name: CaReBench
    data_files:
      - split: test
        path: json/metadata.json

Logo CaReBench: A Fine-grained Benchmark for Video Captioning and Retrieval

Yifan Xu, Xinhao Li, Yichun Yang, Desen Meng, Rui Huang, Limin Wang

๐Ÿค— Model    |    ๐Ÿค— Data   ๏ฝœ    ๐Ÿ“‘ Paper   

๐Ÿ“ Introduction

๐ŸŒŸ CaReBench is a fine-grained benchmark comprising 1,000 high-quality videos with detailed human-annotated captions, including manually separated spatial and temporal descriptions for independent spatiotemporal bias evaluation. CaReBench

๐Ÿ“Š ReBias and CapST Metrics are designed specifically for retrieval and captioning tasks, providing a comprehensive evaluation framework for spatiotemporal understanding in video-language models.

โšก CaRe: A Unified Baseline for fine-grained video retrieval and captioning, achieving competitive performance through two-stage Supervised Fine-Tuning (SFT). CaRe excels in both generating detailed video descriptions and extracting robust video features. CaRe Training Recipe

๐Ÿš€ State-of-the-art performance on both detailed video captioning and fine-grained video retrieval. CaRe outperforms CLIP-based retrieval models and popular MLLMs in captioning tasks. alt text