nielsr HF Staff commited on
Commit
ed8a342
·
verified ·
1 Parent(s): cb132fc

Add GitHub link, paper metadata, and improve model card

Browse files

Hi! I'm Niels, part of the community science team at Hugging Face.

This PR improves the model card for **RankVideo** by:
- Adding the `arxiv` metadata tag to link the repository to its [research paper](https://huggingface.co/papers/2602.02444).
- Adding `library_name: transformers` to the metadata based on the `config.json` architecture.
- Adding a link to the official [GitHub repository](https://github.com/tskow99/RANKVIDEO-Reasoning-Reranker).
- Fixing the formatting of the usage example and the BibTeX block.

These changes make the model more discoverable and provide better context for users.

Files changed (1) hide show
  1. README.md +18 -12
README.md CHANGED
@@ -1,46 +1,51 @@
1
  ---
2
- license: mit
3
  language:
4
  - en
 
 
 
 
5
  tags:
6
  - video
7
  - retrieval
8
  - reranking
9
  - qwen3-vl
10
- base_model: Qwen/Qwen3-VL-8B-Instruct
11
- pipeline_tag: video-text-to-text
12
  ---
13
 
14
  # RankVideo
15
 
16
- Video-native reasoning reranker for text-to-video retrieval. Fine-tuned from Qwen3-VL-8B-Instruct.
17
 
18
- ## Reference
19
- - [arXiv:2602.02444](https://arxiv.org/abs/2602.02444)
20
 
21
- ## Training Data
22
-
23
- This model was trained using the [MultiVENT 2.0 dataset](https://huggingface.co/datasets/hltcoe/MultiVENT2.0 ).
24
 
 
25
 
 
26
 
27
  ## Usage
28
 
29
- ```
30
 
 
31
  from rankvideo import VLMReranker
32
 
33
  reranker = VLMReranker(model_path="hltcoe/RankVideo")
34
 
 
35
  scores = reranker.score_batch(
36
  queries=["person playing guitar"],
37
  video_paths=["/path/to/video.mp4"],
38
  )
39
 
40
  print(f"Relevance score: {scores[0]['logit_delta_yes_minus_no']:.3f}")
41
-
42
  ```
 
43
  ## BibTeX
 
44
  ```bibtex
45
  @misc{skow2026rankvideoreasoningrerankingtexttovideo,
46
  title={RANKVIDEO: Reasoning Reranking for Text-to-Video Retrieval},
@@ -50,4 +55,5 @@ print(f"Relevance score: {scores[0]['logit_delta_yes_minus_no']:.3f}")
50
  archivePrefix={arXiv},
51
  primaryClass={cs.IR},
52
  url={https://arxiv.org/abs/2602.02444},
53
- }
 
 
1
  ---
2
+ base_model: Qwen/Qwen3-VL-8B-Instruct
3
  language:
4
  - en
5
+ license: mit
6
+ pipeline_tag: video-text-to-text
7
+ library_name: transformers
8
+ arxiv: 2602.02444
9
  tags:
10
  - video
11
  - retrieval
12
  - reranking
13
  - qwen3-vl
 
 
14
  ---
15
 
16
  # RankVideo
17
 
18
+ RankVideo is a video-native reasoning reranker for text-to-video retrieval, fine-tuned from [Qwen3-VL-8B-Instruct](https://huggingface.co/Qwen/Qwen3-VL-8B-Instruct).
19
 
20
+ The model explicitly reasons over query-video pairs using video content to assess relevance. It was introduced in the paper [RANKVIDEO: Reasoning Reranking for Text-to-Video Retrieval](https://huggingface.co/papers/2602.02444).
 
21
 
22
+ - **Repository:** [https://github.com/tskow99/RANKVIDEO-Reasoning-Reranker](https://github.com/tskow99/RANKVIDEO-Reasoning-Reranker)
23
+ - **Paper:** [RANKVIDEO: Reasoning Reranking for Text-to-Video Retrieval](https://arxiv.org/abs/2602.02444)
 
24
 
25
+ ## Training Data
26
 
27
+ This model was trained using the [MultiVENT 2.0 dataset](https://huggingface.co/datasets/hltcoe/MultiVENT2.0).
28
 
29
  ## Usage
30
 
31
+ You can use the model for scoring query-video pairs via the `rankvideo` library as follows:
32
 
33
+ ```python
34
  from rankvideo import VLMReranker
35
 
36
  reranker = VLMReranker(model_path="hltcoe/RankVideo")
37
 
38
+ # Score query-video pairs for relevance
39
  scores = reranker.score_batch(
40
  queries=["person playing guitar"],
41
  video_paths=["/path/to/video.mp4"],
42
  )
43
 
44
  print(f"Relevance score: {scores[0]['logit_delta_yes_minus_no']:.3f}")
 
45
  ```
46
+
47
  ## BibTeX
48
+
49
  ```bibtex
50
  @misc{skow2026rankvideoreasoningrerankingtexttovideo,
51
  title={RANKVIDEO: Reasoning Reranking for Text-to-Video Retrieval},
 
55
  archivePrefix={arXiv},
56
  primaryClass={cs.IR},
57
  url={https://arxiv.org/abs/2602.02444},
58
+ }
59
+ ```