nielsr HF Staff commited on
Commit
f0f0eb0
·
verified ·
1 Parent(s): 933c6c3

Add model metadata and links to paper/code

Browse files

Hi! I'm Niels from the Hugging Face community team. I've noticed this model is missing some important metadata and documentation. This PR adds the `library_name` and `pipeline_tag` to the YAML metadata, which helps with model discoverability and enables automated code snippets. I've also added links to the ICLR 2026 paper and the associated GitHub repository in the README, along with the citation information.

Files changed (1) hide show
  1. README.md +35 -3
README.md CHANGED
@@ -1,3 +1,35 @@
1
- ---
2
- license: apache-2.0
3
- ---
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ ---
2
+ license: apache-2.0
3
+ library_name: transformers
4
+ pipeline_tag: image-text-to-text
5
+ ---
6
+
7
+ # TPRU-7B
8
+
9
+ This repository contains the model weights for **TPRU-7B**, presented in the paper [TPRU: Advancing Temporal and Procedural Understanding in Large Multimodal Models](https://huggingface.co/papers/2602.18884).
10
+
11
+ TPRU is a large-scale dataset and training paradigm designed to bridge the gap in Multimodal Large Language Models (MLLMs) regarding temporal and procedural visual data understanding. TPRU-7B leverages reinforcement learning (GRPO) to excel in tasks such as:
12
+ - **Temporal Reordering:** Reconstructing the correct sequence of shuffled frames.
13
+ - **Next-Frame Prediction:** Predicting the immediate future state given a sequence.
14
+ - **Previous-Frame Review:** Deducing the prerequisite state given an outcome.
15
+
16
+ ## Resources
17
+ - **Paper:** [TPRU: Advancing Temporal and Procedural Understanding in Large Multimodal Models](https://huggingface.co/papers/2602.18884)
18
+ - **Code:** [GitHub Repository](https://github.com/Stephen-gzk/TPRU)
19
+ - **Dataset:** [TPRU-25k](https://huggingface.co/datasets/Stephengzk/TPRU-25k)
20
+
21
+ ## Citation
22
+
23
+ If you find this work useful, please consider citing:
24
+
25
+ ```bibtex
26
+ @inproceedings{gao2026tpru,
27
+ title={TPRU: Advancing Temporal and Procedural Understanding in Large Multimodal Models},
28
+ author={Gao, Zhenkun and Wang, Xuhong and Tan, Xin and Xie, Yuan},
29
+ booktitle={Published as a conference paper at ICLR 2026},
30
+ year={2026}
31
+ }
32
+ ```
33
+
34
+ ## Acknowledgements
35
+ We thank the developers of [Qwen2.5-VL](https://github.com/QwenLM/Qwen2.5-VL), [Easy-R1](https://github.com/hiyouga/EasyR1), and [VLMEvalKit](https://github.com/open-compass/VLMEvalKit) for their open-source contributions.