Improve model card: add metadata, links, and project info

#1
by nielsr HF Staff - opened
Files changed (1) hide show
  1. README.md +21 -0
README.md CHANGED
@@ -1,7 +1,28 @@
1
  ---
2
  license: mit
 
 
 
 
 
 
3
  ---
4
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
5
  ## Citation
6
 
7
  If you find this work useful, please cite our paper:
 
1
  ---
2
  license: mit
3
+ pipeline_tag: image-text-to-text
4
+ library_name: transformers
5
+ base_model: Qwen/Qwen2.5-VL-3B-Instruct
6
+ tags:
7
+ - progress-reasoning
8
+ - vlm
9
  ---
10
 
11
+ # ProgressLM-3B
12
+
13
+ [**Website**](https://progresslm.github.io/ProgressLM/) | [**Paper**](https://huggingface.co/papers/2601.15224) | [**Code**](https://github.com/ProgressLM/ProgressLM)
14
+
15
+ ProgressLM-3B is a Vision-Language Model (VLM) specifically designed for **progress reasoning**. While traditional VLMs are proficient at describing static visual content, ProgressLM is trained to infer how far a task has progressed from partial observations.
16
+
17
+ The model is built upon the [Qwen2.5-VL-3B-Instruct](https://huggingface.co/Qwen/Qwen2.5-VL-3B-Instruct) architecture and was fine-tuned on the **ProgressLM-45K** dataset. It employs a human-inspired two-stage reasoning paradigm: episodic retrieval (to locate the observation along a task trajectory) and mental simulation (to imagine the transition from an anchor to the current observation).
18
+
19
+ ## Model Details
20
+ - **Developed by:** Jianshu Zhang, Chengxuan Qian, Haosen Sun, Haoran Lu, Dingcheng Wang, Letian Xue, Han Liu
21
+ - **Model Type:** Vision-Language Model
22
+ - **Base Model:** Qwen2.5-VL-3B-Instruct
23
+ - **Language(s):** English
24
+ - **License:** MIT
25
+
26
  ## Citation
27
 
28
  If you find this work useful, please cite our paper: