Improve model card: add metadata, links, and project info
#1
by
nielsr HF Staff - opened
README.md
CHANGED
|
@@ -1,7 +1,28 @@
|
|
| 1 |
---
|
| 2 |
license: mit
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 3 |
---
|
| 4 |
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 5 |
## Citation
|
| 6 |
|
| 7 |
If you find this work useful, please cite our paper:
|
|
|
|
| 1 |
---
|
| 2 |
license: mit
|
| 3 |
+
pipeline_tag: image-text-to-text
|
| 4 |
+
library_name: transformers
|
| 5 |
+
base_model: Qwen/Qwen2.5-VL-3B-Instruct
|
| 6 |
+
tags:
|
| 7 |
+
- progress-reasoning
|
| 8 |
+
- vlm
|
| 9 |
---
|
| 10 |
|
| 11 |
+
# ProgressLM-3B
|
| 12 |
+
|
| 13 |
+
[**Website**](https://progresslm.github.io/ProgressLM/) | [**Paper**](https://huggingface.co/papers/2601.15224) | [**Code**](https://github.com/ProgressLM/ProgressLM)
|
| 14 |
+
|
| 15 |
+
ProgressLM-3B is a Vision-Language Model (VLM) specifically designed for **progress reasoning**. While traditional VLMs are proficient at describing static visual content, ProgressLM is trained to infer how far a task has progressed from partial observations.
|
| 16 |
+
|
| 17 |
+
The model is built upon the [Qwen2.5-VL-3B-Instruct](https://huggingface.co/Qwen/Qwen2.5-VL-3B-Instruct) architecture and was fine-tuned on the **ProgressLM-45K** dataset. It employs a human-inspired two-stage reasoning paradigm: episodic retrieval (to locate the observation along a task trajectory) and mental simulation (to imagine the transition from an anchor to the current observation).
|
| 18 |
+
|
| 19 |
+
## Model Details
|
| 20 |
+
- **Developed by:** Jianshu Zhang, Chengxuan Qian, Haosen Sun, Haoran Lu, Dingcheng Wang, Letian Xue, Han Liu
|
| 21 |
+
- **Model Type:** Vision-Language Model
|
| 22 |
+
- **Base Model:** Qwen2.5-VL-3B-Instruct
|
| 23 |
+
- **Language(s):** English
|
| 24 |
+
- **License:** MIT
|
| 25 |
+
|
| 26 |
## Citation
|
| 27 |
|
| 28 |
If you find this work useful, please cite our paper:
|