Update README.md
Browse files
README.md
CHANGED
|
@@ -16,6 +16,9 @@ tags:
|
|
| 16 |
|
| 17 |
🤗<a href="https://huggingface.co/collections/long-xing1/caprl-68d64ac32ded31596c36e189">CapRL Collection</a> | 🤗<a href="https://huggingface.co/papers/2509.22647">Daily Paper</a>
|
| 18 |
|
|
|
|
|
|
|
|
|
|
| 19 |
|
| 20 |
|
| 21 |
## Introduction
|
|
@@ -34,10 +37,10 @@ By employing CapRL training framework, initializing with the Qwen2.5-VL-3B model
|
|
| 34 |
filtered 75K QA dataset as the training set, we obtained a highly capable captioner, CapRL-3B.
|
| 35 |
|
| 36 |
<p align="center">
|
| 37 |
-
<img src="./assets/teaser.png"
|
| 38 |
</p>
|
| 39 |
<p align="center">
|
| 40 |
-
<img src="./assets/performance_update.png"
|
| 41 |
</p>
|
| 42 |
|
| 43 |
## Key Features
|
|
|
|
| 16 |
|
| 17 |
🤗<a href="https://huggingface.co/collections/long-xing1/caprl-68d64ac32ded31596c36e189">CapRL Collection</a> | 🤗<a href="https://huggingface.co/papers/2509.22647">Daily Paper</a>
|
| 18 |
|
| 19 |
+
Based on the same recipe as CapRL-3B, we used InternVL3.5-8B as the policy model and obtained **CapRL-InternVL3.5-8B** through CapRL. **Its performance significantly surpasses that of Qwen2.5-VL-72B**.
|
| 20 |
+
|
| 21 |
+
We are working on even stronger base models and upgrading our training recipe — stay tuned!
|
| 22 |
|
| 23 |
|
| 24 |
## Introduction
|
|
|
|
| 37 |
filtered 75K QA dataset as the training set, we obtained a highly capable captioner, CapRL-3B.
|
| 38 |
|
| 39 |
<p align="center">
|
| 40 |
+
<img src="./assets/teaser.png" width="750"/>
|
| 41 |
</p>
|
| 42 |
<p align="center">
|
| 43 |
+
<img src="./assets/performance_update.png" width="750"/>
|
| 44 |
</p>
|
| 45 |
|
| 46 |
## Key Features
|