Update README.md
Browse files
README.md
CHANGED
|
@@ -14,12 +14,14 @@ tags:
|
|
| 14 |
π<a href="https://arxiv.org/abs/2509.22647">Paper</a> | π <a href="https://github.com/InternLM/CapRL">Github</a> |π€<a href="https://huggingface.co/internlm/CapRL-3B">CapRL-3B Model</a> |π€<a href="https://huggingface.co/yuhangzang/CapRL-InternVL3.5-8B">CapRL-InternVL3.5-8B Model</a> |
|
| 15 |
π€<a href="https://huggingface.co/datasets/internlm/CapRL-2M">CapRL-2M Dataset</a>
|
| 16 |
|
| 17 |
-
π€<a href="https://huggingface.co/collections/long-xing1/caprl-68d64ac32ded31596c36e189">CapRL Collection</a> | π€<a href="https://huggingface.co/papers/2509.22647">Daily Paper</a>
|
|
|
|
| 18 |
|
| 19 |
Based on the same recipe as CapRL-3B, we used InternVL3.5-8B as the policy model and obtained **CapRL-InternVL3.5-8B** through CapRL. **Its performance significantly surpasses that of Qwen2.5-VL-72B**.
|
| 20 |
|
| 21 |
We are working on even stronger base models and upgrading our training recipe β stay tuned!
|
| 22 |
|
|
|
|
| 23 |
|
| 24 |
## Introduction
|
| 25 |
We are excited to introduce CapRL-3B, a lightweight 3B image captioner that achieves perception capabilities comparable to Qwen2.5-VL-72B.
|
|
|
|
| 14 |
π<a href="https://arxiv.org/abs/2509.22647">Paper</a> | π <a href="https://github.com/InternLM/CapRL">Github</a> |π€<a href="https://huggingface.co/internlm/CapRL-3B">CapRL-3B Model</a> |π€<a href="https://huggingface.co/yuhangzang/CapRL-InternVL3.5-8B">CapRL-InternVL3.5-8B Model</a> |
|
| 15 |
π€<a href="https://huggingface.co/datasets/internlm/CapRL-2M">CapRL-2M Dataset</a>
|
| 16 |
|
| 17 |
+
π€<a href="https://huggingface.co/collections/long-xing1/caprl-68d64ac32ded31596c36e189">CapRL Collection</a> | π€<a href="https://huggingface.co/papers/2509.22647">Daily Paper</a> ο½π€<a href="https://huggingface.co/mradermacher/CapRL-3B-GGUF">CapRL-3B-GGUF</a> ο½π€<a href="https://huggingface.co/mradermacher/CapRL-3B-i1-GGUF">CapRL-3B-i1-GGUF</a>
|
| 18 |
+
|
| 19 |
|
| 20 |
Based on the same recipe as CapRL-3B, we used InternVL3.5-8B as the policy model and obtained **CapRL-InternVL3.5-8B** through CapRL. **Its performance significantly surpasses that of Qwen2.5-VL-72B**.
|
| 21 |
|
| 22 |
We are working on even stronger base models and upgrading our training recipe β stay tuned!
|
| 23 |
|
| 24 |
+
CapRL-3B-GGUF is static quants version, and CapRL-3B-i1-GGUF is weighted/imatrix quants version. Thanks for their contribution!
|
| 25 |
|
| 26 |
## Introduction
|
| 27 |
We are excited to introduce CapRL-3B, a lightweight 3B image captioner that achieves perception capabilities comparable to Qwen2.5-VL-72B.
|