yuhangzang commited on
Commit
30c515b
·
verified ·
1 Parent(s): 72e3579

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +5 -2
README.md CHANGED
@@ -16,6 +16,9 @@ tags:
16
 
17
  🤗<a href="https://huggingface.co/collections/long-xing1/caprl-68d64ac32ded31596c36e189">CapRL Collection</a> | 🤗<a href="https://huggingface.co/papers/2509.22647">Daily Paper</a>
18
 
 
 
 
19
 
20
 
21
  ## Introduction
@@ -34,10 +37,10 @@ By employing CapRL training framework, initializing with the Qwen2.5-VL-3B model
34
  filtered 75K QA dataset as the training set, we obtained a highly capable captioner, CapRL-3B.
35
 
36
  <p align="center">
37
- <img src="./assets/teaser.png" alt="Main Results on GPT2" width="750"/>
38
  </p>
39
  <p align="center">
40
- <img src="./assets/performance_update.png" alt="Main Results on GPT2" width="750"/>
41
  </p>
42
 
43
  ## Key Features
 
16
 
17
  🤗<a href="https://huggingface.co/collections/long-xing1/caprl-68d64ac32ded31596c36e189">CapRL Collection</a> | 🤗<a href="https://huggingface.co/papers/2509.22647">Daily Paper</a>
18
 
19
+ Based on the same recipe as CapRL-3B, we used InternVL3.5-8B as the policy model and obtained **CapRL-InternVL3.5-8B** through CapRL. **Its performance significantly surpasses that of Qwen2.5-VL-72B**.
20
+
21
+ We are working on even stronger base models and upgrading our training recipe — stay tuned!
22
 
23
 
24
  ## Introduction
 
37
  filtered 75K QA dataset as the training set, we obtained a highly capable captioner, CapRL-3B.
38
 
39
  <p align="center">
40
+ <img src="./assets/teaser.png" width="750"/>
41
  </p>
42
  <p align="center">
43
+ <img src="./assets/performance_update.png" width="750"/>
44
  </p>
45
 
46
  ## Key Features