FYYDCC commited on
Commit
0331d97
·
verified ·
1 Parent(s): 1e53b01

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +7 -4
README.md CHANGED
@@ -10,7 +10,7 @@ Interleaved Vision-Text Latent Reasoning (IVT-LR) is the first VLM framework tha
10
 
11
  ## Usage
12
 
13
- This repository provides pretrained Qwen2-VL models for IVT-LR.
14
 
15
  To see detailed usage, including inference code and scripts for training, please refer to the [GitHub repository](https://github.com/ModalityDance/IVT-LR).
16
 
@@ -20,10 +20,13 @@ To see detailed usage, including inference code and scripts for training, please
20
 
21
  You can download the models directly from Hugging Face using `huggingface_hub`:
22
 
 
23
  from huggingface_hub import hf_hub_download
24
 
25
- # Qwen2-VL on M3CoT
26
  qwen_m3cot_path = hf_hub_download("ModalityDance/IVTLR_QWEN_M3COT", "model.pth")
27
 
28
- # Qwen2-VL on ScienceQA
29
- qwen_sqa_path = hf_hub_download("ModalityDance/IVTLR_QWEN_SQA", "model.pth")
 
 
 
10
 
11
  ## Usage
12
 
13
+ This repository provides pretrained Qwen2-VL models for IVT-LR on **M3CoT** and **ScienceQA** datasets.
14
 
15
  To see detailed usage, including inference code and scripts for training, please refer to the [GitHub repository](https://github.com/ModalityDance/IVT-LR).
16
 
 
20
 
21
  You can download the models directly from Hugging Face using `huggingface_hub`:
22
 
23
+ ```python
24
  from huggingface_hub import hf_hub_download
25
 
26
+ # Download Qwen2-VL model trained on M3CoT
27
  qwen_m3cot_path = hf_hub_download("ModalityDance/IVTLR_QWEN_M3COT", "model.pth")
28
 
29
+ # Download Qwen2-VL model trained on ScienceQA
30
+ qwen_sqa_path = hf_hub_download("ModalityDance/IVTLR_QWEN_SQA", "model.pth")
31
+ ```
32
+