|
|
--- |
|
|
base_model: Qwen2-VL-7B-Instruct |
|
|
library_name: peft |
|
|
--- |
|
|
|
|
|
# Model Card |
|
|
|
|
|
## Model Details |
|
|
- **Finetuned from model:** Qwen2-VL-7B-Instruct |
|
|
- **Finetuned dataset:** VAR, YoucookII |
|
|
- **Paper:** ABDUCTIVEMLLM: Boosting Visual Abductive Reasoning Within MLLMs |
|
|
|
|
|
|
|
|
## How to Get Started with the Model |
|
|
|
|
|
Use the code below to get started with the model. |
|
|
|
|
|
```python |
|
|
model = Qwen2VLForConditionalGeneration.from_pretrained( |
|
|
'path/to/your/Qwen2-VL-7B-Instruct', |
|
|
torch_dtype='auto', |
|
|
attn_implementation="flash_attention_2" |
|
|
) |
|
|
|
|
|
# merge lora adapter from Training Stage-1 |
|
|
lora_model_dir = 'qwen2vl_7b_var_lora1' |
|
|
model = PeftModel.from_pretrained(model, model_id=lora_model_dir) |
|
|
model = model.merge_and_unload() |
|
|
|
|
|
# merge lora adapter from Training Stage-2 |
|
|
lora2_dir = 'qwen2vl_7b_var_select3_lora2' |
|
|
model = PeftModel.from_pretrained(model, model_id=lora2_dir) |
|
|
model = model.merge_and_unload() |
|
|
|
|
|
model.generate() |
|
|
``` |
|
|
|
|
|
|
|
|
## Citation |
|
|
``` |
|
|
@article{chang2026abductivemllm, |
|
|
title={AbductiveMLLM: Boosting Visual Abductive Reasoning Within MLLMs}, |
|
|
author={Chang, Boyu and Wang, Qi and Guo, Xi and Nan, Zhixiong and Yao, Yazhou and Zhou, Tianfei}, |
|
|
journal={arXiv preprint arXiv:2601.02771}, |
|
|
year={2026} |
|
|
} |
|
|
``` |
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|