AbductiveMLLM / README.md
PtRain's picture
Upload folder using huggingface_hub
d35a579 verified
---
base_model: Qwen2-VL-7B-Instruct
library_name: peft
---
# Model Card
## Model Details
- **Finetuned from model:** Qwen2-VL-7B-Instruct
- **Finetuned dataset:** VAR, YoucookII
- **Paper:** ABDUCTIVEMLLM: Boosting Visual Abductive Reasoning Within MLLMs
## How to Get Started with the Model
Use the code below to get started with the model.
```python
model = Qwen2VLForConditionalGeneration.from_pretrained(
'path/to/your/Qwen2-VL-7B-Instruct',
torch_dtype='auto',
attn_implementation="flash_attention_2"
)
# merge lora adapter from Training Stage-1
lora_model_dir = 'qwen2vl_7b_var_lora1'
model = PeftModel.from_pretrained(model, model_id=lora_model_dir)
model = model.merge_and_unload()
# merge lora adapter from Training Stage-2
lora2_dir = 'qwen2vl_7b_var_select3_lora2'
model = PeftModel.from_pretrained(model, model_id=lora2_dir)
model = model.merge_and_unload()
model.generate()
```
## Citation
```
@article{chang2026abductivemllm,
title={AbductiveMLLM: Boosting Visual Abductive Reasoning Within MLLMs},
author={Chang, Boyu and Wang, Qi and Guo, Xi and Nan, Zhixiong and Yao, Yazhou and Zhou, Tianfei},
journal={arXiv preprint arXiv:2601.02771},
year={2026}
}
```