# Load model directly
from transformers import AutoProcessor, AutoModelForImageTextToText
processor = AutoProcessor.from_pretrained("Agents-X/PyVision-Image-7B-RL")
model = AutoModelForImageTextToText.from_pretrained("Agents-X/PyVision-Image-7B-RL")
messages = [
{
"role": "user",
"content": [
{"type": "image", "url": "https://huggingface.co/datasets/huggingface/documentation-images/resolve/main/p-blog/candy.JPG"},
{"type": "text", "text": "What animal is on the candy?"}
]
},
]
inputs = processor.apply_chat_template(
messages,
add_generation_prompt=True,
tokenize=True,
return_dict=True,
return_tensors="pt",
).to(model.device)
outputs = model.generate(**inputs, max_new_tokens=40)
print(processor.decode(outputs[0][inputs["input_ids"].shape[-1]:]))PyVision-Image-7B-RL
PyVision-RL: Forging Open Agentic Vision Models via RL
This is PyVision-Image-7B-RL, a multimodal agentic vision model post-trained from Qwen2.5-VL-7B-Instruct using the PyVision-RL reinforcement learning framework.
- Project Page: https://agent-x.space/pyvision-rl/
- Repository: https://github.com/agents-x-project/PyVision-RL
- Paper: https://arxiv.org/abs/2602.20739
Description
Reinforcement learning for agentic multimodal models often suffers from "interaction collapse," where models learn to reduce tool usage and multi-turn reasoning. PyVision-RL is a framework designed to stabilize training and sustain interaction using an oversampling-filtering-ranking rollout strategy combined with an accumulative tool reward.
PyVision-Image-7B-RL is specifically optimized for image understanding tasks and sustained multi-turn tool interaction, demonstrating strong performance and efficiency for scalable multimodal agents.
Citation
If you find this work useful, please cite the following paper:
@article{pyvisionrl2026,
title={PyVision-RL: Forging Open Agentic Vision Models via RL},
author={Zhao, Shitian and Lin, Shaoheng and Li, Ming and Zhang, Haoquan and Peng, Wenshuo and Zhang, Kaipeng and Wei, Chen},
journal={arXiv:2602.20739},
year={2026}
}
- Downloads last month
- 61
Model tree for Agents-X/PyVision-Image-7B-RL
Base model
Qwen/Qwen2.5-VL-7B-Instruct
# Use a pipeline as a high-level helper from transformers import pipeline pipe = pipeline("image-text-to-text", model="Agents-X/PyVision-Image-7B-RL") messages = [ { "role": "user", "content": [ {"type": "image", "url": "https://huggingface.co/datasets/huggingface/documentation-images/resolve/main/p-blog/candy.JPG"}, {"type": "text", "text": "What animal is on the candy?"} ] }, ] pipe(text=messages)