This model is covnert from HelloKKMe/GTA1-7B by AutoAwq.

Model Description

GTA1-7B is an agent-grounding model that achieves state-of-the-art performance across a variety of GUI benchmarks.

Paper: GTA1: GUI Test-time Scaling Agent

Github: https://github.com/Yan98/GTA1

Model Size Open Source ScreenSpot-V2 ScreenSpotPro OSWORLD-G
OpenAI CUA β€” ❌ 87.9 23.4 β€”
Claude 3.7 β€” ❌ 87.6 27.7 β€”
JEDI-7B 7B βœ… 91.7 39.5 54.1
SE-GUI 7B βœ… 90.3 47.0 β€”
UI-TARS 7B βœ… 91.6 35.7 47.5
UI-TARS-1.5* 7B βœ… 89.7* 42.0* 64.2*
UGround-v1-7B 7B βœ… β€” 31.1 36.4
Qwen2.5-VL-32B-Instruct 32B βœ… 91.9* 48.0 59.6*
UGround-v1-72B 72B βœ… β€” 34.5 β€”
Qwen2.5-VL-72B-Instruct 72B βœ… 94.00* 53.3 62.2*
UI-TARS 72B βœ… 90.3 38.1 β€”
GTA1 (Ours) 7B βœ… 92.4 (βˆ† +2.7) 50.1(βˆ† +8.1) 67.7 (βˆ† +3.5)
GTA1 (Ours) 32B βœ… 93.2 (βˆ† +1.3) 53.6 (βˆ† +5.6) 61.9(βˆ† +2.3)
GTA1 (Ours) 72B βœ… 94.8(βˆ† +0.8) 58.4 (βˆ† +5.1) 66.7(βˆ† +4.5)

Note:

  • Model size is indicated in billions (B) of parameters.
  • A dash (β€”) denotes results that are currently unavailable.
  • A superscript asterisk (οΉ‘) denotes our evaluated result.
  • UI-TARS-1.5 7B, Qwen2.5-VL-32B-Instruct, and Qwen2.5-VL-72B-Instruct are applied as our baseline models.
  • βˆ† indicates the performance improvement (βˆ†) of our model compared to its baseline.
Downloads last month
7
Safetensors
Model size
8B params
Tensor type
I32
Β·
BF16
Β·
F16
Β·
Inference Providers NEW
This model isn't deployed by any Inference Provider. πŸ™‹ Ask for provider support

Model tree for flin775/GTA1-7B-AWQ

Base model

HelloKKMe/GTA1-7B
Quantized
(6)
this model