Commit ·
0fe41f7
1
Parent(s): 58730a8
update README.md
Browse files
README.md
CHANGED
|
@@ -17,11 +17,12 @@ library_name: transformers
|
|
| 17 |
UI-TARS is a next-generation native GUI agent model designed to interact seamlessly with graphical user interfaces (GUIs) using human-like perception, reasoning, and action capabilities. Unlike traditional modular frameworks, UI-TARS integrates all key components—perception, reasoning, grounding, and memory—within a single vision-language model (VLM), enabling end-to-end task automation without predefined workflows or manual rules.
|
| 18 |
<!--  -->
|
| 19 |
<p align="center">
|
| 20 |
-
<img src="https://github.com/bytedance/UI-TARS/blob/main/figures/UI-TARS.png?raw=true" width="
|
| 21 |
<p>
|
| 22 |
<p align="center">
|
| 23 |
-
<img src="https://github.com/bytedance/UI-TARS/blob/main/figures/UI-TARS
|
| 24 |
<p>
|
|
|
|
| 25 |
<!--  -->
|
| 26 |
|
| 27 |
## Core Features
|
|
|
|
| 17 |
UI-TARS is a next-generation native GUI agent model designed to interact seamlessly with graphical user interfaces (GUIs) using human-like perception, reasoning, and action capabilities. Unlike traditional modular frameworks, UI-TARS integrates all key components—perception, reasoning, grounding, and memory—within a single vision-language model (VLM), enabling end-to-end task automation without predefined workflows or manual rules.
|
| 18 |
<!--  -->
|
| 19 |
<p align="center">
|
| 20 |
+
<img src="https://github.com/bytedance/UI-TARS/blob/main/figures/UI-TARS-vs-Previous-SOTA.png?raw=true" width="90%"/>
|
| 21 |
<p>
|
| 22 |
<p align="center">
|
| 23 |
+
<img src="https://github.com/bytedance/UI-TARS/blob/main/figures/UI-TARS.png?raw=true" width="90%"/>
|
| 24 |
<p>
|
| 25 |
+
|
| 26 |
<!--  -->
|
| 27 |
|
| 28 |
## Core Features
|