Add pipeline tag, license, links to paper, project page and Github repo
#1
by
nielsr
HF Staff
- opened
README.md
CHANGED
|
@@ -1,34 +1,36 @@
|
|
| 1 |
---
|
| 2 |
-
library_name: transformers
|
| 3 |
base_model: Qwen/Qwen-VL-Chat
|
|
|
|
|
|
|
|
|
|
| 4 |
---
|
| 5 |
|
| 6 |
# Model Card for Model ID
|
| 7 |
|
| 8 |
<!-- Provide a quick summary of what the model is/does. -->
|
| 9 |
|
| 10 |
-
|
| 11 |
|
| 12 |
## Model Details
|
| 13 |
|
| 14 |
### Model Description
|
| 15 |
|
| 16 |
<!-- Provide a longer summary of what this model is. -->
|
| 17 |
-
The UI-VLM trained with the AutoGUI-625k training data.
|
| 18 |
-
|
| 19 |
|
| 20 |
- **Developed by:** [Hongxin Li]
|
| 21 |
- **Model type:** [Vision-language model]
|
| 22 |
- **Language(s) (NLP):** [English]
|
| 23 |
-
- **License:** [
|
| 24 |
- **Finetuned from model [optional]:** [Qwen-VL-Chat]
|
| 25 |
|
| 26 |
### Model Sources [optional]
|
| 27 |
|
| 28 |
<!-- Provide the basic links for the model. -->
|
| 29 |
|
| 30 |
-
- **Repository:** [
|
| 31 |
-
- **Paper [optional]:** [
|
|
|
|
| 32 |
- **Demo [optional]:** [More Information Needed]
|
| 33 |
|
| 34 |
## Uses
|
|
@@ -87,7 +89,6 @@ Use the code below to get started with the model.
|
|
| 87 |
|
| 88 |
[More Information Needed]
|
| 89 |
|
| 90 |
-
|
| 91 |
#### Training Hyperparameters
|
| 92 |
|
| 93 |
- **Training regime:** [More Information Needed] <!--fp32, fp16 mixed precision, bf16 mixed precision, bf16 non-mixed precision, fp16 non-mixed precision, fp8 mixed precision -->
|
|
@@ -128,8 +129,6 @@ Use the code below to get started with the model.
|
|
| 128 |
|
| 129 |
#### Summary
|
| 130 |
|
| 131 |
-
|
| 132 |
-
|
| 133 |
## Model Examination [optional]
|
| 134 |
|
| 135 |
<!-- Relevant interpretability work for the model goes here -->
|
|
|
|
| 1 |
---
|
|
|
|
| 2 |
base_model: Qwen/Qwen-VL-Chat
|
| 3 |
+
library_name: transformers
|
| 4 |
+
pipeline_tag: image-text-to-text
|
| 5 |
+
license: cc-by-4.0
|
| 6 |
---
|
| 7 |
|
| 8 |
# Model Card for Model ID
|
| 9 |
|
| 10 |
<!-- Provide a quick summary of what the model is/does. -->
|
| 11 |
|
| 12 |
+
This model is a UI-VLM trained with the AutoGUI-625k training data.
|
| 13 |
|
| 14 |
## Model Details
|
| 15 |
|
| 16 |
### Model Description
|
| 17 |
|
| 18 |
<!-- Provide a longer summary of what this model is. -->
|
| 19 |
+
The UI-VLM trained with the AutoGUI-625k training data as described in [AutoGUI: Scaling GUI Grounding with Automatic Functionality Annotations from LLMs](https://huggingface.co/papers/2502.01977).
|
|
|
|
| 20 |
|
| 21 |
- **Developed by:** [Hongxin Li]
|
| 22 |
- **Model type:** [Vision-language model]
|
| 23 |
- **Language(s) (NLP):** [English]
|
| 24 |
+
- **License:** [CC-BY-4.0]
|
| 25 |
- **Finetuned from model [optional]:** [Qwen-VL-Chat]
|
| 26 |
|
| 27 |
### Model Sources [optional]
|
| 28 |
|
| 29 |
<!-- Provide the basic links for the model. -->
|
| 30 |
|
| 31 |
+
- **Repository:** [https://github.com/BraveGroup/AutoGUI](https://github.com/BraveGroup/AutoGUI)
|
| 32 |
+
- **Paper [optional]:** [https://huggingface.co/papers/2502.01977](https://huggingface.co/papers/2502.01977)
|
| 33 |
+
- **Project page [optional]:** [https://autogui-project.github.io/](https://autogui-project.github.io/)
|
| 34 |
- **Demo [optional]:** [More Information Needed]
|
| 35 |
|
| 36 |
## Uses
|
|
|
|
| 89 |
|
| 90 |
[More Information Needed]
|
| 91 |
|
|
|
|
| 92 |
#### Training Hyperparameters
|
| 93 |
|
| 94 |
- **Training regime:** [More Information Needed] <!--fp32, fp16 mixed precision, bf16 mixed precision, bf16 non-mixed precision, fp16 non-mixed precision, fp8 mixed precision -->
|
|
|
|
| 129 |
|
| 130 |
#### Summary
|
| 131 |
|
|
|
|
|
|
|
| 132 |
## Model Examination [optional]
|
| 133 |
|
| 134 |
<!-- Relevant interpretability work for the model goes here -->
|