Add pipeline tag, license, links to paper, project page and Github repo

#1
by nielsr HF Staff - opened
Files changed (1) hide show
  1. README.md +9 -10
README.md CHANGED
@@ -1,34 +1,36 @@
1
  ---
2
- library_name: transformers
3
  base_model: Qwen/Qwen-VL-Chat
 
 
 
4
  ---
5
 
6
  # Model Card for Model ID
7
 
8
  <!-- Provide a quick summary of what the model is/does. -->
9
 
10
-
11
 
12
  ## Model Details
13
 
14
  ### Model Description
15
 
16
  <!-- Provide a longer summary of what this model is. -->
17
- The UI-VLM trained with the AutoGUI-625k training data.
18
-
19
 
20
  - **Developed by:** [Hongxin Li]
21
  - **Model type:** [Vision-language model]
22
  - **Language(s) (NLP):** [English]
23
- - **License:** [More Information Needed]
24
  - **Finetuned from model [optional]:** [Qwen-VL-Chat]
25
 
26
  ### Model Sources [optional]
27
 
28
  <!-- Provide the basic links for the model. -->
29
 
30
- - **Repository:** [More Information Needed]
31
- - **Paper [optional]:** [More Information Needed]
 
32
  - **Demo [optional]:** [More Information Needed]
33
 
34
  ## Uses
@@ -87,7 +89,6 @@ Use the code below to get started with the model.
87
 
88
  [More Information Needed]
89
 
90
-
91
  #### Training Hyperparameters
92
 
93
  - **Training regime:** [More Information Needed] <!--fp32, fp16 mixed precision, bf16 mixed precision, bf16 non-mixed precision, fp16 non-mixed precision, fp8 mixed precision -->
@@ -128,8 +129,6 @@ Use the code below to get started with the model.
128
 
129
  #### Summary
130
 
131
-
132
-
133
  ## Model Examination [optional]
134
 
135
  <!-- Relevant interpretability work for the model goes here -->
 
1
  ---
 
2
  base_model: Qwen/Qwen-VL-Chat
3
+ library_name: transformers
4
+ pipeline_tag: image-text-to-text
5
+ license: cc-by-4.0
6
  ---
7
 
8
  # Model Card for Model ID
9
 
10
  <!-- Provide a quick summary of what the model is/does. -->
11
 
12
+ This model is a UI-VLM trained with the AutoGUI-625k training data.
13
 
14
  ## Model Details
15
 
16
  ### Model Description
17
 
18
  <!-- Provide a longer summary of what this model is. -->
19
+ The UI-VLM trained with the AutoGUI-625k training data as described in [AutoGUI: Scaling GUI Grounding with Automatic Functionality Annotations from LLMs](https://huggingface.co/papers/2502.01977).
 
20
 
21
  - **Developed by:** [Hongxin Li]
22
  - **Model type:** [Vision-language model]
23
  - **Language(s) (NLP):** [English]
24
+ - **License:** [CC-BY-4.0]
25
  - **Finetuned from model [optional]:** [Qwen-VL-Chat]
26
 
27
  ### Model Sources [optional]
28
 
29
  <!-- Provide the basic links for the model. -->
30
 
31
+ - **Repository:** [https://github.com/BraveGroup/AutoGUI](https://github.com/BraveGroup/AutoGUI)
32
+ - **Paper [optional]:** [https://huggingface.co/papers/2502.01977](https://huggingface.co/papers/2502.01977)
33
+ - **Project page [optional]:** [https://autogui-project.github.io/](https://autogui-project.github.io/)
34
  - **Demo [optional]:** [More Information Needed]
35
 
36
  ## Uses
 
89
 
90
  [More Information Needed]
91
 
 
92
  #### Training Hyperparameters
93
 
94
  - **Training regime:** [More Information Needed] <!--fp32, fp16 mixed precision, bf16 mixed precision, bf16 non-mixed precision, fp16 non-mixed precision, fp8 mixed precision -->
 
129
 
130
  #### Summary
131
 
 
 
132
  ## Model Examination [optional]
133
 
134
  <!-- Relevant interpretability work for the model goes here -->