lbourdois commited on
Commit
b059932
·
verified ·
1 Parent(s): ed55cb4

Improve language tag

Browse files

Hi! As the model is multilingual, this is a PR to add other languages than English to the language tag to improve the referencing. Note that 29 languages are announced in the README, but only 13 are explicitly listed. I was therefore only able to add these 13 languages.

Files changed (1) hide show
  1. README.md +174 -160
README.md CHANGED
@@ -1,160 +1,174 @@
1
- ---
2
- license: apache-2.0
3
- datasets:
4
- - agentlans/crash-course
5
- - vicgalle/configurable-system-prompt-multitask
6
- base_model:
7
- - Qwen/Qwen2.5-0.5B-Instruct
8
- model-index:
9
- - name: Qwen2.5-0.5B-Instruct-CrashCourse-dropout
10
- results:
11
- - task:
12
- type: text-generation
13
- name: Text Generation
14
- dataset:
15
- name: IFEval (0-Shot)
16
- type: wis-k/instruction-following-eval
17
- split: train
18
- args:
19
- num_few_shot: 0
20
- metrics:
21
- - type: inst_level_strict_acc and prompt_level_strict_acc
22
- value: 29.49
23
- name: averaged accuracy
24
- source:
25
- url: https://huggingface.co/spaces/open-llm-leaderboard/open_llm_leaderboard#/?search=agentlans%2FQwen2.5-0.5B-Instruct-CrashCourse-dropout
26
- name: Open LLM Leaderboard
27
- - task:
28
- type: text-generation
29
- name: Text Generation
30
- dataset:
31
- name: BBH (3-Shot)
32
- type: SaylorTwift/bbh
33
- split: test
34
- args:
35
- num_few_shot: 3
36
- metrics:
37
- - type: acc_norm
38
- value: 7.23
39
- name: normalized accuracy
40
- source:
41
- url: https://huggingface.co/spaces/open-llm-leaderboard/open_llm_leaderboard#/?search=agentlans%2FQwen2.5-0.5B-Instruct-CrashCourse-dropout
42
- name: Open LLM Leaderboard
43
- - task:
44
- type: text-generation
45
- name: Text Generation
46
- dataset:
47
- name: MATH Lvl 5 (4-Shot)
48
- type: lighteval/MATH-Hard
49
- split: test
50
- args:
51
- num_few_shot: 4
52
- metrics:
53
- - type: exact_match
54
- value: 0.08
55
- name: exact match
56
- source:
57
- url: https://huggingface.co/spaces/open-llm-leaderboard/open_llm_leaderboard#/?search=agentlans%2FQwen2.5-0.5B-Instruct-CrashCourse-dropout
58
- name: Open LLM Leaderboard
59
- - task:
60
- type: text-generation
61
- name: Text Generation
62
- dataset:
63
- name: GPQA (0-shot)
64
- type: Idavidrein/gpqa
65
- split: train
66
- args:
67
- num_few_shot: 0
68
- metrics:
69
- - type: acc_norm
70
- value: 1.79
71
- name: acc_norm
72
- source:
73
- url: https://huggingface.co/spaces/open-llm-leaderboard/open_llm_leaderboard#/?search=agentlans%2FQwen2.5-0.5B-Instruct-CrashCourse-dropout
74
- name: Open LLM Leaderboard
75
- - task:
76
- type: text-generation
77
- name: Text Generation
78
- dataset:
79
- name: MuSR (0-shot)
80
- type: TAUR-Lab/MuSR
81
- args:
82
- num_few_shot: 0
83
- metrics:
84
- - type: acc_norm
85
- value: 1.11
86
- name: acc_norm
87
- source:
88
- url: https://huggingface.co/spaces/open-llm-leaderboard/open_llm_leaderboard#/?search=agentlans%2FQwen2.5-0.5B-Instruct-CrashCourse-dropout
89
- name: Open LLM Leaderboard
90
- - task:
91
- type: text-generation
92
- name: Text Generation
93
- dataset:
94
- name: MMLU-PRO (5-shot)
95
- type: TIGER-Lab/MMLU-Pro
96
- config: main
97
- split: test
98
- args:
99
- num_few_shot: 5
100
- metrics:
101
- - type: acc
102
- value: 6.76
103
- name: accuracy
104
- source:
105
- url: https://huggingface.co/spaces/open-llm-leaderboard/open_llm_leaderboard#/?search=agentlans%2FQwen2.5-0.5B-Instruct-CrashCourse-dropout
106
- name: Open LLM Leaderboard
107
- ---
108
- # Qwen2.5-0.5B-Instruct-CrashCourse-dropout
109
-
110
- This model is a fine-tuned version of [Qwen/Qwen2.5-0.5B-Instruct](https://huggingface.co/Qwen/Qwen2.5-0.5B-Instruct),
111
- specifically adapted for enhanced performance on instructional and multitask scenarios.
112
- It leverages two datasets: [agentlans/crash-course](https://huggingface.co/datasets/agentlans/crash-course) and
113
- [vicgalle/configurable-system-prompt-multitask](https://huggingface.co/datasets/vicgalle/configurable-system-prompt-multitask)
114
- to improve its capabilities in handling diverse tasks and responding to various instruction formats.
115
-
116
- > [!NOTE]
117
- > **Update:** Despite the poor benchmark, the model seems OK at slightly complex prompts. There's more finetuning potential here.
118
-
119
- ## Intended Use
120
-
121
- This model is designed for:
122
-
123
- - Answering questions related to crash course materials
124
- - Handling configurable system prompts for multitask scenarios
125
- - General instruction-following tasks
126
-
127
- ## Training Procedure
128
-
129
- The model was fine-tuned on the specified datasets using the Qwen2.5-0.5B-Instruct as the base model.
130
- More details on the training process will be added here later.
131
-
132
- ## Limitations
133
-
134
- - The model's performance may be biased towards the specific domains covered in the training datasets.
135
- - As with all language models, it may occasionally produce inaccurate or inconsistent outputs.
136
- - The model's knowledge is limited to the information available in its training data and the base model's knowledge cutoff.
137
-
138
- ## Ethical Considerations
139
-
140
- Users should be aware that this model, like all AI models, may reflect biases present in its training data. It's crucial to use the model responsibly and to verify important information from authoritative sources.
141
-
142
- ## Additional Information
143
-
144
- For more details on the base model, please refer to the Qwen/Qwen2.5-0.5B-Instruct model card. For information about the datasets used in fine-tuning, check the respective dataset cards on the Hugging Face Hub.
145
-
146
- # [Open LLM Leaderboard Evaluation Results](https://huggingface.co/spaces/open-llm-leaderboard/open_llm_leaderboard)
147
- Detailed results can be found [here](https://huggingface.co/datasets/open-llm-leaderboard/agentlans__Qwen2.5-0.5B-Instruct-CrashCourse-dropout-details)!
148
- Summarized results can be found [here](https://huggingface.co/datasets/open-llm-leaderboard/contents/viewer/default/train?q=agentlans%2FQwen2.5-0.5B-Instruct-CrashCourse-dropout&sort[column]=Average%20%E2%AC%86%EF%B8%8F&sort[direction]=desc)!
149
-
150
- | Metric | Qwen2.5-0.5B-Instruct-CrashCourse-dropout | Qwen2.5-0.5B-Instruct |
151
- |-------------------|-----------------------------------------:|----------------------:|
152
- | **Average** | 7.74 % | 8.38 % |
153
- | IFEval (0-Shot) | 29.49 % | 31.53 % |
154
- | BBH (3-Shot) | 7.23 % | 8.17 % |
155
- | MATH Lvl 5 (4-Shot)| 0.08 % | 0.00 % |
156
- | GPQA (0-shot) | 1.79 % | 1.23 % |
157
- | MuSR (0-shot) | 1.11 % | 1.37 % |
158
- | MMLU-PRO (5-shot) | 6.76 % | 8.00 % |
159
-
160
-
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ ---
2
+ license: apache-2.0
3
+ datasets:
4
+ - agentlans/crash-course
5
+ - vicgalle/configurable-system-prompt-multitask
6
+ base_model:
7
+ - Qwen/Qwen2.5-0.5B-Instruct
8
+ language:
9
+ - zho
10
+ - eng
11
+ - fra
12
+ - spa
13
+ - por
14
+ - deu
15
+ - ita
16
+ - rus
17
+ - jpn
18
+ - kor
19
+ - vie
20
+ - tha
21
+ - ara
22
+ model-index:
23
+ - name: Qwen2.5-0.5B-Instruct-CrashCourse-dropout
24
+ results:
25
+ - task:
26
+ type: text-generation
27
+ name: Text Generation
28
+ dataset:
29
+ name: IFEval (0-Shot)
30
+ type: wis-k/instruction-following-eval
31
+ split: train
32
+ args:
33
+ num_few_shot: 0
34
+ metrics:
35
+ - type: inst_level_strict_acc and prompt_level_strict_acc
36
+ value: 29.49
37
+ name: averaged accuracy
38
+ source:
39
+ url: https://huggingface.co/spaces/open-llm-leaderboard/open_llm_leaderboard#/?search=agentlans%2FQwen2.5-0.5B-Instruct-CrashCourse-dropout
40
+ name: Open LLM Leaderboard
41
+ - task:
42
+ type: text-generation
43
+ name: Text Generation
44
+ dataset:
45
+ name: BBH (3-Shot)
46
+ type: SaylorTwift/bbh
47
+ split: test
48
+ args:
49
+ num_few_shot: 3
50
+ metrics:
51
+ - type: acc_norm
52
+ value: 7.23
53
+ name: normalized accuracy
54
+ source:
55
+ url: https://huggingface.co/spaces/open-llm-leaderboard/open_llm_leaderboard#/?search=agentlans%2FQwen2.5-0.5B-Instruct-CrashCourse-dropout
56
+ name: Open LLM Leaderboard
57
+ - task:
58
+ type: text-generation
59
+ name: Text Generation
60
+ dataset:
61
+ name: MATH Lvl 5 (4-Shot)
62
+ type: lighteval/MATH-Hard
63
+ split: test
64
+ args:
65
+ num_few_shot: 4
66
+ metrics:
67
+ - type: exact_match
68
+ value: 0.08
69
+ name: exact match
70
+ source:
71
+ url: https://huggingface.co/spaces/open-llm-leaderboard/open_llm_leaderboard#/?search=agentlans%2FQwen2.5-0.5B-Instruct-CrashCourse-dropout
72
+ name: Open LLM Leaderboard
73
+ - task:
74
+ type: text-generation
75
+ name: Text Generation
76
+ dataset:
77
+ name: GPQA (0-shot)
78
+ type: Idavidrein/gpqa
79
+ split: train
80
+ args:
81
+ num_few_shot: 0
82
+ metrics:
83
+ - type: acc_norm
84
+ value: 1.79
85
+ name: acc_norm
86
+ source:
87
+ url: https://huggingface.co/spaces/open-llm-leaderboard/open_llm_leaderboard#/?search=agentlans%2FQwen2.5-0.5B-Instruct-CrashCourse-dropout
88
+ name: Open LLM Leaderboard
89
+ - task:
90
+ type: text-generation
91
+ name: Text Generation
92
+ dataset:
93
+ name: MuSR (0-shot)
94
+ type: TAUR-Lab/MuSR
95
+ args:
96
+ num_few_shot: 0
97
+ metrics:
98
+ - type: acc_norm
99
+ value: 1.11
100
+ name: acc_norm
101
+ source:
102
+ url: https://huggingface.co/spaces/open-llm-leaderboard/open_llm_leaderboard#/?search=agentlans%2FQwen2.5-0.5B-Instruct-CrashCourse-dropout
103
+ name: Open LLM Leaderboard
104
+ - task:
105
+ type: text-generation
106
+ name: Text Generation
107
+ dataset:
108
+ name: MMLU-PRO (5-shot)
109
+ type: TIGER-Lab/MMLU-Pro
110
+ config: main
111
+ split: test
112
+ args:
113
+ num_few_shot: 5
114
+ metrics:
115
+ - type: acc
116
+ value: 6.76
117
+ name: accuracy
118
+ source:
119
+ url: https://huggingface.co/spaces/open-llm-leaderboard/open_llm_leaderboard#/?search=agentlans%2FQwen2.5-0.5B-Instruct-CrashCourse-dropout
120
+ name: Open LLM Leaderboard
121
+ ---
122
+ # Qwen2.5-0.5B-Instruct-CrashCourse-dropout
123
+
124
+ This model is a fine-tuned version of [Qwen/Qwen2.5-0.5B-Instruct](https://huggingface.co/Qwen/Qwen2.5-0.5B-Instruct),
125
+ specifically adapted for enhanced performance on instructional and multitask scenarios.
126
+ It leverages two datasets: [agentlans/crash-course](https://huggingface.co/datasets/agentlans/crash-course) and
127
+ [vicgalle/configurable-system-prompt-multitask](https://huggingface.co/datasets/vicgalle/configurable-system-prompt-multitask)
128
+ to improve its capabilities in handling diverse tasks and responding to various instruction formats.
129
+
130
+ > [!NOTE]
131
+ > **Update:** Despite the poor benchmark, the model seems OK at slightly complex prompts. There's more finetuning potential here.
132
+
133
+ ## Intended Use
134
+
135
+ This model is designed for:
136
+
137
+ - Answering questions related to crash course materials
138
+ - Handling configurable system prompts for multitask scenarios
139
+ - General instruction-following tasks
140
+
141
+ ## Training Procedure
142
+
143
+ The model was fine-tuned on the specified datasets using the Qwen2.5-0.5B-Instruct as the base model.
144
+ More details on the training process will be added here later.
145
+
146
+ ## Limitations
147
+
148
+ - The model's performance may be biased towards the specific domains covered in the training datasets.
149
+ - As with all language models, it may occasionally produce inaccurate or inconsistent outputs.
150
+ - The model's knowledge is limited to the information available in its training data and the base model's knowledge cutoff.
151
+
152
+ ## Ethical Considerations
153
+
154
+ Users should be aware that this model, like all AI models, may reflect biases present in its training data. It's crucial to use the model responsibly and to verify important information from authoritative sources.
155
+
156
+ ## Additional Information
157
+
158
+ For more details on the base model, please refer to the Qwen/Qwen2.5-0.5B-Instruct model card. For information about the datasets used in fine-tuning, check the respective dataset cards on the Hugging Face Hub.
159
+
160
+ # [Open LLM Leaderboard Evaluation Results](https://huggingface.co/spaces/open-llm-leaderboard/open_llm_leaderboard)
161
+ Detailed results can be found [here](https://huggingface.co/datasets/open-llm-leaderboard/agentlans__Qwen2.5-0.5B-Instruct-CrashCourse-dropout-details)!
162
+ Summarized results can be found [here](https://huggingface.co/datasets/open-llm-leaderboard/contents/viewer/default/train?q=agentlans%2FQwen2.5-0.5B-Instruct-CrashCourse-dropout&sort[column]=Average%20%E2%AC%86%EF%B8%8F&sort[direction]=desc)!
163
+
164
+ | Metric | Qwen2.5-0.5B-Instruct-CrashCourse-dropout | Qwen2.5-0.5B-Instruct |
165
+ |-------------------|-----------------------------------------:|----------------------:|
166
+ | **Average** | 7.74 % | 8.38 % |
167
+ | IFEval (0-Shot) | 29.49 % | 31.53 % |
168
+ | BBH (3-Shot) | 7.23 % | 8.17 % |
169
+ | MATH Lvl 5 (4-Shot)| 0.08 % | 0.00 % |
170
+ | GPQA (0-shot) | 1.79 % | 1.23 % |
171
+ | MuSR (0-shot) | 1.11 % | 1.37 % |
172
+ | MMLU-PRO (5-shot) | 6.76 % | 8.00 % |
173
+
174
+