Improve language tag

#2
by lbourdois - opened
Files changed (1) hide show
  1. README.md +182 -168
README.md CHANGED
@@ -1,168 +1,182 @@
1
- ---
2
- library_name: transformers
3
- tags:
4
- - mergekit
5
- - merge
6
- base_model:
7
- - Etherll/Qwen2.5-7B-della-test
8
- - jeffmeloy/Qwen2.5-7B-nerd-uncensored-v1.0
9
- - fblgit/cybertron-v4-qw7B-MGS
10
- - Qwen/Qwen2.5-7B
11
- model-index:
12
- - name: QandoraExp-7B
13
- results:
14
- - task:
15
- type: text-generation
16
- name: Text Generation
17
- dataset:
18
- name: IFEval (0-Shot)
19
- type: HuggingFaceH4/ifeval
20
- args:
21
- num_few_shot: 0
22
- metrics:
23
- - type: inst_level_strict_acc and prompt_level_strict_acc
24
- value: 75.09
25
- name: strict accuracy
26
- source:
27
- url: https://huggingface.co/spaces/open-llm-leaderboard/open_llm_leaderboard?query=bunnycore/QandoraExp-7B
28
- name: Open LLM Leaderboard
29
- - task:
30
- type: text-generation
31
- name: Text Generation
32
- dataset:
33
- name: BBH (3-Shot)
34
- type: BBH
35
- args:
36
- num_few_shot: 3
37
- metrics:
38
- - type: acc_norm
39
- value: 35.92
40
- name: normalized accuracy
41
- source:
42
- url: https://huggingface.co/spaces/open-llm-leaderboard/open_llm_leaderboard?query=bunnycore/QandoraExp-7B
43
- name: Open LLM Leaderboard
44
- - task:
45
- type: text-generation
46
- name: Text Generation
47
- dataset:
48
- name: MATH Lvl 5 (4-Shot)
49
- type: hendrycks/competition_math
50
- args:
51
- num_few_shot: 4
52
- metrics:
53
- - type: exact_match
54
- value: 0.91
55
- name: exact match
56
- source:
57
- url: https://huggingface.co/spaces/open-llm-leaderboard/open_llm_leaderboard?query=bunnycore/QandoraExp-7B
58
- name: Open LLM Leaderboard
59
- - task:
60
- type: text-generation
61
- name: Text Generation
62
- dataset:
63
- name: GPQA (0-shot)
64
- type: Idavidrein/gpqa
65
- args:
66
- num_few_shot: 0
67
- metrics:
68
- - type: acc_norm
69
- value: 8.05
70
- name: acc_norm
71
- source:
72
- url: https://huggingface.co/spaces/open-llm-leaderboard/open_llm_leaderboard?query=bunnycore/QandoraExp-7B
73
- name: Open LLM Leaderboard
74
- - task:
75
- type: text-generation
76
- name: Text Generation
77
- dataset:
78
- name: MuSR (0-shot)
79
- type: TAUR-Lab/MuSR
80
- args:
81
- num_few_shot: 0
82
- metrics:
83
- - type: acc_norm
84
- value: 13.2
85
- name: acc_norm
86
- source:
87
- url: https://huggingface.co/spaces/open-llm-leaderboard/open_llm_leaderboard?query=bunnycore/QandoraExp-7B
88
- name: Open LLM Leaderboard
89
- - task:
90
- type: text-generation
91
- name: Text Generation
92
- dataset:
93
- name: MMLU-PRO (5-shot)
94
- type: TIGER-Lab/MMLU-Pro
95
- config: main
96
- split: test
97
- args:
98
- num_few_shot: 5
99
- metrics:
100
- - type: acc
101
- value: 37.89
102
- name: accuracy
103
- source:
104
- url: https://huggingface.co/spaces/open-llm-leaderboard/open_llm_leaderboard?query=bunnycore/QandoraExp-7B
105
- name: Open LLM Leaderboard
106
- ---
107
- # merge
108
-
109
- This is a merge of pre-trained language models created using [mergekit](https://github.com/cg123/mergekit).
110
-
111
- ## Merge Details
112
- ### Merge Method
113
-
114
- This model was merged using the della merge method using [Qwen/Qwen2.5-7B](https://huggingface.co/Qwen/Qwen2.5-7B) as a base.
115
-
116
- ### Models Merged
117
-
118
- The following models were included in the merge:
119
- * [Etherll/Qwen2.5-7B-della-test](https://huggingface.co/Etherll/Qwen2.5-7B-della-test)
120
- * [jeffmeloy/Qwen2.5-7B-nerd-uncensored-v1.0](https://huggingface.co/jeffmeloy/Qwen2.5-7B-nerd-uncensored-v1.0)
121
- * [fblgit/cybertron-v4-qw7B-MGS](https://huggingface.co/fblgit/cybertron-v4-qw7B-MGS)
122
-
123
- ### Configuration
124
-
125
- The following YAML configuration was used to produce this model:
126
-
127
- ```yaml
128
-
129
- models:
130
- - model: Etherll/Qwen2.5-7B-della-test
131
- parameters:
132
- weight: 1
133
- density: 1
134
- lambda: 0.9
135
- - model: jeffmeloy/Qwen2.5-7B-nerd-uncensored-v1.0
136
- parameters:
137
- weight: 1
138
- density: 1
139
- lambda: 0.9
140
- - model: fblgit/cybertron-v4-qw7B-MGS
141
- parameters:
142
- weight: 1
143
- density: 1
144
- lambda: 0.9
145
- merge_method: della
146
- base_model: Qwen/Qwen2.5-7B
147
- parameters:
148
- weight: 1
149
- density: 1
150
- lambda: 0.9
151
- int8_mask: true
152
- dtype: bfloat16
153
-
154
- ```
155
-
156
- # [Open LLM Leaderboard Evaluation Results](https://huggingface.co/spaces/open-llm-leaderboard/open_llm_leaderboard)
157
- Detailed results can be found [here](https://huggingface.co/datasets/open-llm-leaderboard/details_bunnycore__QandoraExp-7B)
158
-
159
- | Metric |Value|
160
- |-------------------|----:|
161
- |Avg. |28.51|
162
- |IFEval (0-Shot) |75.09|
163
- |BBH (3-Shot) |35.92|
164
- |MATH Lvl 5 (4-Shot)| 0.91|
165
- |GPQA (0-shot) | 8.05|
166
- |MuSR (0-shot) |13.20|
167
- |MMLU-PRO (5-shot) |37.89|
168
-
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ ---
2
+ library_name: transformers
3
+ tags:
4
+ - mergekit
5
+ - merge
6
+ base_model:
7
+ - Etherll/Qwen2.5-7B-della-test
8
+ - jeffmeloy/Qwen2.5-7B-nerd-uncensored-v1.0
9
+ - fblgit/cybertron-v4-qw7B-MGS
10
+ - Qwen/Qwen2.5-7B
11
+ language:
12
+ - zho
13
+ - eng
14
+ - fra
15
+ - spa
16
+ - por
17
+ - deu
18
+ - ita
19
+ - rus
20
+ - jpn
21
+ - kor
22
+ - vie
23
+ - tha
24
+ - ara
25
+ model-index:
26
+ - name: QandoraExp-7B
27
+ results:
28
+ - task:
29
+ type: text-generation
30
+ name: Text Generation
31
+ dataset:
32
+ name: IFEval (0-Shot)
33
+ type: HuggingFaceH4/ifeval
34
+ args:
35
+ num_few_shot: 0
36
+ metrics:
37
+ - type: inst_level_strict_acc and prompt_level_strict_acc
38
+ value: 75.09
39
+ name: strict accuracy
40
+ source:
41
+ url: https://huggingface.co/spaces/open-llm-leaderboard/open_llm_leaderboard?query=bunnycore/QandoraExp-7B
42
+ name: Open LLM Leaderboard
43
+ - task:
44
+ type: text-generation
45
+ name: Text Generation
46
+ dataset:
47
+ name: BBH (3-Shot)
48
+ type: BBH
49
+ args:
50
+ num_few_shot: 3
51
+ metrics:
52
+ - type: acc_norm
53
+ value: 35.92
54
+ name: normalized accuracy
55
+ source:
56
+ url: https://huggingface.co/spaces/open-llm-leaderboard/open_llm_leaderboard?query=bunnycore/QandoraExp-7B
57
+ name: Open LLM Leaderboard
58
+ - task:
59
+ type: text-generation
60
+ name: Text Generation
61
+ dataset:
62
+ name: MATH Lvl 5 (4-Shot)
63
+ type: hendrycks/competition_math
64
+ args:
65
+ num_few_shot: 4
66
+ metrics:
67
+ - type: exact_match
68
+ value: 0.91
69
+ name: exact match
70
+ source:
71
+ url: https://huggingface.co/spaces/open-llm-leaderboard/open_llm_leaderboard?query=bunnycore/QandoraExp-7B
72
+ name: Open LLM Leaderboard
73
+ - task:
74
+ type: text-generation
75
+ name: Text Generation
76
+ dataset:
77
+ name: GPQA (0-shot)
78
+ type: Idavidrein/gpqa
79
+ args:
80
+ num_few_shot: 0
81
+ metrics:
82
+ - type: acc_norm
83
+ value: 8.05
84
+ name: acc_norm
85
+ source:
86
+ url: https://huggingface.co/spaces/open-llm-leaderboard/open_llm_leaderboard?query=bunnycore/QandoraExp-7B
87
+ name: Open LLM Leaderboard
88
+ - task:
89
+ type: text-generation
90
+ name: Text Generation
91
+ dataset:
92
+ name: MuSR (0-shot)
93
+ type: TAUR-Lab/MuSR
94
+ args:
95
+ num_few_shot: 0
96
+ metrics:
97
+ - type: acc_norm
98
+ value: 13.2
99
+ name: acc_norm
100
+ source:
101
+ url: https://huggingface.co/spaces/open-llm-leaderboard/open_llm_leaderboard?query=bunnycore/QandoraExp-7B
102
+ name: Open LLM Leaderboard
103
+ - task:
104
+ type: text-generation
105
+ name: Text Generation
106
+ dataset:
107
+ name: MMLU-PRO (5-shot)
108
+ type: TIGER-Lab/MMLU-Pro
109
+ config: main
110
+ split: test
111
+ args:
112
+ num_few_shot: 5
113
+ metrics:
114
+ - type: acc
115
+ value: 37.89
116
+ name: accuracy
117
+ source:
118
+ url: https://huggingface.co/spaces/open-llm-leaderboard/open_llm_leaderboard?query=bunnycore/QandoraExp-7B
119
+ name: Open LLM Leaderboard
120
+ ---
121
+ # merge
122
+
123
+ This is a merge of pre-trained language models created using [mergekit](https://github.com/cg123/mergekit).
124
+
125
+ ## Merge Details
126
+ ### Merge Method
127
+
128
+ This model was merged using the della merge method using [Qwen/Qwen2.5-7B](https://huggingface.co/Qwen/Qwen2.5-7B) as a base.
129
+
130
+ ### Models Merged
131
+
132
+ The following models were included in the merge:
133
+ * [Etherll/Qwen2.5-7B-della-test](https://huggingface.co/Etherll/Qwen2.5-7B-della-test)
134
+ * [jeffmeloy/Qwen2.5-7B-nerd-uncensored-v1.0](https://huggingface.co/jeffmeloy/Qwen2.5-7B-nerd-uncensored-v1.0)
135
+ * [fblgit/cybertron-v4-qw7B-MGS](https://huggingface.co/fblgit/cybertron-v4-qw7B-MGS)
136
+
137
+ ### Configuration
138
+
139
+ The following YAML configuration was used to produce this model:
140
+
141
+ ```yaml
142
+
143
+ models:
144
+ - model: Etherll/Qwen2.5-7B-della-test
145
+ parameters:
146
+ weight: 1
147
+ density: 1
148
+ lambda: 0.9
149
+ - model: jeffmeloy/Qwen2.5-7B-nerd-uncensored-v1.0
150
+ parameters:
151
+ weight: 1
152
+ density: 1
153
+ lambda: 0.9
154
+ - model: fblgit/cybertron-v4-qw7B-MGS
155
+ parameters:
156
+ weight: 1
157
+ density: 1
158
+ lambda: 0.9
159
+ merge_method: della
160
+ base_model: Qwen/Qwen2.5-7B
161
+ parameters:
162
+ weight: 1
163
+ density: 1
164
+ lambda: 0.9
165
+ int8_mask: true
166
+ dtype: bfloat16
167
+
168
+ ```
169
+
170
+ # [Open LLM Leaderboard Evaluation Results](https://huggingface.co/spaces/open-llm-leaderboard/open_llm_leaderboard)
171
+ Detailed results can be found [here](https://huggingface.co/datasets/open-llm-leaderboard/details_bunnycore__QandoraExp-7B)
172
+
173
+ | Metric |Value|
174
+ |-------------------|----:|
175
+ |Avg. |28.51|
176
+ |IFEval (0-Shot) |75.09|
177
+ |BBH (3-Shot) |35.92|
178
+ |MATH Lvl 5 (4-Shot)| 0.91|
179
+ |GPQA (0-shot) | 8.05|
180
+ |MuSR (0-shot) |13.20|
181
+ |MMLU-PRO (5-shot) |37.89|
182
+