code380 commited on
Commit
25bcdff
·
verified ·
1 Parent(s): 74781da

Upload README.md with huggingface_hub

Browse files
Files changed (1) hide show
  1. README.md +385 -0
README.md ADDED
@@ -0,0 +1,385 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ ---
2
+ base_model: OpenLLM-Ro/RoGemma2-9b-Instruct-DPO
3
+ datasets:
4
+ - OpenLLM-Ro/ro_dpo_helpsteer
5
+ language:
6
+ - ro
7
+ license: cc-by-nc-4.0
8
+ tags:
9
+ - llama-cpp
10
+ - gguf-my-repo
11
+ model-index:
12
+ - name: OpenLLM-Ro/RoGemma2-9b-Instruct-DPO-2024-10-09
13
+ results:
14
+ - task:
15
+ type: text-generation
16
+ dataset:
17
+ name: RoMT-Bench
18
+ type: RoMT-Bench
19
+ metrics:
20
+ - type: Score
21
+ value: 6.77
22
+ name: Score
23
+ - type: Score
24
+ value: 7.24
25
+ name: First turn
26
+ - type: Score
27
+ value: 6.3
28
+ name: Second turn
29
+ - task:
30
+ type: text-generation
31
+ dataset:
32
+ name: RoCulturaBench
33
+ type: RoCulturaBench
34
+ metrics:
35
+ - type: Score
36
+ value: 4.83
37
+ name: Score
38
+ - task:
39
+ type: text-generation
40
+ dataset:
41
+ name: Romanian_Academic_Benchmarks
42
+ type: Romanian_Academic_Benchmarks
43
+ metrics:
44
+ - type: accuracy
45
+ value: 59.08
46
+ name: Average accuracy
47
+ - task:
48
+ type: text-generation
49
+ dataset:
50
+ name: OpenLLM-Ro/ro_arc_challenge
51
+ type: OpenLLM-Ro/ro_arc_challenge
52
+ metrics:
53
+ - type: accuracy
54
+ value: 54.1
55
+ name: Average accuracy
56
+ - type: accuracy
57
+ value: 51.59
58
+ name: 0-shot
59
+ - type: accuracy
60
+ value: 50.99
61
+ name: 1-shot
62
+ - type: accuracy
63
+ value: 53.47
64
+ name: 3-shot
65
+ - type: accuracy
66
+ value: 54.84
67
+ name: 5-shot
68
+ - type: accuracy
69
+ value: 58.1
70
+ name: 10-shot
71
+ - type: accuracy
72
+ value: 55.61
73
+ name: 25-shot
74
+ - task:
75
+ type: text-generation
76
+ dataset:
77
+ name: OpenLLM-Ro/ro_mmlu
78
+ type: OpenLLM-Ro/ro_mmlu
79
+ metrics:
80
+ - type: accuracy
81
+ value: 63.41
82
+ name: Average accuracy
83
+ - type: accuracy
84
+ value: 62.15
85
+ name: 0-shot
86
+ - type: accuracy
87
+ value: 62.78
88
+ name: 1-shot
89
+ - type: accuracy
90
+ value: 64.27
91
+ name: 3-shot
92
+ - type: accuracy
93
+ value: 64.43
94
+ name: 5-shot
95
+ - task:
96
+ type: text-generation
97
+ dataset:
98
+ name: OpenLLM-Ro/ro_winogrande
99
+ type: OpenLLM-Ro/ro_winogrande
100
+ metrics:
101
+ - type: accuracy
102
+ value: 70.02
103
+ name: Average accuracy
104
+ - type: accuracy
105
+ value: 66.69
106
+ name: 0-shot
107
+ - type: accuracy
108
+ value: 68.82
109
+ name: 1-shot
110
+ - type: accuracy
111
+ value: 71.82
112
+ name: 3-shot
113
+ - type: accuracy
114
+ value: 72.77
115
+ name: 5-shot
116
+ - task:
117
+ type: text-generation
118
+ dataset:
119
+ name: OpenLLM-Ro/ro_hellaswag
120
+ type: OpenLLM-Ro/ro_hellaswag
121
+ metrics:
122
+ - type: accuracy
123
+ value: 59.35
124
+ name: Average accuracy
125
+ - type: accuracy
126
+ value: 56.98
127
+ name: 0-shot
128
+ - type: accuracy
129
+ value: 57.73
130
+ name: 1-shot
131
+ - type: accuracy
132
+ value: 59.29
133
+ name: 3-shot
134
+ - type: accuracy
135
+ value: 60.7
136
+ name: 5-shot
137
+ - type: accuracy
138
+ value: 62.03
139
+ name: 10-shot
140
+ - task:
141
+ type: text-generation
142
+ dataset:
143
+ name: OpenLLM-Ro/ro_gsm8k
144
+ type: OpenLLM-Ro/ro_gsm8k
145
+ metrics:
146
+ - type: accuracy
147
+ value: 57.24
148
+ name: Average accuracy
149
+ - type: accuracy
150
+ value: 46.78
151
+ name: 1-shot
152
+ - type: accuracy
153
+ value: 59.97
154
+ name: 3-shot
155
+ - type: accuracy
156
+ value: 64.97
157
+ name: 5-shot
158
+ - task:
159
+ type: text-generation
160
+ dataset:
161
+ name: OpenLLM-Ro/ro_truthfulqa
162
+ type: OpenLLM-Ro/ro_truthfulqa
163
+ metrics:
164
+ - type: accuracy
165
+ value: 50.39
166
+ name: Average accuracy
167
+ - task:
168
+ type: text-generation
169
+ dataset:
170
+ name: LaRoSeDa_binary
171
+ type: LaRoSeDa_binary
172
+ metrics:
173
+ - type: macro-f1
174
+ value: 97.74
175
+ name: Average macro-f1
176
+ - type: macro-f1
177
+ value: 97.3
178
+ name: 0-shot
179
+ - type: macro-f1
180
+ value: 97.5
181
+ name: 1-shot
182
+ - type: macro-f1
183
+ value: 97.83
184
+ name: 3-shot
185
+ - type: macro-f1
186
+ value: 98.33
187
+ name: 5-shot
188
+ - task:
189
+ type: text-generation
190
+ dataset:
191
+ name: LaRoSeDa_multiclass
192
+ type: LaRoSeDa_multiclass
193
+ metrics:
194
+ - type: macro-f1
195
+ value: 67.4
196
+ name: Average macro-f1
197
+ - type: macro-f1
198
+ value: 59.3
199
+ name: 0-shot
200
+ - type: macro-f1
201
+ value: 65.52
202
+ name: 1-shot
203
+ - type: macro-f1
204
+ value: 70.94
205
+ name: 3-shot
206
+ - type: macro-f1
207
+ value: 73.85
208
+ name: 5-shot
209
+ - task:
210
+ type: text-generation
211
+ dataset:
212
+ name: WMT_EN-RO
213
+ type: WMT_EN-RO
214
+ metrics:
215
+ - type: bleu
216
+ value: 27.32
217
+ name: Average bleu
218
+ - type: bleu
219
+ value: 17.49
220
+ name: 0-shot
221
+ - type: bleu
222
+ value: 30.33
223
+ name: 1-shot
224
+ - type: bleu
225
+ value: 30.58
226
+ name: 3-shot
227
+ - type: bleu
228
+ value: 30.88
229
+ name: 5-shot
230
+ - task:
231
+ type: text-generation
232
+ dataset:
233
+ name: WMT_RO-EN
234
+ type: WMT_RO-EN
235
+ metrics:
236
+ - type: bleu
237
+ value: 15.96
238
+ name: Average bleu
239
+ - type: bleu
240
+ value: 2.17
241
+ name: 0-shot
242
+ - type: bleu
243
+ value: 10.69
244
+ name: 1-shot
245
+ - type: bleu
246
+ value: 21.68
247
+ name: 3-shot
248
+ - type: bleu
249
+ value: 29.28
250
+ name: 5-shot
251
+ - task:
252
+ type: text-generation
253
+ dataset:
254
+ name: XQuAD
255
+ type: XQuAD
256
+ metrics:
257
+ - type: exact_match
258
+ value: 32.42
259
+ name: Average exact_match
260
+ - type: f1
261
+ value: 58.68
262
+ name: Average f1
263
+ - task:
264
+ type: text-generation
265
+ dataset:
266
+ name: STS
267
+ type: STS
268
+ metrics:
269
+ - type: spearman
270
+ value: 80.82
271
+ name: Average spearman
272
+ - type: pearson
273
+ value: 81.5
274
+ name: Average pearson
275
+ - task:
276
+ type: text-generation
277
+ dataset:
278
+ name: XQuAD_EM
279
+ type: XQuAD_EM
280
+ metrics:
281
+ - type: exact_match
282
+ value: 23.28
283
+ name: 0-shot
284
+ - type: exact_match
285
+ value: 33.45
286
+ name: 1-shot
287
+ - type: exact_match
288
+ value: 34.37
289
+ name: 3-shot
290
+ - type: exact_match
291
+ value: 38.57
292
+ name: 5-shot
293
+ - task:
294
+ type: text-generation
295
+ dataset:
296
+ name: XQuAD_F1
297
+ type: XQuAD_F1
298
+ metrics:
299
+ - type: f1
300
+ value: 47.16
301
+ name: 0-shot
302
+ - type: f1
303
+ value: 60.28
304
+ name: 1-shot
305
+ - type: f1
306
+ value: 62.09
307
+ name: 3-shot
308
+ - type: f1
309
+ value: 65.2
310
+ name: 5-shot
311
+ - task:
312
+ type: text-generation
313
+ dataset:
314
+ name: STS_Spearman
315
+ type: STS_Spearman
316
+ metrics:
317
+ - type: spearman
318
+ value: 75.34
319
+ name: 1-shot
320
+ - type: spearman
321
+ value: 82.71
322
+ name: 3-shot
323
+ - type: spearman
324
+ value: 84.41
325
+ name: 5-shot
326
+ - task:
327
+ type: text-generation
328
+ dataset:
329
+ name: STS_Pearson
330
+ type: STS_Pearson
331
+ metrics:
332
+ - type: pearson
333
+ value: 77.97
334
+ name: 1-shot
335
+ - type: pearson
336
+ value: 82.49
337
+ name: 3-shot
338
+ - type: pearson
339
+ value: 84.05
340
+ name: 5-shot
341
+ ---
342
+
343
+ # code380/RoGemma2-9b-Instruct-DPO-Q4_K_M-GGUF
344
+ This model was converted to GGUF format from [`OpenLLM-Ro/RoGemma2-9b-Instruct-DPO`](https://huggingface.co/OpenLLM-Ro/RoGemma2-9b-Instruct-DPO) using llama.cpp via the ggml.ai's [GGUF-my-repo](https://huggingface.co/spaces/ggml-org/gguf-my-repo) space.
345
+ Refer to the [original model card](https://huggingface.co/OpenLLM-Ro/RoGemma2-9b-Instruct-DPO) for more details on the model.
346
+
347
+ ## Use with llama.cpp
348
+ Install llama.cpp through brew (works on Mac and Linux)
349
+
350
+ ```bash
351
+ brew install llama.cpp
352
+
353
+ ```
354
+ Invoke the llama.cpp server or the CLI.
355
+
356
+ ### CLI:
357
+ ```bash
358
+ llama-cli --hf-repo code380/RoGemma2-9b-Instruct-DPO-Q4_K_M-GGUF --hf-file rogemma2-9b-instruct-dpo-q4_k_m.gguf -p "The meaning to life and the universe is"
359
+ ```
360
+
361
+ ### Server:
362
+ ```bash
363
+ llama-server --hf-repo code380/RoGemma2-9b-Instruct-DPO-Q4_K_M-GGUF --hf-file rogemma2-9b-instruct-dpo-q4_k_m.gguf -c 2048
364
+ ```
365
+
366
+ Note: You can also use this checkpoint directly through the [usage steps](https://github.com/ggerganov/llama.cpp?tab=readme-ov-file#usage) listed in the Llama.cpp repo as well.
367
+
368
+ Step 1: Clone llama.cpp from GitHub.
369
+ ```
370
+ git clone https://github.com/ggerganov/llama.cpp
371
+ ```
372
+
373
+ Step 2: Move into the llama.cpp folder and build it with `LLAMA_CURL=1` flag along with other hardware-specific flags (for ex: LLAMA_CUDA=1 for Nvidia GPUs on Linux).
374
+ ```
375
+ cd llama.cpp && LLAMA_CURL=1 make
376
+ ```
377
+
378
+ Step 3: Run inference through the main binary.
379
+ ```
380
+ ./llama-cli --hf-repo code380/RoGemma2-9b-Instruct-DPO-Q4_K_M-GGUF --hf-file rogemma2-9b-instruct-dpo-q4_k_m.gguf -p "The meaning to life and the universe is"
381
+ ```
382
+ or
383
+ ```
384
+ ./llama-server --hf-repo code380/RoGemma2-9b-Instruct-DPO-Q4_K_M-GGUF --hf-file rogemma2-9b-instruct-dpo-q4_k_m.gguf -c 2048
385
+ ```