simon-muenker commited on
Commit
2fb4173
·
verified ·
1 Parent(s): 14b2b75

Model save

Browse files
README.md CHANGED
@@ -20,7 +20,7 @@ should probably proofread and complete it, then remove this comment. -->
20
 
21
  This model is a fine-tuned version of [meta-llama/Llama-3.2-3B-Instruct](https://huggingface.co/meta-llama/Llama-3.2-3B-Instruct) on the generator dataset.
22
  It achieves the following results on the evaluation set:
23
- - Loss: 2.9401
24
 
25
  ## Model description
26
 
@@ -51,12 +51,14 @@ The following hyperparameters were used during training:
51
 
52
  | Training Loss | Epoch | Step | Validation Loss |
53
  |:-------------:|:-----:|:----:|:---------------:|
54
- | 3.3716 | 1.25 | 5 | 3.2267 |
55
- | 3.2374 | 2.5 | 10 | 3.1263 |
56
- | 3.1718 | 3.75 | 15 | 3.0495 |
57
- | 3.0543 | 5.0 | 20 | 2.9950 |
58
- | 3.0153 | 6.25 | 25 | 2.9599 |
59
- | 2.9787 | 7.5 | 30 | 2.9422 |
 
 
60
 
61
 
62
  ### Framework versions
 
20
 
21
  This model is a fine-tuned version of [meta-llama/Llama-3.2-3B-Instruct](https://huggingface.co/meta-llama/Llama-3.2-3B-Instruct) on the generator dataset.
22
  It achieves the following results on the evaluation set:
23
+ - Loss: 2.8582
24
 
25
  ## Model description
26
 
 
51
 
52
  | Training Loss | Epoch | Step | Validation Loss |
53
  |:-------------:|:-----:|:----:|:---------------:|
54
+ | 3.4066 | 1.0 | 5 | 3.2770 |
55
+ | 3.2678 | 2.0 | 10 | 3.1435 |
56
+ | 3.1418 | 3.0 | 15 | 3.0416 |
57
+ | 3.0482 | 4.0 | 20 | 2.9683 |
58
+ | 2.9771 | 5.0 | 25 | 2.9152 |
59
+ | 2.9285 | 6.0 | 30 | 2.8812 |
60
+ | 2.8971 | 7.0 | 35 | 2.8638 |
61
+ | 2.884 | 8.0 | 40 | 2.8582 |
62
 
63
 
64
  ### Framework versions
adapter_model.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:1f783db783113281c507bd79b87f05ed4f94fde52397f4b9b91d9aa5c5af0c82
3
  size 9189792
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:3aafd95c9f7491cf2794e305fee1aec53bf5cd1ae465cfa86f679ba74be4024a
3
  size 9189792
runs/Jun25_12-31-59_twon/events.out.tfevents.1750854735.twon.442856.0 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:17c8a79a666daa827b19cd39d27c339e1b4cb8fbf2e32b2f5279b65e5475dae0
3
+ size 6047
runs/Jun25_12-37-18_twon/events.out.tfevents.1750855050.twon.443781.0 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:f02f7ee3c05b2f5dddad825520bd38929f30193ca18578f0e7a1549e2c1f1f05
3
+ size 9972
runs/Jun25_12-37-18_twon/events.out.tfevents.1750855174.twon.443781.1 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:bb8807b1d0c5ceafa8522a45118e6bcb5c38a2c1e166360565db0f6276870d3b
3
+ size 354
test.bleu.json CHANGED
@@ -2,109 +2,109 @@
2
  {
3
  "model":"adapter",
4
  "metric":"1-gram",
5
- "mean":0.1967213115,
6
- "std":null
7
  },
8
  {
9
  "model":"adapter",
10
  "metric":"2-gram",
11
- "mean":0.0178571429,
12
- "std":null
13
  },
14
  {
15
  "model":"adapter",
16
  "metric":"3-gram",
17
- "mean":0.0196078431,
18
- "std":null
19
  },
20
  {
21
  "model":"adapter",
22
  "metric":"4-gram",
23
- "mean":0.0217391304,
24
- "std":null
25
  },
26
  {
27
  "model":"adapter",
28
  "metric":"bleu",
29
- "mean":0.0201823568,
30
- "std":null
31
  },
32
  {
33
  "model":"adapter",
34
  "metric":"brevity_penalty",
35
- "mean":0.5769498104,
36
- "std":null
37
  },
38
  {
39
  "model":"adapter",
40
  "metric":"length_ratio",
41
- "mean":0.6451612903,
42
- "std":null
43
  },
44
  {
45
  "model":"adapter",
46
  "metric":"reference_length",
47
- "mean":93.0,
48
- "std":null
49
  },
50
  {
51
  "model":"adapter",
52
  "metric":"translation_length",
53
- "mean":60.0,
54
- "std":null
55
  },
56
  {
57
  "model":"base",
58
  "metric":"1-gram",
59
- "mean":0.0901960784,
60
- "std":null
61
  },
62
  {
63
  "model":"base",
64
  "metric":"2-gram",
65
- "mean":0.004,
66
- "std":null
67
  },
68
  {
69
  "model":"base",
70
  "metric":"3-gram",
71
- "mean":0.0040816327,
72
- "std":null
73
  },
74
  {
75
  "model":"base",
76
  "metric":"4-gram",
77
- "mean":0.0041666667,
78
- "std":null
79
  },
80
  {
81
  "model":"base",
82
  "metric":"bleu",
83
- "mean":0.0088504954,
84
- "std":null
85
  },
86
  {
87
  "model":"base",
88
  "metric":"brevity_penalty",
89
  "mean":1.0,
90
- "std":null
91
  },
92
  {
93
  "model":"base",
94
  "metric":"length_ratio",
95
- "mean":2.7311827957,
96
- "std":null
97
  },
98
  {
99
  "model":"base",
100
  "metric":"reference_length",
101
- "mean":93.0,
102
- "std":null
103
  },
104
  {
105
  "model":"base",
106
  "metric":"translation_length",
107
- "mean":254.0,
108
- "std":null
109
  }
110
  ]
 
2
  {
3
  "model":"adapter",
4
  "metric":"1-gram",
5
+ "mean":0.1626490066,
6
+ "std":0.0603148036
7
  },
8
  {
9
  "model":"adapter",
10
  "metric":"2-gram",
11
+ "mean":0.0132151491,
12
+ "std":0.0103701397
13
  },
14
  {
15
  "model":"adapter",
16
  "metric":"3-gram",
17
+ "mean":0.0065764023,
18
+ "std":0.0007294461
19
  },
20
  {
21
  "model":"adapter",
22
  "metric":"4-gram",
23
+ "mean":0.0068014706,
24
+ "std":0.0007798972
25
  },
26
  {
27
  "model":"adapter",
28
  "metric":"bleu",
29
+ "mean":0.0172222099,
30
+ "std":0.0062717745
31
  },
32
  {
33
  "model":"adapter",
34
  "metric":"brevity_penalty",
35
+ "mean":1.0,
36
+ "std":0.0
37
  },
38
  {
39
  "model":"adapter",
40
  "metric":"length_ratio",
41
+ "mean":1.1689199689,
42
+ "std":0.1696616737
43
  },
44
  {
45
  "model":"adapter",
46
  "metric":"reference_length",
47
+ "mean":139.0,
48
+ "std":5.6568542495
49
  },
50
  {
51
  "model":"adapter",
52
  "metric":"translation_length",
53
+ "mean":162.0,
54
+ "std":16.9705627485
55
  },
56
  {
57
  "model":"base",
58
  "metric":"1-gram",
59
+ "mean":0.1359168376,
60
+ "std":0.0132010601
61
  },
62
  {
63
  "model":"base",
64
  "metric":"2-gram",
65
+ "mean":0.0121779946,
66
+ "std":0.0009669568
67
  },
68
  {
69
  "model":"base",
70
  "metric":"3-gram",
71
+ "mean":0.0045925668,
72
+ "std":0.0018732573
73
  },
74
  {
75
  "model":"base",
76
  "metric":"4-gram",
77
+ "mean":0.0047099101,
78
+ "std":0.0019624348
79
  },
80
  {
81
  "model":"base",
82
  "metric":"bleu",
83
+ "mean":0.013598395,
84
+ "std":0.00292452
85
  },
86
  {
87
  "model":"base",
88
  "metric":"brevity_penalty",
89
  "mean":1.0,
90
+ "std":0.0
91
  },
92
  {
93
  "model":"base",
94
  "metric":"length_ratio",
95
+ "mean":1.7606578607,
96
+ "std":0.6252794541
97
  },
98
  {
99
  "model":"base",
100
  "metric":"reference_length",
101
+ "mean":139.0,
102
+ "std":5.6568542495
103
  },
104
  {
105
  "model":"base",
106
  "metric":"translation_length",
107
+ "mean":246.5,
108
+ "std":96.8736290226
109
  }
110
  ]
test.calc_semantic_distance.json CHANGED
@@ -2,13 +2,13 @@
2
  {
3
  "model":"adapter",
4
  "metric":"semantic_distance",
5
- "mean":2.4793167114,
6
- "std":null
7
  },
8
  {
9
  "model":"base",
10
  "metric":"semantic_distance",
11
- "mean":4.512901783,
12
- "std":null
13
  }
14
  ]
 
2
  {
3
  "model":"adapter",
4
  "metric":"semantic_distance",
5
+ "mean":2.7020903826,
6
+ "std":0.4905031685
7
  },
8
  {
9
  "model":"base",
10
  "metric":"semantic_distance",
11
+ "mean":4.0621213913,
12
+ "std":2.4974907446
13
  }
14
  ]
test.tweeteval_corr.json CHANGED
@@ -2,469 +2,469 @@
2
  {
3
  "model":"adapter",
4
  "metric":"results.emotions.anger",
5
- "mean":-0.5268230504,
6
- "std":null
7
  },
8
  {
9
  "model":"adapter",
10
  "metric":"results.emotions.anticipation",
11
- "mean":0.136352266,
12
- "std":null
13
  },
14
  {
15
  "model":"adapter",
16
  "metric":"results.emotions.disgust",
17
- "mean":-0.784915379,
18
- "std":null
19
  },
20
  {
21
  "model":"adapter",
22
  "metric":"results.emotions.fear",
23
- "mean":-0.8235635564,
24
- "std":null
25
  },
26
  {
27
  "model":"adapter",
28
  "metric":"results.emotions.joy",
29
- "mean":-0.691358143,
30
- "std":null
31
  },
32
  {
33
  "model":"adapter",
34
  "metric":"results.emotions.love",
35
- "mean":-0.6486570854,
36
- "std":null
37
  },
38
  {
39
  "model":"adapter",
40
  "metric":"results.emotions.optimism",
41
- "mean":-0.3272765906,
42
- "std":null
43
  },
44
  {
45
  "model":"adapter",
46
  "metric":"results.emotions.pessimism",
47
- "mean":0.1835588729,
48
- "std":null
49
  },
50
  {
51
  "model":"adapter",
52
  "metric":"results.emotions.sadness",
53
- "mean":-0.9702203861,
54
- "std":null
55
  },
56
  {
57
  "model":"adapter",
58
  "metric":"results.emotions.surprise",
59
- "mean":-0.070388995,
60
- "std":null
61
  },
62
  {
63
  "model":"adapter",
64
  "metric":"results.emotions.trust",
65
- "mean":-0.24713608,
66
- "std":null
67
  },
68
  {
69
  "model":"adapter",
70
  "metric":"results.hate.HATE",
71
- "mean":-0.7180069259,
72
- "std":null
73
  },
74
  {
75
  "model":"adapter",
76
  "metric":"results.hate.NOT-HATE",
77
- "mean":-0.8052264212,
78
- "std":null
79
  },
80
  {
81
  "model":"adapter",
82
  "metric":"results.irony.irony",
83
- "mean":0.4368692015,
84
- "std":null
85
  },
86
  {
87
  "model":"adapter",
88
  "metric":"results.irony.non_irony",
89
- "mean":0.7627584038,
90
- "std":null
91
  },
92
  {
93
  "model":"adapter",
94
  "metric":"results.offensive.non-offensive",
95
- "mean":0.3843241711,
96
- "std":null
97
  },
98
  {
99
  "model":"adapter",
100
  "metric":"results.offensive.offensive",
101
- "mean":0.4960829201,
102
- "std":null
103
  },
104
  {
105
  "model":"adapter",
106
  "metric":"results.sentiment.negative",
107
- "mean":-0.5483485782,
108
- "std":null
109
  },
110
  {
111
  "model":"adapter",
112
  "metric":"results.sentiment.neutral",
113
- "mean":-0.6908123357,
114
- "std":null
115
  },
116
  {
117
  "model":"adapter",
118
  "metric":"results.sentiment.positive",
119
- "mean":-0.4203216699,
120
- "std":null
121
  },
122
  {
123
  "model":"adapter",
124
  "metric":"results.topics.arts_&_culture",
125
- "mean":-0.2850008481,
126
- "std":null
127
  },
128
  {
129
  "model":"adapter",
130
  "metric":"results.topics.business_&_entrepreneurs",
131
- "mean":0.2184144566,
132
- "std":null
133
  },
134
  {
135
  "model":"adapter",
136
  "metric":"results.topics.celebrity_&_pop_culture",
137
- "mean":-0.3665800952,
138
- "std":null
139
  },
140
  {
141
  "model":"adapter",
142
  "metric":"results.topics.diaries_&_daily_life",
143
- "mean":0.9181839521,
144
- "std":null
145
  },
146
  {
147
  "model":"adapter",
148
  "metric":"results.topics.family",
149
- "mean":-0.512329636,
150
- "std":null
151
  },
152
  {
153
  "model":"adapter",
154
  "metric":"results.topics.fashion_&_style",
155
- "mean":-0.3612212626,
156
- "std":null
157
  },
158
  {
159
  "model":"adapter",
160
  "metric":"results.topics.film_tv_&_video",
161
- "mean":-0.3911099948,
162
- "std":null
163
  },
164
  {
165
  "model":"adapter",
166
  "metric":"results.topics.fitness_&_health",
167
- "mean":0.7116950861,
168
- "std":null
169
  },
170
  {
171
  "model":"adapter",
172
  "metric":"results.topics.food_&_dining",
173
- "mean":0.537486211,
174
- "std":null
175
  },
176
  {
177
  "model":"adapter",
178
  "metric":"results.topics.gaming",
179
- "mean":0.715438698,
180
- "std":null
181
  },
182
  {
183
  "model":"adapter",
184
  "metric":"results.topics.learning_&_educational",
185
- "mean":-0.2387290904,
186
- "std":null
187
  },
188
  {
189
  "model":"adapter",
190
  "metric":"results.topics.music",
191
- "mean":-0.5180386816,
192
- "std":null
193
  },
194
  {
195
  "model":"adapter",
196
  "metric":"results.topics.news_&_social_concern",
197
- "mean":0.0786748779,
198
- "std":null
199
  },
200
  {
201
  "model":"adapter",
202
  "metric":"results.topics.other_hobbies",
203
- "mean":0.7760161899,
204
- "std":null
205
  },
206
  {
207
  "model":"adapter",
208
  "metric":"results.topics.relationships",
209
- "mean":0.5919301099,
210
- "std":null
211
  },
212
  {
213
  "model":"adapter",
214
  "metric":"results.topics.science_&_technology",
215
- "mean":0.8093006883,
216
- "std":null
217
  },
218
  {
219
  "model":"adapter",
220
  "metric":"results.topics.sports",
221
- "mean":0.9973952131,
222
- "std":null
223
  },
224
  {
225
  "model":"adapter",
226
  "metric":"results.topics.travel_&_adventure",
227
- "mean":-0.907770534,
228
- "std":null
229
  },
230
  {
231
  "model":"adapter",
232
  "metric":"results.topics.youth_&_student_life",
233
- "mean":0.7013296485,
234
- "std":null
235
  },
236
  {
237
  "model":"base",
238
  "metric":"results.emotions.anger",
239
- "mean":0.3948842053,
240
- "std":null
241
  },
242
  {
243
  "model":"base",
244
  "metric":"results.emotions.anticipation",
245
- "mean":0.0609444262,
246
- "std":null
247
  },
248
  {
249
  "model":"base",
250
  "metric":"results.emotions.disgust",
251
- "mean":0.7081311043,
252
- "std":null
253
  },
254
  {
255
  "model":"base",
256
  "metric":"results.emotions.fear",
257
- "mean":-0.5757331951,
258
- "std":null
259
  },
260
  {
261
  "model":"base",
262
  "metric":"results.emotions.joy",
263
- "mean":0.8526462202,
264
- "std":null
265
  },
266
  {
267
  "model":"base",
268
  "metric":"results.emotions.love",
269
- "mean":0.2730326751,
270
- "std":null
271
  },
272
  {
273
  "model":"base",
274
  "metric":"results.emotions.optimism",
275
- "mean":-0.0047403378,
276
- "std":null
277
  },
278
  {
279
  "model":"base",
280
  "metric":"results.emotions.pessimism",
281
- "mean":0.9436862612,
282
- "std":null
283
  },
284
  {
285
  "model":"base",
286
  "metric":"results.emotions.sadness",
287
- "mean":0.0733659606,
288
- "std":null
289
  },
290
  {
291
  "model":"base",
292
  "metric":"results.emotions.surprise",
293
- "mean":-0.0047165292,
294
- "std":null
295
  },
296
  {
297
  "model":"base",
298
  "metric":"results.emotions.trust",
299
- "mean":0.0035559544,
300
- "std":null
301
  },
302
  {
303
  "model":"base",
304
  "metric":"results.hate.HATE",
305
- "mean":-0.0753961261,
306
- "std":null
307
  },
308
  {
309
  "model":"base",
310
  "metric":"results.hate.NOT-HATE",
311
- "mean":-0.0086332772,
312
- "std":null
313
  },
314
  {
315
  "model":"base",
316
  "metric":"results.irony.irony",
317
- "mean":-0.5451346785,
318
- "std":null
319
  },
320
  {
321
  "model":"base",
322
  "metric":"results.irony.non_irony",
323
- "mean":-0.6431540259,
324
- "std":null
325
  },
326
  {
327
  "model":"base",
328
  "metric":"results.offensive.non-offensive",
329
- "mean":-0.4269971018,
330
- "std":null
331
  },
332
  {
333
  "model":"base",
334
  "metric":"results.offensive.offensive",
335
- "mean":-0.434692032,
336
- "std":null
337
  },
338
  {
339
  "model":"base",
340
  "metric":"results.sentiment.negative",
341
- "mean":-0.3852661674,
342
- "std":null
343
  },
344
  {
345
  "model":"base",
346
  "metric":"results.sentiment.neutral",
347
- "mean":0.6297353835,
348
- "std":null
349
  },
350
  {
351
  "model":"base",
352
  "metric":"results.sentiment.positive",
353
- "mean":-0.5743319451,
354
- "std":null
355
  },
356
  {
357
  "model":"base",
358
  "metric":"results.topics.arts_&_culture",
359
- "mean":-0.0565709636,
360
- "std":null
361
  },
362
  {
363
  "model":"base",
364
  "metric":"results.topics.business_&_entrepreneurs",
365
- "mean":0.9255617665,
366
- "std":null
367
  },
368
  {
369
  "model":"base",
370
  "metric":"results.topics.celebrity_&_pop_culture",
371
- "mean":0.8184318427,
372
- "std":null
373
  },
374
  {
375
  "model":"base",
376
  "metric":"results.topics.diaries_&_daily_life",
377
- "mean":-0.2536250455,
378
- "std":null
379
  },
380
  {
381
  "model":"base",
382
  "metric":"results.topics.family",
383
- "mean":0.0272948072,
384
- "std":null
385
  },
386
  {
387
  "model":"base",
388
  "metric":"results.topics.fashion_&_style",
389
- "mean":-0.8452099115,
390
- "std":null
391
  },
392
  {
393
  "model":"base",
394
  "metric":"results.topics.film_tv_&_video",
395
- "mean":-0.2549525606,
396
- "std":null
397
  },
398
  {
399
  "model":"base",
400
  "metric":"results.topics.fitness_&_health",
401
- "mean":-0.3832844614,
402
- "std":null
403
  },
404
  {
405
  "model":"base",
406
  "metric":"results.topics.food_&_dining",
407
- "mean":-0.0963838107,
408
- "std":null
409
  },
410
  {
411
  "model":"base",
412
  "metric":"results.topics.gaming",
413
- "mean":0.5778929475,
414
- "std":null
415
  },
416
  {
417
  "model":"base",
418
  "metric":"results.topics.learning_&_educational",
419
- "mean":0.1090421307,
420
- "std":null
421
  },
422
  {
423
  "model":"base",
424
  "metric":"results.topics.music",
425
- "mean":-0.0073864199,
426
- "std":null
427
  },
428
  {
429
  "model":"base",
430
  "metric":"results.topics.news_&_social_concern",
431
- "mean":0.0808063377,
432
- "std":null
433
  },
434
  {
435
  "model":"base",
436
  "metric":"results.topics.other_hobbies",
437
- "mean":-0.123554983,
438
- "std":null
439
  },
440
  {
441
  "model":"base",
442
  "metric":"results.topics.relationships",
443
- "mean":-0.4367711054,
444
- "std":null
445
  },
446
  {
447
  "model":"base",
448
  "metric":"results.topics.science_&_technology",
449
- "mean":0.7922300884,
450
- "std":null
451
  },
452
  {
453
  "model":"base",
454
  "metric":"results.topics.sports",
455
- "mean":0.9998696917,
456
- "std":null
457
  },
458
  {
459
  "model":"base",
460
  "metric":"results.topics.travel_&_adventure",
461
- "mean":0.5795364891,
462
- "std":null
463
  },
464
  {
465
  "model":"base",
466
  "metric":"results.topics.youth_&_student_life",
467
- "mean":0.8595356149,
468
- "std":null
469
  }
470
  ]
 
2
  {
3
  "model":"adapter",
4
  "metric":"results.emotions.anger",
5
+ "mean":-0.2634782801,
6
+ "std":0.231464387
7
  },
8
  {
9
  "model":"adapter",
10
  "metric":"results.emotions.anticipation",
11
+ "mean":0.0158213843,
12
+ "std":0.7282000721
13
  },
14
  {
15
  "model":"adapter",
16
  "metric":"results.emotions.disgust",
17
+ "mean":-0.1954270493,
18
+ "std":0.2348718284
19
  },
20
  {
21
  "model":"adapter",
22
  "metric":"results.emotions.fear",
23
+ "mean":0.360446438,
24
+ "std":0.3989076837
25
  },
26
  {
27
  "model":"adapter",
28
  "metric":"results.emotions.joy",
29
+ "mean":0.452002472,
30
+ "std":0.0691351988
31
  },
32
  {
33
  "model":"adapter",
34
  "metric":"results.emotions.love",
35
+ "mean":0.3477192322,
36
+ "std":0.2212280497
37
  },
38
  {
39
  "model":"adapter",
40
  "metric":"results.emotions.optimism",
41
+ "mean":0.1995318999,
42
+ "std":0.4214880877
43
  },
44
  {
45
  "model":"adapter",
46
  "metric":"results.emotions.pessimism",
47
+ "mean":0.1422411741,
48
+ "std":0.3278830476
49
  },
50
  {
51
  "model":"adapter",
52
  "metric":"results.emotions.sadness",
53
+ "mean":-0.071874181,
54
+ "std":0.3405189847
55
  },
56
  {
57
  "model":"adapter",
58
  "metric":"results.emotions.surprise",
59
+ "mean":-0.2988987546,
60
+ "std":0.0584183935
61
  },
62
  {
63
  "model":"adapter",
64
  "metric":"results.emotions.trust",
65
+ "mean":0.1972516818,
66
+ "std":0.8360275703
67
  },
68
  {
69
  "model":"adapter",
70
  "metric":"results.hate.HATE",
71
+ "mean":0.7693368432,
72
+ "std":0.1325322234
73
  },
74
  {
75
  "model":"adapter",
76
  "metric":"results.hate.NOT-HATE",
77
+ "mean":0.6934967639,
78
+ "std":0.0708291026
79
  },
80
  {
81
  "model":"adapter",
82
  "metric":"results.irony.irony",
83
+ "mean":-0.408708591,
84
+ "std":0.4524896518
85
  },
86
  {
87
  "model":"adapter",
88
  "metric":"results.irony.non_irony",
89
+ "mean":-0.3346725432,
90
+ "std":0.3758579324
91
  },
92
  {
93
  "model":"adapter",
94
  "metric":"results.offensive.non-offensive",
95
+ "mean":0.6837666214,
96
+ "std":0.1490585857
97
  },
98
  {
99
  "model":"adapter",
100
  "metric":"results.offensive.offensive",
101
+ "mean":0.5454805613,
102
+ "std":0.1660513222
103
  },
104
  {
105
  "model":"adapter",
106
  "metric":"results.sentiment.negative",
107
+ "mean":-0.0985854479,
108
+ "std":0.0661983998
109
  },
110
  {
111
  "model":"adapter",
112
  "metric":"results.sentiment.neutral",
113
+ "mean":0.4411701438,
114
+ "std":0.2714188042
115
  },
116
  {
117
  "model":"adapter",
118
  "metric":"results.sentiment.positive",
119
+ "mean":0.2591377753,
120
+ "std":0.9515331155
121
  },
122
  {
123
  "model":"adapter",
124
  "metric":"results.topics.arts_&_culture",
125
+ "mean":0.771723217,
126
+ "std":0.1209096237
127
  },
128
  {
129
  "model":"adapter",
130
  "metric":"results.topics.business_&_entrepreneurs",
131
+ "mean":-0.3745750785,
132
+ "std":0.4921254989
133
  },
134
  {
135
  "model":"adapter",
136
  "metric":"results.topics.celebrity_&_pop_culture",
137
+ "mean":0.580032456,
138
+ "std":0.2564609314
139
  },
140
  {
141
  "model":"adapter",
142
  "metric":"results.topics.diaries_&_daily_life",
143
+ "mean":-0.0215109424,
144
+ "std":1.0244196565
145
  },
146
  {
147
  "model":"adapter",
148
  "metric":"results.topics.family",
149
+ "mean":0.3351652866,
150
+ "std":0.1194140217
151
  },
152
  {
153
  "model":"adapter",
154
  "metric":"results.topics.fashion_&_style",
155
+ "mean":-0.0419560101,
156
+ "std":0.7778513722
157
  },
158
  {
159
  "model":"adapter",
160
  "metric":"results.topics.film_tv_&_video",
161
+ "mean":0.2693774699,
162
+ "std":0.7733902516
163
  },
164
  {
165
  "model":"adapter",
166
  "metric":"results.topics.fitness_&_health",
167
+ "mean":0.0801936849,
168
+ "std":0.3627242271
169
  },
170
  {
171
  "model":"adapter",
172
  "metric":"results.topics.food_&_dining",
173
+ "mean":-0.3229975875,
174
+ "std":0.8888187017
175
  },
176
  {
177
  "model":"adapter",
178
  "metric":"results.topics.gaming",
179
+ "mean":0.1766638513,
180
+ "std":1.0331581003
181
  },
182
  {
183
  "model":"adapter",
184
  "metric":"results.topics.learning_&_educational",
185
+ "mean":0.0384401537,
186
+ "std":0.3507022848
187
  },
188
  {
189
  "model":"adapter",
190
  "metric":"results.topics.music",
191
+ "mean":0.2304459916,
192
+ "std":0.6272646145
193
  },
194
  {
195
  "model":"adapter",
196
  "metric":"results.topics.news_&_social_concern",
197
+ "mean":-0.0469632613,
198
+ "std":0.6048732353
199
  },
200
  {
201
  "model":"adapter",
202
  "metric":"results.topics.other_hobbies",
203
+ "mean":0.2681204647,
204
+ "std":0.7218519746
205
  },
206
  {
207
  "model":"adapter",
208
  "metric":"results.topics.relationships",
209
+ "mean":-0.0346842448,
210
+ "std":0.7513717815
211
  },
212
  {
213
  "model":"adapter",
214
  "metric":"results.topics.science_&_technology",
215
+ "mean":-0.2324337056,
216
+ "std":0.4541593535
217
  },
218
  {
219
  "model":"adapter",
220
  "metric":"results.topics.sports",
221
+ "mean":0.7623055896,
222
+ "std":0.0911501257
223
  },
224
  {
225
  "model":"adapter",
226
  "metric":"results.topics.travel_&_adventure",
227
+ "mean":-0.0859171174,
228
+ "std":0.6724467808
229
  },
230
  {
231
  "model":"adapter",
232
  "metric":"results.topics.youth_&_student_life",
233
+ "mean":0.5490275795,
234
+ "std":0.1476321641
235
  },
236
  {
237
  "model":"base",
238
  "metric":"results.emotions.anger",
239
+ "mean":-0.2393579051,
240
+ "std":0.023630144
241
  },
242
  {
243
  "model":"base",
244
  "metric":"results.emotions.anticipation",
245
+ "mean":-0.2485840928,
246
+ "std":0.5616452868
247
  },
248
  {
249
  "model":"base",
250
  "metric":"results.emotions.disgust",
251
+ "mean":-0.3651939343,
252
+ "std":0.0493475415
253
  },
254
  {
255
  "model":"base",
256
  "metric":"results.emotions.fear",
257
+ "mean":-0.3302927018,
258
+ "std":0.3822707369
259
  },
260
  {
261
  "model":"base",
262
  "metric":"results.emotions.joy",
263
+ "mean":-0.5810791542,
264
+ "std":0.3846805332
265
  },
266
  {
267
  "model":"base",
268
  "metric":"results.emotions.love",
269
+ "mean":-0.5647878827,
270
+ "std":0.2811753843
271
  },
272
  {
273
  "model":"base",
274
  "metric":"results.emotions.optimism",
275
+ "mean":0.2410828358,
276
+ "std":0.7299098559
277
  },
278
  {
279
  "model":"base",
280
  "metric":"results.emotions.pessimism",
281
+ "mean":-0.4917788409,
282
+ "std":0.2287283009
283
  },
284
  {
285
  "model":"base",
286
  "metric":"results.emotions.sadness",
287
+ "mean":-0.5704090472,
288
+ "std":0.0634712245
289
  },
290
  {
291
  "model":"base",
292
  "metric":"results.emotions.surprise",
293
+ "mean":0.0063650273,
294
+ "std":0.0689701594
295
  },
296
  {
297
  "model":"base",
298
  "metric":"results.emotions.trust",
299
+ "mean":-0.1072807721,
300
+ "std":0.2421654136
301
  },
302
  {
303
  "model":"base",
304
  "metric":"results.hate.HATE",
305
+ "mean":0.5510144592,
306
+ "std":0.2694872625
307
  },
308
  {
309
  "model":"base",
310
  "metric":"results.hate.NOT-HATE",
311
+ "mean":0.5237950237,
312
+ "std":0.1433796284
313
  },
314
  {
315
  "model":"base",
316
  "metric":"results.irony.irony",
317
+ "mean":0.4234751291,
318
+ "std":0.4284324793
319
  },
320
  {
321
  "model":"base",
322
  "metric":"results.irony.non_irony",
323
+ "mean":0.7084472491,
324
+ "std":0.0316268926
325
  },
326
  {
327
  "model":"base",
328
  "metric":"results.offensive.non-offensive",
329
+ "mean":0.6389777355,
330
+ "std":0.0075294264
331
  },
332
  {
333
  "model":"base",
334
  "metric":"results.offensive.offensive",
335
+ "mean":0.3525055169,
336
+ "std":0.2163360288
337
  },
338
  {
339
  "model":"base",
340
  "metric":"results.sentiment.negative",
341
+ "mean":0.2720692423,
342
+ "std":0.2242873033
343
  },
344
  {
345
  "model":"base",
346
  "metric":"results.sentiment.neutral",
347
+ "mean":0.0517201238,
348
+ "std":0.2067528262
349
  },
350
  {
351
  "model":"base",
352
  "metric":"results.sentiment.positive",
353
+ "mean":0.2668476854,
354
+ "std":0.5360949201
355
  },
356
  {
357
  "model":"base",
358
  "metric":"results.topics.arts_&_culture",
359
+ "mean":0.3647038471,
360
+ "std":0.6824471237
361
  },
362
  {
363
  "model":"base",
364
  "metric":"results.topics.business_&_entrepreneurs",
365
+ "mean":0.647623194,
366
+ "std":0.185803947
367
  },
368
  {
369
  "model":"base",
370
  "metric":"results.topics.celebrity_&_pop_culture",
371
+ "mean":-0.2452197218,
372
+ "std":0.2939242865
373
  },
374
  {
375
  "model":"base",
376
  "metric":"results.topics.diaries_&_daily_life",
377
+ "mean":0.4238200489,
378
+ "std":0.3900375399
379
  },
380
  {
381
  "model":"base",
382
  "metric":"results.topics.family",
383
+ "mean":0.1575945482,
384
+ "std":0.3913791778
385
  },
386
  {
387
  "model":"base",
388
  "metric":"results.topics.fashion_&_style",
389
+ "mean":0.1995624794,
390
+ "std":0.1044695892
391
  },
392
  {
393
  "model":"base",
394
  "metric":"results.topics.film_tv_&_video",
395
+ "mean":-0.0647391818,
396
+ "std":0.6628295573
397
  },
398
  {
399
  "model":"base",
400
  "metric":"results.topics.fitness_&_health",
401
+ "mean":0.1389187054,
402
+ "std":0.4902649188
403
  },
404
  {
405
  "model":"base",
406
  "metric":"results.topics.food_&_dining",
407
+ "mean":0.553381089,
408
+ "std":0.228152023
409
  },
410
  {
411
  "model":"base",
412
  "metric":"results.topics.gaming",
413
+ "mean":0.3978395165,
414
+ "std":0.6079301593
415
  },
416
  {
417
  "model":"base",
418
  "metric":"results.topics.learning_&_educational",
419
+ "mean":0.6062929768,
420
+ "std":0.2607694925
421
  },
422
  {
423
  "model":"base",
424
  "metric":"results.topics.music",
425
+ "mean":0.1091611032,
426
+ "std":0.548908049
427
  },
428
  {
429
  "model":"base",
430
  "metric":"results.topics.news_&_social_concern",
431
+ "mean":0.0880267301,
432
+ "std":0.2368066687
433
  },
434
  {
435
  "model":"base",
436
  "metric":"results.topics.other_hobbies",
437
+ "mean":0.2516230179,
438
+ "std":0.2954611949
439
  },
440
  {
441
  "model":"base",
442
  "metric":"results.topics.relationships",
443
+ "mean":-0.1903590952,
444
+ "std":0.0953118287
445
  },
446
  {
447
  "model":"base",
448
  "metric":"results.topics.science_&_technology",
449
+ "mean":0.5132086365,
450
+ "std":0.0568449013
451
  },
452
  {
453
  "model":"base",
454
  "metric":"results.topics.sports",
455
+ "mean":-0.3543906543,
456
+ "std":0.3185806976
457
  },
458
  {
459
  "model":"base",
460
  "metric":"results.topics.travel_&_adventure",
461
+ "mean":0.6841729534,
462
+ "std":0.2047344569
463
  },
464
  {
465
  "model":"base",
466
  "metric":"results.topics.youth_&_student_life",
467
+ "mean":0.0780676124,
468
+ "std":0.1020736338
469
  }
470
  ]
training_args.bin CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:3573d5b83134ea6f5c79a2ab577cfe1571a2bf5ac952cdd7147f6472582e0b33
3
  size 5688
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:db13484329cb0d06275787f4f249428ca1791342391e39d078f22f8292d8c1e6
3
  size 5688