xdolez52 commited on
Commit
67cfeac
·
verified ·
1 Parent(s): a31ae04

Upload tournament.json with huggingface_hub

Browse files
Files changed (1) hide show
  1. tournament.json +602 -0
tournament.json CHANGED
@@ -1,5 +1,607 @@
1
  {
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
2
  "CohereF_aya-23-_78291f9_91bd234": {
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
3
  "CohereF_aya-23-_78291f9_91bd234": {
4
  "benczechmark_agree": false,
5
  "benczechmark_belebele": false,
 
1
  {
2
+ "CohereF_aya-23-_696092a_369b1ca": {
3
+ "CohereF_aya-23-_696092a_369b1ca": {
4
+ "benczechmark_agree": false,
5
+ "benczechmark_belebele": false,
6
+ "benczechmark_capek": false,
7
+ "benczechmark_cermat_czech_mc": false,
8
+ "benczechmark_cermat_czech_open": false,
9
+ "benczechmark_cermat_czech_tf": false,
10
+ "benczechmark_cermat_czmath_mc": false,
11
+ "benczechmark_cermat_czmath_open": false,
12
+ "benczechmark_correspondence": false,
13
+ "benczechmark_cs_court_decisions_ner": false,
14
+ "benczechmark_cs_naturalquestions": false,
15
+ "benczechmark_cs_ner": false,
16
+ "benczechmark_cs_sqad32": false,
17
+ "benczechmark_cs_triviaQA": false,
18
+ "benczechmark_csfever_nli": false,
19
+ "benczechmark_ctkfacts_nli": false,
20
+ "benczechmark_czechnews": false,
21
+ "benczechmark_dialect": false,
22
+ "benczechmark_essay": false,
23
+ "benczechmark_fiction": false,
24
+ "benczechmark_grammarerrorcorrection": false,
25
+ "benczechmark_havlicek": false,
26
+ "benczechmark_hellaswag": false,
27
+ "benczechmark_histcorpus": false,
28
+ "benczechmark_history_ir": false,
29
+ "benczechmark_klokan_qa": false,
30
+ "benczechmark_propaganda_argumentace": false,
31
+ "benczechmark_propaganda_demonizace": false,
32
+ "benczechmark_propaganda_emoce": false,
33
+ "benczechmark_propaganda_fabulace": false,
34
+ "benczechmark_propaganda_lokace": false,
35
+ "benczechmark_propaganda_nalepkovani": false,
36
+ "benczechmark_propaganda_nazor": false,
37
+ "benczechmark_propaganda_relativizace": false,
38
+ "benczechmark_propaganda_rusko": false,
39
+ "benczechmark_propaganda_strach": false,
40
+ "benczechmark_propaganda_vina": false,
41
+ "benczechmark_propaganda_zamereni": false,
42
+ "benczechmark_propaganda_zanr": false,
43
+ "benczechmark_sentiment_csfd": false,
44
+ "benczechmark_sentiment_fb": false,
45
+ "benczechmark_sentiment_mall": false,
46
+ "benczechmark_snli": false,
47
+ "benczechmark_speeches": false,
48
+ "benczechmark_spoken": false,
49
+ "benczechmark_subjectivity": false,
50
+ "benczechmark_summarization": false,
51
+ "benczechmark_umimeto_biology": false,
52
+ "benczechmark_umimeto_chemistry": false,
53
+ "benczechmark_umimeto_czech": false,
54
+ "benczechmark_umimeto_history": false,
55
+ "benczechmark_umimeto_informatics": false,
56
+ "benczechmark_umimeto_math": false,
57
+ "benczechmark_umimeto_physics": false
58
+ },
59
+ "CohereF_aya-23-_78291f9_91bd234": {
60
+ "benczechmark_agree": {
61
+ "delta": 0.09932334230071149,
62
+ "p_value": 0.0,
63
+ "significant": true
64
+ },
65
+ "benczechmark_belebele": {
66
+ "delta": 0.01003970490614492,
67
+ "p_value": 0.0,
68
+ "significant": true
69
+ },
70
+ "benczechmark_capek": {
71
+ "delta": 145.99803456352325,
72
+ "p_value": 0.0,
73
+ "significant": true
74
+ },
75
+ "benczechmark_cermat_czech_mc": {
76
+ "delta": 0.1386748844375963,
77
+ "p_value": 7.826110560855631e-10,
78
+ "significant": true
79
+ },
80
+ "benczechmark_cermat_czech_open": {
81
+ "delta": 0.10610932475884244,
82
+ "p_value": 7.341590281192024e-08,
83
+ "significant": true
84
+ },
85
+ "benczechmark_cermat_czech_tf": {
86
+ "delta": 0.103153836254262,
87
+ "p_value": 0.0,
88
+ "significant": true
89
+ },
90
+ "benczechmark_cermat_czmath_mc": {
91
+ "delta": -0.07936507936507936,
92
+ "p_value": 1,
93
+ "significant": false
94
+ },
95
+ "benczechmark_cermat_czmath_open": {
96
+ "delta": 0.025157232704402514,
97
+ "p_value": 0.010342067855694326,
98
+ "significant": true
99
+ },
100
+ "benczechmark_correspondence": {
101
+ "delta": 354.52784074181886,
102
+ "p_value": 0.0,
103
+ "significant": true
104
+ },
105
+ "benczechmark_cs_court_decisions_ner": {
106
+ "delta": 0.16877308963639892,
107
+ "p_value": 3.735308249911242e-108,
108
+ "significant": true
109
+ },
110
+ "benczechmark_cs_naturalquestions": {
111
+ "delta": 0.044044321329639896,
112
+ "p_value": 2.4506839976957746e-12,
113
+ "significant": true
114
+ },
115
+ "benczechmark_cs_ner": {
116
+ "delta": 0.048814120242691605,
117
+ "p_value": 1.256168703135736e-24,
118
+ "significant": true
119
+ },
120
+ "benczechmark_cs_sqad32": {
121
+ "delta": 0.0,
122
+ "p_value": 0.5,
123
+ "significant": false
124
+ },
125
+ "benczechmark_cs_triviaQA": {
126
+ "delta": 0.11835426128773713,
127
+ "p_value": 6.353727722512506e-161,
128
+ "significant": true
129
+ },
130
+ "benczechmark_csfever_nli": {
131
+ "delta": 0.042112138610123195,
132
+ "p_value": 0.0,
133
+ "significant": true
134
+ },
135
+ "benczechmark_ctkfacts_nli": {
136
+ "delta": 0.04442219451754714,
137
+ "p_value": 0.0,
138
+ "significant": true
139
+ },
140
+ "benczechmark_czechnews": {
141
+ "delta": 0.05682547133941318,
142
+ "p_value": 0.0,
143
+ "significant": true
144
+ },
145
+ "benczechmark_dialect": {
146
+ "delta": 31.26411167278839,
147
+ "p_value": 0.0,
148
+ "significant": true
149
+ },
150
+ "benczechmark_essay": {
151
+ "delta": 50.96153218166829,
152
+ "p_value": 0.0,
153
+ "significant": true
154
+ },
155
+ "benczechmark_fiction": {
156
+ "delta": 93.66096815803243,
157
+ "p_value": 0.0,
158
+ "significant": true
159
+ },
160
+ "benczechmark_grammarerrorcorrection": {
161
+ "delta": 0.10454155955441302,
162
+ "p_value": 4.4321835576703555e-33,
163
+ "significant": true
164
+ },
165
+ "benczechmark_havlicek": {
166
+ "delta": 2642.263507830858,
167
+ "p_value": 0.0,
168
+ "significant": true
169
+ },
170
+ "benczechmark_hellaswag": {
171
+ "delta": 0.08046205935072692,
172
+ "p_value": 1.0106674884003879e-99,
173
+ "significant": true
174
+ },
175
+ "benczechmark_histcorpus": {
176
+ "delta": 533.897540517744,
177
+ "p_value": 0.0,
178
+ "significant": true
179
+ },
180
+ "benczechmark_history_ir": {
181
+ "delta": 0.05506216696269983,
182
+ "p_value": 0.0004816202914721512,
183
+ "significant": true
184
+ },
185
+ "benczechmark_klokan_qa": {
186
+ "delta": 0.018587360594795543,
187
+ "p_value": 0.1475434188516723,
188
+ "significant": false
189
+ },
190
+ "benczechmark_propaganda_argumentace": {
191
+ "delta": -0.0030850142969079952,
192
+ "p_value": 1,
193
+ "significant": false
194
+ },
195
+ "benczechmark_propaganda_demonizace": {
196
+ "delta": -0.009153379026243802,
197
+ "p_value": 1,
198
+ "significant": false
199
+ },
200
+ "benczechmark_propaganda_emoce": {
201
+ "delta": 0.03464847620627465,
202
+ "p_value": 0.259409,
203
+ "significant": false
204
+ },
205
+ "benczechmark_propaganda_fabulace": {
206
+ "delta": 0.0016047845787706683,
207
+ "p_value": 0.293885,
208
+ "significant": false
209
+ },
210
+ "benczechmark_propaganda_lokace": {
211
+ "delta": 0.0313878001968273,
212
+ "p_value": 0.027409,
213
+ "significant": true
214
+ },
215
+ "benczechmark_propaganda_nalepkovani": {
216
+ "delta": -0.008238467261904692,
217
+ "p_value": 1,
218
+ "significant": false
219
+ },
220
+ "benczechmark_propaganda_nazor": {
221
+ "delta": -0.03192019841992533,
222
+ "p_value": 1,
223
+ "significant": false
224
+ },
225
+ "benczechmark_propaganda_relativizace": {
226
+ "delta": -0.03643804771178183,
227
+ "p_value": 1,
228
+ "significant": false
229
+ },
230
+ "benczechmark_propaganda_rusko": {
231
+ "delta": 0.018006584693525785,
232
+ "p_value": 0.364675,
233
+ "significant": false
234
+ },
235
+ "benczechmark_propaganda_strach": {
236
+ "delta": -0.03086846846846847,
237
+ "p_value": 1,
238
+ "significant": false
239
+ },
240
+ "benczechmark_propaganda_vina": {
241
+ "delta": -0.0480361610018426,
242
+ "p_value": 1,
243
+ "significant": false
244
+ },
245
+ "benczechmark_propaganda_zamereni": {
246
+ "delta": 0.1702775685245309,
247
+ "p_value": 0.0,
248
+ "significant": true
249
+ },
250
+ "benczechmark_propaganda_zanr": {
251
+ "delta": 0.052606761801089474,
252
+ "p_value": 0.0,
253
+ "significant": true
254
+ },
255
+ "benczechmark_sentiment_csfd": {
256
+ "delta": 0.002391555235682019,
257
+ "p_value": 0.000254,
258
+ "significant": true
259
+ },
260
+ "benczechmark_sentiment_fb": {
261
+ "delta": 0.00980239057576704,
262
+ "p_value": 0.0,
263
+ "significant": true
264
+ },
265
+ "benczechmark_sentiment_mall": {
266
+ "delta": 0.005247331884877027,
267
+ "p_value": 0.0,
268
+ "significant": true
269
+ },
270
+ "benczechmark_snli": {
271
+ "delta": 0.04450086863557401,
272
+ "p_value": 7.956791515529281e-23,
273
+ "significant": true
274
+ },
275
+ "benczechmark_speeches": {
276
+ "delta": 87.1958280629542,
277
+ "p_value": 0.0,
278
+ "significant": true
279
+ },
280
+ "benczechmark_spoken": {
281
+ "delta": 56.86880509541322,
282
+ "p_value": 0.0,
283
+ "significant": true
284
+ },
285
+ "benczechmark_subjectivity": {
286
+ "delta": -0.19795892443189245,
287
+ "p_value": 1,
288
+ "significant": false
289
+ },
290
+ "benczechmark_summarization": {
291
+ "delta": 0.017408604795130046,
292
+ "p_value": 0.0,
293
+ "significant": true
294
+ },
295
+ "benczechmark_umimeto_biology": {
296
+ "delta": 0.06999999999999995,
297
+ "p_value": 0.07263629795146974,
298
+ "significant": false
299
+ },
300
+ "benczechmark_umimeto_chemistry": {
301
+ "delta": 0.06000000000000005,
302
+ "p_value": 0.12061411335992354,
303
+ "significant": false
304
+ },
305
+ "benczechmark_umimeto_czech": {
306
+ "delta": 0.12,
307
+ "p_value": 0.01659962391514575,
308
+ "significant": true
309
+ },
310
+ "benczechmark_umimeto_history": {
311
+ "delta": 0.06999999999999995,
312
+ "p_value": 0.06362955087062457,
313
+ "significant": false
314
+ },
315
+ "benczechmark_umimeto_informatics": {
316
+ "delta": 0.07999999999999996,
317
+ "p_value": 0.03673774082234953,
318
+ "significant": true
319
+ },
320
+ "benczechmark_umimeto_math": {
321
+ "delta": 0.09999999999999998,
322
+ "p_value": 0.02467578814335497,
323
+ "significant": true
324
+ },
325
+ "benczechmark_umimeto_physics": {
326
+ "delta": 0.010000000000000009,
327
+ "p_value": 0.424254885710078,
328
+ "significant": false
329
+ }
330
+ }
331
+ },
332
  "CohereF_aya-23-_78291f9_91bd234": {
333
+ "CohereF_aya-23-_696092a_369b1ca": {
334
+ "benczechmark_agree": {
335
+ "delta": -0.1011368645567794,
336
+ "p_value": 1,
337
+ "significant": false
338
+ },
339
+ "benczechmark_belebele": {
340
+ "delta": -0.010194489328797185,
341
+ "p_value": 1,
342
+ "significant": false
343
+ },
344
+ "benczechmark_capek": {
345
+ "delta": -145.99803456352325,
346
+ "p_value": 1,
347
+ "significant": false
348
+ },
349
+ "benczechmark_cermat_czech_mc": {
350
+ "delta": -0.1386748844375963,
351
+ "p_value": 1,
352
+ "significant": false
353
+ },
354
+ "benczechmark_cermat_czech_open": {
355
+ "delta": -0.10610932475884244,
356
+ "p_value": 1,
357
+ "significant": false
358
+ },
359
+ "benczechmark_cermat_czech_tf": {
360
+ "delta": -0.10381633870005968,
361
+ "p_value": 1,
362
+ "significant": false
363
+ },
364
+ "benczechmark_cermat_czmath_mc": {
365
+ "delta": 0.07936507936507936,
366
+ "p_value": 0.043187570313197934,
367
+ "significant": true
368
+ },
369
+ "benczechmark_cermat_czmath_open": {
370
+ "delta": -0.025157232704402514,
371
+ "p_value": 1,
372
+ "significant": false
373
+ },
374
+ "benczechmark_correspondence": {
375
+ "delta": -354.52784074181886,
376
+ "p_value": 1,
377
+ "significant": false
378
+ },
379
+ "benczechmark_cs_court_decisions_ner": {
380
+ "delta": -0.16877308963639892,
381
+ "p_value": 1,
382
+ "significant": false
383
+ },
384
+ "benczechmark_cs_naturalquestions": {
385
+ "delta": -0.044044321329639896,
386
+ "p_value": 1,
387
+ "significant": false
388
+ },
389
+ "benczechmark_cs_ner": {
390
+ "delta": -0.048814120242691605,
391
+ "p_value": 1,
392
+ "significant": false
393
+ },
394
+ "benczechmark_cs_sqad32": {
395
+ "delta": 0.0,
396
+ "p_value": 0.5,
397
+ "significant": false
398
+ },
399
+ "benczechmark_cs_triviaQA": {
400
+ "delta": -0.11835426128773713,
401
+ "p_value": 1,
402
+ "significant": false
403
+ },
404
+ "benczechmark_csfever_nli": {
405
+ "delta": -0.04215499320129845,
406
+ "p_value": 1,
407
+ "significant": false
408
+ },
409
+ "benczechmark_ctkfacts_nli": {
410
+ "delta": -0.0450729506855726,
411
+ "p_value": 1,
412
+ "significant": false
413
+ },
414
+ "benczechmark_czechnews": {
415
+ "delta": -0.057340479269040956,
416
+ "p_value": 1,
417
+ "significant": false
418
+ },
419
+ "benczechmark_dialect": {
420
+ "delta": -31.26411167278839,
421
+ "p_value": 1,
422
+ "significant": false
423
+ },
424
+ "benczechmark_essay": {
425
+ "delta": -50.96153218166829,
426
+ "p_value": 1,
427
+ "significant": false
428
+ },
429
+ "benczechmark_fiction": {
430
+ "delta": -93.66096815803243,
431
+ "p_value": 1,
432
+ "significant": false
433
+ },
434
+ "benczechmark_grammarerrorcorrection": {
435
+ "delta": -0.10454155955441302,
436
+ "p_value": 1,
437
+ "significant": false
438
+ },
439
+ "benczechmark_havlicek": {
440
+ "delta": -2642.263507830858,
441
+ "p_value": 1,
442
+ "significant": false
443
+ },
444
+ "benczechmark_hellaswag": {
445
+ "delta": -0.08046205935072692,
446
+ "p_value": 1,
447
+ "significant": false
448
+ },
449
+ "benczechmark_histcorpus": {
450
+ "delta": -533.897540517744,
451
+ "p_value": 1,
452
+ "significant": false
453
+ },
454
+ "benczechmark_history_ir": {
455
+ "delta": -0.05506216696269983,
456
+ "p_value": 1,
457
+ "significant": false
458
+ },
459
+ "benczechmark_klokan_qa": {
460
+ "delta": -0.018587360594795543,
461
+ "p_value": 1,
462
+ "significant": false
463
+ },
464
+ "benczechmark_propaganda_argumentace": {
465
+ "delta": 0.003464040694516779,
466
+ "p_value": 0.047219,
467
+ "significant": true
468
+ },
469
+ "benczechmark_propaganda_demonizace": {
470
+ "delta": 0.00797798684626505,
471
+ "p_value": 0.014886,
472
+ "significant": true
473
+ },
474
+ "benczechmark_propaganda_emoce": {
475
+ "delta": -0.03762121039457278,
476
+ "p_value": 1,
477
+ "significant": false
478
+ },
479
+ "benczechmark_propaganda_fabulace": {
480
+ "delta": -0.001570898365147455,
481
+ "p_value": 1,
482
+ "significant": false
483
+ },
484
+ "benczechmark_propaganda_lokace": {
485
+ "delta": -0.026325735696193564,
486
+ "p_value": 1,
487
+ "significant": false
488
+ },
489
+ "benczechmark_propaganda_nalepkovani": {
490
+ "delta": 0.0073865427446212495,
491
+ "p_value": 0.000237,
492
+ "significant": true
493
+ },
494
+ "benczechmark_propaganda_nazor": {
495
+ "delta": 0.031751959916096695,
496
+ "p_value": 0.0,
497
+ "significant": true
498
+ },
499
+ "benczechmark_propaganda_relativizace": {
500
+ "delta": 0.038360149297259505,
501
+ "p_value": 0.0,
502
+ "significant": true
503
+ },
504
+ "benczechmark_propaganda_rusko": {
505
+ "delta": -0.025915776070387975,
506
+ "p_value": 1,
507
+ "significant": false
508
+ },
509
+ "benczechmark_propaganda_strach": {
510
+ "delta": 0.029769737328192325,
511
+ "p_value": 0.0,
512
+ "significant": true
513
+ },
514
+ "benczechmark_propaganda_vina": {
515
+ "delta": 0.04736962052018423,
516
+ "p_value": 0.0,
517
+ "significant": true
518
+ },
519
+ "benczechmark_propaganda_zamereni": {
520
+ "delta": -0.17568408267218205,
521
+ "p_value": 1,
522
+ "significant": false
523
+ },
524
+ "benczechmark_propaganda_zanr": {
525
+ "delta": -0.05551548113797944,
526
+ "p_value": 1,
527
+ "significant": false
528
+ },
529
+ "benczechmark_sentiment_csfd": {
530
+ "delta": -0.002394380596835499,
531
+ "p_value": 1,
532
+ "significant": false
533
+ },
534
+ "benczechmark_sentiment_fb": {
535
+ "delta": -0.009797675830254704,
536
+ "p_value": 1,
537
+ "significant": false
538
+ },
539
+ "benczechmark_sentiment_mall": {
540
+ "delta": -0.005261277272754605,
541
+ "p_value": 1,
542
+ "significant": false
543
+ },
544
+ "benczechmark_snli": {
545
+ "delta": -0.04450086863557401,
546
+ "p_value": 1,
547
+ "significant": false
548
+ },
549
+ "benczechmark_speeches": {
550
+ "delta": -87.1958280629542,
551
+ "p_value": 1,
552
+ "significant": false
553
+ },
554
+ "benczechmark_spoken": {
555
+ "delta": -56.86880509541322,
556
+ "p_value": 1,
557
+ "significant": false
558
+ },
559
+ "benczechmark_subjectivity": {
560
+ "delta": 0.1975800710055951,
561
+ "p_value": 0.0,
562
+ "significant": true
563
+ },
564
+ "benczechmark_summarization": {
565
+ "delta": -0.017408604795130067,
566
+ "p_value": 1,
567
+ "significant": false
568
+ },
569
+ "benczechmark_umimeto_biology": {
570
+ "delta": -0.06999999999999995,
571
+ "p_value": 1,
572
+ "significant": false
573
+ },
574
+ "benczechmark_umimeto_chemistry": {
575
+ "delta": -0.06000000000000005,
576
+ "p_value": 1,
577
+ "significant": false
578
+ },
579
+ "benczechmark_umimeto_czech": {
580
+ "delta": -0.12,
581
+ "p_value": 1,
582
+ "significant": false
583
+ },
584
+ "benczechmark_umimeto_history": {
585
+ "delta": -0.06999999999999995,
586
+ "p_value": 1,
587
+ "significant": false
588
+ },
589
+ "benczechmark_umimeto_informatics": {
590
+ "delta": -0.07999999999999996,
591
+ "p_value": 1,
592
+ "significant": false
593
+ },
594
+ "benczechmark_umimeto_math": {
595
+ "delta": -0.09999999999999998,
596
+ "p_value": 1,
597
+ "significant": false
598
+ },
599
+ "benczechmark_umimeto_physics": {
600
+ "delta": -0.010000000000000009,
601
+ "p_value": 1,
602
+ "significant": false
603
+ }
604
+ },
605
  "CohereF_aya-23-_78291f9_91bd234": {
606
  "benczechmark_agree": false,
607
  "benczechmark_belebele": false,