| { |
| "BUT-FIT_CSTinyL_0316dd3_27729a2": { |
| "BUT-FIT_CSTinyL_0316dd3_27729a2": { |
| "benczechmark_agree": false, |
| "benczechmark_belebele": false, |
| "benczechmark_capek": false, |
| "benczechmark_cermat_czech_mc": false, |
| "benczechmark_cermat_czech_open": false, |
| "benczechmark_cermat_czech_tf": false, |
| "benczechmark_cermat_czmath_mc": false, |
| "benczechmark_cermat_czmath_open": false, |
| "benczechmark_correspondence": false, |
| "benczechmark_cs_court_decisions_ner": false, |
| "benczechmark_cs_naturalquestions": false, |
| "benczechmark_cs_ner": false, |
| "benczechmark_cs_sqad32": false, |
| "benczechmark_cs_triviaQA": false, |
| "benczechmark_csfever_nli": false, |
| "benczechmark_ctkfacts_nli": false, |
| "benczechmark_czechnews": false, |
| "benczechmark_dialect": false, |
| "benczechmark_essay": false, |
| "benczechmark_fiction": false, |
| "benczechmark_grammarerrorcorrection": false, |
| "benczechmark_havlicek": false, |
| "benczechmark_hellaswag": false, |
| "benczechmark_histcorpus": false, |
| "benczechmark_history_ir": false, |
| "benczechmark_klokan_qa": false, |
| "benczechmark_propaganda_argumentace": false, |
| "benczechmark_propaganda_demonizace": false, |
| "benczechmark_propaganda_emoce": false, |
| "benczechmark_propaganda_fabulace": false, |
| "benczechmark_propaganda_lokace": false, |
| "benczechmark_propaganda_nalepkovani": false, |
| "benczechmark_propaganda_nazor": false, |
| "benczechmark_propaganda_relativizace": false, |
| "benczechmark_propaganda_rusko": false, |
| "benczechmark_propaganda_strach": false, |
| "benczechmark_propaganda_vina": false, |
| "benczechmark_propaganda_zamereni": false, |
| "benczechmark_propaganda_zanr": false, |
| "benczechmark_sentiment_csfd": false, |
| "benczechmark_sentiment_fb": false, |
| "benczechmark_sentiment_mall": false, |
| "benczechmark_snli": false, |
| "benczechmark_speeches": false, |
| "benczechmark_spoken": false, |
| "benczechmark_subjectivity": false, |
| "benczechmark_summarization": false, |
| "benczechmark_umimeto_biology": false, |
| "benczechmark_umimeto_chemistry": false, |
| "benczechmark_umimeto_czech": false, |
| "benczechmark_umimeto_history": false, |
| "benczechmark_umimeto_informatics": false, |
| "benczechmark_umimeto_math": false, |
| "benczechmark_umimeto_physics": false |
| }, |
| "BUT-FIT_csmpt7b_bdd7d3f_0f64387": { |
| "benczechmark_agree": { |
| "delta": -0.14165747835557774, |
| "p_value": 1, |
| "significant": false |
| }, |
| "benczechmark_belebele": { |
| "delta": -0.026759509156824124, |
| "p_value": 1, |
| "significant": false |
| }, |
| "benczechmark_capek": { |
| "delta": -38.979784847302284, |
| "p_value": 1, |
| "significant": false |
| }, |
| "benczechmark_cermat_czech_mc": { |
| "delta": 0.01694915254237289, |
| "p_value": 0.18555166154816954, |
| "significant": false |
| }, |
| "benczechmark_cermat_czech_open": { |
| "delta": -0.00964630225080386, |
| "p_value": 1, |
| "significant": false |
| }, |
| "benczechmark_cermat_czech_tf": { |
| "delta": -0.026793877956668655, |
| "p_value": 1, |
| "significant": false |
| }, |
| "benczechmark_cermat_czmath_mc": { |
| "delta": 0.05555555555555558, |
| "p_value": 0.10497823843214694, |
| "significant": false |
| }, |
| "benczechmark_cermat_czmath_open": { |
| "delta": 0.0, |
| "p_value": 0.5, |
| "significant": false |
| }, |
| "benczechmark_correspondence": { |
| "delta": 33.09354071780615, |
| "p_value": 0.0, |
| "significant": true |
| }, |
| "benczechmark_cs_court_decisions_ner": { |
| "delta": -0.03986000388878087, |
| "p_value": 1, |
| "significant": false |
| }, |
| "benczechmark_cs_naturalquestions": { |
| "delta": -0.0011080332409972304, |
| "p_value": 1, |
| "significant": false |
| }, |
| "benczechmark_cs_ner": { |
| "delta": -0.03162346019488879, |
| "p_value": 1, |
| "significant": false |
| }, |
| "benczechmark_cs_sqad32": { |
| "delta": -0.08797445902802413, |
| "p_value": 1, |
| "significant": false |
| }, |
| "benczechmark_cs_triviaQA": { |
| "delta": -0.013091103393000264, |
| "p_value": 1, |
| "significant": false |
| }, |
| "benczechmark_csfever_nli": { |
| "delta": -0.003830251554403419, |
| "p_value": 1, |
| "significant": false |
| }, |
| "benczechmark_ctkfacts_nli": { |
| "delta": -0.004273161536196279, |
| "p_value": 1, |
| "significant": false |
| }, |
| "benczechmark_czechnews": { |
| "delta": -0.06026309731304491, |
| "p_value": 1, |
| "significant": false |
| }, |
| "benczechmark_dialect": { |
| "delta": 10.7015021509294, |
| "p_value": 0.0, |
| "significant": true |
| }, |
| "benczechmark_essay": { |
| "delta": -3.9330187585401006, |
| "p_value": 1, |
| "significant": false |
| }, |
| "benczechmark_fiction": { |
| "delta": -6.922129848214766, |
| "p_value": 1, |
| "significant": false |
| }, |
| "benczechmark_grammarerrorcorrection": { |
| "delta": 0.01778063410454156, |
| "p_value": 0.010885007320822133, |
| "significant": true |
| }, |
| "benczechmark_havlicek": { |
| "delta": -94.27036012767604, |
| "p_value": 1, |
| "significant": false |
| }, |
| "benczechmark_hellaswag": { |
| "delta": -0.06861183031268675, |
| "p_value": 1, |
| "significant": false |
| }, |
| "benczechmark_histcorpus": { |
| "delta": -39.12895305428549, |
| "p_value": 1, |
| "significant": false |
| }, |
| "benczechmark_history_ir": { |
| "delta": 0.009769094138543544, |
| "p_value": 0.24122327701826957, |
| "significant": false |
| }, |
| "benczechmark_klokan_qa": { |
| "delta": -0.006195786864931829, |
| "p_value": 1, |
| "significant": false |
| }, |
| "benczechmark_propaganda_argumentace": { |
| "delta": -0.07912630918243024, |
| "p_value": 1, |
| "significant": false |
| }, |
| "benczechmark_propaganda_demonizace": { |
| "delta": -0.12044106679924715, |
| "p_value": 1, |
| "significant": false |
| }, |
| "benczechmark_propaganda_emoce": { |
| "delta": -0.04872176932862288, |
| "p_value": 1, |
| "significant": false |
| }, |
| "benczechmark_propaganda_fabulace": { |
| "delta": -0.13838059425040722, |
| "p_value": 1, |
| "significant": false |
| }, |
| "benczechmark_propaganda_lokace": { |
| "delta": 0.061524187407254405, |
| "p_value": 0.00026, |
| "significant": true |
| }, |
| "benczechmark_propaganda_nalepkovani": { |
| "delta": -0.23372953869047625, |
| "p_value": 1, |
| "significant": false |
| }, |
| "benczechmark_propaganda_nazor": { |
| "delta": -0.3301246195072455, |
| "p_value": 1, |
| "significant": false |
| }, |
| "benczechmark_propaganda_relativizace": { |
| "delta": -0.26098973547549487, |
| "p_value": 1, |
| "significant": false |
| }, |
| "benczechmark_propaganda_rusko": { |
| "delta": 0.021837435625203794, |
| "p_value": 0.336379, |
| "significant": false |
| }, |
| "benczechmark_propaganda_strach": { |
| "delta": -0.2886666666666665, |
| "p_value": 1, |
| "significant": false |
| }, |
| "benczechmark_propaganda_vina": { |
| "delta": -0.11717872283056463, |
| "p_value": 1, |
| "significant": false |
| }, |
| "benczechmark_propaganda_zamereni": { |
| "delta": -0.01648175334598867, |
| "p_value": 1, |
| "significant": false |
| }, |
| "benczechmark_propaganda_zanr": { |
| "delta": -0.04256309815937209, |
| "p_value": 1, |
| "significant": false |
| }, |
| "benczechmark_sentiment_csfd": { |
| "delta": -0.07886631165326674, |
| "p_value": 1, |
| "significant": false |
| }, |
| "benczechmark_sentiment_fb": { |
| "delta": -0.06492447153155445, |
| "p_value": 1, |
| "significant": false |
| }, |
| "benczechmark_sentiment_mall": { |
| "delta": -0.04787079792912874, |
| "p_value": 1, |
| "significant": false |
| }, |
| "benczechmark_snli": { |
| "delta": -0.0032740879326473293, |
| "p_value": 1, |
| "significant": false |
| }, |
| "benczechmark_speeches": { |
| "delta": -27.949027813696404, |
| "p_value": 1, |
| "significant": false |
| }, |
| "benczechmark_spoken": { |
| "delta": -12.302187723867164, |
| "p_value": 1, |
| "significant": false |
| }, |
| "benczechmark_subjectivity": { |
| "delta": 0.014476799052734135, |
| "p_value": 0.0, |
| "significant": true |
| }, |
| "benczechmark_summarization": { |
| "delta": 0.004130927375898572, |
| "p_value": 0.001, |
| "significant": true |
| }, |
| "benczechmark_umimeto_biology": { |
| "delta": -0.12, |
| "p_value": 1, |
| "significant": false |
| }, |
| "benczechmark_umimeto_chemistry": { |
| "delta": -0.040000000000000036, |
| "p_value": 1, |
| "significant": false |
| }, |
| "benczechmark_umimeto_czech": { |
| "delta": 0.12, |
| "p_value": 0.008916478606159751, |
| "significant": true |
| }, |
| "benczechmark_umimeto_history": { |
| "delta": -0.040000000000000036, |
| "p_value": 1, |
| "significant": false |
| }, |
| "benczechmark_umimeto_informatics": { |
| "delta": -0.10000000000000003, |
| "p_value": 1, |
| "significant": false |
| }, |
| "benczechmark_umimeto_math": { |
| "delta": -0.030000000000000027, |
| "p_value": 1, |
| "significant": false |
| }, |
| "benczechmark_umimeto_physics": { |
| "delta": 0.09999999999999998, |
| "p_value": 0.03381976333305582, |
| "significant": true |
| } |
| }, |
| "CohereF_aya-23-_696092a_369b1ca": { |
| "benczechmark_agree": { |
| "delta": -0.36294214833544125, |
| "p_value": 1, |
| "significant": false |
| }, |
| "benczechmark_belebele": { |
| "delta": -0.03189016759509544, |
| "p_value": 1, |
| "significant": false |
| }, |
| "benczechmark_capek": { |
| "delta": 55.08467756432839, |
| "p_value": 0.0, |
| "significant": true |
| }, |
| "benczechmark_cermat_czech_mc": { |
| "delta": -0.20955315870570107, |
| "p_value": 1, |
| "significant": false |
| }, |
| "benczechmark_cermat_czech_open": { |
| "delta": -0.15112540192926044, |
| "p_value": 1, |
| "significant": false |
| }, |
| "benczechmark_cermat_czech_tf": { |
| "delta": -0.2762075134168158, |
| "p_value": 1, |
| "significant": false |
| }, |
| "benczechmark_cermat_czmath_mc": { |
| "delta": 0.08730158730158732, |
| "p_value": 0.06931765235875581, |
| "significant": false |
| }, |
| "benczechmark_cermat_czmath_open": { |
| "delta": -0.040880503144654086, |
| "p_value": 1, |
| "significant": false |
| }, |
| "benczechmark_correspondence": { |
| "delta": 130.0242963464703, |
| "p_value": 0.0, |
| "significant": true |
| }, |
| "benczechmark_cs_court_decisions_ner": { |
| "delta": -0.5409294186272603, |
| "p_value": 1, |
| "significant": false |
| }, |
| "benczechmark_cs_naturalquestions": { |
| "delta": -0.16842105263157894, |
| "p_value": 1, |
| "significant": false |
| }, |
| "benczechmark_cs_ner": { |
| "delta": -0.4419010847582276, |
| "p_value": 1, |
| "significant": false |
| }, |
| "benczechmark_cs_sqad32": { |
| "delta": -0.4320681092586023, |
| "p_value": 1, |
| "significant": false |
| }, |
| "benczechmark_cs_triviaQA": { |
| "delta": -0.3631668002493543, |
| "p_value": 1, |
| "significant": false |
| }, |
| "benczechmark_csfever_nli": { |
| "delta": -0.32482979335222373, |
| "p_value": 1, |
| "significant": false |
| }, |
| "benczechmark_ctkfacts_nli": { |
| "delta": -0.37197595488843793, |
| "p_value": 1, |
| "significant": false |
| }, |
| "benczechmark_czechnews": { |
| "delta": -0.33850027225167534, |
| "p_value": 1, |
| "significant": false |
| }, |
| "benczechmark_dialect": { |
| "delta": -27.333522181080255, |
| "p_value": 1, |
| "significant": false |
| }, |
| "benczechmark_essay": { |
| "delta": 16.952152053720326, |
| "p_value": 0.0, |
| "significant": true |
| }, |
| "benczechmark_fiction": { |
| "delta": 64.23837479968857, |
| "p_value": 0.0, |
| "significant": true |
| }, |
| "benczechmark_grammarerrorcorrection": { |
| "delta": -0.1450299914310197, |
| "p_value": 1, |
| "significant": false |
| }, |
| "benczechmark_havlicek": { |
| "delta": 1254.7772124189864, |
| "p_value": 0.0, |
| "significant": true |
| }, |
| "benczechmark_hellaswag": { |
| "delta": -0.1996614220274846, |
| "p_value": 1, |
| "significant": false |
| }, |
| "benczechmark_histcorpus": { |
| "delta": 465.6520804217422, |
| "p_value": 0.0, |
| "significant": true |
| }, |
| "benczechmark_history_ir": { |
| "delta": -0.10390763765541738, |
| "p_value": 1, |
| "significant": false |
| }, |
| "benczechmark_klokan_qa": { |
| "delta": -0.02106567534076828, |
| "p_value": 1, |
| "significant": false |
| }, |
| "benczechmark_propaganda_argumentace": { |
| "delta": -0.028479974072752157, |
| "p_value": 1, |
| "significant": false |
| }, |
| "benczechmark_propaganda_demonizace": { |
| "delta": -0.1334386874533896, |
| "p_value": 1, |
| "significant": false |
| }, |
| "benczechmark_propaganda_emoce": { |
| "delta": -0.22489194847344945, |
| "p_value": 1, |
| "significant": false |
| }, |
| "benczechmark_propaganda_fabulace": { |
| "delta": -0.1358818348931649, |
| "p_value": 1, |
| "significant": false |
| }, |
| "benczechmark_propaganda_lokace": { |
| "delta": -0.08284072590271263, |
| "p_value": 1, |
| "significant": false |
| }, |
| "benczechmark_propaganda_nalepkovani": { |
| "delta": -0.25483816964285716, |
| "p_value": 1, |
| "significant": false |
| }, |
| "benczechmark_propaganda_nazor": { |
| "delta": -0.34254971338380547, |
| "p_value": 1, |
| "significant": false |
| }, |
| "benczechmark_propaganda_relativizace": { |
| "delta": -0.2411098263550795, |
| "p_value": 1, |
| "significant": false |
| }, |
| "benczechmark_propaganda_rusko": { |
| "delta": -0.06682622705208752, |
| "p_value": 1, |
| "significant": false |
| }, |
| "benczechmark_propaganda_strach": { |
| "delta": -0.23289729729729725, |
| "p_value": 1, |
| "significant": false |
| }, |
| "benczechmark_propaganda_vina": { |
| "delta": -0.06493701035097432, |
| "p_value": 1, |
| "significant": false |
| }, |
| "benczechmark_propaganda_zamereni": { |
| "delta": -0.29783103789668997, |
| "p_value": 1, |
| "significant": false |
| }, |
| "benczechmark_propaganda_zanr": { |
| "delta": -0.2618941770600769, |
| "p_value": 1, |
| "significant": false |
| }, |
| "benczechmark_sentiment_csfd": { |
| "delta": -0.18848773376632122, |
| "p_value": 1, |
| "significant": false |
| }, |
| "benczechmark_sentiment_fb": { |
| "delta": -0.3015689677941502, |
| "p_value": 1, |
| "significant": false |
| }, |
| "benczechmark_sentiment_mall": { |
| "delta": -0.1506104449281993, |
| "p_value": 1, |
| "significant": false |
| }, |
| "benczechmark_snli": { |
| "delta": -0.3344246959775491, |
| "p_value": 1, |
| "significant": false |
| }, |
| "benczechmark_speeches": { |
| "delta": 21.296491459250547, |
| "p_value": 0.0, |
| "significant": true |
| }, |
| "benczechmark_spoken": { |
| "delta": -14.976728132910694, |
| "p_value": 1, |
| "significant": false |
| }, |
| "benczechmark_subjectivity": { |
| "delta": 0.14572260219315808, |
| "p_value": 0.0, |
| "significant": true |
| }, |
| "benczechmark_summarization": { |
| "delta": -0.035038764236280204, |
| "p_value": 1, |
| "significant": false |
| }, |
| "benczechmark_umimeto_biology": { |
| "delta": -0.4, |
| "p_value": 1, |
| "significant": false |
| }, |
| "benczechmark_umimeto_chemistry": { |
| "delta": -0.16000000000000003, |
| "p_value": 1, |
| "significant": false |
| }, |
| "benczechmark_umimeto_czech": { |
| "delta": -0.010000000000000009, |
| "p_value": 1, |
| "significant": false |
| }, |
| "benczechmark_umimeto_history": { |
| "delta": -0.22999999999999998, |
| "p_value": 1, |
| "significant": false |
| }, |
| "benczechmark_umimeto_informatics": { |
| "delta": -0.46, |
| "p_value": 1, |
| "significant": false |
| }, |
| "benczechmark_umimeto_math": { |
| "delta": -0.26, |
| "p_value": 1, |
| "significant": false |
| }, |
| "benczechmark_umimeto_physics": { |
| "delta": -0.13, |
| "p_value": 1, |
| "significant": false |
| } |
| }, |
| "CohereF_aya-23-_78291f9_91bd234": { |
| "benczechmark_agree": { |
| "delta": -0.26180528377866186, |
| "p_value": 1, |
| "significant": false |
| }, |
| "benczechmark_belebele": { |
| "delta": -0.021695678266298257, |
| "p_value": 1, |
| "significant": false |
| }, |
| "benczechmark_capek": { |
| "delta": 201.08271212785164, |
| "p_value": 0.0, |
| "significant": true |
| }, |
| "benczechmark_cermat_czech_mc": { |
| "delta": -0.07087827426810478, |
| "p_value": 1, |
| "significant": false |
| }, |
| "benczechmark_cermat_czech_open": { |
| "delta": -0.04501607717041801, |
| "p_value": 1, |
| "significant": false |
| }, |
| "benczechmark_cermat_czech_tf": { |
| "delta": -0.17239117471675613, |
| "p_value": 1, |
| "significant": false |
| }, |
| "benczechmark_cermat_czmath_mc": { |
| "delta": 0.007936507936507964, |
| "p_value": 0.44851403278277635, |
| "significant": false |
| }, |
| "benczechmark_cermat_czmath_open": { |
| "delta": -0.015723270440251572, |
| "p_value": 1, |
| "significant": false |
| }, |
| "benczechmark_correspondence": { |
| "delta": 484.55213708828916, |
| "p_value": 0.0, |
| "significant": true |
| }, |
| "benczechmark_cs_court_decisions_ner": { |
| "delta": -0.3721563289908614, |
| "p_value": 1, |
| "significant": false |
| }, |
| "benczechmark_cs_naturalquestions": { |
| "delta": -0.12437673130193905, |
| "p_value": 1, |
| "significant": false |
| }, |
| "benczechmark_cs_ner": { |
| "delta": -0.393086964515536, |
| "p_value": 1, |
| "significant": false |
| }, |
| "benczechmark_cs_sqad32": { |
| "delta": -0.4320681092586023, |
| "p_value": 1, |
| "significant": false |
| }, |
| "benczechmark_cs_triviaQA": { |
| "delta": -0.24481253896161723, |
| "p_value": 1, |
| "significant": false |
| }, |
| "benczechmark_csfever_nli": { |
| "delta": -0.2826748001509253, |
| "p_value": 1, |
| "significant": false |
| }, |
| "benczechmark_ctkfacts_nli": { |
| "delta": -0.32690300420286533, |
| "p_value": 1, |
| "significant": false |
| }, |
| "benczechmark_czechnews": { |
| "delta": -0.2811597929826344, |
| "p_value": 1, |
| "significant": false |
| }, |
| "benczechmark_dialect": { |
| "delta": 3.930589491708133, |
| "p_value": 0.002, |
| "significant": true |
| }, |
| "benczechmark_essay": { |
| "delta": 67.91368423538862, |
| "p_value": 0.0, |
| "significant": true |
| }, |
| "benczechmark_fiction": { |
| "delta": 157.899342957721, |
| "p_value": 0.0, |
| "significant": true |
| }, |
| "benczechmark_grammarerrorcorrection": { |
| "delta": -0.04048843187660667, |
| "p_value": 1, |
| "significant": false |
| }, |
| "benczechmark_havlicek": { |
| "delta": 3897.0407202498445, |
| "p_value": 0.0, |
| "significant": true |
| }, |
| "benczechmark_hellaswag": { |
| "delta": -0.11919936267675768, |
| "p_value": 1, |
| "significant": false |
| }, |
| "benczechmark_histcorpus": { |
| "delta": 999.5496209394862, |
| "p_value": 0.0, |
| "significant": true |
| }, |
| "benczechmark_history_ir": { |
| "delta": -0.04884547069271755, |
| "p_value": 1, |
| "significant": false |
| }, |
| "benczechmark_klokan_qa": { |
| "delta": -0.0024783147459727373, |
| "p_value": 1, |
| "significant": false |
| }, |
| "benczechmark_propaganda_argumentace": { |
| "delta": -0.03156498836966015, |
| "p_value": 1, |
| "significant": false |
| }, |
| "benczechmark_propaganda_demonizace": { |
| "delta": -0.1425920664796334, |
| "p_value": 1, |
| "significant": false |
| }, |
| "benczechmark_propaganda_emoce": { |
| "delta": -0.18727073807887668, |
| "p_value": 1, |
| "significant": false |
| }, |
| "benczechmark_propaganda_fabulace": { |
| "delta": -0.13431093652801745, |
| "p_value": 1, |
| "significant": false |
| }, |
| "benczechmark_propaganda_lokace": { |
| "delta": -0.05651499020651907, |
| "p_value": 1, |
| "significant": false |
| }, |
| "benczechmark_propaganda_nalepkovani": { |
| "delta": -0.26307663690476185, |
| "p_value": 1, |
| "significant": false |
| }, |
| "benczechmark_propaganda_nazor": { |
| "delta": -0.3744699118037308, |
| "p_value": 1, |
| "significant": false |
| }, |
| "benczechmark_propaganda_relativizace": { |
| "delta": -0.2775478740668613, |
| "p_value": 1, |
| "significant": false |
| }, |
| "benczechmark_propaganda_rusko": { |
| "delta": -0.04091045098169954, |
| "p_value": 1, |
| "significant": false |
| }, |
| "benczechmark_propaganda_strach": { |
| "delta": -0.2637657657657657, |
| "p_value": 1, |
| "significant": false |
| }, |
| "benczechmark_propaganda_vina": { |
| "delta": -0.11297317135281693, |
| "p_value": 1, |
| "significant": false |
| }, |
| "benczechmark_propaganda_zamereni": { |
| "delta": -0.12214695522450791, |
| "p_value": 1, |
| "significant": false |
| }, |
| "benczechmark_propaganda_zanr": { |
| "delta": -0.20637869592209745, |
| "p_value": 1, |
| "significant": false |
| }, |
| "benczechmark_sentiment_csfd": { |
| "delta": -0.18609335316948572, |
| "p_value": 1, |
| "significant": false |
| }, |
| "benczechmark_sentiment_fb": { |
| "delta": -0.2917712919638955, |
| "p_value": 1, |
| "significant": false |
| }, |
| "benczechmark_sentiment_mall": { |
| "delta": -0.1453491676554447, |
| "p_value": 1, |
| "significant": false |
| }, |
| "benczechmark_snli": { |
| "delta": -0.2899238273419751, |
| "p_value": 1, |
| "significant": false |
| }, |
| "benczechmark_speeches": { |
| "delta": 108.49231952220475, |
| "p_value": 0.0, |
| "significant": true |
| }, |
| "benczechmark_spoken": { |
| "delta": 41.89207696250253, |
| "p_value": 0.0, |
| "significant": true |
| }, |
| "benczechmark_subjectivity": { |
| "delta": -0.05193266782029515, |
| "p_value": 1, |
| "significant": false |
| }, |
| "benczechmark_summarization": { |
| "delta": -0.017630159441150133, |
| "p_value": 1, |
| "significant": false |
| }, |
| "benczechmark_umimeto_biology": { |
| "delta": -0.33000000000000007, |
| "p_value": 1, |
| "significant": false |
| }, |
| "benczechmark_umimeto_chemistry": { |
| "delta": -0.09999999999999998, |
| "p_value": 1, |
| "significant": false |
| }, |
| "benczechmark_umimeto_czech": { |
| "delta": 0.10999999999999999, |
| "p_value": 0.06939870829413583, |
| "significant": false |
| }, |
| "benczechmark_umimeto_history": { |
| "delta": -0.16000000000000003, |
| "p_value": 1, |
| "significant": false |
| }, |
| "benczechmark_umimeto_informatics": { |
| "delta": -0.38000000000000006, |
| "p_value": 1, |
| "significant": false |
| }, |
| "benczechmark_umimeto_math": { |
| "delta": -0.16000000000000003, |
| "p_value": 1, |
| "significant": false |
| }, |
| "benczechmark_umimeto_physics": { |
| "delta": -0.12, |
| "p_value": 1, |
| "significant": false |
| } |
| } |
| }, |
| "BUT-FIT_csmpt7b_bdd7d3f_0f64387": { |
| "BUT-FIT_CSTinyL_0316dd3_27729a2": { |
| "benczechmark_agree": { |
| "delta": 0.1391477775034693, |
| "p_value": 0.0, |
| "significant": true |
| }, |
| "benczechmark_belebele": { |
| "delta": 0.02656033647114403, |
| "p_value": 0.0, |
| "significant": true |
| }, |
| "benczechmark_capek": { |
| "delta": 38.979784847302284, |
| "p_value": 0.0, |
| "significant": true |
| }, |
| "benczechmark_cermat_czech_mc": { |
| "delta": -0.01694915254237289, |
| "p_value": 1, |
| "significant": false |
| }, |
| "benczechmark_cermat_czech_open": { |
| "delta": 0.00964630225080386, |
| "p_value": 0.041631453549706425, |
| "significant": true |
| }, |
| "benczechmark_cermat_czech_tf": { |
| "delta": 0.026876360533836796, |
| "p_value": 0.0, |
| "significant": true |
| }, |
| "benczechmark_cermat_czmath_mc": { |
| "delta": -0.05555555555555558, |
| "p_value": 1, |
| "significant": false |
| }, |
| "benczechmark_cermat_czmath_open": { |
| "delta": 0.0, |
| "p_value": 0.5, |
| "significant": false |
| }, |
| "benczechmark_correspondence": { |
| "delta": -33.09354071780615, |
| "p_value": 1, |
| "significant": false |
| }, |
| "benczechmark_cs_court_decisions_ner": { |
| "delta": 0.03986000388878087, |
| "p_value": 3.141662273662873e-26, |
| "significant": true |
| }, |
| "benczechmark_cs_naturalquestions": { |
| "delta": 0.0011080332409972304, |
| "p_value": 0.3028224407785538, |
| "significant": false |
| }, |
| "benczechmark_cs_ner": { |
| "delta": 0.03162346019488879, |
| "p_value": 5.985879102095067e-16, |
| "significant": true |
| }, |
| "benczechmark_cs_sqad32": { |
| "delta": 0.08797445902802413, |
| "p_value": 3.5494303052424245e-33, |
| "significant": true |
| }, |
| "benczechmark_cs_triviaQA": { |
| "delta": 0.013091103393000264, |
| "p_value": 2.6608687310336288e-08, |
| "significant": true |
| }, |
| "benczechmark_csfever_nli": { |
| "delta": 0.003818640544250962, |
| "p_value": 0.0, |
| "significant": true |
| }, |
| "benczechmark_ctkfacts_nli": { |
| "delta": 0.004170903395295911, |
| "p_value": 0.037647, |
| "significant": true |
| }, |
| "benczechmark_czechnews": { |
| "delta": 0.05966454267819399, |
| "p_value": 0.0, |
| "significant": true |
| }, |
| "benczechmark_dialect": { |
| "delta": -10.7015021509294, |
| "p_value": 1, |
| "significant": false |
| }, |
| "benczechmark_essay": { |
| "delta": 3.9330187585401006, |
| "p_value": 0.0, |
| "significant": true |
| }, |
| "benczechmark_fiction": { |
| "delta": 6.922129848214766, |
| "p_value": 0.0, |
| "significant": true |
| }, |
| "benczechmark_grammarerrorcorrection": { |
| "delta": -0.01778063410454156, |
| "p_value": 1, |
| "significant": false |
| }, |
| "benczechmark_havlicek": { |
| "delta": 94.27036012767604, |
| "p_value": 0.0, |
| "significant": true |
| }, |
| "benczechmark_hellaswag": { |
| "delta": 0.06861183031268675, |
| "p_value": 5.273965869163969e-78, |
| "significant": true |
| }, |
| "benczechmark_histcorpus": { |
| "delta": 39.12895305428549, |
| "p_value": 0.0, |
| "significant": true |
| }, |
| "benczechmark_history_ir": { |
| "delta": -0.009769094138543544, |
| "p_value": 1, |
| "significant": false |
| }, |
| "benczechmark_klokan_qa": { |
| "delta": 0.006195786864931829, |
| "p_value": 0.37480585816362966, |
| "significant": false |
| }, |
| "benczechmark_propaganda_argumentace": { |
| "delta": 0.07832602810018413, |
| "p_value": 0.0, |
| "significant": true |
| }, |
| "benczechmark_propaganda_demonizace": { |
| "delta": 0.11595278741310344, |
| "p_value": 0.0, |
| "significant": true |
| }, |
| "benczechmark_propaganda_emoce": { |
| "delta": 0.03238807903753932, |
| "p_value": 0.265751, |
| "significant": false |
| }, |
| "benczechmark_propaganda_fabulace": { |
| "delta": 0.13781558734314608, |
| "p_value": 0.0, |
| "significant": true |
| }, |
| "benczechmark_propaganda_lokace": { |
| "delta": -0.06723935520825841, |
| "p_value": 1, |
| "significant": false |
| }, |
| "benczechmark_propaganda_nalepkovani": { |
| "delta": 0.2324187566983621, |
| "p_value": 0.0, |
| "significant": true |
| }, |
| "benczechmark_propaganda_nazor": { |
| "delta": 0.32710595697154116, |
| "p_value": 0.0, |
| "significant": true |
| }, |
| "benczechmark_propaganda_relativizace": { |
| "delta": 0.25665370421111533, |
| "p_value": 0.0, |
| "significant": true |
| }, |
| "benczechmark_propaganda_rusko": { |
| "delta": -0.026343686641658426, |
| "p_value": 1, |
| "significant": false |
| }, |
| "benczechmark_propaganda_strach": { |
| "delta": 0.28447721360556466, |
| "p_value": 0.0, |
| "significant": true |
| }, |
| "benczechmark_propaganda_vina": { |
| "delta": 0.1167885712714975, |
| "p_value": 0.0, |
| "significant": true |
| }, |
| "benczechmark_propaganda_zamereni": { |
| "delta": 0.015422573424564878, |
| "p_value": 0.003682, |
| "significant": true |
| }, |
| "benczechmark_propaganda_zanr": { |
| "delta": 0.04012011329606524, |
| "p_value": 0.0, |
| "significant": true |
| }, |
| "benczechmark_sentiment_csfd": { |
| "delta": 0.0785678319896399, |
| "p_value": 0.0, |
| "significant": true |
| }, |
| "benczechmark_sentiment_fb": { |
| "delta": 0.06459448378330879, |
| "p_value": 0.0, |
| "significant": true |
| }, |
| "benczechmark_sentiment_mall": { |
| "delta": 0.04764692606689502, |
| "p_value": 0.0, |
| "significant": true |
| }, |
| "benczechmark_snli": { |
| "delta": 0.0032740879326473293, |
| "p_value": 0.1269896687688912, |
| "significant": false |
| }, |
| "benczechmark_speeches": { |
| "delta": 27.949027813696404, |
| "p_value": 0.0, |
| "significant": true |
| }, |
| "benczechmark_spoken": { |
| "delta": 12.302187723867164, |
| "p_value": 0.0, |
| "significant": true |
| }, |
| "benczechmark_subjectivity": { |
| "delta": -0.014490927419354982, |
| "p_value": 1, |
| "significant": false |
| }, |
| "benczechmark_summarization": { |
| "delta": -0.004130927375898558, |
| "p_value": 1, |
| "significant": false |
| }, |
| "benczechmark_umimeto_biology": { |
| "delta": 0.12, |
| "p_value": 0.03836462002636626, |
| "significant": true |
| }, |
| "benczechmark_umimeto_chemistry": { |
| "delta": 0.040000000000000036, |
| "p_value": 0.24113123538615433, |
| "significant": false |
| }, |
| "benczechmark_umimeto_czech": { |
| "delta": -0.12, |
| "p_value": 1, |
| "significant": false |
| }, |
| "benczechmark_umimeto_history": { |
| "delta": 0.040000000000000036, |
| "p_value": 0.2648946017749204, |
| "significant": false |
| }, |
| "benczechmark_umimeto_informatics": { |
| "delta": 0.10000000000000003, |
| "p_value": 0.052533804480738064, |
| "significant": false |
| }, |
| "benczechmark_umimeto_math": { |
| "delta": 0.030000000000000027, |
| "p_value": 0.2962723644298042, |
| "significant": false |
| }, |
| "benczechmark_umimeto_physics": { |
| "delta": -0.09999999999999998, |
| "p_value": 1, |
| "significant": false |
| } |
| }, |
| "BUT-FIT_csmpt7b_bdd7d3f_0f64387": { |
| "benczechmark_agree": false, |
| "benczechmark_belebele": false, |
| "benczechmark_capek": false, |
| "benczechmark_cermat_czech_mc": false, |
| "benczechmark_cermat_czech_open": false, |
| "benczechmark_cermat_czech_tf": false, |
| "benczechmark_cermat_czmath_mc": false, |
| "benczechmark_cermat_czmath_open": false, |
| "benczechmark_correspondence": false, |
| "benczechmark_cs_court_decisions_ner": false, |
| "benczechmark_cs_naturalquestions": false, |
| "benczechmark_cs_ner": false, |
| "benczechmark_cs_sqad32": false, |
| "benczechmark_cs_triviaQA": false, |
| "benczechmark_csfever_nli": false, |
| "benczechmark_ctkfacts_nli": false, |
| "benczechmark_czechnews": false, |
| "benczechmark_dialect": false, |
| "benczechmark_essay": false, |
| "benczechmark_fiction": false, |
| "benczechmark_grammarerrorcorrection": false, |
| "benczechmark_havlicek": false, |
| "benczechmark_hellaswag": false, |
| "benczechmark_histcorpus": false, |
| "benczechmark_history_ir": false, |
| "benczechmark_klokan_qa": false, |
| "benczechmark_propaganda_argumentace": false, |
| "benczechmark_propaganda_demonizace": false, |
| "benczechmark_propaganda_emoce": false, |
| "benczechmark_propaganda_fabulace": false, |
| "benczechmark_propaganda_lokace": false, |
| "benczechmark_propaganda_nalepkovani": false, |
| "benczechmark_propaganda_nazor": false, |
| "benczechmark_propaganda_relativizace": false, |
| "benczechmark_propaganda_rusko": false, |
| "benczechmark_propaganda_strach": false, |
| "benczechmark_propaganda_vina": false, |
| "benczechmark_propaganda_zamereni": false, |
| "benczechmark_propaganda_zanr": false, |
| "benczechmark_sentiment_csfd": false, |
| "benczechmark_sentiment_fb": false, |
| "benczechmark_sentiment_mall": false, |
| "benczechmark_snli": false, |
| "benczechmark_speeches": false, |
| "benczechmark_spoken": false, |
| "benczechmark_subjectivity": false, |
| "benczechmark_summarization": false, |
| "benczechmark_umimeto_biology": false, |
| "benczechmark_umimeto_chemistry": false, |
| "benczechmark_umimeto_czech": false, |
| "benczechmark_umimeto_history": false, |
| "benczechmark_umimeto_informatics": false, |
| "benczechmark_umimeto_math": false, |
| "benczechmark_umimeto_physics": false |
| }, |
| "CohereF_aya-23-_696092a_369b1ca": { |
| "benczechmark_agree": { |
| "delta": -0.2212846699798635, |
| "p_value": 1, |
| "significant": false |
| }, |
| "benczechmark_belebele": { |
| "delta": -0.005130658438271318, |
| "p_value": 1, |
| "significant": false |
| }, |
| "benczechmark_capek": { |
| "delta": 94.06446241163067, |
| "p_value": 0.0, |
| "significant": true |
| }, |
| "benczechmark_cermat_czech_mc": { |
| "delta": -0.22650231124807396, |
| "p_value": 1, |
| "significant": false |
| }, |
| "benczechmark_cermat_czech_open": { |
| "delta": -0.14147909967845657, |
| "p_value": 1, |
| "significant": false |
| }, |
| "benczechmark_cermat_czech_tf": { |
| "delta": -0.24941363546014717, |
| "p_value": 1, |
| "significant": false |
| }, |
| "benczechmark_cermat_czmath_mc": { |
| "delta": 0.031746031746031744, |
| "p_value": 0.2743003072106708, |
| "significant": false |
| }, |
| "benczechmark_cermat_czmath_open": { |
| "delta": -0.040880503144654086, |
| "p_value": 1, |
| "significant": false |
| }, |
| "benczechmark_correspondence": { |
| "delta": 96.93075562866414, |
| "p_value": 0.0, |
| "significant": true |
| }, |
| "benczechmark_cs_court_decisions_ner": { |
| "delta": -0.5010694147384794, |
| "p_value": 1, |
| "significant": false |
| }, |
| "benczechmark_cs_naturalquestions": { |
| "delta": -0.1673130193905817, |
| "p_value": 1, |
| "significant": false |
| }, |
| "benczechmark_cs_ner": { |
| "delta": -0.4102776245633388, |
| "p_value": 1, |
| "significant": false |
| }, |
| "benczechmark_cs_sqad32": { |
| "delta": -0.34409365023057814, |
| "p_value": 1, |
| "significant": false |
| }, |
| "benczechmark_cs_triviaQA": { |
| "delta": -0.3500756968563541, |
| "p_value": 1, |
| "significant": false |
| }, |
| "benczechmark_csfever_nli": { |
| "delta": -0.3209995417978203, |
| "p_value": 1, |
| "significant": false |
| }, |
| "benczechmark_ctkfacts_nli": { |
| "delta": -0.36770279335224165, |
| "p_value": 1, |
| "significant": false |
| }, |
| "benczechmark_czechnews": { |
| "delta": -0.2782371749386304, |
| "p_value": 1, |
| "significant": false |
| }, |
| "benczechmark_dialect": { |
| "delta": -38.035024332009655, |
| "p_value": 1, |
| "significant": false |
| }, |
| "benczechmark_essay": { |
| "delta": 20.885170812260426, |
| "p_value": 0.0, |
| "significant": true |
| }, |
| "benczechmark_fiction": { |
| "delta": 71.16050464790334, |
| "p_value": 0.0, |
| "significant": true |
| }, |
| "benczechmark_grammarerrorcorrection": { |
| "delta": -0.16281062553556125, |
| "p_value": 1, |
| "significant": false |
| }, |
| "benczechmark_havlicek": { |
| "delta": 1349.0475725466624, |
| "p_value": 0.0, |
| "significant": true |
| }, |
| "benczechmark_hellaswag": { |
| "delta": -0.13104959171479785, |
| "p_value": 1, |
| "significant": false |
| }, |
| "benczechmark_histcorpus": { |
| "delta": 504.7810334760277, |
| "p_value": 0.0, |
| "significant": true |
| }, |
| "benczechmark_history_ir": { |
| "delta": -0.11367673179396093, |
| "p_value": 1, |
| "significant": false |
| }, |
| "benczechmark_klokan_qa": { |
| "delta": -0.014869888475836451, |
| "p_value": 1, |
| "significant": false |
| }, |
| "benczechmark_propaganda_argumentace": { |
| "delta": 0.050709338332311615, |
| "p_value": 0.0, |
| "significant": true |
| }, |
| "benczechmark_propaganda_demonizace": { |
| "delta": -0.01299762065414245, |
| "p_value": 1, |
| "significant": false |
| }, |
| "benczechmark_propaganda_emoce": { |
| "delta": -0.17617017914482658, |
| "p_value": 1, |
| "significant": false |
| }, |
| "benczechmark_propaganda_fabulace": { |
| "delta": 0.002715496897753855, |
| "p_value": 0.281788, |
| "significant": false |
| }, |
| "benczechmark_propaganda_lokace": { |
| "delta": -0.15008008111097104, |
| "p_value": 1, |
| "significant": false |
| }, |
| "benczechmark_propaganda_nalepkovani": { |
| "delta": -0.021108630952380913, |
| "p_value": 1, |
| "significant": false |
| }, |
| "benczechmark_propaganda_nazor": { |
| "delta": -0.01242509387655999, |
| "p_value": 1, |
| "significant": false |
| }, |
| "benczechmark_propaganda_relativizace": { |
| "delta": 0.020524331670466966, |
| "p_value": 0.000172, |
| "significant": true |
| }, |
| "benczechmark_propaganda_rusko": { |
| "delta": -0.09316991369374594, |
| "p_value": 1, |
| "significant": false |
| }, |
| "benczechmark_propaganda_strach": { |
| "delta": 0.054414764217387224, |
| "p_value": 0.0, |
| "significant": true |
| }, |
| "benczechmark_propaganda_vina": { |
| "delta": 0.05156045755302363, |
| "p_value": 0.0, |
| "significant": true |
| }, |
| "benczechmark_propaganda_zamereni": { |
| "delta": -0.2813492845507013, |
| "p_value": 1, |
| "significant": false |
| }, |
| "benczechmark_propaganda_zanr": { |
| "delta": -0.2193310789007048, |
| "p_value": 1, |
| "significant": false |
| }, |
| "benczechmark_sentiment_csfd": { |
| "delta": -0.10962142211305448, |
| "p_value": 1, |
| "significant": false |
| }, |
| "benczechmark_sentiment_fb": { |
| "delta": -0.23664449626259576, |
| "p_value": 1, |
| "significant": false |
| }, |
| "benczechmark_sentiment_mall": { |
| "delta": -0.10273964699907057, |
| "p_value": 1, |
| "significant": false |
| }, |
| "benczechmark_snli": { |
| "delta": -0.33115060804490176, |
| "p_value": 1, |
| "significant": false |
| }, |
| "benczechmark_speeches": { |
| "delta": 49.24551927294695, |
| "p_value": 0.0, |
| "significant": true |
| }, |
| "benczechmark_spoken": { |
| "delta": -2.67454040904353, |
| "p_value": 1, |
| "significant": false |
| }, |
| "benczechmark_subjectivity": { |
| "delta": 0.1312609769763992, |
| "p_value": 0.0, |
| "significant": true |
| }, |
| "benczechmark_summarization": { |
| "delta": -0.03916969161217876, |
| "p_value": 1, |
| "significant": false |
| }, |
| "benczechmark_umimeto_biology": { |
| "delta": -0.28, |
| "p_value": 1, |
| "significant": false |
| }, |
| "benczechmark_umimeto_chemistry": { |
| "delta": -0.12, |
| "p_value": 1, |
| "significant": false |
| }, |
| "benczechmark_umimeto_czech": { |
| "delta": -0.13, |
| "p_value": 1, |
| "significant": false |
| }, |
| "benczechmark_umimeto_history": { |
| "delta": -0.18999999999999995, |
| "p_value": 1, |
| "significant": false |
| }, |
| "benczechmark_umimeto_informatics": { |
| "delta": -0.36, |
| "p_value": 1, |
| "significant": false |
| }, |
| "benczechmark_umimeto_math": { |
| "delta": -0.22999999999999998, |
| "p_value": 1, |
| "significant": false |
| }, |
| "benczechmark_umimeto_physics": { |
| "delta": -0.22999999999999998, |
| "p_value": 1, |
| "significant": false |
| } |
| }, |
| "CohereF_aya-23-_78291f9_91bd234": { |
| "benczechmark_agree": { |
| "delta": -0.12014780542308412, |
| "p_value": 1, |
| "significant": false |
| }, |
| "benczechmark_belebele": { |
| "delta": 0.004999309419651543, |
| "p_value": 0.000102, |
| "significant": true |
| }, |
| "benczechmark_capek": { |
| "delta": 240.06249697515392, |
| "p_value": 0.0, |
| "significant": true |
| }, |
| "benczechmark_cermat_czech_mc": { |
| "delta": -0.08782742681047767, |
| "p_value": 1, |
| "significant": false |
| }, |
| "benczechmark_cermat_czech_open": { |
| "delta": -0.03536977491961415, |
| "p_value": 1, |
| "significant": false |
| }, |
| "benczechmark_cermat_czech_tf": { |
| "delta": -0.14559729676008748, |
| "p_value": 1, |
| "significant": false |
| }, |
| "benczechmark_cermat_czmath_mc": { |
| "delta": -0.047619047619047616, |
| "p_value": 1, |
| "significant": false |
| }, |
| "benczechmark_cermat_czmath_open": { |
| "delta": -0.015723270440251572, |
| "p_value": 1, |
| "significant": false |
| }, |
| "benczechmark_correspondence": { |
| "delta": 451.458596370483, |
| "p_value": 0.0, |
| "significant": true |
| }, |
| "benczechmark_cs_court_decisions_ner": { |
| "delta": -0.3322963251020805, |
| "p_value": 1, |
| "significant": false |
| }, |
| "benczechmark_cs_naturalquestions": { |
| "delta": -0.12326869806094182, |
| "p_value": 1, |
| "significant": false |
| }, |
| "benczechmark_cs_ner": { |
| "delta": -0.3614635043206472, |
| "p_value": 1, |
| "significant": false |
| }, |
| "benczechmark_cs_sqad32": { |
| "delta": -0.34409365023057814, |
| "p_value": 1, |
| "significant": false |
| }, |
| "benczechmark_cs_triviaQA": { |
| "delta": -0.23172143556861696, |
| "p_value": 1, |
| "significant": false |
| }, |
| "benczechmark_csfever_nli": { |
| "delta": -0.27884454859652186, |
| "p_value": 1, |
| "significant": false |
| }, |
| "benczechmark_ctkfacts_nli": { |
| "delta": -0.32262984266666905, |
| "p_value": 1, |
| "significant": false |
| }, |
| "benczechmark_czechnews": { |
| "delta": -0.22089669566958947, |
| "p_value": 1, |
| "significant": false |
| }, |
| "benczechmark_dialect": { |
| "delta": -6.770912659221267, |
| "p_value": 1, |
| "significant": false |
| }, |
| "benczechmark_essay": { |
| "delta": 71.84670299392872, |
| "p_value": 0.0, |
| "significant": true |
| }, |
| "benczechmark_fiction": { |
| "delta": 164.82147280593577, |
| "p_value": 0.0, |
| "significant": true |
| }, |
| "benczechmark_grammarerrorcorrection": { |
| "delta": -0.05826906598114823, |
| "p_value": 1, |
| "significant": false |
| }, |
| "benczechmark_havlicek": { |
| "delta": 3991.3110803775207, |
| "p_value": 0.0, |
| "significant": true |
| }, |
| "benczechmark_hellaswag": { |
| "delta": -0.05058753236407093, |
| "p_value": 1, |
| "significant": false |
| }, |
| "benczechmark_histcorpus": { |
| "delta": 1038.6785739937716, |
| "p_value": 0.0, |
| "significant": true |
| }, |
| "benczechmark_history_ir": { |
| "delta": -0.058614564831261096, |
| "p_value": 1, |
| "significant": false |
| }, |
| "benczechmark_klokan_qa": { |
| "delta": 0.003717472118959092, |
| "p_value": 0.42668593366680074, |
| "significant": false |
| }, |
| "benczechmark_propaganda_argumentace": { |
| "delta": 0.04714157308984501, |
| "p_value": 0.0, |
| "significant": true |
| }, |
| "benczechmark_propaganda_demonizace": { |
| "delta": -0.022150999680386252, |
| "p_value": 1, |
| "significant": false |
| }, |
| "benczechmark_propaganda_emoce": { |
| "delta": -0.1385489687502538, |
| "p_value": 1, |
| "significant": false |
| }, |
| "benczechmark_propaganda_fabulace": { |
| "delta": 0.004528070710645649, |
| "p_value": 0.181323, |
| "significant": false |
| }, |
| "benczechmark_propaganda_lokace": { |
| "delta": -0.12375434541477748, |
| "p_value": 1, |
| "significant": false |
| }, |
| "benczechmark_propaganda_nalepkovani": { |
| "delta": -0.029347098214285605, |
| "p_value": 1, |
| "significant": false |
| }, |
| "benczechmark_propaganda_nazor": { |
| "delta": -0.04434529229648532, |
| "p_value": 1, |
| "significant": false |
| }, |
| "benczechmark_propaganda_relativizace": { |
| "delta": -0.016558138591366456, |
| "p_value": 1, |
| "significant": false |
| }, |
| "benczechmark_propaganda_rusko": { |
| "delta": -0.06725413762335797, |
| "p_value": 1, |
| "significant": false |
| }, |
| "benczechmark_propaganda_strach": { |
| "delta": 0.024566371060218195, |
| "p_value": 0.0, |
| "significant": true |
| }, |
| "benczechmark_propaganda_vina": { |
| "delta": 0.004232313862501669, |
| "p_value": 0.11531, |
| "significant": false |
| }, |
| "benczechmark_propaganda_zamereni": { |
| "delta": -0.10566520187851924, |
| "p_value": 1, |
| "significant": false |
| }, |
| "benczechmark_propaganda_zanr": { |
| "delta": -0.16381559776272536, |
| "p_value": 1, |
| "significant": false |
| }, |
| "benczechmark_sentiment_csfd": { |
| "delta": -0.10722704151621898, |
| "p_value": 1, |
| "significant": false |
| }, |
| "benczechmark_sentiment_fb": { |
| "delta": -0.22684682043234106, |
| "p_value": 1, |
| "significant": false |
| }, |
| "benczechmark_sentiment_mall": { |
| "delta": -0.09747836972631596, |
| "p_value": 1, |
| "significant": false |
| }, |
| "benczechmark_snli": { |
| "delta": -0.28664973940932775, |
| "p_value": 1, |
| "significant": false |
| }, |
| "benczechmark_speeches": { |
| "delta": 136.44134733590116, |
| "p_value": 0.0, |
| "significant": true |
| }, |
| "benczechmark_spoken": { |
| "delta": 54.19426468636969, |
| "p_value": 0.0, |
| "significant": true |
| }, |
| "benczechmark_subjectivity": { |
| "delta": -0.06642359523965014, |
| "p_value": 1, |
| "significant": false |
| }, |
| "benczechmark_summarization": { |
| "delta": -0.02176108681704869, |
| "p_value": 1, |
| "significant": false |
| }, |
| "benczechmark_umimeto_biology": { |
| "delta": -0.21000000000000008, |
| "p_value": 1, |
| "significant": false |
| }, |
| "benczechmark_umimeto_chemistry": { |
| "delta": -0.05999999999999994, |
| "p_value": 1, |
| "significant": false |
| }, |
| "benczechmark_umimeto_czech": { |
| "delta": -0.010000000000000009, |
| "p_value": 1, |
| "significant": false |
| }, |
| "benczechmark_umimeto_history": { |
| "delta": -0.12, |
| "p_value": 1, |
| "significant": false |
| }, |
| "benczechmark_umimeto_informatics": { |
| "delta": -0.28, |
| "p_value": 1, |
| "significant": false |
| }, |
| "benczechmark_umimeto_math": { |
| "delta": -0.13, |
| "p_value": 1, |
| "significant": false |
| }, |
| "benczechmark_umimeto_physics": { |
| "delta": -0.21999999999999997, |
| "p_value": 1, |
| "significant": false |
| } |
| } |
| }, |
| "CohereF_aya-23-_696092a_369b1ca": { |
| "BUT-FIT_CSTinyL_0316dd3_27729a2": { |
| "benczechmark_agree": { |
| "delta": 0.35568280685942955, |
| "p_value": 0.0, |
| "significant": true |
| }, |
| "benczechmark_belebele": { |
| "delta": 0.031635475340863306, |
| "p_value": 0.0, |
| "significant": true |
| }, |
| "benczechmark_capek": { |
| "delta": -55.08467756432839, |
| "p_value": 1, |
| "significant": false |
| }, |
| "benczechmark_cermat_czech_mc": { |
| "delta": 0.20955315870570107, |
| "p_value": 8.482518960068804e-15, |
| "significant": true |
| }, |
| "benczechmark_cermat_czech_open": { |
| "delta": 0.15112540192926044, |
| "p_value": 5.351331562656212e-13, |
| "significant": true |
| }, |
| "benczechmark_cermat_czech_tf": { |
| "delta": 0.2743637899110441, |
| "p_value": 0.0, |
| "significant": true |
| }, |
| "benczechmark_cermat_czmath_mc": { |
| "delta": -0.08730158730158732, |
| "p_value": 1, |
| "significant": false |
| }, |
| "benczechmark_cermat_czmath_open": { |
| "delta": 0.040880503144654086, |
| "p_value": 0.0003638643907341183, |
| "significant": true |
| }, |
| "benczechmark_correspondence": { |
| "delta": -130.0242963464703, |
| "p_value": 1, |
| "significant": false |
| }, |
| "benczechmark_cs_court_decisions_ner": { |
| "delta": 0.5409294186272603, |
| "p_value": 0.0, |
| "significant": true |
| }, |
| "benczechmark_cs_naturalquestions": { |
| "delta": 0.16842105263157894, |
| "p_value": 7.550520451956127e-137, |
| "significant": true |
| }, |
| "benczechmark_cs_ner": { |
| "delta": 0.4419010847582276, |
| "p_value": 0.0, |
| "significant": true |
| }, |
| "benczechmark_cs_sqad32": { |
| "delta": 0.4320681092586023, |
| "p_value": 0.0, |
| "significant": true |
| }, |
| "benczechmark_cs_triviaQA": { |
| "delta": 0.3631668002493543, |
| "p_value": 0.0, |
| "significant": true |
| }, |
| "benczechmark_csfever_nli": { |
| "delta": 0.32454144663822215, |
| "p_value": 0.0, |
| "significant": true |
| }, |
| "benczechmark_ctkfacts_nli": { |
| "delta": 0.3675370523333683, |
| "p_value": 0.0, |
| "significant": true |
| }, |
| "benczechmark_czechnews": { |
| "delta": 0.335012947274522, |
| "p_value": 0.0, |
| "significant": true |
| }, |
| "benczechmark_dialect": { |
| "delta": 27.333522181080255, |
| "p_value": 0.0, |
| "significant": true |
| }, |
| "benczechmark_essay": { |
| "delta": -16.952152053720326, |
| "p_value": 1, |
| "significant": false |
| }, |
| "benczechmark_fiction": { |
| "delta": -64.23837479968857, |
| "p_value": 1, |
| "significant": false |
| }, |
| "benczechmark_grammarerrorcorrection": { |
| "delta": 0.1450299914310197, |
| "p_value": 1.5661300833616094e-56, |
| "significant": true |
| }, |
| "benczechmark_havlicek": { |
| "delta": -1254.7772124189864, |
| "p_value": 1, |
| "significant": false |
| }, |
| "benczechmark_hellaswag": { |
| "delta": 0.1996614220274846, |
| "p_value": 0.0, |
| "significant": true |
| }, |
| "benczechmark_histcorpus": { |
| "delta": -465.6520804217422, |
| "p_value": 1, |
| "significant": false |
| }, |
| "benczechmark_history_ir": { |
| "delta": 0.10390763765541738, |
| "p_value": 1.2837017756524212e-07, |
| "significant": true |
| }, |
| "benczechmark_klokan_qa": { |
| "delta": 0.02106567534076828, |
| "p_value": 0.1446145300205421, |
| "significant": false |
| }, |
| "benczechmark_propaganda_argumentace": { |
| "delta": 0.027887474507235888, |
| "p_value": 0.0, |
| "significant": true |
| }, |
| "benczechmark_propaganda_demonizace": { |
| "delta": 0.12943210140388406, |
| "p_value": 0.0, |
| "significant": true |
| }, |
| "benczechmark_propaganda_emoce": { |
| "delta": 0.18883119627908612, |
| "p_value": 0.000207, |
| "significant": true |
| }, |
| "benczechmark_propaganda_fabulace": { |
| "delta": 0.1353426219396045, |
| "p_value": 0.0, |
| "significant": true |
| }, |
| "benczechmark_propaganda_lokace": { |
| "delta": 0.08569572489459942, |
| "p_value": 0.0, |
| "significant": true |
| }, |
| "benczechmark_propaganda_nalepkovani": { |
| "delta": 0.25419499903477977, |
| "p_value": 0.0, |
| "significant": true |
| }, |
| "benczechmark_propaganda_nazor": { |
| "delta": 0.33963003333588887, |
| "p_value": 0.0, |
| "significant": true |
| }, |
| "benczechmark_propaganda_relativizace": { |
| "delta": 0.23601272935887607, |
| "p_value": 0.0, |
| "significant": true |
| }, |
| "benczechmark_propaganda_rusko": { |
| "delta": 0.06290630715389178, |
| "p_value": 0.11315, |
| "significant": false |
| }, |
| "benczechmark_propaganda_strach": { |
| "delta": 0.23007277275203292, |
| "p_value": 0.0, |
| "significant": true |
| }, |
| "benczechmark_propaganda_vina": { |
| "delta": 0.06506900018876055, |
| "p_value": 0.0, |
| "significant": true |
| }, |
| "benczechmark_propaganda_zamereni": { |
| "delta": 0.29017788773152586, |
| "p_value": 0.0, |
| "significant": true |
| }, |
| "benczechmark_propaganda_zanr": { |
| "delta": 0.2478208718284458, |
| "p_value": 0.0, |
| "significant": true |
| }, |
| "benczechmark_sentiment_csfd": { |
| "delta": 0.1877280254029049, |
| "p_value": 0.0, |
| "significant": true |
| }, |
| "benczechmark_sentiment_fb": { |
| "delta": 0.30035600482172786, |
| "p_value": 0.0, |
| "significant": true |
| }, |
| "benczechmark_sentiment_mall": { |
| "delta": 0.14998168933294065, |
| "p_value": 0.0, |
| "significant": true |
| }, |
| "benczechmark_snli": { |
| "delta": 0.3344246959775491, |
| "p_value": 0.0, |
| "significant": true |
| }, |
| "benczechmark_speeches": { |
| "delta": -21.296491459250547, |
| "p_value": 1, |
| "significant": false |
| }, |
| "benczechmark_spoken": { |
| "delta": 14.976728132910694, |
| "p_value": 0.0, |
| "significant": true |
| }, |
| "benczechmark_subjectivity": { |
| "delta": -0.1460262566115973, |
| "p_value": 1, |
| "significant": false |
| }, |
| "benczechmark_summarization": { |
| "delta": 0.03503876423628012, |
| "p_value": 0.0, |
| "significant": true |
| }, |
| "benczechmark_umimeto_biology": { |
| "delta": 0.4, |
| "p_value": 3.547341178289203e-10, |
| "significant": true |
| }, |
| "benczechmark_umimeto_chemistry": { |
| "delta": 0.16000000000000003, |
| "p_value": 0.005348305652243378, |
| "significant": true |
| }, |
| "benczechmark_umimeto_czech": { |
| "delta": 0.010000000000000009, |
| "p_value": 0.44244405899297884, |
| "significant": false |
| }, |
| "benczechmark_umimeto_history": { |
| "delta": 0.22999999999999998, |
| "p_value": 4.839420830895372e-05, |
| "significant": true |
| }, |
| "benczechmark_umimeto_informatics": { |
| "delta": 0.46, |
| "p_value": 1.1399696115411409e-11, |
| "significant": true |
| }, |
| "benczechmark_umimeto_math": { |
| "delta": 0.26, |
| "p_value": 0.00014188421172165448, |
| "significant": true |
| }, |
| "benczechmark_umimeto_physics": { |
| "delta": 0.13, |
| "p_value": 0.02876569748641526, |
| "significant": true |
| } |
| }, |
| "BUT-FIT_csmpt7b_bdd7d3f_0f64387": { |
| "benczechmark_agree": { |
| "delta": 0.21674023542523613, |
| "p_value": 0.0, |
| "significant": true |
| }, |
| "benczechmark_belebele": { |
| "delta": 0.0050773045215843915, |
| "p_value": 4.9e-05, |
| "significant": true |
| }, |
| "benczechmark_capek": { |
| "delta": -94.06446241163067, |
| "p_value": 1, |
| "significant": false |
| }, |
| "benczechmark_cermat_czech_mc": { |
| "delta": 0.22650231124807396, |
| "p_value": 2.368350059943889e-17, |
| "significant": true |
| }, |
| "benczechmark_cermat_czech_open": { |
| "delta": 0.14147909967845657, |
| "p_value": 3.183060036488657e-12, |
| "significant": true |
| }, |
| "benczechmark_cermat_czech_tf": { |
| "delta": 0.24781444922986706, |
| "p_value": 0.0, |
| "significant": true |
| }, |
| "benczechmark_cermat_czmath_mc": { |
| "delta": -0.031746031746031744, |
| "p_value": 1, |
| "significant": false |
| }, |
| "benczechmark_cermat_czmath_open": { |
| "delta": 0.040880503144654086, |
| "p_value": 0.00013917959396887876, |
| "significant": true |
| }, |
| "benczechmark_correspondence": { |
| "delta": -96.93075562866414, |
| "p_value": 1, |
| "significant": false |
| }, |
| "benczechmark_cs_court_decisions_ner": { |
| "delta": 0.5010694147384794, |
| "p_value": 0.0, |
| "significant": true |
| }, |
| "benczechmark_cs_naturalquestions": { |
| "delta": 0.1673130193905817, |
| "p_value": 6.315259337806227e-135, |
| "significant": true |
| }, |
| "benczechmark_cs_ner": { |
| "delta": 0.4102776245633388, |
| "p_value": 0.0, |
| "significant": true |
| }, |
| "benczechmark_cs_sqad32": { |
| "delta": 0.34409365023057814, |
| "p_value": 1.42463318344124e-219, |
| "significant": true |
| }, |
| "benczechmark_cs_triviaQA": { |
| "delta": 0.3500756968563541, |
| "p_value": 0.0, |
| "significant": true |
| }, |
| "benczechmark_csfever_nli": { |
| "delta": 0.3207201344231242, |
| "p_value": 0.0, |
| "significant": true |
| }, |
| "benczechmark_ctkfacts_nli": { |
| "delta": 0.3630250467919519, |
| "p_value": 0.0, |
| "significant": true |
| }, |
| "benczechmark_czechnews": { |
| "delta": 0.2754138682026783, |
| "p_value": 0.0, |
| "significant": true |
| }, |
| "benczechmark_dialect": { |
| "delta": 38.035024332009655, |
| "p_value": 0.0, |
| "significant": true |
| }, |
| "benczechmark_essay": { |
| "delta": -20.885170812260426, |
| "p_value": 1, |
| "significant": false |
| }, |
| "benczechmark_fiction": { |
| "delta": -71.16050464790334, |
| "p_value": 1, |
| "significant": false |
| }, |
| "benczechmark_grammarerrorcorrection": { |
| "delta": 0.16281062553556125, |
| "p_value": 9.892976866938815e-70, |
| "significant": true |
| }, |
| "benczechmark_havlicek": { |
| "delta": -1349.0475725466624, |
| "p_value": 1, |
| "significant": false |
| }, |
| "benczechmark_hellaswag": { |
| "delta": 0.13104959171479785, |
| "p_value": 5.031176711843308e-153, |
| "significant": true |
| }, |
| "benczechmark_histcorpus": { |
| "delta": -504.7810334760277, |
| "p_value": 1, |
| "significant": false |
| }, |
| "benczechmark_history_ir": { |
| "delta": 0.11367673179396093, |
| "p_value": 5.810355525191201e-09, |
| "significant": true |
| }, |
| "benczechmark_klokan_qa": { |
| "delta": 0.014869888475836451, |
| "p_value": 0.23760176980187475, |
| "significant": false |
| }, |
| "benczechmark_propaganda_argumentace": { |
| "delta": -0.05064633510967809, |
| "p_value": 1, |
| "significant": false |
| }, |
| "benczechmark_propaganda_demonizace": { |
| "delta": 0.013204602572310598, |
| "p_value": 0.001694, |
| "significant": true |
| }, |
| "benczechmark_propaganda_emoce": { |
| "delta": 0.1547432859225376, |
| "p_value": 0.000942, |
| "significant": true |
| }, |
| "benczechmark_propaganda_fabulace": { |
| "delta": -0.0024987593572423084, |
| "p_value": 1, |
| "significant": false |
| }, |
| "benczechmark_propaganda_lokace": { |
| "delta": 0.14714591040155145, |
| "p_value": 0.0, |
| "significant": true |
| }, |
| "benczechmark_propaganda_nalepkovani": { |
| "delta": 0.02170608184996614, |
| "p_value": 0.0, |
| "significant": true |
| }, |
| "benczechmark_propaganda_nazor": { |
| "delta": 0.012349791215253259, |
| "p_value": 0.000458, |
| "significant": true |
| }, |
| "benczechmark_propaganda_relativizace": { |
| "delta": -0.019879909120415373, |
| "p_value": 1, |
| "significant": false |
| }, |
| "benczechmark_propaganda_rusko": { |
| "delta": 0.08400752994845484, |
| "p_value": 0.046179, |
| "significant": true |
| }, |
| "benczechmark_propaganda_strach": { |
| "delta": -0.05576936936936927, |
| "p_value": 1, |
| "significant": false |
| }, |
| "benczechmark_propaganda_vina": { |
| "delta": -0.05224171247959031, |
| "p_value": 1, |
| "significant": false |
| }, |
| "benczechmark_propaganda_zamereni": { |
| "delta": 0.2739630179102599, |
| "p_value": 0.0, |
| "significant": true |
| }, |
| "benczechmark_propaganda_zanr": { |
| "delta": 0.20792430121004446, |
| "p_value": 0.0, |
| "significant": true |
| }, |
| "benczechmark_sentiment_csfd": { |
| "delta": 0.10918610516968574, |
| "p_value": 0.0, |
| "significant": true |
| }, |
| "benczechmark_sentiment_fb": { |
| "delta": 0.23563994544153777, |
| "p_value": 0.0, |
| "significant": true |
| }, |
| "benczechmark_sentiment_mall": { |
| "delta": 0.10235167763400144, |
| "p_value": 0.0, |
| "significant": true |
| }, |
| "benczechmark_snli": { |
| "delta": 0.33115060804490176, |
| "p_value": 0.0, |
| "significant": true |
| }, |
| "benczechmark_speeches": { |
| "delta": -49.24551927294695, |
| "p_value": 1, |
| "significant": false |
| }, |
| "benczechmark_spoken": { |
| "delta": 2.67454040904353, |
| "p_value": 0.0, |
| "significant": true |
| }, |
| "benczechmark_subjectivity": { |
| "delta": -0.1315353291922423, |
| "p_value": 1, |
| "significant": false |
| }, |
| "benczechmark_summarization": { |
| "delta": 0.03916969161217869, |
| "p_value": 0.0, |
| "significant": true |
| }, |
| "benczechmark_umimeto_biology": { |
| "delta": 0.28, |
| "p_value": 1.7778129553630544e-06, |
| "significant": true |
| }, |
| "benczechmark_umimeto_chemistry": { |
| "delta": 0.12, |
| "p_value": 0.04162431262248055, |
| "significant": true |
| }, |
| "benczechmark_umimeto_czech": { |
| "delta": 0.13, |
| "p_value": 0.034233089073427016, |
| "significant": true |
| }, |
| "benczechmark_umimeto_history": { |
| "delta": 0.18999999999999995, |
| "p_value": 0.00023945997506528077, |
| "significant": true |
| }, |
| "benczechmark_umimeto_informatics": { |
| "delta": 0.36, |
| "p_value": 5.3597415612773694e-08, |
| "significant": true |
| }, |
| "benczechmark_umimeto_math": { |
| "delta": 0.22999999999999998, |
| "p_value": 0.000799208324426571, |
| "significant": true |
| }, |
| "benczechmark_umimeto_physics": { |
| "delta": 0.22999999999999998, |
| "p_value": 0.0011752908310954041, |
| "significant": true |
| } |
| }, |
| "CohereF_aya-23-_696092a_369b1ca": { |
| "benczechmark_agree": false, |
| "benczechmark_belebele": false, |
| "benczechmark_capek": false, |
| "benczechmark_cermat_czech_mc": false, |
| "benczechmark_cermat_czech_open": false, |
| "benczechmark_cermat_czech_tf": false, |
| "benczechmark_cermat_czmath_mc": false, |
| "benczechmark_cermat_czmath_open": false, |
| "benczechmark_correspondence": false, |
| "benczechmark_cs_court_decisions_ner": false, |
| "benczechmark_cs_naturalquestions": false, |
| "benczechmark_cs_ner": false, |
| "benczechmark_cs_sqad32": false, |
| "benczechmark_cs_triviaQA": false, |
| "benczechmark_csfever_nli": false, |
| "benczechmark_ctkfacts_nli": false, |
| "benczechmark_czechnews": false, |
| "benczechmark_dialect": false, |
| "benczechmark_essay": false, |
| "benczechmark_fiction": false, |
| "benczechmark_grammarerrorcorrection": false, |
| "benczechmark_havlicek": false, |
| "benczechmark_hellaswag": false, |
| "benczechmark_histcorpus": false, |
| "benczechmark_history_ir": false, |
| "benczechmark_klokan_qa": false, |
| "benczechmark_propaganda_argumentace": false, |
| "benczechmark_propaganda_demonizace": false, |
| "benczechmark_propaganda_emoce": false, |
| "benczechmark_propaganda_fabulace": false, |
| "benczechmark_propaganda_lokace": false, |
| "benczechmark_propaganda_nalepkovani": false, |
| "benczechmark_propaganda_nazor": false, |
| "benczechmark_propaganda_relativizace": false, |
| "benczechmark_propaganda_rusko": false, |
| "benczechmark_propaganda_strach": false, |
| "benczechmark_propaganda_vina": false, |
| "benczechmark_propaganda_zamereni": false, |
| "benczechmark_propaganda_zanr": false, |
| "benczechmark_sentiment_csfd": false, |
| "benczechmark_sentiment_fb": false, |
| "benczechmark_sentiment_mall": false, |
| "benczechmark_snli": false, |
| "benczechmark_speeches": false, |
| "benczechmark_spoken": false, |
| "benczechmark_subjectivity": false, |
| "benczechmark_summarization": false, |
| "benczechmark_umimeto_biology": false, |
| "benczechmark_umimeto_chemistry": false, |
| "benczechmark_umimeto_czech": false, |
| "benczechmark_umimeto_history": false, |
| "benczechmark_umimeto_informatics": false, |
| "benczechmark_umimeto_math": false, |
| "benczechmark_umimeto_physics": false |
| }, |
| "CohereF_aya-23-_78291f9_91bd234": { |
| "benczechmark_agree": { |
| "delta": 0.09932334230071149, |
| "p_value": 0.0, |
| "significant": true |
| }, |
| "benczechmark_belebele": { |
| "delta": 0.01003970490614492, |
| "p_value": 0.0, |
| "significant": true |
| }, |
| "benczechmark_capek": { |
| "delta": 145.99803456352325, |
| "p_value": 0.0, |
| "significant": true |
| }, |
| "benczechmark_cermat_czech_mc": { |
| "delta": 0.1386748844375963, |
| "p_value": 7.826110560855631e-10, |
| "significant": true |
| }, |
| "benczechmark_cermat_czech_open": { |
| "delta": 0.10610932475884244, |
| "p_value": 7.341590281192024e-08, |
| "significant": true |
| }, |
| "benczechmark_cermat_czech_tf": { |
| "delta": 0.103153836254262, |
| "p_value": 0.0, |
| "significant": true |
| }, |
| "benczechmark_cermat_czmath_mc": { |
| "delta": -0.07936507936507936, |
| "p_value": 1, |
| "significant": false |
| }, |
| "benczechmark_cermat_czmath_open": { |
| "delta": 0.025157232704402514, |
| "p_value": 0.010342067855694326, |
| "significant": true |
| }, |
| "benczechmark_correspondence": { |
| "delta": 354.52784074181886, |
| "p_value": 0.0, |
| "significant": true |
| }, |
| "benczechmark_cs_court_decisions_ner": { |
| "delta": 0.16877308963639892, |
| "p_value": 3.735308249911242e-108, |
| "significant": true |
| }, |
| "benczechmark_cs_naturalquestions": { |
| "delta": 0.044044321329639896, |
| "p_value": 2.4506839976957746e-12, |
| "significant": true |
| }, |
| "benczechmark_cs_ner": { |
| "delta": 0.048814120242691605, |
| "p_value": 1.256168703135736e-24, |
| "significant": true |
| }, |
| "benczechmark_cs_sqad32": { |
| "delta": 0.0, |
| "p_value": 0.5, |
| "significant": false |
| }, |
| "benczechmark_cs_triviaQA": { |
| "delta": 0.11835426128773713, |
| "p_value": 6.353727722512506e-161, |
| "significant": true |
| }, |
| "benczechmark_csfever_nli": { |
| "delta": 0.042112138610123195, |
| "p_value": 0.0, |
| "significant": true |
| }, |
| "benczechmark_ctkfacts_nli": { |
| "delta": 0.04442219451754714, |
| "p_value": 0.0, |
| "significant": true |
| }, |
| "benczechmark_czechnews": { |
| "delta": 0.05682547133941318, |
| "p_value": 0.0, |
| "significant": true |
| }, |
| "benczechmark_dialect": { |
| "delta": 31.26411167278839, |
| "p_value": 0.0, |
| "significant": true |
| }, |
| "benczechmark_essay": { |
| "delta": 50.96153218166829, |
| "p_value": 0.0, |
| "significant": true |
| }, |
| "benczechmark_fiction": { |
| "delta": 93.66096815803243, |
| "p_value": 0.0, |
| "significant": true |
| }, |
| "benczechmark_grammarerrorcorrection": { |
| "delta": 0.10454155955441302, |
| "p_value": 4.4321835576703555e-33, |
| "significant": true |
| }, |
| "benczechmark_havlicek": { |
| "delta": 2642.263507830858, |
| "p_value": 0.0, |
| "significant": true |
| }, |
| "benczechmark_hellaswag": { |
| "delta": 0.08046205935072692, |
| "p_value": 1.0106674884003879e-99, |
| "significant": true |
| }, |
| "benczechmark_histcorpus": { |
| "delta": 533.897540517744, |
| "p_value": 0.0, |
| "significant": true |
| }, |
| "benczechmark_history_ir": { |
| "delta": 0.05506216696269983, |
| "p_value": 0.0004816202914721512, |
| "significant": true |
| }, |
| "benczechmark_klokan_qa": { |
| "delta": 0.018587360594795543, |
| "p_value": 0.1475434188516723, |
| "significant": false |
| }, |
| "benczechmark_propaganda_argumentace": { |
| "delta": -0.0030850142969079952, |
| "p_value": 1, |
| "significant": false |
| }, |
| "benczechmark_propaganda_demonizace": { |
| "delta": -0.009153379026243802, |
| "p_value": 1, |
| "significant": false |
| }, |
| "benczechmark_propaganda_emoce": { |
| "delta": 0.03464847620627465, |
| "p_value": 0.259409, |
| "significant": false |
| }, |
| "benczechmark_propaganda_fabulace": { |
| "delta": 0.0016047845787706683, |
| "p_value": 0.293885, |
| "significant": false |
| }, |
| "benczechmark_propaganda_lokace": { |
| "delta": 0.0313878001968273, |
| "p_value": 0.027409, |
| "significant": true |
| }, |
| "benczechmark_propaganda_nalepkovani": { |
| "delta": -0.008238467261904692, |
| "p_value": 1, |
| "significant": false |
| }, |
| "benczechmark_propaganda_nazor": { |
| "delta": -0.03192019841992533, |
| "p_value": 1, |
| "significant": false |
| }, |
| "benczechmark_propaganda_relativizace": { |
| "delta": -0.03643804771178183, |
| "p_value": 1, |
| "significant": false |
| }, |
| "benczechmark_propaganda_rusko": { |
| "delta": 0.018006584693525785, |
| "p_value": 0.364675, |
| "significant": false |
| }, |
| "benczechmark_propaganda_strach": { |
| "delta": -0.03086846846846847, |
| "p_value": 1, |
| "significant": false |
| }, |
| "benczechmark_propaganda_vina": { |
| "delta": -0.0480361610018426, |
| "p_value": 1, |
| "significant": false |
| }, |
| "benczechmark_propaganda_zamereni": { |
| "delta": 0.1702775685245309, |
| "p_value": 0.0, |
| "significant": true |
| }, |
| "benczechmark_propaganda_zanr": { |
| "delta": 0.052606761801089474, |
| "p_value": 0.0, |
| "significant": true |
| }, |
| "benczechmark_sentiment_csfd": { |
| "delta": 0.002391555235682019, |
| "p_value": 0.000254, |
| "significant": true |
| }, |
| "benczechmark_sentiment_fb": { |
| "delta": 0.00980239057576704, |
| "p_value": 0.0, |
| "significant": true |
| }, |
| "benczechmark_sentiment_mall": { |
| "delta": 0.005247331884877027, |
| "p_value": 0.0, |
| "significant": true |
| }, |
| "benczechmark_snli": { |
| "delta": 0.04450086863557401, |
| "p_value": 7.956791515529281e-23, |
| "significant": true |
| }, |
| "benczechmark_speeches": { |
| "delta": 87.1958280629542, |
| "p_value": 0.0, |
| "significant": true |
| }, |
| "benczechmark_spoken": { |
| "delta": 56.86880509541322, |
| "p_value": 0.0, |
| "significant": true |
| }, |
| "benczechmark_subjectivity": { |
| "delta": -0.19795892443189245, |
| "p_value": 1, |
| "significant": false |
| }, |
| "benczechmark_summarization": { |
| "delta": 0.017408604795130046, |
| "p_value": 0.0, |
| "significant": true |
| }, |
| "benczechmark_umimeto_biology": { |
| "delta": 0.06999999999999995, |
| "p_value": 0.07263629795146974, |
| "significant": false |
| }, |
| "benczechmark_umimeto_chemistry": { |
| "delta": 0.06000000000000005, |
| "p_value": 0.12061411335992354, |
| "significant": false |
| }, |
| "benczechmark_umimeto_czech": { |
| "delta": 0.12, |
| "p_value": 0.01659962391514575, |
| "significant": true |
| }, |
| "benczechmark_umimeto_history": { |
| "delta": 0.06999999999999995, |
| "p_value": 0.06362955087062457, |
| "significant": false |
| }, |
| "benczechmark_umimeto_informatics": { |
| "delta": 0.07999999999999996, |
| "p_value": 0.03673774082234953, |
| "significant": true |
| }, |
| "benczechmark_umimeto_math": { |
| "delta": 0.09999999999999998, |
| "p_value": 0.02467578814335497, |
| "significant": true |
| }, |
| "benczechmark_umimeto_physics": { |
| "delta": 0.010000000000000009, |
| "p_value": 0.424254885710078, |
| "significant": false |
| } |
| } |
| }, |
| "CohereF_aya-23-_78291f9_91bd234": { |
| "BUT-FIT_CSTinyL_0316dd3_27729a2": { |
| "benczechmark_agree": { |
| "delta": 0.25614463056675396, |
| "p_value": 0.0, |
| "significant": true |
| }, |
| "benczechmark_belebele": { |
| "delta": 0.021413738920653957, |
| "p_value": 0.0, |
| "significant": true |
| }, |
| "benczechmark_capek": { |
| "delta": -201.08271212785164, |
| "p_value": 1, |
| "significant": false |
| }, |
| "benczechmark_cermat_czech_mc": { |
| "delta": 0.07087827426810478, |
| "p_value": 0.0008499984221900466, |
| "significant": true |
| }, |
| "benczechmark_cermat_czech_open": { |
| "delta": 0.04501607717041801, |
| "p_value": 7.977922032401883e-05, |
| "significant": true |
| }, |
| "benczechmark_cermat_czech_tf": { |
| "delta": 0.17157351562075968, |
| "p_value": 0.0, |
| "significant": true |
| }, |
| "benczechmark_cermat_czmath_mc": { |
| "delta": -0.007936507936507964, |
| "p_value": 1, |
| "significant": false |
| }, |
| "benczechmark_cermat_czmath_open": { |
| "delta": 0.015723270440251572, |
| "p_value": 0.02933296403798324, |
| "significant": true |
| }, |
| "benczechmark_correspondence": { |
| "delta": -484.55213708828916, |
| "p_value": 1, |
| "significant": false |
| }, |
| "benczechmark_cs_court_decisions_ner": { |
| "delta": 0.3721563289908614, |
| "p_value": 0.0, |
| "significant": true |
| }, |
| "benczechmark_cs_naturalquestions": { |
| "delta": 0.12437673130193905, |
| "p_value": 2.640473105680222e-98, |
| "significant": true |
| }, |
| "benczechmark_cs_ner": { |
| "delta": 0.393086964515536, |
| "p_value": 0.0, |
| "significant": true |
| }, |
| "benczechmark_cs_sqad32": { |
| "delta": 0.4320681092586023, |
| "p_value": 0.0, |
| "significant": true |
| }, |
| "benczechmark_cs_triviaQA": { |
| "delta": 0.24481253896161723, |
| "p_value": 0.0, |
| "significant": true |
| }, |
| "benczechmark_csfever_nli": { |
| "delta": 0.28242557534980917, |
| "p_value": 0.0, |
| "significant": true |
| }, |
| "benczechmark_ctkfacts_nli": { |
| "delta": 0.3229439817639449, |
| "p_value": 0.0, |
| "significant": true |
| }, |
| "benczechmark_czechnews": { |
| "delta": 0.2783240543158273, |
| "p_value": 0.0, |
| "significant": true |
| }, |
| "benczechmark_dialect": { |
| "delta": -3.930589491708133, |
| "p_value": 1, |
| "significant": false |
| }, |
| "benczechmark_essay": { |
| "delta": -67.91368423538862, |
| "p_value": 1, |
| "significant": false |
| }, |
| "benczechmark_fiction": { |
| "delta": -157.899342957721, |
| "p_value": 1, |
| "significant": false |
| }, |
| "benczechmark_grammarerrorcorrection": { |
| "delta": 0.04048843187660667, |
| "p_value": 2.0632788855429967e-09, |
| "significant": true |
| }, |
| "benczechmark_havlicek": { |
| "delta": -3897.0407202498445, |
| "p_value": 1, |
| "significant": false |
| }, |
| "benczechmark_hellaswag": { |
| "delta": 0.11919936267675768, |
| "p_value": 3.36511221021426e-128, |
| "significant": true |
| }, |
| "benczechmark_histcorpus": { |
| "delta": -999.5496209394862, |
| "p_value": 1, |
| "significant": false |
| }, |
| "benczechmark_history_ir": { |
| "delta": 0.04884547069271755, |
| "p_value": 0.008361229992533544, |
| "significant": true |
| }, |
| "benczechmark_klokan_qa": { |
| "delta": 0.0024783147459727373, |
| "p_value": 0.4486867028903291, |
| "significant": false |
| }, |
| "benczechmark_propaganda_argumentace": { |
| "delta": 0.031286618997990234, |
| "p_value": 0.0, |
| "significant": true |
| }, |
| "benczechmark_propaganda_demonizace": { |
| "delta": 0.137235880173473, |
| "p_value": 0.0, |
| "significant": true |
| }, |
| "benczechmark_propaganda_emoce": { |
| "delta": 0.15377936749080162, |
| "p_value": 0.000791, |
| "significant": true |
| }, |
| "benczechmark_propaganda_fabulace": { |
| "delta": 0.1335419193343514, |
| "p_value": 0.0, |
| "significant": true |
| }, |
| "benczechmark_propaganda_lokace": { |
| "delta": 0.05703366915634944, |
| "p_value": 0.000521, |
| "significant": true |
| }, |
| "benczechmark_propaganda_nalepkovani": { |
| "delta": 0.26146580941850567, |
| "p_value": 0.0, |
| "significant": true |
| }, |
| "benczechmark_propaganda_nazor": { |
| "delta": 0.3712025066352686, |
| "p_value": 0.0, |
| "significant": true |
| }, |
| "benczechmark_propaganda_relativizace": { |
| "delta": 0.27423801330676295, |
| "p_value": 0.0, |
| "significant": true |
| }, |
| "benczechmark_propaganda_rusko": { |
| "delta": 0.04487910774872861, |
| "p_value": 0.193682, |
| "significant": false |
| }, |
| "benczechmark_propaganda_strach": { |
| "delta": 0.2595679846781028, |
| "p_value": 0.0, |
| "significant": true |
| }, |
| "benczechmark_propaganda_vina": { |
| "delta": 0.11254494210808363, |
| "p_value": 0.0, |
| "significant": true |
| }, |
| "benczechmark_propaganda_zamereni": { |
| "delta": 0.11932268864355511, |
| "p_value": 0.0, |
| "significant": true |
| }, |
| "benczechmark_propaganda_zanr": { |
| "delta": 0.1959233872560372, |
| "p_value": 0.0, |
| "significant": true |
| }, |
| "benczechmark_sentiment_csfd": { |
| "delta": 0.18535551528039917, |
| "p_value": 0.0, |
| "significant": true |
| }, |
| "benczechmark_sentiment_fb": { |
| "delta": 0.2905784996117139, |
| "p_value": 0.0, |
| "significant": true |
| }, |
| "benczechmark_sentiment_mall": { |
| "delta": 0.14473971232796545, |
| "p_value": 0.0, |
| "significant": true |
| }, |
| "benczechmark_snli": { |
| "delta": 0.2899238273419751, |
| "p_value": 0.0, |
| "significant": true |
| }, |
| "benczechmark_speeches": { |
| "delta": -108.49231952220475, |
| "p_value": 1, |
| "significant": false |
| }, |
| "benczechmark_spoken": { |
| "delta": -41.89207696250253, |
| "p_value": 1, |
| "significant": false |
| }, |
| "benczechmark_subjectivity": { |
| "delta": 0.0518251874848793, |
| "p_value": 0.0, |
| "significant": true |
| }, |
| "benczechmark_summarization": { |
| "delta": 0.017630159441150074, |
| "p_value": 0.0, |
| "significant": true |
| }, |
| "benczechmark_umimeto_biology": { |
| "delta": 0.33000000000000007, |
| "p_value": 9.676867892513548e-07, |
| "significant": true |
| }, |
| "benczechmark_umimeto_chemistry": { |
| "delta": 0.09999999999999998, |
| "p_value": 0.08333061452982428, |
| "significant": false |
| }, |
| "benczechmark_umimeto_czech": { |
| "delta": -0.10999999999999999, |
| "p_value": 1, |
| "significant": false |
| }, |
| "benczechmark_umimeto_history": { |
| "delta": 0.16000000000000003, |
| "p_value": 0.008775649916884726, |
| "significant": true |
| }, |
| "benczechmark_umimeto_informatics": { |
| "delta": 0.38000000000000006, |
| "p_value": 3.4297885989720324e-09, |
| "significant": true |
| }, |
| "benczechmark_umimeto_math": { |
| "delta": 0.16000000000000003, |
| "p_value": 0.022467362606315162, |
| "significant": true |
| }, |
| "benczechmark_umimeto_physics": { |
| "delta": 0.12, |
| "p_value": 0.028690585123921343, |
| "significant": true |
| } |
| }, |
| "BUT-FIT_csmpt7b_bdd7d3f_0f64387": { |
| "benczechmark_agree": { |
| "delta": 0.11716403228835615, |
| "p_value": 0.0, |
| "significant": true |
| }, |
| "benczechmark_belebele": { |
| "delta": -0.005063830890525867, |
| "p_value": 1, |
| "significant": false |
| }, |
| "benczechmark_capek": { |
| "delta": -240.06249697515392, |
| "p_value": 1, |
| "significant": false |
| }, |
| "benczechmark_cermat_czech_mc": { |
| "delta": 0.08782742681047767, |
| "p_value": 0.00010039276369893295, |
| "significant": true |
| }, |
| "benczechmark_cermat_czech_open": { |
| "delta": 0.03536977491961415, |
| "p_value": 0.0010823918862926496, |
| "significant": true |
| }, |
| "benczechmark_cermat_czech_tf": { |
| "delta": 0.14475309201624298, |
| "p_value": 0.0, |
| "significant": true |
| }, |
| "benczechmark_cermat_czmath_mc": { |
| "delta": 0.047619047619047616, |
| "p_value": 0.18389967039468963, |
| "significant": false |
| }, |
| "benczechmark_cermat_czmath_open": { |
| "delta": 0.015723270440251572, |
| "p_value": 0.012557280661300385, |
| "significant": true |
| }, |
| "benczechmark_correspondence": { |
| "delta": -451.458596370483, |
| "p_value": 1, |
| "significant": false |
| }, |
| "benczechmark_cs_court_decisions_ner": { |
| "delta": 0.3322963251020805, |
| "p_value": 0.0, |
| "significant": true |
| }, |
| "benczechmark_cs_naturalquestions": { |
| "delta": 0.12326869806094182, |
| "p_value": 4.8683355521684166e-99, |
| "significant": true |
| }, |
| "benczechmark_cs_ner": { |
| "delta": 0.3614635043206472, |
| "p_value": 0.0, |
| "significant": true |
| }, |
| "benczechmark_cs_sqad32": { |
| "delta": 0.34409365023057814, |
| "p_value": 2.6379948220638095e-226, |
| "significant": true |
| }, |
| "benczechmark_cs_triviaQA": { |
| "delta": 0.23172143556861696, |
| "p_value": 0.0, |
| "significant": true |
| }, |
| "benczechmark_csfever_nli": { |
| "delta": 0.278598997779703, |
| "p_value": 0.0, |
| "significant": true |
| }, |
| "benczechmark_ctkfacts_nli": { |
| "delta": 0.31877310946855864, |
| "p_value": 0.0, |
| "significant": true |
| }, |
| "benczechmark_czechnews": { |
| "delta": 0.21859078006947974, |
| "p_value": 0.0, |
| "significant": true |
| }, |
| "benczechmark_dialect": { |
| "delta": 6.770912659221267, |
| "p_value": 0.0, |
| "significant": true |
| }, |
| "benczechmark_essay": { |
| "delta": -71.84670299392872, |
| "p_value": 1, |
| "significant": false |
| }, |
| "benczechmark_fiction": { |
| "delta": -164.82147280593577, |
| "p_value": 1, |
| "significant": false |
| }, |
| "benczechmark_grammarerrorcorrection": { |
| "delta": 0.05826906598114823, |
| "p_value": 6.47368850844115e-12, |
| "significant": true |
| }, |
| "benczechmark_havlicek": { |
| "delta": -3991.3110803775207, |
| "p_value": 1, |
| "significant": false |
| }, |
| "benczechmark_hellaswag": { |
| "delta": 0.05058753236407093, |
| "p_value": 3.0180088113385055e-25, |
| "significant": true |
| }, |
| "benczechmark_histcorpus": { |
| "delta": -1038.6785739937716, |
| "p_value": 1, |
| "significant": false |
| }, |
| "benczechmark_history_ir": { |
| "delta": 0.058614564831261096, |
| "p_value": 0.0015614555519679043, |
| "significant": true |
| }, |
| "benczechmark_klokan_qa": { |
| "delta": -0.003717472118959092, |
| "p_value": 1, |
| "significant": false |
| }, |
| "benczechmark_propaganda_argumentace": { |
| "delta": -0.04756132081277009, |
| "p_value": 1, |
| "significant": false |
| }, |
| "benczechmark_propaganda_demonizace": { |
| "delta": 0.021372363344823397, |
| "p_value": 0.0, |
| "significant": true |
| }, |
| "benczechmark_propaganda_emoce": { |
| "delta": 0.11992853179012164, |
| "p_value": 0.011483, |
| "significant": true |
| }, |
| "benczechmark_propaganda_fabulace": { |
| "delta": -0.004069657722389763, |
| "p_value": 1, |
| "significant": false |
| }, |
| "benczechmark_propaganda_lokace": { |
| "delta": 0.11736471795874959, |
| "p_value": 0.0, |
| "significant": true |
| }, |
| "benczechmark_propaganda_nalepkovani": { |
| "delta": 0.02933064466516, |
| "p_value": 0.0, |
| "significant": true |
| }, |
| "benczechmark_propaganda_nazor": { |
| "delta": 0.04409677028457526, |
| "p_value": 0.0, |
| "significant": true |
| }, |
| "benczechmark_propaganda_relativizace": { |
| "delta": 0.017982540539464686, |
| "p_value": 0.000238, |
| "significant": true |
| }, |
| "benczechmark_propaganda_rusko": { |
| "delta": 0.06299738387017295, |
| "p_value": 0.113578, |
| "significant": false |
| }, |
| "benczechmark_propaganda_strach": { |
| "delta": -0.024900900900900802, |
| "p_value": 1, |
| "significant": false |
| }, |
| "benczechmark_propaganda_vina": { |
| "delta": -0.004205551477747704, |
| "p_value": 1, |
| "significant": false |
| }, |
| "benczechmark_propaganda_zamereni": { |
| "delta": 0.10334637032721083, |
| "p_value": 0.0, |
| "significant": true |
| }, |
| "benczechmark_propaganda_zanr": { |
| "delta": 0.15587237299853673, |
| "p_value": 0.0, |
| "significant": true |
| }, |
| "benczechmark_sentiment_csfd": { |
| "delta": 0.1067887820906519, |
| "p_value": 0.0, |
| "significant": true |
| }, |
| "benczechmark_sentiment_fb": { |
| "delta": 0.22585191094330992, |
| "p_value": 0.0, |
| "significant": true |
| }, |
| "benczechmark_sentiment_mall": { |
| "delta": 0.0970698289268569, |
| "p_value": 0.0, |
| "significant": true |
| }, |
| "benczechmark_snli": { |
| "delta": 0.28664973940932775, |
| "p_value": 0.0, |
| "significant": true |
| }, |
| "benczechmark_speeches": { |
| "delta": -136.44134733590116, |
| "p_value": 1, |
| "significant": false |
| }, |
| "benczechmark_spoken": { |
| "delta": -54.19426468636969, |
| "p_value": 1, |
| "significant": false |
| }, |
| "benczechmark_subjectivity": { |
| "delta": 0.06631240185610632, |
| "p_value": 0.0, |
| "significant": true |
| }, |
| "benczechmark_summarization": { |
| "delta": 0.021761086817048646, |
| "p_value": 0.0, |
| "significant": true |
| }, |
| "benczechmark_umimeto_biology": { |
| "delta": 0.21000000000000008, |
| "p_value": 0.0007176269279768656, |
| "significant": true |
| }, |
| "benczechmark_umimeto_chemistry": { |
| "delta": 0.05999999999999994, |
| "p_value": 0.19942729605267112, |
| "significant": false |
| }, |
| "benczechmark_umimeto_czech": { |
| "delta": 0.010000000000000009, |
| "p_value": 0.4457787095226334, |
| "significant": false |
| }, |
| "benczechmark_umimeto_history": { |
| "delta": 0.12, |
| "p_value": 0.025545229104680277, |
| "significant": true |
| }, |
| "benczechmark_umimeto_informatics": { |
| "delta": 0.28, |
| "p_value": 4.229216349661341e-05, |
| "significant": true |
| }, |
| "benczechmark_umimeto_math": { |
| "delta": 0.13, |
| "p_value": 0.05085721137150555, |
| "significant": false |
| }, |
| "benczechmark_umimeto_physics": { |
| "delta": 0.21999999999999997, |
| "p_value": 0.0009603912042542502, |
| "significant": true |
| } |
| }, |
| "CohereF_aya-23-_696092a_369b1ca": { |
| "benczechmark_agree": { |
| "delta": -0.1011368645567794, |
| "p_value": 1, |
| "significant": false |
| }, |
| "benczechmark_belebele": { |
| "delta": -0.010194489328797185, |
| "p_value": 1, |
| "significant": false |
| }, |
| "benczechmark_capek": { |
| "delta": -145.99803456352325, |
| "p_value": 1, |
| "significant": false |
| }, |
| "benczechmark_cermat_czech_mc": { |
| "delta": -0.1386748844375963, |
| "p_value": 1, |
| "significant": false |
| }, |
| "benczechmark_cermat_czech_open": { |
| "delta": -0.10610932475884244, |
| "p_value": 1, |
| "significant": false |
| }, |
| "benczechmark_cermat_czech_tf": { |
| "delta": -0.10381633870005968, |
| "p_value": 1, |
| "significant": false |
| }, |
| "benczechmark_cermat_czmath_mc": { |
| "delta": 0.07936507936507936, |
| "p_value": 0.043187570313197934, |
| "significant": true |
| }, |
| "benczechmark_cermat_czmath_open": { |
| "delta": -0.025157232704402514, |
| "p_value": 1, |
| "significant": false |
| }, |
| "benczechmark_correspondence": { |
| "delta": -354.52784074181886, |
| "p_value": 1, |
| "significant": false |
| }, |
| "benczechmark_cs_court_decisions_ner": { |
| "delta": -0.16877308963639892, |
| "p_value": 1, |
| "significant": false |
| }, |
| "benczechmark_cs_naturalquestions": { |
| "delta": -0.044044321329639896, |
| "p_value": 1, |
| "significant": false |
| }, |
| "benczechmark_cs_ner": { |
| "delta": -0.048814120242691605, |
| "p_value": 1, |
| "significant": false |
| }, |
| "benczechmark_cs_sqad32": { |
| "delta": 0.0, |
| "p_value": 0.5, |
| "significant": false |
| }, |
| "benczechmark_cs_triviaQA": { |
| "delta": -0.11835426128773713, |
| "p_value": 1, |
| "significant": false |
| }, |
| "benczechmark_csfever_nli": { |
| "delta": -0.04215499320129845, |
| "p_value": 1, |
| "significant": false |
| }, |
| "benczechmark_ctkfacts_nli": { |
| "delta": -0.0450729506855726, |
| "p_value": 1, |
| "significant": false |
| }, |
| "benczechmark_czechnews": { |
| "delta": -0.057340479269040956, |
| "p_value": 1, |
| "significant": false |
| }, |
| "benczechmark_dialect": { |
| "delta": -31.26411167278839, |
| "p_value": 1, |
| "significant": false |
| }, |
| "benczechmark_essay": { |
| "delta": -50.96153218166829, |
| "p_value": 1, |
| "significant": false |
| }, |
| "benczechmark_fiction": { |
| "delta": -93.66096815803243, |
| "p_value": 1, |
| "significant": false |
| }, |
| "benczechmark_grammarerrorcorrection": { |
| "delta": -0.10454155955441302, |
| "p_value": 1, |
| "significant": false |
| }, |
| "benczechmark_havlicek": { |
| "delta": -2642.263507830858, |
| "p_value": 1, |
| "significant": false |
| }, |
| "benczechmark_hellaswag": { |
| "delta": -0.08046205935072692, |
| "p_value": 1, |
| "significant": false |
| }, |
| "benczechmark_histcorpus": { |
| "delta": -533.897540517744, |
| "p_value": 1, |
| "significant": false |
| }, |
| "benczechmark_history_ir": { |
| "delta": -0.05506216696269983, |
| "p_value": 1, |
| "significant": false |
| }, |
| "benczechmark_klokan_qa": { |
| "delta": -0.018587360594795543, |
| "p_value": 1, |
| "significant": false |
| }, |
| "benczechmark_propaganda_argumentace": { |
| "delta": 0.003464040694516779, |
| "p_value": 0.047219, |
| "significant": true |
| }, |
| "benczechmark_propaganda_demonizace": { |
| "delta": 0.00797798684626505, |
| "p_value": 0.014886, |
| "significant": true |
| }, |
| "benczechmark_propaganda_emoce": { |
| "delta": -0.03762121039457278, |
| "p_value": 1, |
| "significant": false |
| }, |
| "benczechmark_propaganda_fabulace": { |
| "delta": -0.001570898365147455, |
| "p_value": 1, |
| "significant": false |
| }, |
| "benczechmark_propaganda_lokace": { |
| "delta": -0.026325735696193564, |
| "p_value": 1, |
| "significant": false |
| }, |
| "benczechmark_propaganda_nalepkovani": { |
| "delta": 0.0073865427446212495, |
| "p_value": 0.000237, |
| "significant": true |
| }, |
| "benczechmark_propaganda_nazor": { |
| "delta": 0.031751959916096695, |
| "p_value": 0.0, |
| "significant": true |
| }, |
| "benczechmark_propaganda_relativizace": { |
| "delta": 0.038360149297259505, |
| "p_value": 0.0, |
| "significant": true |
| }, |
| "benczechmark_propaganda_rusko": { |
| "delta": -0.025915776070387975, |
| "p_value": 1, |
| "significant": false |
| }, |
| "benczechmark_propaganda_strach": { |
| "delta": 0.029769737328192325, |
| "p_value": 0.0, |
| "significant": true |
| }, |
| "benczechmark_propaganda_vina": { |
| "delta": 0.04736962052018423, |
| "p_value": 0.0, |
| "significant": true |
| }, |
| "benczechmark_propaganda_zamereni": { |
| "delta": -0.17568408267218205, |
| "p_value": 1, |
| "significant": false |
| }, |
| "benczechmark_propaganda_zanr": { |
| "delta": -0.05551548113797944, |
| "p_value": 1, |
| "significant": false |
| }, |
| "benczechmark_sentiment_csfd": { |
| "delta": -0.002394380596835499, |
| "p_value": 1, |
| "significant": false |
| }, |
| "benczechmark_sentiment_fb": { |
| "delta": -0.009797675830254704, |
| "p_value": 1, |
| "significant": false |
| }, |
| "benczechmark_sentiment_mall": { |
| "delta": -0.005261277272754605, |
| "p_value": 1, |
| "significant": false |
| }, |
| "benczechmark_snli": { |
| "delta": -0.04450086863557401, |
| "p_value": 1, |
| "significant": false |
| }, |
| "benczechmark_speeches": { |
| "delta": -87.1958280629542, |
| "p_value": 1, |
| "significant": false |
| }, |
| "benczechmark_spoken": { |
| "delta": -56.86880509541322, |
| "p_value": 1, |
| "significant": false |
| }, |
| "benczechmark_subjectivity": { |
| "delta": 0.1975800710055951, |
| "p_value": 0.0, |
| "significant": true |
| }, |
| "benczechmark_summarization": { |
| "delta": -0.017408604795130067, |
| "p_value": 1, |
| "significant": false |
| }, |
| "benczechmark_umimeto_biology": { |
| "delta": -0.06999999999999995, |
| "p_value": 1, |
| "significant": false |
| }, |
| "benczechmark_umimeto_chemistry": { |
| "delta": -0.06000000000000005, |
| "p_value": 1, |
| "significant": false |
| }, |
| "benczechmark_umimeto_czech": { |
| "delta": -0.12, |
| "p_value": 1, |
| "significant": false |
| }, |
| "benczechmark_umimeto_history": { |
| "delta": -0.06999999999999995, |
| "p_value": 1, |
| "significant": false |
| }, |
| "benczechmark_umimeto_informatics": { |
| "delta": -0.07999999999999996, |
| "p_value": 1, |
| "significant": false |
| }, |
| "benczechmark_umimeto_math": { |
| "delta": -0.09999999999999998, |
| "p_value": 1, |
| "significant": false |
| }, |
| "benczechmark_umimeto_physics": { |
| "delta": -0.010000000000000009, |
| "p_value": 1, |
| "significant": false |
| } |
| }, |
| "CohereF_aya-23-_78291f9_91bd234": { |
| "benczechmark_agree": false, |
| "benczechmark_belebele": false, |
| "benczechmark_capek": false, |
| "benczechmark_cermat_czech_mc": false, |
| "benczechmark_cermat_czech_open": false, |
| "benczechmark_cermat_czech_tf": false, |
| "benczechmark_cermat_czmath_mc": false, |
| "benczechmark_cermat_czmath_open": false, |
| "benczechmark_correspondence": false, |
| "benczechmark_cs_court_decisions_ner": false, |
| "benczechmark_cs_naturalquestions": false, |
| "benczechmark_cs_ner": false, |
| "benczechmark_cs_sqad32": false, |
| "benczechmark_cs_triviaQA": false, |
| "benczechmark_csfever_nli": false, |
| "benczechmark_ctkfacts_nli": false, |
| "benczechmark_czechnews": false, |
| "benczechmark_dialect": false, |
| "benczechmark_essay": false, |
| "benczechmark_fiction": false, |
| "benczechmark_grammarerrorcorrection": false, |
| "benczechmark_havlicek": false, |
| "benczechmark_hellaswag": false, |
| "benczechmark_histcorpus": false, |
| "benczechmark_history_ir": false, |
| "benczechmark_klokan_qa": false, |
| "benczechmark_propaganda_argumentace": false, |
| "benczechmark_propaganda_demonizace": false, |
| "benczechmark_propaganda_emoce": false, |
| "benczechmark_propaganda_fabulace": false, |
| "benczechmark_propaganda_lokace": false, |
| "benczechmark_propaganda_nalepkovani": false, |
| "benczechmark_propaganda_nazor": false, |
| "benczechmark_propaganda_relativizace": false, |
| "benczechmark_propaganda_rusko": false, |
| "benczechmark_propaganda_strach": false, |
| "benczechmark_propaganda_vina": false, |
| "benczechmark_propaganda_zamereni": false, |
| "benczechmark_propaganda_zanr": false, |
| "benczechmark_sentiment_csfd": false, |
| "benczechmark_sentiment_fb": false, |
| "benczechmark_sentiment_mall": false, |
| "benczechmark_snli": false, |
| "benczechmark_speeches": false, |
| "benczechmark_spoken": false, |
| "benczechmark_subjectivity": false, |
| "benczechmark_summarization": false, |
| "benczechmark_umimeto_biology": false, |
| "benczechmark_umimeto_chemistry": false, |
| "benczechmark_umimeto_czech": false, |
| "benczechmark_umimeto_history": false, |
| "benczechmark_umimeto_informatics": false, |
| "benczechmark_umimeto_math": false, |
| "benczechmark_umimeto_physics": false |
| } |
| } |
| } |