Upload folder using huggingface_hub
Browse files- results_Qwen__Qwen2.5-0.5B-Instruct_few_shot.json +30 -0
- results_Qwen__Qwen2.5-0.5B-Instruct_zero_shot.json +30 -0
- results_Qwen__Qwen2.5-1.5B-Instruct_few_shot.json +30 -0
- results_Qwen__Qwen2.5-1.5B-Instruct_zero_shot.json +30 -0
- results_Qwen__Qwen2.5-3B-Instruct_few_shot.json +30 -0
- results_Qwen__Qwen2.5-3B-Instruct_zero_shot.json +30 -0
- results_Qwen__Qwen2.5-7B-Instruct_few_shot.json +30 -0
- results_Qwen__Qwen2.5-7B-Instruct_zero_shot.json +30 -0
- results_Qwen__Qwen3-0.6B_few_shot.json +30 -0
- results_Qwen__Qwen3-0.6B_zero_shot.json +30 -0
- results_Qwen__Qwen3-1.7B_few_shot.json +30 -0
- results_Qwen__Qwen3-1.7B_zero_shot.json +30 -0
- results_Qwen__Qwen3-4B_few_shot.json +30 -0
- results_Qwen__Qwen3-4B_zero_shot.json +30 -0
- results_Qwen__Qwen3-8B_few_shot.json +30 -0
- results_Qwen__Qwen3-8B_zero_shot.json +30 -0
- results_google__gemma-3-1b-it_few_shot.json +30 -0
- results_google__gemma-3-1b-it_zero_shot.json +30 -0
- results_google__gemma-3-270m_few_shot.json +30 -0
- results_google__gemma-3-270m_zero_shot.json +30 -0
- results_google__gemma-3-4b-it_few_shot.json +30 -0
- results_google__gemma-3-4b-it_zero_shot.json +30 -0
- results_meta-llama__Llama-3.1-8B-Instruct_few_shot.json +30 -0
- results_meta-llama__Llama-3.1-8B-Instruct_zero_shot.json +30 -0
- results_meta-llama__Llama-3.2-1B-Instruct_few_shot.json +30 -0
- results_meta-llama__Llama-3.2-1B-Instruct_zero_shot.json +30 -0
- results_meta-llama__Llama-3.2-3B-Instruct_few_shot.json +30 -0
- results_meta-llama__Llama-3.2-3B-Instruct_zero_shot.json +30 -0
results_Qwen__Qwen2.5-0.5B-Instruct_few_shot.json
ADDED
|
@@ -0,0 +1,30 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
{
|
| 2 |
+
"config": {
|
| 3 |
+
"model_name": "Qwen/Qwen2.5-0.5B-Instruct",
|
| 4 |
+
"model_dtype": "float16",
|
| 5 |
+
"model_sha": "main"
|
| 6 |
+
},
|
| 7 |
+
"results": {
|
| 8 |
+
"KyrgyzMMLU": {
|
| 9 |
+
"metric_name": 0.254
|
| 10 |
+
},
|
| 11 |
+
"KyrgyzRC": {
|
| 12 |
+
"metric_name": 0.54
|
| 13 |
+
},
|
| 14 |
+
"WinoGrande": {
|
| 15 |
+
"metric_name": 0.49700000000000005
|
| 16 |
+
},
|
| 17 |
+
"BoolQ": {
|
| 18 |
+
"metric_name": 0.61
|
| 19 |
+
},
|
| 20 |
+
"HellaSwag": {
|
| 21 |
+
"metric_name": 0.259
|
| 22 |
+
},
|
| 23 |
+
"GSM8K": {
|
| 24 |
+
"metric_name": 0.022000000000000002
|
| 25 |
+
},
|
| 26 |
+
"TruthfulQA": {
|
| 27 |
+
"metric_name": 0.33399999999999996
|
| 28 |
+
}
|
| 29 |
+
}
|
| 30 |
+
}
|
results_Qwen__Qwen2.5-0.5B-Instruct_zero_shot.json
ADDED
|
@@ -0,0 +1,30 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
{
|
| 2 |
+
"config": {
|
| 3 |
+
"model_name": "Qwen/Qwen2.5-0.5B-Instruct",
|
| 4 |
+
"model_dtype": "float16",
|
| 5 |
+
"model_sha": "main"
|
| 6 |
+
},
|
| 7 |
+
"results": {
|
| 8 |
+
"KyrgyzMMLU": {
|
| 9 |
+
"metric_name": 0.27399999999999997
|
| 10 |
+
},
|
| 11 |
+
"KyrgyzRC": {
|
| 12 |
+
"metric_name": 0.532
|
| 13 |
+
},
|
| 14 |
+
"WinoGrande": {
|
| 15 |
+
"metric_name": 0.515
|
| 16 |
+
},
|
| 17 |
+
"BoolQ": {
|
| 18 |
+
"metric_name": 0.379
|
| 19 |
+
},
|
| 20 |
+
"HellaSwag": {
|
| 21 |
+
"metric_name": 0.146
|
| 22 |
+
},
|
| 23 |
+
"GSM8K": {
|
| 24 |
+
"metric_name": 0.006999999999999999
|
| 25 |
+
},
|
| 26 |
+
"TruthfulQA": {
|
| 27 |
+
"metric_name": 0.335
|
| 28 |
+
}
|
| 29 |
+
}
|
| 30 |
+
}
|
results_Qwen__Qwen2.5-1.5B-Instruct_few_shot.json
ADDED
|
@@ -0,0 +1,30 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
{
|
| 2 |
+
"config": {
|
| 3 |
+
"model_name": "Qwen/Qwen2.5-1.5B-Instruct",
|
| 4 |
+
"model_dtype": "float16",
|
| 5 |
+
"model_sha": "main"
|
| 6 |
+
},
|
| 7 |
+
"results": {
|
| 8 |
+
"KyrgyzMMLU": {
|
| 9 |
+
"metric_name": 0.287
|
| 10 |
+
},
|
| 11 |
+
"KyrgyzRC": {
|
| 12 |
+
"metric_name": 0.675
|
| 13 |
+
},
|
| 14 |
+
"WinoGrande": {
|
| 15 |
+
"metric_name": 0.501
|
| 16 |
+
},
|
| 17 |
+
"BoolQ": {
|
| 18 |
+
"metric_name": 0.58
|
| 19 |
+
},
|
| 20 |
+
"HellaSwag": {
|
| 21 |
+
"metric_name": 0.265
|
| 22 |
+
},
|
| 23 |
+
"GSM8K": {
|
| 24 |
+
"metric_name": 0.061
|
| 25 |
+
},
|
| 26 |
+
"TruthfulQA": {
|
| 27 |
+
"metric_name": 0.32899999999999996
|
| 28 |
+
}
|
| 29 |
+
}
|
| 30 |
+
}
|
results_Qwen__Qwen2.5-1.5B-Instruct_zero_shot.json
ADDED
|
@@ -0,0 +1,30 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
{
|
| 2 |
+
"config": {
|
| 3 |
+
"model_name": "Qwen/Qwen2.5-1.5B-Instruct",
|
| 4 |
+
"model_dtype": "float16",
|
| 5 |
+
"model_sha": "main"
|
| 6 |
+
},
|
| 7 |
+
"results": {
|
| 8 |
+
"KyrgyzMMLU": {
|
| 9 |
+
"metric_name": 0.27899999999999997
|
| 10 |
+
},
|
| 11 |
+
"KyrgyzRC": {
|
| 12 |
+
"metric_name": 0.605
|
| 13 |
+
},
|
| 14 |
+
"WinoGrande": {
|
| 15 |
+
"metric_name": 0.501
|
| 16 |
+
},
|
| 17 |
+
"BoolQ": {
|
| 18 |
+
"metric_name": 0.386
|
| 19 |
+
},
|
| 20 |
+
"HellaSwag": {
|
| 21 |
+
"metric_name": 0.22899999999999998
|
| 22 |
+
},
|
| 23 |
+
"GSM8K": {
|
| 24 |
+
"metric_name": 0.006999999999999999
|
| 25 |
+
},
|
| 26 |
+
"TruthfulQA": {
|
| 27 |
+
"metric_name": 0.325
|
| 28 |
+
}
|
| 29 |
+
}
|
| 30 |
+
}
|
results_Qwen__Qwen2.5-3B-Instruct_few_shot.json
ADDED
|
@@ -0,0 +1,30 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
{
|
| 2 |
+
"config": {
|
| 3 |
+
"model_name": "Qwen/Qwen2.5-3B-Instruct",
|
| 4 |
+
"model_dtype": "float16",
|
| 5 |
+
"model_sha": "main"
|
| 6 |
+
},
|
| 7 |
+
"results": {
|
| 8 |
+
"KyrgyzMMLU": {
|
| 9 |
+
"metric_name": 0.34
|
| 10 |
+
},
|
| 11 |
+
"KyrgyzRC": {
|
| 12 |
+
"metric_name": 0.732
|
| 13 |
+
},
|
| 14 |
+
"WinoGrande": {
|
| 15 |
+
"metric_name": 0.513
|
| 16 |
+
},
|
| 17 |
+
"BoolQ": {
|
| 18 |
+
"metric_name": 0.574
|
| 19 |
+
},
|
| 20 |
+
"HellaSwag": {
|
| 21 |
+
"metric_name": 0.237
|
| 22 |
+
},
|
| 23 |
+
"GSM8K": {
|
| 24 |
+
"metric_name": 0.095
|
| 25 |
+
},
|
| 26 |
+
"TruthfulQA": {
|
| 27 |
+
"metric_name": 0.344
|
| 28 |
+
}
|
| 29 |
+
}
|
| 30 |
+
}
|
results_Qwen__Qwen2.5-3B-Instruct_zero_shot.json
ADDED
|
@@ -0,0 +1,30 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
{
|
| 2 |
+
"config": {
|
| 3 |
+
"model_name": "Qwen/Qwen2.5-3B-Instruct",
|
| 4 |
+
"model_dtype": "float16",
|
| 5 |
+
"model_sha": "main"
|
| 6 |
+
},
|
| 7 |
+
"results": {
|
| 8 |
+
"KyrgyzMMLU": {
|
| 9 |
+
"metric_name": 0.28600000000000003
|
| 10 |
+
},
|
| 11 |
+
"KyrgyzRC": {
|
| 12 |
+
"metric_name": 0.66
|
| 13 |
+
},
|
| 14 |
+
"WinoGrande": {
|
| 15 |
+
"metric_name": 0.505
|
| 16 |
+
},
|
| 17 |
+
"BoolQ": {
|
| 18 |
+
"metric_name": 0.594
|
| 19 |
+
},
|
| 20 |
+
"HellaSwag": {
|
| 21 |
+
"metric_name": 0.22
|
| 22 |
+
},
|
| 23 |
+
"GSM8K": {
|
| 24 |
+
"metric_name": 0.006999999999999999
|
| 25 |
+
},
|
| 26 |
+
"TruthfulQA": {
|
| 27 |
+
"metric_name": 0.342
|
| 28 |
+
}
|
| 29 |
+
}
|
| 30 |
+
}
|
results_Qwen__Qwen2.5-7B-Instruct_few_shot.json
ADDED
|
@@ -0,0 +1,30 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
{
|
| 2 |
+
"config": {
|
| 3 |
+
"model_name": "Qwen/Qwen2.5-7B-Instruct",
|
| 4 |
+
"model_dtype": "float16",
|
| 5 |
+
"model_sha": "main"
|
| 6 |
+
},
|
| 7 |
+
"results": {
|
| 8 |
+
"KyrgyzMMLU": {
|
| 9 |
+
"metric_name": 0.385
|
| 10 |
+
},
|
| 11 |
+
"KyrgyzRC": {
|
| 12 |
+
"metric_name": 0.748
|
| 13 |
+
},
|
| 14 |
+
"WinoGrande": {
|
| 15 |
+
"metric_name": 0.504
|
| 16 |
+
},
|
| 17 |
+
"BoolQ": {
|
| 18 |
+
"metric_name": 0.6459999999999999
|
| 19 |
+
},
|
| 20 |
+
"HellaSwag": {
|
| 21 |
+
"metric_name": 0.17800000000000002
|
| 22 |
+
},
|
| 23 |
+
"GSM8K": {
|
| 24 |
+
"metric_name": 0.321
|
| 25 |
+
},
|
| 26 |
+
"TruthfulQA": {
|
| 27 |
+
"metric_name": 0.36200000000000004
|
| 28 |
+
}
|
| 29 |
+
}
|
| 30 |
+
}
|
results_Qwen__Qwen2.5-7B-Instruct_zero_shot.json
ADDED
|
@@ -0,0 +1,30 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
{
|
| 2 |
+
"config": {
|
| 3 |
+
"model_name": "Qwen/Qwen2.5-7B-Instruct",
|
| 4 |
+
"model_dtype": "float16",
|
| 5 |
+
"model_sha": "main"
|
| 6 |
+
},
|
| 7 |
+
"results": {
|
| 8 |
+
"KyrgyzMMLU": {
|
| 9 |
+
"metric_name": 0.315
|
| 10 |
+
},
|
| 11 |
+
"KyrgyzRC": {
|
| 12 |
+
"metric_name": 0.7
|
| 13 |
+
},
|
| 14 |
+
"WinoGrande": {
|
| 15 |
+
"metric_name": 0.48700000000000004
|
| 16 |
+
},
|
| 17 |
+
"BoolQ": {
|
| 18 |
+
"metric_name": 0.563
|
| 19 |
+
},
|
| 20 |
+
"HellaSwag": {
|
| 21 |
+
"metric_name": 0.1
|
| 22 |
+
},
|
| 23 |
+
"GSM8K": {
|
| 24 |
+
"metric_name": 0.011000000000000001
|
| 25 |
+
},
|
| 26 |
+
"TruthfulQA": {
|
| 27 |
+
"metric_name": 0.341
|
| 28 |
+
}
|
| 29 |
+
}
|
| 30 |
+
}
|
results_Qwen__Qwen3-0.6B_few_shot.json
ADDED
|
@@ -0,0 +1,30 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
{
|
| 2 |
+
"config": {
|
| 3 |
+
"model_name": "Qwen/Qwen3-0.6B",
|
| 4 |
+
"model_dtype": "float16",
|
| 5 |
+
"model_sha": "main"
|
| 6 |
+
},
|
| 7 |
+
"results": {
|
| 8 |
+
"KyrgyzMMLU": {
|
| 9 |
+
"metric_name": 0.268
|
| 10 |
+
},
|
| 11 |
+
"KyrgyzRC": {
|
| 12 |
+
"metric_name": 0.595
|
| 13 |
+
},
|
| 14 |
+
"WinoGrande": {
|
| 15 |
+
"metric_name": 0.501
|
| 16 |
+
},
|
| 17 |
+
"BoolQ": {
|
| 18 |
+
"metric_name": 0.601
|
| 19 |
+
},
|
| 20 |
+
"HellaSwag": {
|
| 21 |
+
"metric_name": 0.264
|
| 22 |
+
},
|
| 23 |
+
"GSM8K": {
|
| 24 |
+
"metric_name": 0.043
|
| 25 |
+
},
|
| 26 |
+
"TruthfulQA": {
|
| 27 |
+
"metric_name": 0.3
|
| 28 |
+
}
|
| 29 |
+
}
|
| 30 |
+
}
|
results_Qwen__Qwen3-0.6B_zero_shot.json
ADDED
|
@@ -0,0 +1,30 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
{
|
| 2 |
+
"config": {
|
| 3 |
+
"model_name": "Qwen/Qwen3-0.6B",
|
| 4 |
+
"model_dtype": "float16",
|
| 5 |
+
"model_sha": "main"
|
| 6 |
+
},
|
| 7 |
+
"results": {
|
| 8 |
+
"KyrgyzMMLU": {
|
| 9 |
+
"metric_name": 0.26
|
| 10 |
+
},
|
| 11 |
+
"KyrgyzRC": {
|
| 12 |
+
"metric_name": 0.618
|
| 13 |
+
},
|
| 14 |
+
"WinoGrande": {
|
| 15 |
+
"metric_name": 0.498
|
| 16 |
+
},
|
| 17 |
+
"BoolQ": {
|
| 18 |
+
"metric_name": 0.38
|
| 19 |
+
},
|
| 20 |
+
"HellaSwag": {
|
| 21 |
+
"metric_name": 0.111
|
| 22 |
+
},
|
| 23 |
+
"GSM8K": {
|
| 24 |
+
"metric_name": 0.006999999999999999
|
| 25 |
+
},
|
| 26 |
+
"TruthfulQA": {
|
| 27 |
+
"metric_name": 0.299
|
| 28 |
+
}
|
| 29 |
+
}
|
| 30 |
+
}
|
results_Qwen__Qwen3-1.7B_few_shot.json
ADDED
|
@@ -0,0 +1,30 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
{
|
| 2 |
+
"config": {
|
| 3 |
+
"model_name": "Qwen/Qwen3-1.7B",
|
| 4 |
+
"model_dtype": "float16",
|
| 5 |
+
"model_sha": "main"
|
| 6 |
+
},
|
| 7 |
+
"results": {
|
| 8 |
+
"KyrgyzMMLU": {
|
| 9 |
+
"metric_name": 0.308
|
| 10 |
+
},
|
| 11 |
+
"KyrgyzRC": {
|
| 12 |
+
"metric_name": 0.7120000000000001
|
| 13 |
+
},
|
| 14 |
+
"WinoGrande": {
|
| 15 |
+
"metric_name": 0.486
|
| 16 |
+
},
|
| 17 |
+
"BoolQ": {
|
| 18 |
+
"metric_name": 0.62
|
| 19 |
+
},
|
| 20 |
+
"HellaSwag": {
|
| 21 |
+
"metric_name": 0.252
|
| 22 |
+
},
|
| 23 |
+
"GSM8K": {
|
| 24 |
+
"metric_name": 0.185
|
| 25 |
+
},
|
| 26 |
+
"TruthfulQA": {
|
| 27 |
+
"metric_name": 0.303
|
| 28 |
+
}
|
| 29 |
+
}
|
| 30 |
+
}
|
results_Qwen__Qwen3-1.7B_zero_shot.json
ADDED
|
@@ -0,0 +1,30 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
{
|
| 2 |
+
"config": {
|
| 3 |
+
"model_name": "Qwen/Qwen3-1.7B",
|
| 4 |
+
"model_dtype": "float16",
|
| 5 |
+
"model_sha": "main"
|
| 6 |
+
},
|
| 7 |
+
"results": {
|
| 8 |
+
"KyrgyzMMLU": {
|
| 9 |
+
"metric_name": 0.27899999999999997
|
| 10 |
+
},
|
| 11 |
+
"KyrgyzRC": {
|
| 12 |
+
"metric_name": 0.618
|
| 13 |
+
},
|
| 14 |
+
"WinoGrande": {
|
| 15 |
+
"metric_name": 0.489
|
| 16 |
+
},
|
| 17 |
+
"BoolQ": {
|
| 18 |
+
"metric_name": 0.40399999999999997
|
| 19 |
+
},
|
| 20 |
+
"HellaSwag": {
|
| 21 |
+
"metric_name": 0.24600000000000002
|
| 22 |
+
},
|
| 23 |
+
"GSM8K": {
|
| 24 |
+
"metric_name": 0.006999999999999999
|
| 25 |
+
},
|
| 26 |
+
"TruthfulQA": {
|
| 27 |
+
"metric_name": 0.29600000000000004
|
| 28 |
+
}
|
| 29 |
+
}
|
| 30 |
+
}
|
results_Qwen__Qwen3-4B_few_shot.json
ADDED
|
@@ -0,0 +1,30 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
{
|
| 2 |
+
"config": {
|
| 3 |
+
"model_name": "Qwen/Qwen3-4B",
|
| 4 |
+
"model_dtype": "float16",
|
| 5 |
+
"model_sha": "main"
|
| 6 |
+
},
|
| 7 |
+
"results": {
|
| 8 |
+
"KyrgyzMMLU": {
|
| 9 |
+
"metric_name": 0.385
|
| 10 |
+
},
|
| 11 |
+
"KyrgyzRC": {
|
| 12 |
+
"metric_name": 0.772
|
| 13 |
+
},
|
| 14 |
+
"WinoGrande": {
|
| 15 |
+
"metric_name": 0.48100000000000004
|
| 16 |
+
},
|
| 17 |
+
"BoolQ": {
|
| 18 |
+
"metric_name": 0.74
|
| 19 |
+
},
|
| 20 |
+
"HellaSwag": {
|
| 21 |
+
"metric_name": 0.247
|
| 22 |
+
},
|
| 23 |
+
"GSM8K": {
|
| 24 |
+
"metric_name": 0.515
|
| 25 |
+
},
|
| 26 |
+
"TruthfulQA": {
|
| 27 |
+
"metric_name": 0.325
|
| 28 |
+
}
|
| 29 |
+
}
|
| 30 |
+
}
|
results_Qwen__Qwen3-4B_zero_shot.json
ADDED
|
@@ -0,0 +1,30 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
{
|
| 2 |
+
"config": {
|
| 3 |
+
"model_name": "Qwen/Qwen3-4B",
|
| 4 |
+
"model_dtype": "float16",
|
| 5 |
+
"model_sha": "main"
|
| 6 |
+
},
|
| 7 |
+
"results": {
|
| 8 |
+
"KyrgyzMMLU": {
|
| 9 |
+
"metric_name": 0.303
|
| 10 |
+
},
|
| 11 |
+
"KyrgyzRC": {
|
| 12 |
+
"metric_name": 0.682
|
| 13 |
+
},
|
| 14 |
+
"WinoGrande": {
|
| 15 |
+
"metric_name": 0.49
|
| 16 |
+
},
|
| 17 |
+
"BoolQ": {
|
| 18 |
+
"metric_name": 0.38299999999999995
|
| 19 |
+
},
|
| 20 |
+
"HellaSwag": {
|
| 21 |
+
"metric_name": 0.245
|
| 22 |
+
},
|
| 23 |
+
"GSM8K": {
|
| 24 |
+
"metric_name": 0.006999999999999999
|
| 25 |
+
},
|
| 26 |
+
"TruthfulQA": {
|
| 27 |
+
"metric_name": 0.32899999999999996
|
| 28 |
+
}
|
| 29 |
+
}
|
| 30 |
+
}
|
results_Qwen__Qwen3-8B_few_shot.json
ADDED
|
@@ -0,0 +1,30 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
{
|
| 2 |
+
"config": {
|
| 3 |
+
"model_name": "Qwen/Qwen3-8B",
|
| 4 |
+
"model_dtype": "float16",
|
| 5 |
+
"model_sha": "main"
|
| 6 |
+
},
|
| 7 |
+
"results": {
|
| 8 |
+
"KyrgyzMMLU": {
|
| 9 |
+
"metric_name": 0.445
|
| 10 |
+
},
|
| 11 |
+
"KyrgyzRC": {
|
| 12 |
+
"metric_name": 0.818
|
| 13 |
+
},
|
| 14 |
+
"WinoGrande": {
|
| 15 |
+
"metric_name": 0.506
|
| 16 |
+
},
|
| 17 |
+
"BoolQ": {
|
| 18 |
+
"metric_name": 0.769
|
| 19 |
+
},
|
| 20 |
+
"HellaSwag": {
|
| 21 |
+
"metric_name": 0.264
|
| 22 |
+
},
|
| 23 |
+
"GSM8K": {
|
| 24 |
+
"metric_name": 0.6
|
| 25 |
+
},
|
| 26 |
+
"TruthfulQA": {
|
| 27 |
+
"metric_name": 0.358
|
| 28 |
+
}
|
| 29 |
+
}
|
| 30 |
+
}
|
results_Qwen__Qwen3-8B_zero_shot.json
ADDED
|
@@ -0,0 +1,30 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
{
|
| 2 |
+
"config": {
|
| 3 |
+
"model_name": "Qwen/Qwen3-8B",
|
| 4 |
+
"model_dtype": "float16",
|
| 5 |
+
"model_sha": "main"
|
| 6 |
+
},
|
| 7 |
+
"results": {
|
| 8 |
+
"KyrgyzMMLU": {
|
| 9 |
+
"metric_name": 0.321
|
| 10 |
+
},
|
| 11 |
+
"KyrgyzRC": {
|
| 12 |
+
"metric_name": 0.718
|
| 13 |
+
},
|
| 14 |
+
"WinoGrande": {
|
| 15 |
+
"metric_name": 0.51
|
| 16 |
+
},
|
| 17 |
+
"BoolQ": {
|
| 18 |
+
"metric_name": 0.392
|
| 19 |
+
},
|
| 20 |
+
"HellaSwag": {
|
| 21 |
+
"metric_name": 0.24600000000000002
|
| 22 |
+
},
|
| 23 |
+
"GSM8K": {
|
| 24 |
+
"metric_name": 0.006999999999999999
|
| 25 |
+
},
|
| 26 |
+
"TruthfulQA": {
|
| 27 |
+
"metric_name": 0.34700000000000003
|
| 28 |
+
}
|
| 29 |
+
}
|
| 30 |
+
}
|
results_google__gemma-3-1b-it_few_shot.json
ADDED
|
@@ -0,0 +1,30 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
{
|
| 2 |
+
"config": {
|
| 3 |
+
"model_name": "google/gemma-3-1b-it",
|
| 4 |
+
"model_dtype": "float16",
|
| 5 |
+
"model_sha": "main"
|
| 6 |
+
},
|
| 7 |
+
"results": {
|
| 8 |
+
"KyrgyzMMLU": {
|
| 9 |
+
"metric_name": 0.265
|
| 10 |
+
},
|
| 11 |
+
"KyrgyzRC": {
|
| 12 |
+
"metric_name": 0.38
|
| 13 |
+
},
|
| 14 |
+
"WinoGrande": {
|
| 15 |
+
"metric_name": 0.489
|
| 16 |
+
},
|
| 17 |
+
"BoolQ": {
|
| 18 |
+
"metric_name": 0.628
|
| 19 |
+
},
|
| 20 |
+
"HellaSwag": {
|
| 21 |
+
"metric_name": 0.235
|
| 22 |
+
},
|
| 23 |
+
"GSM8K": {
|
| 24 |
+
"metric_name": 0.032
|
| 25 |
+
},
|
| 26 |
+
"TruthfulQA": {
|
| 27 |
+
"metric_name": 0.313
|
| 28 |
+
}
|
| 29 |
+
}
|
| 30 |
+
}
|
results_google__gemma-3-1b-it_zero_shot.json
ADDED
|
@@ -0,0 +1,30 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
{
|
| 2 |
+
"config": {
|
| 3 |
+
"model_name": "google/gemma-3-1b-it",
|
| 4 |
+
"model_dtype": "float16",
|
| 5 |
+
"model_sha": "main"
|
| 6 |
+
},
|
| 7 |
+
"results": {
|
| 8 |
+
"KyrgyzMMLU": {
|
| 9 |
+
"metric_name": 0.267
|
| 10 |
+
},
|
| 11 |
+
"KyrgyzRC": {
|
| 12 |
+
"metric_name": 0.5820000000000001
|
| 13 |
+
},
|
| 14 |
+
"WinoGrande": {
|
| 15 |
+
"metric_name": 0.5
|
| 16 |
+
},
|
| 17 |
+
"BoolQ": {
|
| 18 |
+
"metric_name": 0.379
|
| 19 |
+
},
|
| 20 |
+
"HellaSwag": {
|
| 21 |
+
"metric_name": 0.244
|
| 22 |
+
},
|
| 23 |
+
"GSM8K": {
|
| 24 |
+
"metric_name": 0.006999999999999999
|
| 25 |
+
},
|
| 26 |
+
"TruthfulQA": {
|
| 27 |
+
"metric_name": 0.34
|
| 28 |
+
}
|
| 29 |
+
}
|
| 30 |
+
}
|
results_google__gemma-3-270m_few_shot.json
ADDED
|
@@ -0,0 +1,30 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
{
|
| 2 |
+
"config": {
|
| 3 |
+
"model_name": "google/gemma-3-270m",
|
| 4 |
+
"model_dtype": "float16",
|
| 5 |
+
"model_sha": "main"
|
| 6 |
+
},
|
| 7 |
+
"results": {
|
| 8 |
+
"KyrgyzMMLU": {
|
| 9 |
+
"metric_name": 0.27
|
| 10 |
+
},
|
| 11 |
+
"KyrgyzRC": {
|
| 12 |
+
"metric_name": 0.532
|
| 13 |
+
},
|
| 14 |
+
"WinoGrande": {
|
| 15 |
+
"metric_name": 0.48700000000000004
|
| 16 |
+
},
|
| 17 |
+
"BoolQ": {
|
| 18 |
+
"metric_name": 0.615
|
| 19 |
+
},
|
| 20 |
+
"HellaSwag": {
|
| 21 |
+
"metric_name": 0.276
|
| 22 |
+
},
|
| 23 |
+
"GSM8K": {
|
| 24 |
+
"metric_name": 0.013999999999999999
|
| 25 |
+
},
|
| 26 |
+
"TruthfulQA": {
|
| 27 |
+
"metric_name": 0.366
|
| 28 |
+
}
|
| 29 |
+
}
|
| 30 |
+
}
|
results_google__gemma-3-270m_zero_shot.json
ADDED
|
@@ -0,0 +1,30 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
{
|
| 2 |
+
"config": {
|
| 3 |
+
"model_name": "google/gemma-3-270m",
|
| 4 |
+
"model_dtype": "float16",
|
| 5 |
+
"model_sha": "main"
|
| 6 |
+
},
|
| 7 |
+
"results": {
|
| 8 |
+
"KyrgyzMMLU": {
|
| 9 |
+
"metric_name": 0.275
|
| 10 |
+
},
|
| 11 |
+
"KyrgyzRC": {
|
| 12 |
+
"metric_name": 0.568
|
| 13 |
+
},
|
| 14 |
+
"WinoGrande": {
|
| 15 |
+
"metric_name": 0.483
|
| 16 |
+
},
|
| 17 |
+
"BoolQ": {
|
| 18 |
+
"metric_name": 0.379
|
| 19 |
+
},
|
| 20 |
+
"HellaSwag": {
|
| 21 |
+
"metric_name": 0.174
|
| 22 |
+
},
|
| 23 |
+
"GSM8K": {
|
| 24 |
+
"metric_name": 0.006999999999999999
|
| 25 |
+
},
|
| 26 |
+
"TruthfulQA": {
|
| 27 |
+
"metric_name": 0.34700000000000003
|
| 28 |
+
}
|
| 29 |
+
}
|
| 30 |
+
}
|
results_google__gemma-3-4b-it_few_shot.json
ADDED
|
@@ -0,0 +1,30 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
{
|
| 2 |
+
"config": {
|
| 3 |
+
"model_name": "google/gemma-3-4b-it",
|
| 4 |
+
"model_dtype": "float16",
|
| 5 |
+
"model_sha": "main"
|
| 6 |
+
},
|
| 7 |
+
"results": {
|
| 8 |
+
"KyrgyzMMLU": {
|
| 9 |
+
"metric_name": 0.295
|
| 10 |
+
},
|
| 11 |
+
"KyrgyzRC": {
|
| 12 |
+
"metric_name": 0.25
|
| 13 |
+
},
|
| 14 |
+
"WinoGrande": {
|
| 15 |
+
"metric_name": 0.496
|
| 16 |
+
},
|
| 17 |
+
"BoolQ": {
|
| 18 |
+
"metric_name": 0.621
|
| 19 |
+
},
|
| 20 |
+
"HellaSwag": {
|
| 21 |
+
"metric_name": 0.24600000000000002
|
| 22 |
+
},
|
| 23 |
+
"GSM8K": {
|
| 24 |
+
"metric_name": 0.0
|
| 25 |
+
},
|
| 26 |
+
"TruthfulQA": {
|
| 27 |
+
"metric_name": 0.5
|
| 28 |
+
}
|
| 29 |
+
}
|
| 30 |
+
}
|
results_google__gemma-3-4b-it_zero_shot.json
ADDED
|
@@ -0,0 +1,30 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
{
|
| 2 |
+
"config": {
|
| 3 |
+
"model_name": "google/gemma-3-4b-it",
|
| 4 |
+
"model_dtype": "float16",
|
| 5 |
+
"model_sha": "main"
|
| 6 |
+
},
|
| 7 |
+
"results": {
|
| 8 |
+
"KyrgyzMMLU": {
|
| 9 |
+
"metric_name": 0.303
|
| 10 |
+
},
|
| 11 |
+
"KyrgyzRC": {
|
| 12 |
+
"metric_name": 0.7020000000000001
|
| 13 |
+
},
|
| 14 |
+
"WinoGrande": {
|
| 15 |
+
"metric_name": 0.506
|
| 16 |
+
},
|
| 17 |
+
"BoolQ": {
|
| 18 |
+
"metric_name": 0.583
|
| 19 |
+
},
|
| 20 |
+
"HellaSwag": {
|
| 21 |
+
"metric_name": 0.24600000000000002
|
| 22 |
+
},
|
| 23 |
+
"GSM8K": {
|
| 24 |
+
"metric_name": 0.006999999999999999
|
| 25 |
+
},
|
| 26 |
+
"TruthfulQA": {
|
| 27 |
+
"metric_name": 0.34700000000000003
|
| 28 |
+
}
|
| 29 |
+
}
|
| 30 |
+
}
|
results_meta-llama__Llama-3.1-8B-Instruct_few_shot.json
ADDED
|
@@ -0,0 +1,30 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
{
|
| 2 |
+
"config": {
|
| 3 |
+
"model_name": "meta-llama/Llama-3.1-8B-Instruct",
|
| 4 |
+
"model_dtype": "float16",
|
| 5 |
+
"model_sha": "main"
|
| 6 |
+
},
|
| 7 |
+
"results": {
|
| 8 |
+
"KyrgyzMMLU": {
|
| 9 |
+
"metric_name": 0.381
|
| 10 |
+
},
|
| 11 |
+
"KyrgyzRC": {
|
| 12 |
+
"metric_name": 0.805
|
| 13 |
+
},
|
| 14 |
+
"WinoGrande": {
|
| 15 |
+
"metric_name": 0.516
|
| 16 |
+
},
|
| 17 |
+
"BoolQ": {
|
| 18 |
+
"metric_name": 0.755
|
| 19 |
+
},
|
| 20 |
+
"HellaSwag": {
|
| 21 |
+
"metric_name": 0.21899999999999997
|
| 22 |
+
},
|
| 23 |
+
"GSM8K": {
|
| 24 |
+
"metric_name": 0.37
|
| 25 |
+
},
|
| 26 |
+
"TruthfulQA": {
|
| 27 |
+
"metric_name": 0.344
|
| 28 |
+
}
|
| 29 |
+
}
|
| 30 |
+
}
|
results_meta-llama__Llama-3.1-8B-Instruct_zero_shot.json
ADDED
|
@@ -0,0 +1,30 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
{
|
| 2 |
+
"config": {
|
| 3 |
+
"model_name": "meta-llama/Llama-3.1-8B-Instruct",
|
| 4 |
+
"model_dtype": "float16",
|
| 5 |
+
"model_sha": "main"
|
| 6 |
+
},
|
| 7 |
+
"results": {
|
| 8 |
+
"KyrgyzMMLU": {
|
| 9 |
+
"metric_name": 0.31
|
| 10 |
+
},
|
| 11 |
+
"KyrgyzRC": {
|
| 12 |
+
"metric_name": 0.752
|
| 13 |
+
},
|
| 14 |
+
"WinoGrande": {
|
| 15 |
+
"metric_name": 0.506
|
| 16 |
+
},
|
| 17 |
+
"BoolQ": {
|
| 18 |
+
"metric_name": 0.503
|
| 19 |
+
},
|
| 20 |
+
"HellaSwag": {
|
| 21 |
+
"metric_name": 0.266
|
| 22 |
+
},
|
| 23 |
+
"GSM8K": {
|
| 24 |
+
"metric_name": 0.006999999999999999
|
| 25 |
+
},
|
| 26 |
+
"TruthfulQA": {
|
| 27 |
+
"metric_name": 0.337
|
| 28 |
+
}
|
| 29 |
+
}
|
| 30 |
+
}
|
results_meta-llama__Llama-3.2-1B-Instruct_few_shot.json
ADDED
|
@@ -0,0 +1,30 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
{
|
| 2 |
+
"config": {
|
| 3 |
+
"model_name": "meta-llama/Llama-3.2-1B-Instruct",
|
| 4 |
+
"model_dtype": "float16",
|
| 5 |
+
"model_sha": "main"
|
| 6 |
+
},
|
| 7 |
+
"results": {
|
| 8 |
+
"KyrgyzMMLU": {
|
| 9 |
+
"metric_name": 0.261
|
| 10 |
+
},
|
| 11 |
+
"KyrgyzRC": {
|
| 12 |
+
"metric_name": 0.45799999999999996
|
| 13 |
+
},
|
| 14 |
+
"WinoGrande": {
|
| 15 |
+
"metric_name": 0.49700000000000005
|
| 16 |
+
},
|
| 17 |
+
"BoolQ": {
|
| 18 |
+
"metric_name": 0.62
|
| 19 |
+
},
|
| 20 |
+
"HellaSwag": {
|
| 21 |
+
"metric_name": 0.258
|
| 22 |
+
},
|
| 23 |
+
"GSM8K": {
|
| 24 |
+
"metric_name": 0.027000000000000003
|
| 25 |
+
},
|
| 26 |
+
"TruthfulQA": {
|
| 27 |
+
"metric_name": 0.303
|
| 28 |
+
}
|
| 29 |
+
}
|
| 30 |
+
}
|
results_meta-llama__Llama-3.2-1B-Instruct_zero_shot.json
ADDED
|
@@ -0,0 +1,30 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
{
|
| 2 |
+
"config": {
|
| 3 |
+
"model_name": "meta-llama/Llama-3.2-1B-Instruct",
|
| 4 |
+
"model_dtype": "float16",
|
| 5 |
+
"model_sha": "main"
|
| 6 |
+
},
|
| 7 |
+
"results": {
|
| 8 |
+
"KyrgyzMMLU": {
|
| 9 |
+
"metric_name": 0.263
|
| 10 |
+
},
|
| 11 |
+
"KyrgyzRC": {
|
| 12 |
+
"metric_name": 0.5820000000000001
|
| 13 |
+
},
|
| 14 |
+
"WinoGrande": {
|
| 15 |
+
"metric_name": 0.494
|
| 16 |
+
},
|
| 17 |
+
"BoolQ": {
|
| 18 |
+
"metric_name": 0.38299999999999995
|
| 19 |
+
},
|
| 20 |
+
"HellaSwag": {
|
| 21 |
+
"metric_name": 0.002
|
| 22 |
+
},
|
| 23 |
+
"GSM8K": {
|
| 24 |
+
"metric_name": 0.006999999999999999
|
| 25 |
+
},
|
| 26 |
+
"TruthfulQA": {
|
| 27 |
+
"metric_name": 0.301
|
| 28 |
+
}
|
| 29 |
+
}
|
| 30 |
+
}
|
results_meta-llama__Llama-3.2-3B-Instruct_few_shot.json
ADDED
|
@@ -0,0 +1,30 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
{
|
| 2 |
+
"config": {
|
| 3 |
+
"model_name": "meta-llama/Llama-3.2-3B-Instruct",
|
| 4 |
+
"model_dtype": "float16",
|
| 5 |
+
"model_sha": "main"
|
| 6 |
+
},
|
| 7 |
+
"results": {
|
| 8 |
+
"KyrgyzMMLU": {
|
| 9 |
+
"metric_name": 0.294
|
| 10 |
+
},
|
| 11 |
+
"KyrgyzRC": {
|
| 12 |
+
"metric_name": 0.648
|
| 13 |
+
},
|
| 14 |
+
"WinoGrande": {
|
| 15 |
+
"metric_name": 0.489
|
| 16 |
+
},
|
| 17 |
+
"BoolQ": {
|
| 18 |
+
"metric_name": 0.623
|
| 19 |
+
},
|
| 20 |
+
"HellaSwag": {
|
| 21 |
+
"metric_name": 0.253
|
| 22 |
+
},
|
| 23 |
+
"GSM8K": {
|
| 24 |
+
"metric_name": 0.129
|
| 25 |
+
},
|
| 26 |
+
"TruthfulQA": {
|
| 27 |
+
"metric_name": 0.32899999999999996
|
| 28 |
+
}
|
| 29 |
+
}
|
| 30 |
+
}
|
results_meta-llama__Llama-3.2-3B-Instruct_zero_shot.json
ADDED
|
@@ -0,0 +1,30 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
{
|
| 2 |
+
"config": {
|
| 3 |
+
"model_name": "meta-llama/Llama-3.2-3B-Instruct",
|
| 4 |
+
"model_dtype": "float16",
|
| 5 |
+
"model_sha": "main"
|
| 6 |
+
},
|
| 7 |
+
"results": {
|
| 8 |
+
"KyrgyzMMLU": {
|
| 9 |
+
"metric_name": 0.278
|
| 10 |
+
},
|
| 11 |
+
"KyrgyzRC": {
|
| 12 |
+
"metric_name": 0.642
|
| 13 |
+
},
|
| 14 |
+
"WinoGrande": {
|
| 15 |
+
"metric_name": 0.491
|
| 16 |
+
},
|
| 17 |
+
"BoolQ": {
|
| 18 |
+
"metric_name": 0.431
|
| 19 |
+
},
|
| 20 |
+
"HellaSwag": {
|
| 21 |
+
"metric_name": 0.245
|
| 22 |
+
},
|
| 23 |
+
"GSM8K": {
|
| 24 |
+
"metric_name": 0.006999999999999999
|
| 25 |
+
},
|
| 26 |
+
"TruthfulQA": {
|
| 27 |
+
"metric_name": 0.315
|
| 28 |
+
}
|
| 29 |
+
}
|
| 30 |
+
}
|