Datasets:

davidadamczyk commited on
Commit
d231d56
·
verified ·
1 Parent(s): 2946918

Delete davidadamczyk

Browse files
davidadamczyk/friday_3_eval_request.json DELETED
@@ -1 +0,0 @@
1
- {"eval_name": "friday_3", "agree": 0.6842105263157895, "anli": 0.7307692307692307, "anli_en": 1.0, "arc_challenge": 1.0, "arc_easy": 1.0, "belebele": 1.0, "ctkfacts": 1.0, "ctkfacts_en": 0.0, "czechnews": 1.0, "facebook_comments": 1.0, "gsm8k": 0.0, "klok\u00e1nek": 1.0, "mall_reviews": 0.5, "mmlu": 0.5, "snli": 0.5, "snli_en": 0.7307692307692307, "squad": 0.5, "subjectivity": 1.0, "subjectivity_en": 0.5, "truthfulqa": 0.5}
 
 
davidadamczyk/last_submit_3_eval_request.json DELETED
@@ -1 +0,0 @@
1
- {"eval_name": "last_submit_3", "agree": 1.0, "anli": 1.77, "anli_en": 0.8, "arc_challenge": 0.45, "arc_easy": 1.97, "belebele": 1.21, "ctkfacts": 1.39, "ctkfacts_en": 0.27, "czechnews": 0.16, "facebook_comments": 1.49, "gsm8k": 1.47, "klok\u00e1nek": 1.45, "mall_reviews": 0.5, "mmlu": 1.68, "snli": 1.49, "snli_en": 1.12, "squad": 0.3, "subjectivity": 1.04, "subjectivity_en": 0.66, "truthfulqa": 1.63}
 
 
davidadamczyk/mojmodel_eval_request.json DELETED
@@ -1 +0,0 @@
1
- {"eval_name": "mojmodel", "agree": 1.71, "anli": 0.95, "anli_en": 1.35, "arc_challenge": 0.6, "arc_easy": 1.63, "belebele": 0.88, "ctkfacts": 0.82, "ctkfacts_en": 1.24, "czechnews": 1.83, "facebook_comments": 1.7, "gsm8k": 1.62, "klok\u00e1nek": 0.88, "mall_reviews": 0.68, "mmlu": 0.83, "snli": 0.49, "snli_en": 0.61, "squad": 0.96, "subjectivity": 0.9, "subjectivity_en": 1.79, "truthfulqa": 1.27}
 
 
davidadamczyk/nazev_eval_request.json DELETED
@@ -1 +0,0 @@
1
- {"eval_name": "nazev", "agree": 1.44, "anli": 0.58, "anli_en": 1.23, "arc_challenge": 1.31, "arc_easy": 0.88, "belebele": 1.22, "ctkfacts": 1.43, "ctkfacts_en": 1.98, "czechnews": 1.16, "facebook_comments": 0.35, "gsm8k": 1.55, "klok\u00e1nek": 0.88, "mall_reviews": 1.95, "mmlu": 1.66, "snli": 2.02, "snli_en": 1.51, "squad": 0.58, "subjectivity": 0.31, "subjectivity_en": 0.03, "truthfulqa": 1.73}
 
 
davidadamczyk/rrr_eval_request.json DELETED
@@ -1 +0,0 @@
1
- {"eval_name": "rrr", "agree": 1.46, "anli": 1.08, "anli_en": 1.31, "arc_challenge": 1.38, "arc_easy": 0.74, "belebele": 0.52, "ctkfacts": 0.79, "ctkfacts_en": 2.07, "czechnews": 0.72, "facebook_comments": 0.76, "gsm8k": 1.81, "klok\u00e1nek": 0.32, "mall_reviews": 1.64, "mmlu": 0.59, "snli": 0.87, "snli_en": 2.03, "squad": 1.01, "subjectivity": 1.09, "subjectivity_en": 0.94, "truthfulqa": 1.41}
 
 
davidadamczyk/submut_123_eval_request.json DELETED
@@ -1 +0,0 @@
1
- {"eval_name": "submut_123", "agree": 1.46, "anli": 1.32, "anli_en": 0.34, "arc_challenge": 0.92, "arc_easy": 0.44, "belebele": 1.33, "ctkfacts": 0.4, "ctkfacts_en": 1.51, "czechnews": 1.23, "facebook_comments": 1.25, "gsm8k": 0.75, "klok\u00e1nek": 0.79, "mall_reviews": 0.43, "mmlu": 0.81, "snli": 0.79, "snli_en": 1.72, "squad": 0.74, "subjectivity": 1.44, "subjectivity_en": 1.06, "truthfulqa": 1.0}
 
 
davidadamczyk/test3_eval_request.json DELETED
@@ -1 +0,0 @@
1
- {"eval_name": "test3", "agree": 1.18, "anli": 1.29, "anli_en": 0.2, "arc_challenge": 0.93, "arc_easy": 0.27, "belebele": 1.12, "ctkfacts": 0.15, "ctkfacts_en": 1.5, "czechnews": -0.12, "facebook_comments": 1.02, "gsm8k": 1.82, "klok\u00e1nek": 0.8, "mall_reviews": 0.51, "mmlu": 1.12, "snli": 0.51, "snli_en": 1.24, "squad": 1.38, "subjectivity": 0.36, "subjectivity_en": 0.22, "truthfulqa": 0.64}
 
 
davidadamczyk/test4_eval_request.json DELETED
@@ -1 +0,0 @@
1
- {"eval_name": "test4", "agree": 0.91, "anli": 0.81, "anli_en": 1.63, "arc_challenge": 1.12, "arc_easy": 0.74, "belebele": 1.05, "ctkfacts": 1.61, "ctkfacts_en": 1.0, "czechnews": 0.72, "facebook_comments": 0.88, "gsm8k": 1.98, "klok\u00e1nek": 0.45, "mall_reviews": 0.53, "mmlu": 0.92, "snli": 0.74, "snli_en": 0.65, "squad": 1.07, "subjectivity": 1.42, "subjectivity_en": -0.11, "truthfulqa": 0.56}
 
 
davidadamczyk/test_eval_request.json DELETED
@@ -1 +0,0 @@
1
- {"eval_name": "test", "agree": 0.86, "anli": 1.66, "anli_en": 1.93, "arc_challenge": 0.42, "arc_easy": 1.35, "belebele": 0.72, "ctkfacts": 1.24, "ctkfacts_en": 0.73, "czechnews": 0.9, "facebook_comments": 1.04, "gsm8k": 0.75, "klok\u00e1nek": 1.25, "mall_reviews": 1.73, "mmlu": 0.57, "snli": 0.46, "snli_en": 1.29, "squad": 0.32, "subjectivity": 1.36, "subjectivity_en": 0.74, "truthfulqa": 0.49}
 
 
davidadamczyk/test_friday_2_eval_request.json DELETED
@@ -1 +0,0 @@
1
- {"eval_name": "test_friday_2", "agree": 0.99, "anli": 1.25, "anli_en": 0.98, "arc_challenge": 1.19, "arc_easy": 1.12, "belebele": 0.78, "ctkfacts": 1.31, "ctkfacts_en": 1.34, "czechnews": 1.03, "facebook_comments": 0.79, "gsm8k": 0.9, "klok\u00e1nek": 0.45, "mall_reviews": 0.64, "mmlu": 1.26, "snli": 1.62, "snli_en": 0.92, "squad": 0.87, "subjectivity": 0.79, "subjectivity_en": 0.34, "truthfulqa": 0.73}
 
 
davidadamczyk/test_friday_eval_request.json DELETED
@@ -1 +0,0 @@
1
- {"eval_name": "test_friday", "agree": 1.11, "anli": 0.51, "anli_en": 0.72, "arc_challenge": 0.27, "arc_easy": 0.68, "belebele": 0.94, "ctkfacts": 0.81, "ctkfacts_en": 1.71, "czechnews": 1.35, "facebook_comments": 0.86, "gsm8k": 0.45, "klok\u00e1nek": 1.06, "mall_reviews": 1.21, "mmlu": 0.41, "snli": 1.69, "snli_en": 0.19, "squad": 0.77, "subjectivity": 0.56, "subjectivity_en": 0.22, "truthfulqa": 1.14}
 
 
davidadamczyk/test_hf_eval_request.json DELETED
@@ -1 +0,0 @@
1
- {"eval_name": "test_hf", "agree": 1.51, "anli": 0.27, "anli_en": 1.33, "arc_challenge": 0.89, "arc_easy": -0.08, "belebele": 0.91, "ctkfacts": 0.88, "ctkfacts_en": 0.44, "czechnews": 0.63, "facebook_comments": 2.16, "gsm8k": 1.07, "klok\u00e1nek": 1.74, "mall_reviews": 1.43, "mmlu": 1.19, "snli": 0.14, "snli_en": 1.67, "squad": 0.57, "subjectivity": 1.45, "subjectivity_en": 0.52, "truthfulqa": 1.56}
 
 
davidadamczyk/togethercomputer_Llama-2-7B-32K-Instruct_float16_44_eval_request.json DELETED
@@ -1 +0,0 @@
1
- {"eval_name": "togethercomputer_Llama-2-7B-32K-Instruct_float16_44", "agree": 0.6842105263157895, "anli": 0.7307692307692307, "anli_en": 1.0, "arc_challenge": 1.0, "arc_easy": 1.0, "belebele": 1.0, "ctkfacts": 1.0, "ctkfacts_en": 0.0, "czechnews": 1.0, "facebook_comments": 1.0, "gsm8k": 0.0, "klok\u00e1nek": 1.0, "mall_reviews": 0.5, "mmlu": 0.5, "snli": 0.5, "snli_en": 0.7307692307692307, "squad": 0.5, "subjectivity": 1.0, "subjectivity_en": 0.5, "truthfulqa": 0.5}
 
 
davidadamczyk/togethercomputer_Llama-2-7B-32K-Instruct_float16_45_eval_request.json DELETED
@@ -1 +0,0 @@
1
- {"eval_name": "togethercomputer_Llama-2-7B-32K-Instruct_float16_45", "agree": 0.6842105263157895, "anli": 0.7307692307692307, "anli_en": 1.0, "arc_challenge": 1.0, "arc_easy": 1.0, "belebele": 1.0, "ctkfacts": 1.0, "ctkfacts_en": 0.0, "czechnews": 1.0, "facebook_comments": 1.0, "gsm8k": 0.0, "klok\u00e1nek": 1.0, "mall_reviews": 0.5, "mmlu": 0.5, "snli": 0.5, "snli_en": 0.7307692307692307, "squad": 0.5, "subjectivity": 1.0, "subjectivity_en": 0.5, "truthfulqa": 0.5}
 
 
davidadamczyk/vicgalle_Roleplay-Llama-3-8B_float16_44_eval_request.json DELETED
@@ -1 +0,0 @@
1
- {"eval_name": "vicgalle_Roleplay-Llama-3-8B_float16_44", "agree": 0.6842105263157895, "anli": 0.7307692307692307, "anli_en": 1.0, "arc_challenge": 1.0, "arc_easy": 1.0, "belebele": 1.0, "ctkfacts": 1.0, "ctkfacts_en": 0.0, "czechnews": 1.0, "facebook_comments": 1.0, "gsm8k": 0.0, "klok\u00e1nek": 1.0, "mall_reviews": 0.5, "mmlu": 0.5, "snli": 0.5, "snli_en": 0.7307692307692307, "squad": 0.5, "subjectivity": 1.0, "subjectivity_en": 0.5, "truthfulqa": 0.5}
 
 
davidadamczyk/vicgalle_Roleplay-Llama-3-8B_float16_47_eval_request.json DELETED
@@ -1 +0,0 @@
1
- {"eval_name": "vicgalle_Roleplay-Llama-3-8B_float16_47", "agree": 0.6842105263157895, "anli": 0.7307692307692307, "anli_en": 1.0, "arc_challenge": 1.0, "arc_easy": 1.0, "belebele": 1.0, "ctkfacts": 1.0, "ctkfacts_en": 0.0, "czechnews": 1.0, "facebook_comments": 1.0, "gsm8k": 0.0, "klok\u00e1nek": 1.0, "mall_reviews": 0.5, "mmlu": 0.5, "snli": 0.5, "snli_en": 0.7307692307692307, "squad": 0.5, "subjectivity": 1.0, "subjectivity_en": 0.5, "truthfulqa": 0.5}