Upload folder using huggingface_hub
Browse filesThis view is limited to 50 files because it contains too many changes.
See raw diff
- llama1_guanaco-7B-HF_fs0_mmlu_316695dc.json +1 -0
- llama1_guanaco-7B-HF_fs0_mmlu_316695dc_group_cache.npy +3 -0
- llama1_guanaco-7B-HF_fs1_mmlu_d6011d37.json +1 -0
- llama1_guanaco-7B-HF_fs1_mmlu_d6011d37_group_cache.npy +3 -0
- llama1_guanaco-7B-HF_fs2_mmlu_f2bd887d.json +1 -0
- llama1_guanaco-7B-HF_fs2_mmlu_f2bd887d_group_cache.npy +3 -0
- llama1_guanaco-7B-HF_fs3_mmlu_61b8a8e3.json +1 -0
- llama1_guanaco-7B-HF_fs3_mmlu_61b8a8e3_group_cache.npy +3 -0
- llama1_guanaco-7B-HF_fs4_mmlu_28dff3fc.json +1 -0
- llama1_guanaco-7B-HF_fs4_mmlu_28dff3fc_group_cache.npy +3 -0
- llama1_guanaco-7B-HF_fs5_mmlu_b583442d.json +1 -0
- llama1_guanaco-7B-HF_fs5_mmlu_b583442d_group_cache.npy +3 -0
- llama1_llama-7b_fs0_mmlu_dafad174.json +1 -0
- llama1_llama-7b_fs0_mmlu_dafad174_group_cache.npy +3 -0
- llama1_llama-7b_fs1_mmlu_cf545325.json +1 -0
- llama1_llama-7b_fs1_mmlu_cf545325_group_cache.npy +3 -0
- llama1_llama-7b_fs2_mmlu_a62f4c1e.json +1 -0
- llama1_llama-7b_fs2_mmlu_a62f4c1e_group_cache.npy +3 -0
- llama1_llama-7b_fs3_mmlu_374ca650.json +1 -0
- llama1_llama-7b_fs3_mmlu_374ca650_group_cache.npy +3 -0
- llama1_llama-7b_fs4_mmlu_72b24354.json +1 -0
- llama1_llama-7b_fs4_mmlu_72b24354_group_cache.npy +3 -0
- llama1_llama-7b_fs5_mmlu_479ca4ec.json +1 -0
- llama1_llama-7b_fs5_mmlu_479ca4ec_group_cache.npy +3 -0
- llama2_Llama-2-7b-chat_fs0_mmlu_767c4ad0.json +1 -0
- llama2_Llama-2-7b-chat_fs0_mmlu_767c4ad0_group_cache.npy +3 -0
- llama2_Llama-2-7b-chat_fs1_mmlu_7c7ae7be.json +1 -0
- llama2_Llama-2-7b-chat_fs1_mmlu_7c7ae7be_group_cache.npy +3 -0
- llama2_Llama-2-7b-chat_fs2_mmlu_31b67daa.json +1 -0
- llama2_Llama-2-7b-chat_fs2_mmlu_31b67daa_group_cache.npy +3 -0
- llama2_Llama-2-7b-chat_fs3_mmlu_34207011.json +1 -0
- llama2_Llama-2-7b-chat_fs3_mmlu_34207011_group_cache.npy +3 -0
- llama2_Llama-2-7b-chat_fs4_mmlu_7fabd9ca.json +1 -0
- llama2_Llama-2-7b-chat_fs4_mmlu_7fabd9ca_group_cache.npy +3 -0
- llama2_Llama-2-7b-chat_fs5_mmlu_952c2ce7.json +1 -0
- llama2_Llama-2-7b-chat_fs5_mmlu_952c2ce7_group_cache.npy +3 -0
- llama2_Llama-2-7b-hf_fs0_mmlu_31148294.json +1 -0
- llama2_Llama-2-7b-hf_fs0_mmlu_31148294_group_cache.npy +3 -0
- llama2_Llama-2-7b-hf_fs1_mmlu_afa46a6c.json +1 -0
- llama2_Llama-2-7b-hf_fs1_mmlu_afa46a6c_group_cache.npy +3 -0
- llama2_Llama-2-7b-hf_fs2_mmlu_d9dcdc1b.json +1 -0
- llama2_Llama-2-7b-hf_fs2_mmlu_d9dcdc1b_group_cache.npy +3 -0
- llama2_Llama-2-7b-hf_fs3_mmlu_71925834.json +1 -0
- llama2_Llama-2-7b-hf_fs3_mmlu_71925834_group_cache.npy +3 -0
- llama2_Llama-2-7b-hf_fs4_mmlu_f23ef2af.json +1 -0
- llama2_Llama-2-7b-hf_fs4_mmlu_f23ef2af_group_cache.npy +3 -0
- llama2_Llama-2-7b-hf_fs5_mmlu_3a7720c8.json +1 -0
- llama2_Llama-2-7b-hf_fs5_mmlu_3a7720c8_group_cache.npy +3 -0
- llama2_tulu-2-7b_fs0_mmlu_6d638552.json +1 -0
- llama2_tulu-2-7b_fs0_mmlu_6d638552_group_cache.npy +3 -0
llama1_guanaco-7B-HF_fs0_mmlu_316695dc.json
ADDED
|
@@ -0,0 +1 @@
|
|
|
|
|
|
|
| 1 |
+
{"uid": "316695dc", "model": "TheBloke/guanaco-7B-HF", "model_name": "guanaco-7B-HF", "group": "llama1", "model_args": "pretrained=TheBloke/guanaco-7B-HF,dtype=auto,gpu_memory_utilization=0.7,tensor_parallel_size=1,data_parallel_size=4,max_model_len=2048", "fewshot": 0, "tasks": "mmlu"}
|
llama1_guanaco-7B-HF_fs0_mmlu_316695dc_group_cache.npy
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:70152bc658660735ea4adbcd67d86ceae5b02447faff0ce399c9deb69b0a40c5
|
| 3 |
+
size 5069770
|
llama1_guanaco-7B-HF_fs1_mmlu_d6011d37.json
ADDED
|
@@ -0,0 +1 @@
|
|
|
|
|
|
|
| 1 |
+
{"uid": "d6011d37", "model": "TheBloke/guanaco-7B-HF", "model_name": "guanaco-7B-HF", "group": "llama1", "model_args": "pretrained=TheBloke/guanaco-7B-HF,dtype=auto,gpu_memory_utilization=0.7,tensor_parallel_size=1,data_parallel_size=4,max_model_len=2048", "fewshot": 1, "tasks": "mmlu"}
|
llama1_guanaco-7B-HF_fs1_mmlu_d6011d37_group_cache.npy
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:4a05d110db5f101af6e9d70501ccf3c2246d2bfe8534d52cf2b00a35295608d1
|
| 3 |
+
size 4608861
|
llama1_guanaco-7B-HF_fs2_mmlu_f2bd887d.json
ADDED
|
@@ -0,0 +1 @@
|
|
|
|
|
|
|
| 1 |
+
{"uid": "f2bd887d", "model": "TheBloke/guanaco-7B-HF", "model_name": "guanaco-7B-HF", "group": "llama1", "model_args": "pretrained=TheBloke/guanaco-7B-HF,dtype=auto,gpu_memory_utilization=0.7,tensor_parallel_size=1,data_parallel_size=4,max_model_len=2048", "fewshot": 2, "tasks": "mmlu"}
|
llama1_guanaco-7B-HF_fs2_mmlu_f2bd887d_group_cache.npy
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:593748f3670894064ed1b6aca5ee46f25f1998d7351209b4483c947a0f4a4cf9
|
| 3 |
+
size 5530679
|
llama1_guanaco-7B-HF_fs3_mmlu_61b8a8e3.json
ADDED
|
@@ -0,0 +1 @@
|
|
|
|
|
|
|
| 1 |
+
{"uid": "61b8a8e3", "model": "TheBloke/guanaco-7B-HF", "model_name": "guanaco-7B-HF", "group": "llama1", "model_args": "pretrained=TheBloke/guanaco-7B-HF,dtype=auto,gpu_memory_utilization=0.7,tensor_parallel_size=1,data_parallel_size=4,max_model_len=2048", "fewshot": 3, "tasks": "mmlu"}
|
llama1_guanaco-7B-HF_fs3_mmlu_61b8a8e3_group_cache.npy
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:c71f2a48e5e248148b5264c72c25f5c2cbfae5177866eca1ff7683ea07bb108f
|
| 3 |
+
size 2304338
|
llama1_guanaco-7B-HF_fs4_mmlu_28dff3fc.json
ADDED
|
@@ -0,0 +1 @@
|
|
|
|
|
|
|
| 1 |
+
{"uid": "28dff3fc", "model": "TheBloke/guanaco-7B-HF", "model_name": "guanaco-7B-HF", "group": "llama1", "model_args": "pretrained=TheBloke/guanaco-7B-HF,dtype=auto,gpu_memory_utilization=0.7,tensor_parallel_size=1,data_parallel_size=4,max_model_len=2048", "fewshot": 4, "tasks": "mmlu"}
|
llama1_guanaco-7B-HF_fs4_mmlu_28dff3fc_group_cache.npy
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:18ede2837c523bd2c369ee73aea3b817d0b798f261274800488d19d6ff5672cc
|
| 3 |
+
size 3226156
|
llama1_guanaco-7B-HF_fs5_mmlu_b583442d.json
ADDED
|
@@ -0,0 +1 @@
|
|
|
|
|
|
|
| 1 |
+
{"uid": "b583442d", "model": "TheBloke/guanaco-7B-HF", "model_name": "guanaco-7B-HF", "group": "llama1", "model_args": "pretrained=TheBloke/guanaco-7B-HF,dtype=auto,gpu_memory_utilization=0.7,tensor_parallel_size=1,data_parallel_size=4,max_model_len=2048", "fewshot": 5, "tasks": "mmlu"}
|
llama1_guanaco-7B-HF_fs5_mmlu_b583442d_group_cache.npy
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:5149127c6155eb438fbd81f067a94b5df51463bb3e9223468b53d3ce97988352
|
| 3 |
+
size 2765247
|
llama1_llama-7b_fs0_mmlu_dafad174.json
ADDED
|
@@ -0,0 +1 @@
|
|
|
|
|
|
|
| 1 |
+
{"uid": "dafad174", "model": "huggyllama/llama-7b", "model_name": "llama-7b", "group": "llama1", "model_args": "pretrained=huggyllama/llama-7b,dtype=auto,gpu_memory_utilization=0.7,tensor_parallel_size=1,data_parallel_size=4,max_model_len=2048", "fewshot": 0, "tasks": "mmlu"}
|
llama1_llama-7b_fs0_mmlu_dafad174_group_cache.npy
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:356637341b1ba24f905d14022123928535d9835d04d6107dd7fce77ef9e47529
|
| 3 |
+
size 3687054
|
llama1_llama-7b_fs1_mmlu_cf545325.json
ADDED
|
@@ -0,0 +1 @@
|
|
|
|
|
|
|
| 1 |
+
{"uid": "cf545325", "model": "huggyllama/llama-7b", "model_name": "llama-7b", "group": "llama1", "model_args": "pretrained=huggyllama/llama-7b,dtype=auto,gpu_memory_utilization=0.7,tensor_parallel_size=1,data_parallel_size=4,max_model_len=2048", "fewshot": 1, "tasks": "mmlu"}
|
llama1_llama-7b_fs1_mmlu_cf545325_group_cache.npy
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:17ce112c391e2bd69a930b9c58307259d7319320673440858a7834731e88a12c
|
| 3 |
+
size 460734
|
llama1_llama-7b_fs2_mmlu_a62f4c1e.json
ADDED
|
@@ -0,0 +1 @@
|
|
|
|
|
|
|
| 1 |
+
{"uid": "a62f4c1e", "model": "huggyllama/llama-7b", "model_name": "llama-7b", "group": "llama1", "model_args": "pretrained=huggyllama/llama-7b,dtype=auto,gpu_memory_utilization=0.7,tensor_parallel_size=1,data_parallel_size=4,max_model_len=2048", "fewshot": 2, "tasks": "mmlu"}
|
llama1_llama-7b_fs2_mmlu_a62f4c1e_group_cache.npy
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:582b681dcb7ce95506add7b355a9709993c10495f69049c40287e523e64575b3
|
| 3 |
+
size 1843429
|
llama1_llama-7b_fs3_mmlu_374ca650.json
ADDED
|
@@ -0,0 +1 @@
|
|
|
|
|
|
|
| 1 |
+
{"uid": "374ca650", "model": "huggyllama/llama-7b", "model_name": "llama-7b", "group": "llama1", "model_args": "pretrained=huggyllama/llama-7b,dtype=auto,gpu_memory_utilization=0.7,tensor_parallel_size=1,data_parallel_size=4,max_model_len=2048", "fewshot": 3, "tasks": "mmlu"}
|
llama1_llama-7b_fs3_mmlu_374ca650_group_cache.npy
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:4bdee5f4922163eec5a47bab1918ce44dc6a789c1704a21c7cf215268c444a82
|
| 3 |
+
size 921633
|
llama1_llama-7b_fs4_mmlu_72b24354.json
ADDED
|
@@ -0,0 +1 @@
|
|
|
|
|
|
|
| 1 |
+
{"uid": "72b24354", "model": "huggyllama/llama-7b", "model_name": "llama-7b", "group": "llama1", "model_args": "pretrained=huggyllama/llama-7b,dtype=auto,gpu_memory_utilization=0.7,tensor_parallel_size=1,data_parallel_size=4,max_model_len=2048", "fewshot": 4, "tasks": "mmlu"}
|
llama1_llama-7b_fs4_mmlu_72b24354_group_cache.npy
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:3717e22443b9bfd5bf86d02bd7c8aa4fb7ef6e7dfb287648336e0b21dced985a
|
| 3 |
+
size 1382531
|
llama1_llama-7b_fs5_mmlu_479ca4ec.json
ADDED
|
@@ -0,0 +1 @@
|
|
|
|
|
|
|
| 1 |
+
{"uid": "479ca4ec", "model": "huggyllama/llama-7b", "model_name": "llama-7b", "group": "llama1", "model_args": "pretrained=huggyllama/llama-7b,dtype=auto,gpu_memory_utilization=0.7,tensor_parallel_size=1,data_parallel_size=4,max_model_len=2048", "fewshot": 5, "tasks": "mmlu"}
|
llama1_llama-7b_fs5_mmlu_479ca4ec_group_cache.npy
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:1c5d2b377695b037dc9d59b1c431a9b1c7520b00424ba81ca6b2ed39d9a60c90
|
| 3 |
+
size 4147952
|
llama2_Llama-2-7b-chat_fs0_mmlu_767c4ad0.json
ADDED
|
@@ -0,0 +1 @@
|
|
|
|
|
|
|
| 1 |
+
{"uid": "767c4ad0", "model": "meta-llama/Llama-2-7b-chat", "model_name": "Llama-2-7b-chat", "group": "llama2", "model_args": "pretrained=meta-llama/Llama-2-7b-chat,dtype=auto,gpu_memory_utilization=0.7,tensor_parallel_size=1,data_parallel_size=0,max_model_len=4096", "fewshot": 0, "tasks": "mmlu"}
|
llama2_Llama-2-7b-chat_fs0_mmlu_767c4ad0_group_cache.npy
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:79edee3c6dc0f6b880ea77ec109cc346c8c7e43cbb35598de87dcb06e7204249
|
| 3 |
+
size 9678896
|
llama2_Llama-2-7b-chat_fs1_mmlu_7c7ae7be.json
ADDED
|
@@ -0,0 +1 @@
|
|
|
|
|
|
|
| 1 |
+
{"uid": "7c7ae7be", "model": "meta-llama/Llama-2-7b-chat", "model_name": "Llama-2-7b-chat", "group": "llama2", "model_args": "pretrained=meta-llama/Llama-2-7b-chat,dtype=auto,gpu_memory_utilization=0.7,tensor_parallel_size=1,data_parallel_size=0,max_model_len=4096", "fewshot": 1, "tasks": "mmlu"}
|
llama2_Llama-2-7b-chat_fs1_mmlu_7c7ae7be_group_cache.npy
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:7a215d49916fce04d60267a8d555144b17b5f7d4330b179ea953622f4ff70b54
|
| 3 |
+
size 10600734
|
llama2_Llama-2-7b-chat_fs2_mmlu_31b67daa.json
ADDED
|
@@ -0,0 +1 @@
|
|
|
|
|
|
|
| 1 |
+
{"uid": "31b67daa", "model": "meta-llama/Llama-2-7b-chat", "model_name": "Llama-2-7b-chat", "group": "llama2", "model_args": "pretrained=meta-llama/Llama-2-7b-chat,dtype=auto,gpu_memory_utilization=0.7,tensor_parallel_size=1,data_parallel_size=0,max_model_len=4096", "fewshot": 2, "tasks": "mmlu"}
|
llama2_Llama-2-7b-chat_fs2_mmlu_31b67daa_group_cache.npy
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:08af54bab49ec1257b6430a6490622331c34d16b8b8821f5e992c32bf0bd9b15
|
| 3 |
+
size 10139815
|
llama2_Llama-2-7b-chat_fs3_mmlu_34207011.json
ADDED
|
@@ -0,0 +1 @@
|
|
|
|
|
|
|
| 1 |
+
{"uid": "34207011", "model": "meta-llama/Llama-2-7b-chat", "model_name": "Llama-2-7b-chat", "group": "llama2", "model_args": "pretrained=meta-llama/Llama-2-7b-chat,dtype=auto,gpu_memory_utilization=0.7,tensor_parallel_size=1,data_parallel_size=0,max_model_len=4096", "fewshot": 3, "tasks": "mmlu"}
|
llama2_Llama-2-7b-chat_fs3_mmlu_34207011_group_cache.npy
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:29c37201eabcc3bae197b9d10c85e77259fe39a7557756974a9f1d85035cba5e
|
| 3 |
+
size 8296151
|
llama2_Llama-2-7b-chat_fs4_mmlu_7fabd9ca.json
ADDED
|
@@ -0,0 +1 @@
|
|
|
|
|
|
|
| 1 |
+
{"uid": "7fabd9ca", "model": "meta-llama/Llama-2-7b-chat", "model_name": "Llama-2-7b-chat", "group": "llama2", "model_args": "pretrained=meta-llama/Llama-2-7b-chat,dtype=auto,gpu_memory_utilization=0.7,tensor_parallel_size=1,data_parallel_size=0,max_model_len=4096", "fewshot": 4, "tasks": "mmlu"}
|
llama2_Llama-2-7b-chat_fs4_mmlu_7fabd9ca_group_cache.npy
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:df52b412e7c6b708f6e2cf44697e07316d153ac06642c63f3d066d3ab3504036
|
| 3 |
+
size 6913424
|
llama2_Llama-2-7b-chat_fs5_mmlu_952c2ce7.json
ADDED
|
@@ -0,0 +1 @@
|
|
|
|
|
|
|
| 1 |
+
{"uid": "952c2ce7", "model": "meta-llama/Llama-2-7b-chat", "model_name": "Llama-2-7b-chat", "group": "llama2", "model_args": "pretrained=meta-llama/Llama-2-7b-chat,dtype=auto,gpu_memory_utilization=0.7,tensor_parallel_size=1,data_parallel_size=0,max_model_len=4096", "fewshot": 5, "tasks": "mmlu"}
|
llama2_Llama-2-7b-chat_fs5_mmlu_952c2ce7_group_cache.npy
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:1344c5e9e7bf5d87c8d8aaa17f212b37ccac4c2a7d16a9319e57db541a060295
|
| 3 |
+
size 9217977
|
llama2_Llama-2-7b-hf_fs0_mmlu_31148294.json
ADDED
|
@@ -0,0 +1 @@
|
|
|
|
|
|
|
| 1 |
+
{"uid": "31148294", "model": "meta-llama/Llama-2-7b-hf", "model_name": "Llama-2-7b-hf", "group": "llama2", "model_args": "pretrained=meta-llama/Llama-2-7b-hf,dtype=auto,gpu_memory_utilization=0.7,tensor_parallel_size=1,data_parallel_size=0,max_model_len=4096", "fewshot": 0, "tasks": "mmlu"}
|
llama2_Llama-2-7b-hf_fs0_mmlu_31148294_group_cache.npy
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:e0d1b4cab918ea25214676366098ded3406a801797a694ddf8ae8cf84a184ede
|
| 3 |
+
size 2765285
|
llama2_Llama-2-7b-hf_fs1_mmlu_afa46a6c.json
ADDED
|
@@ -0,0 +1 @@
|
|
|
|
|
|
|
| 1 |
+
{"uid": "afa46a6c", "model": "meta-llama/Llama-2-7b-hf", "model_name": "Llama-2-7b-hf", "group": "llama2", "model_args": "pretrained=meta-llama/Llama-2-7b-hf,dtype=auto,gpu_memory_utilization=0.7,tensor_parallel_size=1,data_parallel_size=4,max_model_len=4096", "fewshot": 1, "tasks": "mmlu"}
|
llama2_Llama-2-7b-hf_fs1_mmlu_afa46a6c_group_cache.npy
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:60545517d76b3024862f7c908888263ce2986ce853647b3e13afd30eb0b605af
|
| 3 |
+
size 5069796
|
llama2_Llama-2-7b-hf_fs2_mmlu_d9dcdc1b.json
ADDED
|
@@ -0,0 +1 @@
|
|
|
|
|
|
|
| 1 |
+
{"uid": "d9dcdc1b", "model": "meta-llama/Llama-2-7b-hf", "model_name": "Llama-2-7b-hf", "group": "llama2", "model_args": "pretrained=meta-llama/Llama-2-7b-hf,dtype=auto,gpu_memory_utilization=0.7,tensor_parallel_size=1,data_parallel_size=4,max_model_len=4096", "fewshot": 2, "tasks": "mmlu"}
|
llama2_Llama-2-7b-hf_fs2_mmlu_d9dcdc1b_group_cache.npy
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:57b4b21679aa758d98aae02db4505d03d7468d485c350c55b2f58d687f51c29a
|
| 3 |
+
size 3687093
|
llama2_Llama-2-7b-hf_fs3_mmlu_71925834.json
ADDED
|
@@ -0,0 +1 @@
|
|
|
|
|
|
|
| 1 |
+
{"uid": "71925834", "model": "meta-llama/Llama-2-7b-hf", "model_name": "Llama-2-7b-hf", "group": "llama2", "model_args": "pretrained=meta-llama/Llama-2-7b-hf,dtype=auto,gpu_memory_utilization=0.7,tensor_parallel_size=1,data_parallel_size=4,max_model_len=4096", "fewshot": 3, "tasks": "mmlu"}
|
llama2_Llama-2-7b-hf_fs3_mmlu_71925834_group_cache.npy
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:30aba90b1e30199cd25b05651f293551ea64e9b1341a7af0fdb310e82a47f594
|
| 3 |
+
size 921657
|
llama2_Llama-2-7b-hf_fs4_mmlu_f23ef2af.json
ADDED
|
@@ -0,0 +1 @@
|
|
|
|
|
|
|
| 1 |
+
{"uid": "f23ef2af", "model": "meta-llama/Llama-2-7b-hf", "model_name": "Llama-2-7b-hf", "group": "llama2", "model_args": "pretrained=meta-llama/Llama-2-7b-hf,dtype=auto,gpu_memory_utilization=0.7,tensor_parallel_size=1,data_parallel_size=4,max_model_len=4096", "fewshot": 4, "tasks": "mmlu"}
|
llama2_Llama-2-7b-hf_fs4_mmlu_f23ef2af_group_cache.npy
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:4959a6a4ca4a80b227679d14ffe84a8ec948bc979c47664c2fcbcc61c54a00ee
|
| 3 |
+
size 7374337
|
llama2_Llama-2-7b-hf_fs5_mmlu_3a7720c8.json
ADDED
|
@@ -0,0 +1 @@
|
|
|
|
|
|
|
| 1 |
+
{"uid": "3a7720c8", "model": "meta-llama/Llama-2-7b-hf", "model_name": "Llama-2-7b-hf", "group": "llama2", "model_args": "pretrained=meta-llama/Llama-2-7b-hf,dtype=auto,gpu_memory_utilization=0.7,tensor_parallel_size=1,data_parallel_size=4,max_model_len=4096", "fewshot": 5, "tasks": "mmlu"}
|
llama2_Llama-2-7b-hf_fs5_mmlu_3a7720c8_group_cache.npy
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:14729da46563d9f29fe57c3d77b4e1aece09c260afba78f886d8ece828acb953
|
| 3 |
+
size 2304372
|
llama2_tulu-2-7b_fs0_mmlu_6d638552.json
ADDED
|
@@ -0,0 +1 @@
|
|
|
|
|
|
|
| 1 |
+
{"uid": "6d638552", "model": "allenai/tulu-2-7b", "model_name": "tulu-2-7b", "group": "llama2", "model_args": "pretrained=allenai/tulu-2-7b,dtype=auto,gpu_memory_utilization=0.7,tensor_parallel_size=1,data_parallel_size=4,max_model_len=4096", "fewshot": 0, "tasks": "mmlu"}
|
llama2_tulu-2-7b_fs0_mmlu_6d638552_group_cache.npy
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:3dcadab5f46e260e585f615a0cc400ba6f0cf9b325790b40e5a497d879cf2737
|
| 3 |
+
size 4608883
|