diff --git a/llama1_guanaco-7B-HF_fs0_mmlu_316695dc.json b/llama1_guanaco-7B-HF_fs0_mmlu_316695dc.json new file mode 100644 index 0000000000000000000000000000000000000000..9ee6aa690f444d23479402744c492a6ec270569b --- /dev/null +++ b/llama1_guanaco-7B-HF_fs0_mmlu_316695dc.json @@ -0,0 +1 @@ +{"uid": "316695dc", "model": "TheBloke/guanaco-7B-HF", "model_name": "guanaco-7B-HF", "group": "llama1", "model_args": "pretrained=TheBloke/guanaco-7B-HF,dtype=auto,gpu_memory_utilization=0.7,tensor_parallel_size=1,data_parallel_size=4,max_model_len=2048", "fewshot": 0, "tasks": "mmlu"} \ No newline at end of file diff --git a/llama1_guanaco-7B-HF_fs0_mmlu_316695dc_group_cache.npy b/llama1_guanaco-7B-HF_fs0_mmlu_316695dc_group_cache.npy new file mode 100644 index 0000000000000000000000000000000000000000..a33d49964676e0c16a199d3740d4a9fb04a06b30 --- /dev/null +++ b/llama1_guanaco-7B-HF_fs0_mmlu_316695dc_group_cache.npy @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:70152bc658660735ea4adbcd67d86ceae5b02447faff0ce399c9deb69b0a40c5 +size 5069770 diff --git a/llama1_guanaco-7B-HF_fs1_mmlu_d6011d37.json b/llama1_guanaco-7B-HF_fs1_mmlu_d6011d37.json new file mode 100644 index 0000000000000000000000000000000000000000..f72b617e15331df7ac53ebcb0159c62e12d85a2c --- /dev/null +++ b/llama1_guanaco-7B-HF_fs1_mmlu_d6011d37.json @@ -0,0 +1 @@ +{"uid": "d6011d37", "model": "TheBloke/guanaco-7B-HF", "model_name": "guanaco-7B-HF", "group": "llama1", "model_args": "pretrained=TheBloke/guanaco-7B-HF,dtype=auto,gpu_memory_utilization=0.7,tensor_parallel_size=1,data_parallel_size=4,max_model_len=2048", "fewshot": 1, "tasks": "mmlu"} \ No newline at end of file diff --git a/llama1_guanaco-7B-HF_fs1_mmlu_d6011d37_group_cache.npy b/llama1_guanaco-7B-HF_fs1_mmlu_d6011d37_group_cache.npy new file mode 100644 index 0000000000000000000000000000000000000000..21743f6972acd770f390079a2ff2e3c15139f187 --- /dev/null +++ b/llama1_guanaco-7B-HF_fs1_mmlu_d6011d37_group_cache.npy @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:4a05d110db5f101af6e9d70501ccf3c2246d2bfe8534d52cf2b00a35295608d1 +size 4608861 diff --git a/llama1_guanaco-7B-HF_fs2_mmlu_f2bd887d.json b/llama1_guanaco-7B-HF_fs2_mmlu_f2bd887d.json new file mode 100644 index 0000000000000000000000000000000000000000..be6b8dc6ccb7f16c1640dfc21cfcf84970c1f32e --- /dev/null +++ b/llama1_guanaco-7B-HF_fs2_mmlu_f2bd887d.json @@ -0,0 +1 @@ +{"uid": "f2bd887d", "model": "TheBloke/guanaco-7B-HF", "model_name": "guanaco-7B-HF", "group": "llama1", "model_args": "pretrained=TheBloke/guanaco-7B-HF,dtype=auto,gpu_memory_utilization=0.7,tensor_parallel_size=1,data_parallel_size=4,max_model_len=2048", "fewshot": 2, "tasks": "mmlu"} \ No newline at end of file diff --git a/llama1_guanaco-7B-HF_fs2_mmlu_f2bd887d_group_cache.npy b/llama1_guanaco-7B-HF_fs2_mmlu_f2bd887d_group_cache.npy new file mode 100644 index 0000000000000000000000000000000000000000..fafc850d90c5c7790437cd0bb9531aa7d7d534e8 --- /dev/null +++ b/llama1_guanaco-7B-HF_fs2_mmlu_f2bd887d_group_cache.npy @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:593748f3670894064ed1b6aca5ee46f25f1998d7351209b4483c947a0f4a4cf9 +size 5530679 diff --git a/llama1_guanaco-7B-HF_fs3_mmlu_61b8a8e3.json b/llama1_guanaco-7B-HF_fs3_mmlu_61b8a8e3.json new file mode 100644 index 0000000000000000000000000000000000000000..456b56fcd0aa865ae67b8b3ea92a8d47e451fb97 --- /dev/null +++ b/llama1_guanaco-7B-HF_fs3_mmlu_61b8a8e3.json @@ -0,0 +1 @@ +{"uid": "61b8a8e3", "model": "TheBloke/guanaco-7B-HF", "model_name": "guanaco-7B-HF", "group": "llama1", "model_args": "pretrained=TheBloke/guanaco-7B-HF,dtype=auto,gpu_memory_utilization=0.7,tensor_parallel_size=1,data_parallel_size=4,max_model_len=2048", "fewshot": 3, "tasks": "mmlu"} \ No newline at end of file diff --git a/llama1_guanaco-7B-HF_fs3_mmlu_61b8a8e3_group_cache.npy b/llama1_guanaco-7B-HF_fs3_mmlu_61b8a8e3_group_cache.npy new file mode 100644 index 0000000000000000000000000000000000000000..dd5c66c3cfb62e7f6780aac5dd755b3fd7569c4c --- /dev/null +++ b/llama1_guanaco-7B-HF_fs3_mmlu_61b8a8e3_group_cache.npy @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:c71f2a48e5e248148b5264c72c25f5c2cbfae5177866eca1ff7683ea07bb108f +size 2304338 diff --git a/llama1_guanaco-7B-HF_fs4_mmlu_28dff3fc.json b/llama1_guanaco-7B-HF_fs4_mmlu_28dff3fc.json new file mode 100644 index 0000000000000000000000000000000000000000..d772e3be802046019c9273620082ad40b8850ebb --- /dev/null +++ b/llama1_guanaco-7B-HF_fs4_mmlu_28dff3fc.json @@ -0,0 +1 @@ +{"uid": "28dff3fc", "model": "TheBloke/guanaco-7B-HF", "model_name": "guanaco-7B-HF", "group": "llama1", "model_args": "pretrained=TheBloke/guanaco-7B-HF,dtype=auto,gpu_memory_utilization=0.7,tensor_parallel_size=1,data_parallel_size=4,max_model_len=2048", "fewshot": 4, "tasks": "mmlu"} \ No newline at end of file diff --git a/llama1_guanaco-7B-HF_fs4_mmlu_28dff3fc_group_cache.npy b/llama1_guanaco-7B-HF_fs4_mmlu_28dff3fc_group_cache.npy new file mode 100644 index 0000000000000000000000000000000000000000..962a17987690ffb77233f32c3d1c0e1ec4623149 --- /dev/null +++ b/llama1_guanaco-7B-HF_fs4_mmlu_28dff3fc_group_cache.npy @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:18ede2837c523bd2c369ee73aea3b817d0b798f261274800488d19d6ff5672cc +size 3226156 diff --git a/llama1_guanaco-7B-HF_fs5_mmlu_b583442d.json b/llama1_guanaco-7B-HF_fs5_mmlu_b583442d.json new file mode 100644 index 0000000000000000000000000000000000000000..ccd81e2f1b7473fe53eb5c66950b6ffea3829eb0 --- /dev/null +++ b/llama1_guanaco-7B-HF_fs5_mmlu_b583442d.json @@ -0,0 +1 @@ +{"uid": "b583442d", "model": "TheBloke/guanaco-7B-HF", "model_name": "guanaco-7B-HF", "group": "llama1", "model_args": "pretrained=TheBloke/guanaco-7B-HF,dtype=auto,gpu_memory_utilization=0.7,tensor_parallel_size=1,data_parallel_size=4,max_model_len=2048", "fewshot": 5, "tasks": "mmlu"} \ No newline at end of file diff --git a/llama1_guanaco-7B-HF_fs5_mmlu_b583442d_group_cache.npy b/llama1_guanaco-7B-HF_fs5_mmlu_b583442d_group_cache.npy new file mode 100644 index 0000000000000000000000000000000000000000..e37bd72b06faebce8195ed79a15fbee8b276af79 --- /dev/null +++ b/llama1_guanaco-7B-HF_fs5_mmlu_b583442d_group_cache.npy @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:5149127c6155eb438fbd81f067a94b5df51463bb3e9223468b53d3ce97988352 +size 2765247 diff --git a/llama1_llama-7b_fs0_mmlu_dafad174.json b/llama1_llama-7b_fs0_mmlu_dafad174.json new file mode 100644 index 0000000000000000000000000000000000000000..ffa92048b8a43810052f1bfda5dbc0e32d425489 --- /dev/null +++ b/llama1_llama-7b_fs0_mmlu_dafad174.json @@ -0,0 +1 @@ +{"uid": "dafad174", "model": "huggyllama/llama-7b", "model_name": "llama-7b", "group": "llama1", "model_args": "pretrained=huggyllama/llama-7b,dtype=auto,gpu_memory_utilization=0.7,tensor_parallel_size=1,data_parallel_size=4,max_model_len=2048", "fewshot": 0, "tasks": "mmlu"} \ No newline at end of file diff --git a/llama1_llama-7b_fs0_mmlu_dafad174_group_cache.npy b/llama1_llama-7b_fs0_mmlu_dafad174_group_cache.npy new file mode 100644 index 0000000000000000000000000000000000000000..d5a5b8edfd34dfcfa7a7420d0b014247368d7855 --- /dev/null +++ b/llama1_llama-7b_fs0_mmlu_dafad174_group_cache.npy @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:356637341b1ba24f905d14022123928535d9835d04d6107dd7fce77ef9e47529 +size 3687054 diff --git a/llama1_llama-7b_fs1_mmlu_cf545325.json b/llama1_llama-7b_fs1_mmlu_cf545325.json new file mode 100644 index 0000000000000000000000000000000000000000..283b78b30d1549c3249dd2b66b8609f8ac99b8f2 --- /dev/null +++ b/llama1_llama-7b_fs1_mmlu_cf545325.json @@ -0,0 +1 @@ +{"uid": "cf545325", "model": "huggyllama/llama-7b", "model_name": "llama-7b", "group": "llama1", "model_args": "pretrained=huggyllama/llama-7b,dtype=auto,gpu_memory_utilization=0.7,tensor_parallel_size=1,data_parallel_size=4,max_model_len=2048", "fewshot": 1, "tasks": "mmlu"} \ No newline at end of file diff --git a/llama1_llama-7b_fs1_mmlu_cf545325_group_cache.npy b/llama1_llama-7b_fs1_mmlu_cf545325_group_cache.npy new file mode 100644 index 0000000000000000000000000000000000000000..23adc06a064b1ec342001cecb20e15cfd61d5610 --- /dev/null +++ b/llama1_llama-7b_fs1_mmlu_cf545325_group_cache.npy @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:17ce112c391e2bd69a930b9c58307259d7319320673440858a7834731e88a12c +size 460734 diff --git a/llama1_llama-7b_fs2_mmlu_a62f4c1e.json b/llama1_llama-7b_fs2_mmlu_a62f4c1e.json new file mode 100644 index 0000000000000000000000000000000000000000..c8c68fa976e56c321214657d38e9cb8c971296a9 --- /dev/null +++ b/llama1_llama-7b_fs2_mmlu_a62f4c1e.json @@ -0,0 +1 @@ +{"uid": "a62f4c1e", "model": "huggyllama/llama-7b", "model_name": "llama-7b", "group": "llama1", "model_args": "pretrained=huggyllama/llama-7b,dtype=auto,gpu_memory_utilization=0.7,tensor_parallel_size=1,data_parallel_size=4,max_model_len=2048", "fewshot": 2, "tasks": "mmlu"} \ No newline at end of file diff --git a/llama1_llama-7b_fs2_mmlu_a62f4c1e_group_cache.npy b/llama1_llama-7b_fs2_mmlu_a62f4c1e_group_cache.npy new file mode 100644 index 0000000000000000000000000000000000000000..94930ce14014bb8042bad5722e68a49b215e3fff --- /dev/null +++ b/llama1_llama-7b_fs2_mmlu_a62f4c1e_group_cache.npy @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:582b681dcb7ce95506add7b355a9709993c10495f69049c40287e523e64575b3 +size 1843429 diff --git a/llama1_llama-7b_fs3_mmlu_374ca650.json b/llama1_llama-7b_fs3_mmlu_374ca650.json new file mode 100644 index 0000000000000000000000000000000000000000..898b970eab8a028aef3504736249309330049ee1 --- /dev/null +++ b/llama1_llama-7b_fs3_mmlu_374ca650.json @@ -0,0 +1 @@ +{"uid": "374ca650", "model": "huggyllama/llama-7b", "model_name": "llama-7b", "group": "llama1", "model_args": "pretrained=huggyllama/llama-7b,dtype=auto,gpu_memory_utilization=0.7,tensor_parallel_size=1,data_parallel_size=4,max_model_len=2048", "fewshot": 3, "tasks": "mmlu"} \ No newline at end of file diff --git a/llama1_llama-7b_fs3_mmlu_374ca650_group_cache.npy b/llama1_llama-7b_fs3_mmlu_374ca650_group_cache.npy new file mode 100644 index 0000000000000000000000000000000000000000..4f1ee6c8ddf8558324cc49cd8f65f097dc7b14e5 --- /dev/null +++ b/llama1_llama-7b_fs3_mmlu_374ca650_group_cache.npy @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:4bdee5f4922163eec5a47bab1918ce44dc6a789c1704a21c7cf215268c444a82 +size 921633 diff --git a/llama1_llama-7b_fs4_mmlu_72b24354.json b/llama1_llama-7b_fs4_mmlu_72b24354.json new file mode 100644 index 0000000000000000000000000000000000000000..47f2393a04f51ccaec9235cfdeacbaca0a7fcf05 --- /dev/null +++ b/llama1_llama-7b_fs4_mmlu_72b24354.json @@ -0,0 +1 @@ +{"uid": "72b24354", "model": "huggyllama/llama-7b", "model_name": "llama-7b", "group": "llama1", "model_args": "pretrained=huggyllama/llama-7b,dtype=auto,gpu_memory_utilization=0.7,tensor_parallel_size=1,data_parallel_size=4,max_model_len=2048", "fewshot": 4, "tasks": "mmlu"} \ No newline at end of file diff --git a/llama1_llama-7b_fs4_mmlu_72b24354_group_cache.npy b/llama1_llama-7b_fs4_mmlu_72b24354_group_cache.npy new file mode 100644 index 0000000000000000000000000000000000000000..26802b330f495eb403d5d05b144f84bdec3394d7 --- /dev/null +++ b/llama1_llama-7b_fs4_mmlu_72b24354_group_cache.npy @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:3717e22443b9bfd5bf86d02bd7c8aa4fb7ef6e7dfb287648336e0b21dced985a +size 1382531 diff --git a/llama1_llama-7b_fs5_mmlu_479ca4ec.json b/llama1_llama-7b_fs5_mmlu_479ca4ec.json new file mode 100644 index 0000000000000000000000000000000000000000..3e4a9f49b6a45d545ec71542d57a0422e3e80e58 --- /dev/null +++ b/llama1_llama-7b_fs5_mmlu_479ca4ec.json @@ -0,0 +1 @@ +{"uid": "479ca4ec", "model": "huggyllama/llama-7b", "model_name": "llama-7b", "group": "llama1", "model_args": "pretrained=huggyllama/llama-7b,dtype=auto,gpu_memory_utilization=0.7,tensor_parallel_size=1,data_parallel_size=4,max_model_len=2048", "fewshot": 5, "tasks": "mmlu"} \ No newline at end of file diff --git a/llama1_llama-7b_fs5_mmlu_479ca4ec_group_cache.npy b/llama1_llama-7b_fs5_mmlu_479ca4ec_group_cache.npy new file mode 100644 index 0000000000000000000000000000000000000000..75a7d349158adf12b668cdfe9e5286bf2ff69b64 --- /dev/null +++ b/llama1_llama-7b_fs5_mmlu_479ca4ec_group_cache.npy @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:1c5d2b377695b037dc9d59b1c431a9b1c7520b00424ba81ca6b2ed39d9a60c90 +size 4147952 diff --git a/llama2_Llama-2-7b-chat_fs0_mmlu_767c4ad0.json b/llama2_Llama-2-7b-chat_fs0_mmlu_767c4ad0.json new file mode 100644 index 0000000000000000000000000000000000000000..38586e7ed2a53b7e4b55d0c7cb4708be83aff4ee --- /dev/null +++ b/llama2_Llama-2-7b-chat_fs0_mmlu_767c4ad0.json @@ -0,0 +1 @@ +{"uid": "767c4ad0", "model": "meta-llama/Llama-2-7b-chat", "model_name": "Llama-2-7b-chat", "group": "llama2", "model_args": "pretrained=meta-llama/Llama-2-7b-chat,dtype=auto,gpu_memory_utilization=0.7,tensor_parallel_size=1,data_parallel_size=0,max_model_len=4096", "fewshot": 0, "tasks": "mmlu"} \ No newline at end of file diff --git a/llama2_Llama-2-7b-chat_fs0_mmlu_767c4ad0_group_cache.npy b/llama2_Llama-2-7b-chat_fs0_mmlu_767c4ad0_group_cache.npy new file mode 100644 index 0000000000000000000000000000000000000000..c719d808907b9160a237c2890fa83cf9b4bd062f --- /dev/null +++ b/llama2_Llama-2-7b-chat_fs0_mmlu_767c4ad0_group_cache.npy @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:79edee3c6dc0f6b880ea77ec109cc346c8c7e43cbb35598de87dcb06e7204249 +size 9678896 diff --git a/llama2_Llama-2-7b-chat_fs1_mmlu_7c7ae7be.json b/llama2_Llama-2-7b-chat_fs1_mmlu_7c7ae7be.json new file mode 100644 index 0000000000000000000000000000000000000000..8ad875fc8c4f51262ac2d15a02a60b06a8505961 --- /dev/null +++ b/llama2_Llama-2-7b-chat_fs1_mmlu_7c7ae7be.json @@ -0,0 +1 @@ +{"uid": "7c7ae7be", "model": "meta-llama/Llama-2-7b-chat", "model_name": "Llama-2-7b-chat", "group": "llama2", "model_args": "pretrained=meta-llama/Llama-2-7b-chat,dtype=auto,gpu_memory_utilization=0.7,tensor_parallel_size=1,data_parallel_size=0,max_model_len=4096", "fewshot": 1, "tasks": "mmlu"} \ No newline at end of file diff --git a/llama2_Llama-2-7b-chat_fs1_mmlu_7c7ae7be_group_cache.npy b/llama2_Llama-2-7b-chat_fs1_mmlu_7c7ae7be_group_cache.npy new file mode 100644 index 0000000000000000000000000000000000000000..663167cafa4f03cbd6287c50b75204e736c56a10 --- /dev/null +++ b/llama2_Llama-2-7b-chat_fs1_mmlu_7c7ae7be_group_cache.npy @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:7a215d49916fce04d60267a8d555144b17b5f7d4330b179ea953622f4ff70b54 +size 10600734 diff --git a/llama2_Llama-2-7b-chat_fs2_mmlu_31b67daa.json b/llama2_Llama-2-7b-chat_fs2_mmlu_31b67daa.json new file mode 100644 index 0000000000000000000000000000000000000000..7219c629d9d49b1fee399207f88273c77c3e81fa --- /dev/null +++ b/llama2_Llama-2-7b-chat_fs2_mmlu_31b67daa.json @@ -0,0 +1 @@ +{"uid": "31b67daa", "model": "meta-llama/Llama-2-7b-chat", "model_name": "Llama-2-7b-chat", "group": "llama2", "model_args": "pretrained=meta-llama/Llama-2-7b-chat,dtype=auto,gpu_memory_utilization=0.7,tensor_parallel_size=1,data_parallel_size=0,max_model_len=4096", "fewshot": 2, "tasks": "mmlu"} \ No newline at end of file diff --git a/llama2_Llama-2-7b-chat_fs2_mmlu_31b67daa_group_cache.npy b/llama2_Llama-2-7b-chat_fs2_mmlu_31b67daa_group_cache.npy new file mode 100644 index 0000000000000000000000000000000000000000..9edcb40066007df8c9150cf1344c210e2e539854 --- /dev/null +++ b/llama2_Llama-2-7b-chat_fs2_mmlu_31b67daa_group_cache.npy @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:08af54bab49ec1257b6430a6490622331c34d16b8b8821f5e992c32bf0bd9b15 +size 10139815 diff --git a/llama2_Llama-2-7b-chat_fs3_mmlu_34207011.json b/llama2_Llama-2-7b-chat_fs3_mmlu_34207011.json new file mode 100644 index 0000000000000000000000000000000000000000..11009b196ec74be3327c5e33415457a1ee86afab --- /dev/null +++ b/llama2_Llama-2-7b-chat_fs3_mmlu_34207011.json @@ -0,0 +1 @@ +{"uid": "34207011", "model": "meta-llama/Llama-2-7b-chat", "model_name": "Llama-2-7b-chat", "group": "llama2", "model_args": "pretrained=meta-llama/Llama-2-7b-chat,dtype=auto,gpu_memory_utilization=0.7,tensor_parallel_size=1,data_parallel_size=0,max_model_len=4096", "fewshot": 3, "tasks": "mmlu"} \ No newline at end of file diff --git a/llama2_Llama-2-7b-chat_fs3_mmlu_34207011_group_cache.npy b/llama2_Llama-2-7b-chat_fs3_mmlu_34207011_group_cache.npy new file mode 100644 index 0000000000000000000000000000000000000000..6bce36b117020e254a5dc0fcba90e8a2d712f4f1 --- /dev/null +++ b/llama2_Llama-2-7b-chat_fs3_mmlu_34207011_group_cache.npy @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:29c37201eabcc3bae197b9d10c85e77259fe39a7557756974a9f1d85035cba5e +size 8296151 diff --git a/llama2_Llama-2-7b-chat_fs4_mmlu_7fabd9ca.json b/llama2_Llama-2-7b-chat_fs4_mmlu_7fabd9ca.json new file mode 100644 index 0000000000000000000000000000000000000000..db8f61ba1a7ec822c54e8314081754cb77b419b4 --- /dev/null +++ b/llama2_Llama-2-7b-chat_fs4_mmlu_7fabd9ca.json @@ -0,0 +1 @@ +{"uid": "7fabd9ca", "model": "meta-llama/Llama-2-7b-chat", "model_name": "Llama-2-7b-chat", "group": "llama2", "model_args": "pretrained=meta-llama/Llama-2-7b-chat,dtype=auto,gpu_memory_utilization=0.7,tensor_parallel_size=1,data_parallel_size=0,max_model_len=4096", "fewshot": 4, "tasks": "mmlu"} \ No newline at end of file diff --git a/llama2_Llama-2-7b-chat_fs4_mmlu_7fabd9ca_group_cache.npy b/llama2_Llama-2-7b-chat_fs4_mmlu_7fabd9ca_group_cache.npy new file mode 100644 index 0000000000000000000000000000000000000000..7449d61fe8b2f8791114874b9970b79b8006ed05 --- /dev/null +++ b/llama2_Llama-2-7b-chat_fs4_mmlu_7fabd9ca_group_cache.npy @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:df52b412e7c6b708f6e2cf44697e07316d153ac06642c63f3d066d3ab3504036 +size 6913424 diff --git a/llama2_Llama-2-7b-chat_fs5_mmlu_952c2ce7.json b/llama2_Llama-2-7b-chat_fs5_mmlu_952c2ce7.json new file mode 100644 index 0000000000000000000000000000000000000000..7087403b7066104d30c62a5ef0c504c581e723c8 --- /dev/null +++ b/llama2_Llama-2-7b-chat_fs5_mmlu_952c2ce7.json @@ -0,0 +1 @@ +{"uid": "952c2ce7", "model": "meta-llama/Llama-2-7b-chat", "model_name": "Llama-2-7b-chat", "group": "llama2", "model_args": "pretrained=meta-llama/Llama-2-7b-chat,dtype=auto,gpu_memory_utilization=0.7,tensor_parallel_size=1,data_parallel_size=0,max_model_len=4096", "fewshot": 5, "tasks": "mmlu"} \ No newline at end of file diff --git a/llama2_Llama-2-7b-chat_fs5_mmlu_952c2ce7_group_cache.npy b/llama2_Llama-2-7b-chat_fs5_mmlu_952c2ce7_group_cache.npy new file mode 100644 index 0000000000000000000000000000000000000000..3afe602f9fc638e7f5613d9a3b4d6c389aa781b2 --- /dev/null +++ b/llama2_Llama-2-7b-chat_fs5_mmlu_952c2ce7_group_cache.npy @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:1344c5e9e7bf5d87c8d8aaa17f212b37ccac4c2a7d16a9319e57db541a060295 +size 9217977 diff --git a/llama2_Llama-2-7b-hf_fs0_mmlu_31148294.json b/llama2_Llama-2-7b-hf_fs0_mmlu_31148294.json new file mode 100644 index 0000000000000000000000000000000000000000..34a74cc396e4e3609eb18bb613bb448fc6787dd8 --- /dev/null +++ b/llama2_Llama-2-7b-hf_fs0_mmlu_31148294.json @@ -0,0 +1 @@ +{"uid": "31148294", "model": "meta-llama/Llama-2-7b-hf", "model_name": "Llama-2-7b-hf", "group": "llama2", "model_args": "pretrained=meta-llama/Llama-2-7b-hf,dtype=auto,gpu_memory_utilization=0.7,tensor_parallel_size=1,data_parallel_size=0,max_model_len=4096", "fewshot": 0, "tasks": "mmlu"} \ No newline at end of file diff --git a/llama2_Llama-2-7b-hf_fs0_mmlu_31148294_group_cache.npy b/llama2_Llama-2-7b-hf_fs0_mmlu_31148294_group_cache.npy new file mode 100644 index 0000000000000000000000000000000000000000..e4eda613c78607bc44eb5d7c76274ec2ec208b6f --- /dev/null +++ b/llama2_Llama-2-7b-hf_fs0_mmlu_31148294_group_cache.npy @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:e0d1b4cab918ea25214676366098ded3406a801797a694ddf8ae8cf84a184ede +size 2765285 diff --git a/llama2_Llama-2-7b-hf_fs1_mmlu_afa46a6c.json b/llama2_Llama-2-7b-hf_fs1_mmlu_afa46a6c.json new file mode 100644 index 0000000000000000000000000000000000000000..02de1468e3d3ee9e0ea2633846b2759f33d483bf --- /dev/null +++ b/llama2_Llama-2-7b-hf_fs1_mmlu_afa46a6c.json @@ -0,0 +1 @@ +{"uid": "afa46a6c", "model": "meta-llama/Llama-2-7b-hf", "model_name": "Llama-2-7b-hf", "group": "llama2", "model_args": "pretrained=meta-llama/Llama-2-7b-hf,dtype=auto,gpu_memory_utilization=0.7,tensor_parallel_size=1,data_parallel_size=4,max_model_len=4096", "fewshot": 1, "tasks": "mmlu"} \ No newline at end of file diff --git a/llama2_Llama-2-7b-hf_fs1_mmlu_afa46a6c_group_cache.npy b/llama2_Llama-2-7b-hf_fs1_mmlu_afa46a6c_group_cache.npy new file mode 100644 index 0000000000000000000000000000000000000000..a702fafbe674fc71e52b4722d7b41f253aa24d9a --- /dev/null +++ b/llama2_Llama-2-7b-hf_fs1_mmlu_afa46a6c_group_cache.npy @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:60545517d76b3024862f7c908888263ce2986ce853647b3e13afd30eb0b605af +size 5069796 diff --git a/llama2_Llama-2-7b-hf_fs2_mmlu_d9dcdc1b.json b/llama2_Llama-2-7b-hf_fs2_mmlu_d9dcdc1b.json new file mode 100644 index 0000000000000000000000000000000000000000..1da0452acd01098cdbe76d9226a174076997f1cb --- /dev/null +++ b/llama2_Llama-2-7b-hf_fs2_mmlu_d9dcdc1b.json @@ -0,0 +1 @@ +{"uid": "d9dcdc1b", "model": "meta-llama/Llama-2-7b-hf", "model_name": "Llama-2-7b-hf", "group": "llama2", "model_args": "pretrained=meta-llama/Llama-2-7b-hf,dtype=auto,gpu_memory_utilization=0.7,tensor_parallel_size=1,data_parallel_size=4,max_model_len=4096", "fewshot": 2, "tasks": "mmlu"} \ No newline at end of file diff --git a/llama2_Llama-2-7b-hf_fs2_mmlu_d9dcdc1b_group_cache.npy b/llama2_Llama-2-7b-hf_fs2_mmlu_d9dcdc1b_group_cache.npy new file mode 100644 index 0000000000000000000000000000000000000000..f531a9947a9368ca8f8720ce24fb3669e88758db --- /dev/null +++ b/llama2_Llama-2-7b-hf_fs2_mmlu_d9dcdc1b_group_cache.npy @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:57b4b21679aa758d98aae02db4505d03d7468d485c350c55b2f58d687f51c29a +size 3687093 diff --git a/llama2_Llama-2-7b-hf_fs3_mmlu_71925834.json b/llama2_Llama-2-7b-hf_fs3_mmlu_71925834.json new file mode 100644 index 0000000000000000000000000000000000000000..fb2f1713a70d447958ba878d8dea7704add07403 --- /dev/null +++ b/llama2_Llama-2-7b-hf_fs3_mmlu_71925834.json @@ -0,0 +1 @@ +{"uid": "71925834", "model": "meta-llama/Llama-2-7b-hf", "model_name": "Llama-2-7b-hf", "group": "llama2", "model_args": "pretrained=meta-llama/Llama-2-7b-hf,dtype=auto,gpu_memory_utilization=0.7,tensor_parallel_size=1,data_parallel_size=4,max_model_len=4096", "fewshot": 3, "tasks": "mmlu"} \ No newline at end of file diff --git a/llama2_Llama-2-7b-hf_fs3_mmlu_71925834_group_cache.npy b/llama2_Llama-2-7b-hf_fs3_mmlu_71925834_group_cache.npy new file mode 100644 index 0000000000000000000000000000000000000000..7fd356590872d19be470f85e309beabc0607d3c0 --- /dev/null +++ b/llama2_Llama-2-7b-hf_fs3_mmlu_71925834_group_cache.npy @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:30aba90b1e30199cd25b05651f293551ea64e9b1341a7af0fdb310e82a47f594 +size 921657 diff --git a/llama2_Llama-2-7b-hf_fs4_mmlu_f23ef2af.json b/llama2_Llama-2-7b-hf_fs4_mmlu_f23ef2af.json new file mode 100644 index 0000000000000000000000000000000000000000..3644b99540def9df16362ebcc9c578964ecd34f1 --- /dev/null +++ b/llama2_Llama-2-7b-hf_fs4_mmlu_f23ef2af.json @@ -0,0 +1 @@ +{"uid": "f23ef2af", "model": "meta-llama/Llama-2-7b-hf", "model_name": "Llama-2-7b-hf", "group": "llama2", "model_args": "pretrained=meta-llama/Llama-2-7b-hf,dtype=auto,gpu_memory_utilization=0.7,tensor_parallel_size=1,data_parallel_size=4,max_model_len=4096", "fewshot": 4, "tasks": "mmlu"} \ No newline at end of file diff --git a/llama2_Llama-2-7b-hf_fs4_mmlu_f23ef2af_group_cache.npy b/llama2_Llama-2-7b-hf_fs4_mmlu_f23ef2af_group_cache.npy new file mode 100644 index 0000000000000000000000000000000000000000..3d0e3641c156df22e41a5cda0be87a2256340471 --- /dev/null +++ b/llama2_Llama-2-7b-hf_fs4_mmlu_f23ef2af_group_cache.npy @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:4959a6a4ca4a80b227679d14ffe84a8ec948bc979c47664c2fcbcc61c54a00ee +size 7374337 diff --git a/llama2_Llama-2-7b-hf_fs5_mmlu_3a7720c8.json b/llama2_Llama-2-7b-hf_fs5_mmlu_3a7720c8.json new file mode 100644 index 0000000000000000000000000000000000000000..7f72736bb8d3fce721b396ab95e9a16d0607c0c4 --- /dev/null +++ b/llama2_Llama-2-7b-hf_fs5_mmlu_3a7720c8.json @@ -0,0 +1 @@ +{"uid": "3a7720c8", "model": "meta-llama/Llama-2-7b-hf", "model_name": "Llama-2-7b-hf", "group": "llama2", "model_args": "pretrained=meta-llama/Llama-2-7b-hf,dtype=auto,gpu_memory_utilization=0.7,tensor_parallel_size=1,data_parallel_size=4,max_model_len=4096", "fewshot": 5, "tasks": "mmlu"} \ No newline at end of file diff --git a/llama2_Llama-2-7b-hf_fs5_mmlu_3a7720c8_group_cache.npy b/llama2_Llama-2-7b-hf_fs5_mmlu_3a7720c8_group_cache.npy new file mode 100644 index 0000000000000000000000000000000000000000..1b0428a84e2e0766e1173d56849e2abd727bb804 --- /dev/null +++ b/llama2_Llama-2-7b-hf_fs5_mmlu_3a7720c8_group_cache.npy @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:14729da46563d9f29fe57c3d77b4e1aece09c260afba78f886d8ece828acb953 +size 2304372 diff --git a/llama2_tulu-2-7b_fs0_mmlu_6d638552.json b/llama2_tulu-2-7b_fs0_mmlu_6d638552.json new file mode 100644 index 0000000000000000000000000000000000000000..e0d7fbd940f336a01f8a908b05f16e95fe700d26 --- /dev/null +++ b/llama2_tulu-2-7b_fs0_mmlu_6d638552.json @@ -0,0 +1 @@ +{"uid": "6d638552", "model": "allenai/tulu-2-7b", "model_name": "tulu-2-7b", "group": "llama2", "model_args": "pretrained=allenai/tulu-2-7b,dtype=auto,gpu_memory_utilization=0.7,tensor_parallel_size=1,data_parallel_size=4,max_model_len=4096", "fewshot": 0, "tasks": "mmlu"} \ No newline at end of file diff --git a/llama2_tulu-2-7b_fs0_mmlu_6d638552_group_cache.npy b/llama2_tulu-2-7b_fs0_mmlu_6d638552_group_cache.npy new file mode 100644 index 0000000000000000000000000000000000000000..14a9fabd8f51e635aaa68c0739b43008ad612df4 --- /dev/null +++ b/llama2_tulu-2-7b_fs0_mmlu_6d638552_group_cache.npy @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:3dcadab5f46e260e585f615a0cc400ba6f0cf9b325790b40e5a497d879cf2737 +size 4608883 diff --git a/llama2_tulu-2-7b_fs1_mmlu_61837a56.json b/llama2_tulu-2-7b_fs1_mmlu_61837a56.json new file mode 100644 index 0000000000000000000000000000000000000000..9bfd67195e2706c78999c27e6f237f0a974f3907 --- /dev/null +++ b/llama2_tulu-2-7b_fs1_mmlu_61837a56.json @@ -0,0 +1 @@ +{"uid": "61837a56", "model": "allenai/tulu-2-7b", "model_name": "tulu-2-7b", "group": "llama2", "model_args": "pretrained=allenai/tulu-2-7b,dtype=auto,gpu_memory_utilization=0.7,tensor_parallel_size=1,data_parallel_size=4,max_model_len=4096", "fewshot": 1, "tasks": "mmlu"} \ No newline at end of file diff --git a/llama2_tulu-2-7b_fs1_mmlu_61837a56_group_cache.npy b/llama2_tulu-2-7b_fs1_mmlu_61837a56_group_cache.npy new file mode 100644 index 0000000000000000000000000000000000000000..e4ce7f50dc173424c37a56db49195e99b8667d3f --- /dev/null +++ b/llama2_tulu-2-7b_fs1_mmlu_61837a56_group_cache.npy @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:c19f0107d8662caab7f2e8b4723eeb93dc21c6a5b0584d21a5823a2e74d91263 +size 1382552 diff --git a/llama2_tulu-2-7b_fs2_mmlu_7e8c7ce7.json b/llama2_tulu-2-7b_fs2_mmlu_7e8c7ce7.json new file mode 100644 index 0000000000000000000000000000000000000000..2210016641c8e437537189f41666ef875befe55e --- /dev/null +++ b/llama2_tulu-2-7b_fs2_mmlu_7e8c7ce7.json @@ -0,0 +1 @@ +{"uid": "7e8c7ce7", "model": "allenai/tulu-2-7b", "model_name": "tulu-2-7b", "group": "llama2", "model_args": "pretrained=allenai/tulu-2-7b,dtype=auto,gpu_memory_utilization=0.7,tensor_parallel_size=1,data_parallel_size=4,max_model_len=4096", "fewshot": 2, "tasks": "mmlu"} \ No newline at end of file diff --git a/llama2_tulu-2-7b_fs2_mmlu_7e8c7ce7_group_cache.npy b/llama2_tulu-2-7b_fs2_mmlu_7e8c7ce7_group_cache.npy new file mode 100644 index 0000000000000000000000000000000000000000..d0f42475508ad18700c420dcb658d4362ae473bf --- /dev/null +++ b/llama2_tulu-2-7b_fs2_mmlu_7e8c7ce7_group_cache.npy @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:d7cc5a01714b627acaf1f59619674511e8bb4a291bafe4f39051055d61985f73 +size 5530691 diff --git a/llama2_tulu-2-7b_fs3_mmlu_8d239629.json b/llama2_tulu-2-7b_fs3_mmlu_8d239629.json new file mode 100644 index 0000000000000000000000000000000000000000..c8a1132d2f3d49b106ac24a16318dfc0fc9527de --- /dev/null +++ b/llama2_tulu-2-7b_fs3_mmlu_8d239629.json @@ -0,0 +1 @@ +{"uid": "8d239629", "model": "allenai/tulu-2-7b", "model_name": "tulu-2-7b", "group": "llama2", "model_args": "pretrained=allenai/tulu-2-7b,dtype=auto,gpu_memory_utilization=0.7,tensor_parallel_size=1,data_parallel_size=4,max_model_len=4096", "fewshot": 3, "tasks": "mmlu"} \ No newline at end of file diff --git a/llama2_tulu-2-7b_fs3_mmlu_8d239629_group_cache.npy b/llama2_tulu-2-7b_fs3_mmlu_8d239629_group_cache.npy new file mode 100644 index 0000000000000000000000000000000000000000..32325d21c9f8969ca894f2a19fb639d0d3dd4a42 --- /dev/null +++ b/llama2_tulu-2-7b_fs3_mmlu_8d239629_group_cache.npy @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:46a04c8c423d85b4971b7c2feabc1035b81dbe14920e493bd0d00123fe82a81c +size 3226180 diff --git a/llama2_tulu-2-7b_fs4_mmlu_9a815e96.json b/llama2_tulu-2-7b_fs4_mmlu_9a815e96.json new file mode 100644 index 0000000000000000000000000000000000000000..5f1568852455ddd3a07ce5064a67340e2b96f8b0 --- /dev/null +++ b/llama2_tulu-2-7b_fs4_mmlu_9a815e96.json @@ -0,0 +1 @@ +{"uid": "9a815e96", "model": "allenai/tulu-2-7b", "model_name": "tulu-2-7b", "group": "llama2", "model_args": "pretrained=allenai/tulu-2-7b,dtype=auto,gpu_memory_utilization=0.7,tensor_parallel_size=1,data_parallel_size=4,max_model_len=4096", "fewshot": 4, "tasks": "mmlu"} \ No newline at end of file diff --git a/llama2_tulu-2-7b_fs4_mmlu_9a815e96_group_cache.npy b/llama2_tulu-2-7b_fs4_mmlu_9a815e96_group_cache.npy new file mode 100644 index 0000000000000000000000000000000000000000..5fc5ad1ee03cdb0e3f24fc978d55e88f1a1fe882 --- /dev/null +++ b/llama2_tulu-2-7b_fs4_mmlu_9a815e96_group_cache.npy @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:25361446d7127aa23b240db467db5d17be9fbf8df62ff9dc12de6bf0c8bb69c6 +size 7835232 diff --git a/llama2_tulu-2-7b_fs5_mmlu_0ae648b3.json b/llama2_tulu-2-7b_fs5_mmlu_0ae648b3.json new file mode 100644 index 0000000000000000000000000000000000000000..f822f8c065f16ed2c804d4814e1096b212355597 --- /dev/null +++ b/llama2_tulu-2-7b_fs5_mmlu_0ae648b3.json @@ -0,0 +1 @@ +{"uid": "0ae648b3", "model": "allenai/tulu-2-7b", "model_name": "tulu-2-7b", "group": "llama2", "model_args": "pretrained=allenai/tulu-2-7b,dtype=auto,gpu_memory_utilization=0.7,tensor_parallel_size=1,data_parallel_size=4,max_model_len=4096", "fewshot": 5, "tasks": "mmlu"} \ No newline at end of file diff --git a/llama2_tulu-2-7b_fs5_mmlu_0ae648b3_group_cache.npy b/llama2_tulu-2-7b_fs5_mmlu_0ae648b3_group_cache.npy new file mode 100644 index 0000000000000000000000000000000000000000..4bf1f27d7180dd08dc4e2bbd72be666a695c5d7c --- /dev/null +++ b/llama2_tulu-2-7b_fs5_mmlu_0ae648b3_group_cache.npy @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:4476f7984dec97335507ec257d1ff1e56ed66195d7f024123c9392f2e3bf8229 +size 4147988 diff --git a/llama2_tulu-2-dpo-7b_fs0_mmlu_a7679a40.json b/llama2_tulu-2-dpo-7b_fs0_mmlu_a7679a40.json new file mode 100644 index 0000000000000000000000000000000000000000..397ecf9a908cfeb21ade43072515af1aa04691ec --- /dev/null +++ b/llama2_tulu-2-dpo-7b_fs0_mmlu_a7679a40.json @@ -0,0 +1 @@ +{"uid": "a7679a40", "model": "allenai/tulu-2-dpo-7b", "model_name": "tulu-2-dpo-7b", "group": "llama2", "model_args": "pretrained=allenai/tulu-2-dpo-7b,dtype=auto,gpu_memory_utilization=0.7,tensor_parallel_size=1,data_parallel_size=4,max_model_len=4096", "fewshot": 0, "tasks": "mmlu"} \ No newline at end of file diff --git a/llama2_tulu-2-dpo-7b_fs0_mmlu_a7679a40_group_cache.npy b/llama2_tulu-2-dpo-7b_fs0_mmlu_a7679a40_group_cache.npy new file mode 100644 index 0000000000000000000000000000000000000000..b9b66d9de7b6ac49a44e19ce2520888edd848dd4 --- /dev/null +++ b/llama2_tulu-2-dpo-7b_fs0_mmlu_a7679a40_group_cache.npy @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:5bdd50a2319bceff04318872446a5baee90436d2fb507fc2aae0703fcca2e1af +size 6452505 diff --git a/llama2_tulu-2-dpo-7b_fs1_mmlu_e4d3a50e.json b/llama2_tulu-2-dpo-7b_fs1_mmlu_e4d3a50e.json new file mode 100644 index 0000000000000000000000000000000000000000..5d9fcb5b0555d4f46cf20f611358f91bdc375ac8 --- /dev/null +++ b/llama2_tulu-2-dpo-7b_fs1_mmlu_e4d3a50e.json @@ -0,0 +1 @@ +{"uid": "e4d3a50e", "model": "allenai/tulu-2-dpo-7b", "model_name": "tulu-2-dpo-7b", "group": "llama2", "model_args": "pretrained=allenai/tulu-2-dpo-7b,dtype=auto,gpu_memory_utilization=0.7,tensor_parallel_size=1,data_parallel_size=4,max_model_len=4096", "fewshot": 1, "tasks": "mmlu"} \ No newline at end of file diff --git a/llama2_tulu-2-dpo-7b_fs1_mmlu_e4d3a50e_group_cache.npy b/llama2_tulu-2-dpo-7b_fs1_mmlu_e4d3a50e_group_cache.npy new file mode 100644 index 0000000000000000000000000000000000000000..b945c2c30bc629ac762db89f4539adb07bcd2351 --- /dev/null +++ b/llama2_tulu-2-dpo-7b_fs1_mmlu_e4d3a50e_group_cache.npy @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:cc70ac1556ac9ef773252a2a12d87d50ac84f33c440d6100604550beb3498987 +size 5991598 diff --git a/llama2_tulu-2-dpo-7b_fs2_mmlu_90a670f7.json b/llama2_tulu-2-dpo-7b_fs2_mmlu_90a670f7.json new file mode 100644 index 0000000000000000000000000000000000000000..262e9b9130183d40145d4645e1bfe95c7ec1a534 --- /dev/null +++ b/llama2_tulu-2-dpo-7b_fs2_mmlu_90a670f7.json @@ -0,0 +1 @@ +{"uid": "90a670f7", "model": "allenai/tulu-2-dpo-7b", "model_name": "tulu-2-dpo-7b", "group": "llama2", "model_args": "pretrained=allenai/tulu-2-dpo-7b,dtype=auto,gpu_memory_utilization=0.7,tensor_parallel_size=1,data_parallel_size=4,max_model_len=4096", "fewshot": 2, "tasks": "mmlu"} \ No newline at end of file diff --git a/llama2_tulu-2-dpo-7b_fs2_mmlu_90a670f7_group_cache.npy b/llama2_tulu-2-dpo-7b_fs2_mmlu_90a670f7_group_cache.npy new file mode 100644 index 0000000000000000000000000000000000000000..4bbde63000bc8465491117bba64b5c8dba838475 --- /dev/null +++ b/llama2_tulu-2-dpo-7b_fs2_mmlu_90a670f7_group_cache.npy @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:22d422f4a5a5f6a0c87d47203310b236e27c07d2bb3c5e02ef7544675a82bc21 +size 11061641 diff --git a/llama2_tulu-2-dpo-7b_fs3_mmlu_28bf1664.json b/llama2_tulu-2-dpo-7b_fs3_mmlu_28bf1664.json new file mode 100644 index 0000000000000000000000000000000000000000..55d43663038c7cde5f31f06a4456be59319a0a95 --- /dev/null +++ b/llama2_tulu-2-dpo-7b_fs3_mmlu_28bf1664.json @@ -0,0 +1 @@ +{"uid": "28bf1664", "model": "allenai/tulu-2-dpo-7b", "model_name": "tulu-2-dpo-7b", "group": "llama2", "model_args": "pretrained=allenai/tulu-2-dpo-7b,dtype=auto,gpu_memory_utilization=0.7,tensor_parallel_size=1,data_parallel_size=4,max_model_len=4096", "fewshot": 3, "tasks": "mmlu"} \ No newline at end of file diff --git a/llama2_tulu-2-dpo-7b_fs3_mmlu_28bf1664_group_cache.npy b/llama2_tulu-2-dpo-7b_fs3_mmlu_28bf1664_group_cache.npy new file mode 100644 index 0000000000000000000000000000000000000000..4872f09b637776e587a7775d01510fb47b397de6 --- /dev/null +++ b/llama2_tulu-2-dpo-7b_fs3_mmlu_28bf1664_group_cache.npy @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:4cd388ec91e3079daed2097b4c6386271cf00a190bb5351af27b34b1809c4fe3 +size 1843459 diff --git a/llama2_tulu-2-dpo-7b_fs4_mmlu_446c48d3.json b/llama2_tulu-2-dpo-7b_fs4_mmlu_446c48d3.json new file mode 100644 index 0000000000000000000000000000000000000000..637dbaab4bb7d138fa8874b2a8284a30ffdb7fa6 --- /dev/null +++ b/llama2_tulu-2-dpo-7b_fs4_mmlu_446c48d3.json @@ -0,0 +1 @@ +{"uid": "446c48d3", "model": "allenai/tulu-2-dpo-7b", "model_name": "tulu-2-dpo-7b", "group": "llama2", "model_args": "pretrained=allenai/tulu-2-dpo-7b,dtype=auto,gpu_memory_utilization=0.7,tensor_parallel_size=1,data_parallel_size=4,max_model_len=4096", "fewshot": 4, "tasks": "mmlu"} \ No newline at end of file diff --git a/llama2_tulu-2-dpo-7b_fs4_mmlu_446c48d3_group_cache.npy b/llama2_tulu-2-dpo-7b_fs4_mmlu_446c48d3_group_cache.npy new file mode 100644 index 0000000000000000000000000000000000000000..0c57007f5f53085da3a1836ba193b5e99129f86c --- /dev/null +++ b/llama2_tulu-2-dpo-7b_fs4_mmlu_446c48d3_group_cache.npy @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:68ed0b3621009b4d8bb9e9b0337383c4641c03ab887e1fa25b2a03b239079aeb +size 460743 diff --git a/llama2_tulu-2-dpo-7b_fs5_mmlu_134bbfc5.json b/llama2_tulu-2-dpo-7b_fs5_mmlu_134bbfc5.json new file mode 100644 index 0000000000000000000000000000000000000000..e0772ed502e3b76655570eba43c07dafd46f7fd2 --- /dev/null +++ b/llama2_tulu-2-dpo-7b_fs5_mmlu_134bbfc5.json @@ -0,0 +1 @@ +{"uid": "134bbfc5", "model": "allenai/tulu-2-dpo-7b", "model_name": "tulu-2-dpo-7b", "group": "llama2", "model_args": "pretrained=allenai/tulu-2-dpo-7b,dtype=auto,gpu_memory_utilization=0.7,tensor_parallel_size=1,data_parallel_size=4,max_model_len=4096", "fewshot": 5, "tasks": "mmlu"} \ No newline at end of file diff --git a/llama2_tulu-2-dpo-7b_fs5_mmlu_134bbfc5_group_cache.npy b/llama2_tulu-2-dpo-7b_fs5_mmlu_134bbfc5_group_cache.npy new file mode 100644 index 0000000000000000000000000000000000000000..2106e6c82debc70327b33f5dbcf234e17a7043e4 --- /dev/null +++ b/llama2_tulu-2-dpo-7b_fs5_mmlu_134bbfc5_group_cache.npy @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:7512a0f3c8ae6572119d006cccf31f0981579de469e5ebd96eecab9c0c7d6ceb +size 8757058 diff --git a/llama3_Llama-3.1-8B-Instruct-AQLM-PV-2Bit-1x16-hf_fs0_mmlu_140d6f07.json b/llama3_Llama-3.1-8B-Instruct-AQLM-PV-2Bit-1x16-hf_fs0_mmlu_140d6f07.json new file mode 100644 index 0000000000000000000000000000000000000000..f133cf3df55884a90ba5cb649a71ffad885c632a --- /dev/null +++ b/llama3_Llama-3.1-8B-Instruct-AQLM-PV-2Bit-1x16-hf_fs0_mmlu_140d6f07.json @@ -0,0 +1 @@ +{"uid": "140d6f07", "model": "ISTA-DASLab/Llama-3.1-8B-Instruct-AQLM-PV-2Bit-1x16-hf", "model_name": "Llama-3.1-8B-Instruct-AQLM-PV-2Bit-1x16-hf", "group": "llama3", "model_args": "pretrained=ISTA-DASLab/Llama-3.1-8B-Instruct-AQLM-PV-2Bit-1x16-hf,dtype=auto,gpu_memory_utilization=0.7,tensor_parallel_size=1,data_parallel_size=4,max_model_len=4096", "fewshot": 0, "tasks": "mmlu"} \ No newline at end of file diff --git a/llama3_Llama-3.1-8B-Instruct-AQLM-PV-2Bit-1x16-hf_fs0_mmlu_140d6f07_group_cache.npy b/llama3_Llama-3.1-8B-Instruct-AQLM-PV-2Bit-1x16-hf_fs0_mmlu_140d6f07_group_cache.npy new file mode 100644 index 0000000000000000000000000000000000000000..485b105c5545a2f11e4e813d58af403ff380fe6d --- /dev/null +++ b/llama3_Llama-3.1-8B-Instruct-AQLM-PV-2Bit-1x16-hf_fs0_mmlu_140d6f07_group_cache.npy @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:398be4fbf253415ed4829cf64a9d0875c8a8d360424685a401052c3323eff09f +size 7836308 diff --git a/llama3_Llama-3.1-8B-Instruct-AQLM-PV-2Bit-1x16-hf_fs0_mmlu_6932ee51.json b/llama3_Llama-3.1-8B-Instruct-AQLM-PV-2Bit-1x16-hf_fs0_mmlu_6932ee51.json new file mode 100644 index 0000000000000000000000000000000000000000..059a03783ef31eb8aa941a328607332e1fc2bede --- /dev/null +++ b/llama3_Llama-3.1-8B-Instruct-AQLM-PV-2Bit-1x16-hf_fs0_mmlu_6932ee51.json @@ -0,0 +1 @@ +{"uid": "6932ee51", "model": "ISTA-DASLab/Llama-3.1-8B-Instruct-AQLM-PV-2Bit-1x16-hf", "model_name": "Llama-3.1-8B-Instruct-AQLM-PV-2Bit-1x16-hf", "group": "llama3", "model_args": "pretrained=ISTA-DASLab/Llama-3.1-8B-Instruct-AQLM-PV-2Bit-1x16-hf,dtype=auto,gpu_memory_utilization=0.7,tensor_parallel_size=1,data_parallel_size=4,max_model_len=4096", "fewshot": 0, "tasks": "mmlu"} \ No newline at end of file diff --git a/llama3_Llama-3.1-8B-Instruct-AQLM-PV-2Bit-1x16-hf_fs0_mmlu_6932ee51_group_cache.npy b/llama3_Llama-3.1-8B-Instruct-AQLM-PV-2Bit-1x16-hf_fs0_mmlu_6932ee51_group_cache.npy new file mode 100644 index 0000000000000000000000000000000000000000..64fae9401d7337fcb6eed7c1c739e59c1c636d64 --- /dev/null +++ b/llama3_Llama-3.1-8B-Instruct-AQLM-PV-2Bit-1x16-hf_fs0_mmlu_6932ee51_group_cache.npy @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:6bce8e5f1206cae374a30720daf0bb4f587192eba7da85a91e4e6dd70a7e7e49 +size 460838 diff --git a/llama3_Llama-3.1-8B-Instruct-AQLM-PV-2Bit-1x16-hf_fs1_mmlu_1c74f19b.json b/llama3_Llama-3.1-8B-Instruct-AQLM-PV-2Bit-1x16-hf_fs1_mmlu_1c74f19b.json new file mode 100644 index 0000000000000000000000000000000000000000..ddf7896518558bcc411c19b48b7388203003c8b4 --- /dev/null +++ b/llama3_Llama-3.1-8B-Instruct-AQLM-PV-2Bit-1x16-hf_fs1_mmlu_1c74f19b.json @@ -0,0 +1 @@ +{"uid": "1c74f19b", "model": "ISTA-DASLab/Llama-3.1-8B-Instruct-AQLM-PV-2Bit-1x16-hf", "model_name": "Llama-3.1-8B-Instruct-AQLM-PV-2Bit-1x16-hf", "group": "llama3", "model_args": "pretrained=ISTA-DASLab/Llama-3.1-8B-Instruct-AQLM-PV-2Bit-1x16-hf,dtype=auto,gpu_memory_utilization=0.7,tensor_parallel_size=1,data_parallel_size=4,max_model_len=4096", "fewshot": 1, "tasks": "mmlu"} \ No newline at end of file diff --git a/llama3_Llama-3.1-8B-Instruct-AQLM-PV-2Bit-1x16-hf_fs1_mmlu_1c74f19b_group_cache.npy b/llama3_Llama-3.1-8B-Instruct-AQLM-PV-2Bit-1x16-hf_fs1_mmlu_1c74f19b_group_cache.npy new file mode 100644 index 0000000000000000000000000000000000000000..c787c05fae59b30d74407033a800adc7600d5e5d --- /dev/null +++ b/llama3_Llama-3.1-8B-Instruct-AQLM-PV-2Bit-1x16-hf_fs1_mmlu_1c74f19b_group_cache.npy @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:ea8b31d32549917e46f7e231b81a3f9af9893a5551ca872340f0bcdd7f40259a +size 5070496 diff --git a/llama3_Llama-3.1-8B-Instruct-AQLM-PV-2Bit-1x16-hf_fs1_mmlu_2004e02f.json b/llama3_Llama-3.1-8B-Instruct-AQLM-PV-2Bit-1x16-hf_fs1_mmlu_2004e02f.json new file mode 100644 index 0000000000000000000000000000000000000000..a0cd083b5a1108d52ac711ccfbcc4241dffba418 --- /dev/null +++ b/llama3_Llama-3.1-8B-Instruct-AQLM-PV-2Bit-1x16-hf_fs1_mmlu_2004e02f.json @@ -0,0 +1 @@ +{"uid": "2004e02f", "model": "ISTA-DASLab/Llama-3.1-8B-Instruct-AQLM-PV-2Bit-1x16-hf", "model_name": "Llama-3.1-8B-Instruct-AQLM-PV-2Bit-1x16-hf", "group": "llama3", "model_args": "pretrained=ISTA-DASLab/Llama-3.1-8B-Instruct-AQLM-PV-2Bit-1x16-hf,dtype=auto,gpu_memory_utilization=0.7,tensor_parallel_size=1,data_parallel_size=4,max_model_len=4096", "fewshot": 1, "tasks": "mmlu"} \ No newline at end of file diff --git a/llama3_Llama-3.1-8B-Instruct-AQLM-PV-2Bit-1x16-hf_fs1_mmlu_2004e02f_group_cache.npy b/llama3_Llama-3.1-8B-Instruct-AQLM-PV-2Bit-1x16-hf_fs1_mmlu_2004e02f_group_cache.npy new file mode 100644 index 0000000000000000000000000000000000000000..f1fa8fc5fb48a4a429f77e407118b5b05106cad8 --- /dev/null +++ b/llama3_Llama-3.1-8B-Instruct-AQLM-PV-2Bit-1x16-hf_fs1_mmlu_2004e02f_group_cache.npy @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:1ec15004682fe2ef510c4f46e44bb354897d0721f55e86d0cdcf4f8ce4f73228 +size 17977396 diff --git a/llama3_Llama-3.1-8B-Instruct-AQLM-PV-2Bit-1x16-hf_fs2_mmlu_29dec1ec.json b/llama3_Llama-3.1-8B-Instruct-AQLM-PV-2Bit-1x16-hf_fs2_mmlu_29dec1ec.json new file mode 100644 index 0000000000000000000000000000000000000000..2f2fb02dd0df17eadadc3d03c67b262a554d473f --- /dev/null +++ b/llama3_Llama-3.1-8B-Instruct-AQLM-PV-2Bit-1x16-hf_fs2_mmlu_29dec1ec.json @@ -0,0 +1 @@ +{"uid": "29dec1ec", "model": "ISTA-DASLab/Llama-3.1-8B-Instruct-AQLM-PV-2Bit-1x16-hf", "model_name": "Llama-3.1-8B-Instruct-AQLM-PV-2Bit-1x16-hf", "group": "llama3", "model_args": "pretrained=ISTA-DASLab/Llama-3.1-8B-Instruct-AQLM-PV-2Bit-1x16-hf,dtype=auto,gpu_memory_utilization=0.7,tensor_parallel_size=1,data_parallel_size=4,max_model_len=4096", "fewshot": 2, "tasks": "mmlu"} \ No newline at end of file diff --git a/llama3_Llama-3.1-8B-Instruct-AQLM-PV-2Bit-1x16-hf_fs2_mmlu_29dec1ec_group_cache.npy b/llama3_Llama-3.1-8B-Instruct-AQLM-PV-2Bit-1x16-hf_fs2_mmlu_29dec1ec_group_cache.npy new file mode 100644 index 0000000000000000000000000000000000000000..df9e8b02a37db279fcce705bcd3ee812ddf18069 --- /dev/null +++ b/llama3_Llama-3.1-8B-Instruct-AQLM-PV-2Bit-1x16-hf_fs2_mmlu_29dec1ec_group_cache.npy @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:90d304d3610c9709a0ca1b50ee6c1538a3dedab8db05c254e16a3ffea5204bc4 +size 3226646 diff --git a/llama3_Llama-3.1-8B-Instruct-AQLM-PV-2Bit-1x16-hf_fs2_mmlu_8a87e2f4.json b/llama3_Llama-3.1-8B-Instruct-AQLM-PV-2Bit-1x16-hf_fs2_mmlu_8a87e2f4.json new file mode 100644 index 0000000000000000000000000000000000000000..b5d1baf5542af0a83ce96a9a7061fa55d72971f2 --- /dev/null +++ b/llama3_Llama-3.1-8B-Instruct-AQLM-PV-2Bit-1x16-hf_fs2_mmlu_8a87e2f4.json @@ -0,0 +1 @@ +{"uid": "8a87e2f4", "model": "ISTA-DASLab/Llama-3.1-8B-Instruct-AQLM-PV-2Bit-1x16-hf", "model_name": "Llama-3.1-8B-Instruct-AQLM-PV-2Bit-1x16-hf", "group": "llama3", "model_args": "pretrained=ISTA-DASLab/Llama-3.1-8B-Instruct-AQLM-PV-2Bit-1x16-hf,dtype=auto,gpu_memory_utilization=0.7,tensor_parallel_size=1,data_parallel_size=4,max_model_len=4096", "fewshot": 2, "tasks": "mmlu"} \ No newline at end of file diff --git a/llama3_Llama-3.1-8B-Instruct-AQLM-PV-2Bit-1x16-hf_fs2_mmlu_8a87e2f4_group_cache.npy b/llama3_Llama-3.1-8B-Instruct-AQLM-PV-2Bit-1x16-hf_fs2_mmlu_8a87e2f4_group_cache.npy new file mode 100644 index 0000000000000000000000000000000000000000..95504027be98f3903421fcb2591065ca44d0aa49 --- /dev/null +++ b/llama3_Llama-3.1-8B-Instruct-AQLM-PV-2Bit-1x16-hf_fs2_mmlu_8a87e2f4_group_cache.npy @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:8a0f45ed0e04652d417e6cb472abf9977f3dd54ba601ca86b9e2c1832cb6f15e +size 12445986 diff --git a/llama3_Llama-3.1-8B-Instruct-AQLM-PV-2Bit-1x16-hf_fs3_mmlu_be723051.json b/llama3_Llama-3.1-8B-Instruct-AQLM-PV-2Bit-1x16-hf_fs3_mmlu_be723051.json new file mode 100644 index 0000000000000000000000000000000000000000..e4ec3c0e6851266bafc324bb4ac2ab4eac15a5a4 --- /dev/null +++ b/llama3_Llama-3.1-8B-Instruct-AQLM-PV-2Bit-1x16-hf_fs3_mmlu_be723051.json @@ -0,0 +1 @@ +{"uid": "be723051", "model": "ISTA-DASLab/Llama-3.1-8B-Instruct-AQLM-PV-2Bit-1x16-hf", "model_name": "Llama-3.1-8B-Instruct-AQLM-PV-2Bit-1x16-hf", "group": "llama3", "model_args": "pretrained=ISTA-DASLab/Llama-3.1-8B-Instruct-AQLM-PV-2Bit-1x16-hf,dtype=auto,gpu_memory_utilization=0.7,tensor_parallel_size=1,data_parallel_size=4,max_model_len=4096", "fewshot": 3, "tasks": "mmlu"} \ No newline at end of file diff --git a/llama3_Llama-3.1-8B-Instruct-AQLM-PV-2Bit-1x16-hf_fs3_mmlu_be723051_group_cache.npy b/llama3_Llama-3.1-8B-Instruct-AQLM-PV-2Bit-1x16-hf_fs3_mmlu_be723051_group_cache.npy new file mode 100644 index 0000000000000000000000000000000000000000..c06d557fd206ad4cf2954708b8504d6b5544aa3f --- /dev/null +++ b/llama3_Llama-3.1-8B-Instruct-AQLM-PV-2Bit-1x16-hf_fs3_mmlu_be723051_group_cache.npy @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:cfe7594ff870d1498c8fed2cc041d74ea2ec75e5eb7779527cdea781a4bdb716 +size 11063063 diff --git a/llama3_Llama-3.1-8B-Instruct-AQLM-PV-2Bit-1x16-hf_fs3_mmlu_f3ed2b72.json b/llama3_Llama-3.1-8B-Instruct-AQLM-PV-2Bit-1x16-hf_fs3_mmlu_f3ed2b72.json new file mode 100644 index 0000000000000000000000000000000000000000..68d2bf2f78db6f3fc116fb9ffd51e15144a26ce0 --- /dev/null +++ b/llama3_Llama-3.1-8B-Instruct-AQLM-PV-2Bit-1x16-hf_fs3_mmlu_f3ed2b72.json @@ -0,0 +1 @@ +{"uid": "f3ed2b72", "model": "ISTA-DASLab/Llama-3.1-8B-Instruct-AQLM-PV-2Bit-1x16-hf", "model_name": "Llama-3.1-8B-Instruct-AQLM-PV-2Bit-1x16-hf", "group": "llama3", "model_args": "pretrained=ISTA-DASLab/Llama-3.1-8B-Instruct-AQLM-PV-2Bit-1x16-hf,dtype=auto,gpu_memory_utilization=0.7,tensor_parallel_size=1,data_parallel_size=4,max_model_len=4096", "fewshot": 3, "tasks": "mmlu"} \ No newline at end of file diff --git a/llama3_Llama-3.1-8B-Instruct-AQLM-PV-2Bit-1x16-hf_fs3_mmlu_f3ed2b72_group_cache.npy b/llama3_Llama-3.1-8B-Instruct-AQLM-PV-2Bit-1x16-hf_fs3_mmlu_f3ed2b72_group_cache.npy new file mode 100644 index 0000000000000000000000000000000000000000..62d43b9c83a60779e421e0d1e13f76da611f3af4 --- /dev/null +++ b/llama3_Llama-3.1-8B-Instruct-AQLM-PV-2Bit-1x16-hf_fs3_mmlu_f3ed2b72_group_cache.npy @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:1513b7bfa15748bc844fb0d979a4dd3a77e358652a049a5f7f0140ec7a0885d5 +size 11984984 diff --git a/llama3_Llama-3.1-8B-Instruct-AQLM-PV-2Bit-1x16-hf_fs4_mmlu_1e1bcc72.json b/llama3_Llama-3.1-8B-Instruct-AQLM-PV-2Bit-1x16-hf_fs4_mmlu_1e1bcc72.json new file mode 100644 index 0000000000000000000000000000000000000000..d657e5848ac9cb2be912c50338f4a3a2eebc8d75 --- /dev/null +++ b/llama3_Llama-3.1-8B-Instruct-AQLM-PV-2Bit-1x16-hf_fs4_mmlu_1e1bcc72.json @@ -0,0 +1 @@ +{"uid": "1e1bcc72", "model": "ISTA-DASLab/Llama-3.1-8B-Instruct-AQLM-PV-2Bit-1x16-hf", "model_name": "Llama-3.1-8B-Instruct-AQLM-PV-2Bit-1x16-hf", "group": "llama3", "model_args": "pretrained=ISTA-DASLab/Llama-3.1-8B-Instruct-AQLM-PV-2Bit-1x16-hf,dtype=auto,gpu_memory_utilization=0.7,tensor_parallel_size=1,data_parallel_size=4,max_model_len=4096", "fewshot": 4, "tasks": "mmlu"} \ No newline at end of file diff --git a/llama3_Llama-3.1-8B-Instruct-AQLM-PV-2Bit-1x16-hf_fs4_mmlu_1e1bcc72_group_cache.npy b/llama3_Llama-3.1-8B-Instruct-AQLM-PV-2Bit-1x16-hf_fs4_mmlu_1e1bcc72_group_cache.npy new file mode 100644 index 0000000000000000000000000000000000000000..b3bb1eabba2bc883f564515a878793b4853af6b5 --- /dev/null +++ b/llama3_Llama-3.1-8B-Instruct-AQLM-PV-2Bit-1x16-hf_fs4_mmlu_1e1bcc72_group_cache.npy @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:5f4d06bddfa07a6a556e375d36369997f8a5d2694f4df19287583f56b83e46df +size 5992444 diff --git a/llama3_Llama-3.1-8B-Instruct-AQLM-PV-2Bit-1x16-hf_fs4_mmlu_dfef5673.json b/llama3_Llama-3.1-8B-Instruct-AQLM-PV-2Bit-1x16-hf_fs4_mmlu_dfef5673.json new file mode 100644 index 0000000000000000000000000000000000000000..98b9054438fb5a0aed4ae702f5145ac060c221de --- /dev/null +++ b/llama3_Llama-3.1-8B-Instruct-AQLM-PV-2Bit-1x16-hf_fs4_mmlu_dfef5673.json @@ -0,0 +1 @@ +{"uid": "dfef5673", "model": "ISTA-DASLab/Llama-3.1-8B-Instruct-AQLM-PV-2Bit-1x16-hf", "model_name": "Llama-3.1-8B-Instruct-AQLM-PV-2Bit-1x16-hf", "group": "llama3", "model_args": "pretrained=ISTA-DASLab/Llama-3.1-8B-Instruct-AQLM-PV-2Bit-1x16-hf,dtype=auto,gpu_memory_utilization=0.7,tensor_parallel_size=1,data_parallel_size=4,max_model_len=4096", "fewshot": 4, "tasks": "mmlu"} \ No newline at end of file diff --git a/llama3_Llama-3.1-8B-Instruct-AQLM-PV-2Bit-1x16-hf_fs4_mmlu_dfef5673_group_cache.npy b/llama3_Llama-3.1-8B-Instruct-AQLM-PV-2Bit-1x16-hf_fs4_mmlu_dfef5673_group_cache.npy new file mode 100644 index 0000000000000000000000000000000000000000..4dbc2ae75ca2027f6f2aeab4a609ebea43e641dc --- /dev/null +++ b/llama3_Llama-3.1-8B-Instruct-AQLM-PV-2Bit-1x16-hf_fs4_mmlu_dfef5673_group_cache.npy @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:917a960f9e49a072d94b00265b70b24eb4dd6d0f20d4315a926b7da0cd2ce60d +size 8297310 diff --git a/llama3_Llama-3.1-8B-Instruct-AQLM-PV-2Bit-1x16-hf_fs5_mmlu_c3d16d70.json b/llama3_Llama-3.1-8B-Instruct-AQLM-PV-2Bit-1x16-hf_fs5_mmlu_c3d16d70.json new file mode 100644 index 0000000000000000000000000000000000000000..7358303f68b2a7c894c1501c9a5b76568c648aeb --- /dev/null +++ b/llama3_Llama-3.1-8B-Instruct-AQLM-PV-2Bit-1x16-hf_fs5_mmlu_c3d16d70.json @@ -0,0 +1 @@ +{"uid": "c3d16d70", "model": "ISTA-DASLab/Llama-3.1-8B-Instruct-AQLM-PV-2Bit-1x16-hf", "model_name": "Llama-3.1-8B-Instruct-AQLM-PV-2Bit-1x16-hf", "group": "llama3", "model_args": "pretrained=ISTA-DASLab/Llama-3.1-8B-Instruct-AQLM-PV-2Bit-1x16-hf,dtype=auto,gpu_memory_utilization=0.7,tensor_parallel_size=1,data_parallel_size=4,max_model_len=4096", "fewshot": 5, "tasks": "mmlu"} \ No newline at end of file diff --git a/llama3_Llama-3.1-8B-Instruct-AQLM-PV-2Bit-1x16-hf_fs5_mmlu_c3d16d70_group_cache.npy b/llama3_Llama-3.1-8B-Instruct-AQLM-PV-2Bit-1x16-hf_fs5_mmlu_c3d16d70_group_cache.npy new file mode 100644 index 0000000000000000000000000000000000000000..fadd3573f2b54843f12da9a95f03dc4781802238 --- /dev/null +++ b/llama3_Llama-3.1-8B-Instruct-AQLM-PV-2Bit-1x16-hf_fs5_mmlu_c3d16d70_group_cache.npy @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:05e65bd50bcf47d2ce019c1c7986a78aa9e25b293abdeb54a1cd9d9797c6b7ea +size 14750738 diff --git a/llama3_Llama-3.1-8B-Instruct-AQLM-PV-2Bit-1x16-hf_fs5_mmlu_fb5b03cd.json b/llama3_Llama-3.1-8B-Instruct-AQLM-PV-2Bit-1x16-hf_fs5_mmlu_fb5b03cd.json new file mode 100644 index 0000000000000000000000000000000000000000..841c178d7038f43631401b753b3c66b7e3794f8c --- /dev/null +++ b/llama3_Llama-3.1-8B-Instruct-AQLM-PV-2Bit-1x16-hf_fs5_mmlu_fb5b03cd.json @@ -0,0 +1 @@ +{"uid": "fb5b03cd", "model": "ISTA-DASLab/Llama-3.1-8B-Instruct-AQLM-PV-2Bit-1x16-hf", "model_name": "Llama-3.1-8B-Instruct-AQLM-PV-2Bit-1x16-hf", "group": "llama3", "model_args": "pretrained=ISTA-DASLab/Llama-3.1-8B-Instruct-AQLM-PV-2Bit-1x16-hf,dtype=auto,gpu_memory_utilization=0.7,tensor_parallel_size=1,data_parallel_size=4,max_model_len=4096", "fewshot": 5, "tasks": "mmlu"} \ No newline at end of file diff --git a/llama3_Llama-3.1-8B-Instruct-AQLM-PV-2Bit-1x16-hf_fs5_mmlu_fb5b03cd_group_cache.npy b/llama3_Llama-3.1-8B-Instruct-AQLM-PV-2Bit-1x16-hf_fs5_mmlu_fb5b03cd_group_cache.npy new file mode 100644 index 0000000000000000000000000000000000000000..d8950248b5db8559e5bac7a4a0b8b53ea1c7256d --- /dev/null +++ b/llama3_Llama-3.1-8B-Instruct-AQLM-PV-2Bit-1x16-hf_fs5_mmlu_fb5b03cd_group_cache.npy @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:0c37583a79bf560babc6714fd8400712c18f6a56503b8709c7420df2eab459f4 +size 19360283 diff --git a/llama3_Meta-Llama-3-8B-Instruct_fs0_mmlu_64232f35.json b/llama3_Meta-Llama-3-8B-Instruct_fs0_mmlu_64232f35.json new file mode 100644 index 0000000000000000000000000000000000000000..a3bf86cbf938d61c8700ae49233b59196a03adcd --- /dev/null +++ b/llama3_Meta-Llama-3-8B-Instruct_fs0_mmlu_64232f35.json @@ -0,0 +1 @@ +{"uid": "64232f35", "model": "meta-llama/Meta-Llama-3-8B-Instruct", "model_name": "Meta-Llama-3-8B-Instruct", "group": "llama3", "model_args": "pretrained=meta-llama/Meta-Llama-3-8B-Instruct,dtype=auto,gpu_memory_utilization=0.7,tensor_parallel_size=1,data_parallel_size=0,max_model_len=4096", "fewshot": 0, "tasks": "mmlu"} \ No newline at end of file diff --git a/llama3_Meta-Llama-3-8B-Instruct_fs0_mmlu_64232f35_group_cache.npy b/llama3_Meta-Llama-3-8B-Instruct_fs0_mmlu_64232f35_group_cache.npy new file mode 100644 index 0000000000000000000000000000000000000000..422dbe24d0e9779183f184a97396b0010c2b0f77 --- /dev/null +++ b/llama3_Meta-Llama-3-8B-Instruct_fs0_mmlu_64232f35_group_cache.npy @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:8707f16b358bef4c4622ba212c03a712eae5d9fa405733dbc2784cd19405dadd +size 3687592 diff --git a/llama3_Meta-Llama-3-8B-Instruct_fs1_mmlu_3d9ac678.json b/llama3_Meta-Llama-3-8B-Instruct_fs1_mmlu_3d9ac678.json new file mode 100644 index 0000000000000000000000000000000000000000..2f72d08860e1cb74a65c9801ffda42fddd700cd5 --- /dev/null +++ b/llama3_Meta-Llama-3-8B-Instruct_fs1_mmlu_3d9ac678.json @@ -0,0 +1 @@ +{"uid": "3d9ac678", "model": "meta-llama/Meta-Llama-3-8B-Instruct", "model_name": "Meta-Llama-3-8B-Instruct", "group": "llama3", "model_args": "pretrained=meta-llama/Meta-Llama-3-8B-Instruct,dtype=auto,gpu_memory_utilization=0.7,tensor_parallel_size=1,data_parallel_size=4,max_model_len=4096", "fewshot": 1, "tasks": "mmlu"} \ No newline at end of file diff --git a/llama3_Meta-Llama-3-8B-Instruct_fs1_mmlu_3d9ac678_group_cache.npy b/llama3_Meta-Llama-3-8B-Instruct_fs1_mmlu_3d9ac678_group_cache.npy new file mode 100644 index 0000000000000000000000000000000000000000..cfad6ea3639d71aac0715d0f53a47c293ef40803 --- /dev/null +++ b/llama3_Meta-Llama-3-8B-Instruct_fs1_mmlu_3d9ac678_group_cache.npy @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:324ae5903a74e5bc355022f1a00e06b183c18ae92b62169126f760b6e57d3463 +size 6914319 diff --git a/llama3_Meta-Llama-3-8B-Instruct_fs2_mmlu_e05c458f.json b/llama3_Meta-Llama-3-8B-Instruct_fs2_mmlu_e05c458f.json new file mode 100644 index 0000000000000000000000000000000000000000..cacc96fa1aeaede24c24f944cf27cb4498947e2a --- /dev/null +++ b/llama3_Meta-Llama-3-8B-Instruct_fs2_mmlu_e05c458f.json @@ -0,0 +1 @@ +{"uid": "e05c458f", "model": "meta-llama/Meta-Llama-3-8B-Instruct", "model_name": "Meta-Llama-3-8B-Instruct", "group": "llama3", "model_args": "pretrained=meta-llama/Meta-Llama-3-8B-Instruct,dtype=auto,gpu_memory_utilization=0.7,tensor_parallel_size=1,data_parallel_size=4,max_model_len=4096", "fewshot": 2, "tasks": "mmlu"} \ No newline at end of file diff --git a/llama3_Meta-Llama-3-8B-Instruct_fs2_mmlu_e05c458f_group_cache.npy b/llama3_Meta-Llama-3-8B-Instruct_fs2_mmlu_e05c458f_group_cache.npy new file mode 100644 index 0000000000000000000000000000000000000000..c090b486c2e221658e45db45bfb14970e1a716f2 --- /dev/null +++ b/llama3_Meta-Llama-3-8B-Instruct_fs2_mmlu_e05c458f_group_cache.npy @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:55f77f1a517d77a747bc5163cae80af807b861e25c7f44d0de630ccfac9f4035 +size 4148538 diff --git a/llama3_Meta-Llama-3-8B-Instruct_fs3_mmlu_69cc3199.json b/llama3_Meta-Llama-3-8B-Instruct_fs3_mmlu_69cc3199.json new file mode 100644 index 0000000000000000000000000000000000000000..08684a9af60adf65927ca4149b03e19c6367ee98 --- /dev/null +++ b/llama3_Meta-Llama-3-8B-Instruct_fs3_mmlu_69cc3199.json @@ -0,0 +1 @@ +{"uid": "69cc3199", "model": "meta-llama/Meta-Llama-3-8B-Instruct", "model_name": "Meta-Llama-3-8B-Instruct", "group": "llama3", "model_args": "pretrained=meta-llama/Meta-Llama-3-8B-Instruct,dtype=auto,gpu_memory_utilization=0.7,tensor_parallel_size=1,data_parallel_size=4,max_model_len=4096", "fewshot": 3, "tasks": "mmlu"} \ No newline at end of file diff --git a/llama3_Meta-Llama-3-8B-Instruct_fs3_mmlu_69cc3199_group_cache.npy b/llama3_Meta-Llama-3-8B-Instruct_fs3_mmlu_69cc3199_group_cache.npy new file mode 100644 index 0000000000000000000000000000000000000000..416741e39b94c9a4591fd07dd5f129d3069e8830 --- /dev/null +++ b/llama3_Meta-Llama-3-8B-Instruct_fs3_mmlu_69cc3199_group_cache.npy @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:de3ad0d7f540360969a02b6a8763c4c5f94d440d58433ec794942bc22344269e +size 15211684 diff --git a/llama3_Meta-Llama-3-8B-Instruct_fs4_mmlu_7d8a1a90.json b/llama3_Meta-Llama-3-8B-Instruct_fs4_mmlu_7d8a1a90.json new file mode 100644 index 0000000000000000000000000000000000000000..cd2bbe5bff354f2e19712d2cc1f3461348742d1c --- /dev/null +++ b/llama3_Meta-Llama-3-8B-Instruct_fs4_mmlu_7d8a1a90.json @@ -0,0 +1 @@ +{"uid": "7d8a1a90", "model": "meta-llama/Meta-Llama-3-8B-Instruct", "model_name": "Meta-Llama-3-8B-Instruct", "group": "llama3", "model_args": "pretrained=meta-llama/Meta-Llama-3-8B-Instruct,dtype=auto,gpu_memory_utilization=0.7,tensor_parallel_size=1,data_parallel_size=4,max_model_len=4096", "fewshot": 4, "tasks": "mmlu"} \ No newline at end of file diff --git a/llama3_Meta-Llama-3-8B-Instruct_fs4_mmlu_7d8a1a90_group_cache.npy b/llama3_Meta-Llama-3-8B-Instruct_fs4_mmlu_7d8a1a90_group_cache.npy new file mode 100644 index 0000000000000000000000000000000000000000..d6804f7c7928b02dd940c503770b4c353cf1a962 --- /dev/null +++ b/llama3_Meta-Llama-3-8B-Instruct_fs4_mmlu_7d8a1a90_group_cache.npy @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:74d18726989afb7355b3b4d8744023c5a33749ba4fee5ff364b52bec3c3d0464 +size 1843728 diff --git a/llama3_Meta-Llama-3-8B-Instruct_fs5_mmlu_3c30d105.json b/llama3_Meta-Llama-3-8B-Instruct_fs5_mmlu_3c30d105.json new file mode 100644 index 0000000000000000000000000000000000000000..67f33ed8d6c6aed76b287c865d9cd7f60255bfb3 --- /dev/null +++ b/llama3_Meta-Llama-3-8B-Instruct_fs5_mmlu_3c30d105.json @@ -0,0 +1 @@ +{"uid": "3c30d105", "model": "meta-llama/Meta-Llama-3-8B-Instruct", "model_name": "Meta-Llama-3-8B-Instruct", "group": "llama3", "model_args": "pretrained=meta-llama/Meta-Llama-3-8B-Instruct,dtype=auto,gpu_memory_utilization=0.7,tensor_parallel_size=1,data_parallel_size=4,max_model_len=4096", "fewshot": 5, "tasks": "mmlu"} \ No newline at end of file diff --git a/llama3_Meta-Llama-3-8B-Instruct_fs5_mmlu_3c30d105_group_cache.npy b/llama3_Meta-Llama-3-8B-Instruct_fs5_mmlu_3c30d105_group_cache.npy new file mode 100644 index 0000000000000000000000000000000000000000..cc9494bc758e90486f9de009092f0ff1c3d03acf --- /dev/null +++ b/llama3_Meta-Llama-3-8B-Instruct_fs5_mmlu_3c30d105_group_cache.npy @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:06e1da3314360f4c2fd0ae299a00a597e2e03846eb5036e03dc750e684689408 +size 5531442 diff --git a/llama3_Meta-Llama-3-8B_fs0_mmlu_c86bd333.json b/llama3_Meta-Llama-3-8B_fs0_mmlu_c86bd333.json new file mode 100644 index 0000000000000000000000000000000000000000..1a30487981a8086398a0580ef41be9dc673f91fe --- /dev/null +++ b/llama3_Meta-Llama-3-8B_fs0_mmlu_c86bd333.json @@ -0,0 +1 @@ +{"uid": "c86bd333", "model": "meta-llama/Meta-Llama-3-8B", "model_name": "Meta-Llama-3-8B", "group": "llama3", "model_args": "pretrained=meta-llama/Meta-Llama-3-8B,dtype=auto,gpu_memory_utilization=0.7,tensor_parallel_size=1,data_parallel_size=0,max_model_len=4096", "fewshot": 0, "tasks": "mmlu"} \ No newline at end of file diff --git a/llama3_Meta-Llama-3-8B_fs0_mmlu_c86bd333_group_cache.npy b/llama3_Meta-Llama-3-8B_fs0_mmlu_c86bd333_group_cache.npy new file mode 100644 index 0000000000000000000000000000000000000000..9a2bfd983df3ada0d625026bb2fe48ddf02e108c --- /dev/null +++ b/llama3_Meta-Llama-3-8B_fs0_mmlu_c86bd333_group_cache.npy @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:d55ff5c01105cbd01a1aadbee7a3bfc05f0fd2bd44331119f103b13dc56a2cf5 +size 11523982 diff --git a/llama3_Meta-Llama-3-8B_fs1_mmlu_8cbd1f0c.json b/llama3_Meta-Llama-3-8B_fs1_mmlu_8cbd1f0c.json new file mode 100644 index 0000000000000000000000000000000000000000..6c233bf0fabc1850533fe5664da49d3d0b832630 --- /dev/null +++ b/llama3_Meta-Llama-3-8B_fs1_mmlu_8cbd1f0c.json @@ -0,0 +1 @@ +{"uid": "8cbd1f0c", "model": "meta-llama/Meta-Llama-3-8B", "model_name": "Meta-Llama-3-8B", "group": "llama3", "model_args": "pretrained=meta-llama/Meta-Llama-3-8B,dtype=auto,gpu_memory_utilization=0.7,tensor_parallel_size=1,data_parallel_size=4,max_model_len=4096", "fewshot": 1, "tasks": "mmlu"} \ No newline at end of file diff --git a/llama3_Meta-Llama-3-8B_fs1_mmlu_8cbd1f0c_group_cache.npy b/llama3_Meta-Llama-3-8B_fs1_mmlu_8cbd1f0c_group_cache.npy new file mode 100644 index 0000000000000000000000000000000000000000..0f38a2216c9f7beda03662c9a5cbe214d4056c03 --- /dev/null +++ b/llama3_Meta-Llama-3-8B_fs1_mmlu_8cbd1f0c_group_cache.npy @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:f286ca7398e02f9581023498964f7969ba194915f7430d258f9cbb06c3781d30 +size 14289736 diff --git a/llama3_Meta-Llama-3-8B_fs2_mmlu_2b9588b0.json b/llama3_Meta-Llama-3-8B_fs2_mmlu_2b9588b0.json new file mode 100644 index 0000000000000000000000000000000000000000..f873251060e967e9ecdcbb90206304245b5f3dd4 --- /dev/null +++ b/llama3_Meta-Llama-3-8B_fs2_mmlu_2b9588b0.json @@ -0,0 +1 @@ +{"uid": "2b9588b0", "model": "meta-llama/Meta-Llama-3-8B", "model_name": "Meta-Llama-3-8B", "group": "llama3", "model_args": "pretrained=meta-llama/Meta-Llama-3-8B,dtype=auto,gpu_memory_utilization=0.7,tensor_parallel_size=1,data_parallel_size=4,max_model_len=4096", "fewshot": 2, "tasks": "mmlu"} \ No newline at end of file diff --git a/llama3_Meta-Llama-3-8B_fs2_mmlu_2b9588b0_group_cache.npy b/llama3_Meta-Llama-3-8B_fs2_mmlu_2b9588b0_group_cache.npy new file mode 100644 index 0000000000000000000000000000000000000000..a49ae511b5eb7b9cd6c99d4e40a7d238c18f2a11 --- /dev/null +++ b/llama3_Meta-Llama-3-8B_fs2_mmlu_2b9588b0_group_cache.npy @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:d33c97b956ffdd395087e3a18ea02798b9c8677b42eb7d18dc0b4a990b9d1359 +size 17055465 diff --git a/llama3_Meta-Llama-3-8B_fs3_mmlu_ab004344.json b/llama3_Meta-Llama-3-8B_fs3_mmlu_ab004344.json new file mode 100644 index 0000000000000000000000000000000000000000..8c07356ea513784d7e07c38ba962f2cab7f2b0bc --- /dev/null +++ b/llama3_Meta-Llama-3-8B_fs3_mmlu_ab004344.json @@ -0,0 +1 @@ +{"uid": "ab004344", "model": "meta-llama/Meta-Llama-3-8B", "model_name": "Meta-Llama-3-8B", "group": "llama3", "model_args": "pretrained=meta-llama/Meta-Llama-3-8B,dtype=auto,gpu_memory_utilization=0.7,tensor_parallel_size=1,data_parallel_size=4,max_model_len=4096", "fewshot": 3, "tasks": "mmlu"} \ No newline at end of file diff --git a/llama3_Meta-Llama-3-8B_fs3_mmlu_ab004344_group_cache.npy b/llama3_Meta-Llama-3-8B_fs3_mmlu_ab004344_group_cache.npy new file mode 100644 index 0000000000000000000000000000000000000000..a3176e7717852bd2f6fbb901939a0f4edc118298 --- /dev/null +++ b/llama3_Meta-Llama-3-8B_fs3_mmlu_ab004344_group_cache.npy @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:1eca32c254bc9602db83e1adddba987e53c244946f75ff649730b8a00fb5de2c +size 12906905 diff --git a/llama3_Meta-Llama-3-8B_fs4_mmlu_cfb1540b.json b/llama3_Meta-Llama-3-8B_fs4_mmlu_cfb1540b.json new file mode 100644 index 0000000000000000000000000000000000000000..bd7537c06414bc4057ea76eb0293f1ac0f8256a1 --- /dev/null +++ b/llama3_Meta-Llama-3-8B_fs4_mmlu_cfb1540b.json @@ -0,0 +1 @@ +{"uid": "cfb1540b", "model": "meta-llama/Meta-Llama-3-8B", "model_name": "Meta-Llama-3-8B", "group": "llama3", "model_args": "pretrained=meta-llama/Meta-Llama-3-8B,dtype=auto,gpu_memory_utilization=0.7,tensor_parallel_size=1,data_parallel_size=4,max_model_len=4096", "fewshot": 4, "tasks": "mmlu"} \ No newline at end of file diff --git a/llama3_Meta-Llama-3-8B_fs4_mmlu_cfb1540b_group_cache.npy b/llama3_Meta-Llama-3-8B_fs4_mmlu_cfb1540b_group_cache.npy new file mode 100644 index 0000000000000000000000000000000000000000..7a37d05afa06938033ecd63a58e2dd6e3a382019 --- /dev/null +++ b/llama3_Meta-Llama-3-8B_fs4_mmlu_cfb1540b_group_cache.npy @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:75f4e5db16ce6c15959c3a2f5d0247f39fffc437e6c75696d03ebc770f5af9df +size 8758229 diff --git a/llama3_Meta-Llama-3-8B_fs5_mmlu_c36b3564.json b/llama3_Meta-Llama-3-8B_fs5_mmlu_c36b3564.json new file mode 100644 index 0000000000000000000000000000000000000000..59e64d1e2e5491514b08f096abe30a0751cc8b24 --- /dev/null +++ b/llama3_Meta-Llama-3-8B_fs5_mmlu_c36b3564.json @@ -0,0 +1 @@ +{"uid": "c36b3564", "model": "meta-llama/Meta-Llama-3-8B", "model_name": "Meta-Llama-3-8B", "group": "llama3", "model_args": "pretrained=meta-llama/Meta-Llama-3-8B,dtype=auto,gpu_memory_utilization=0.7,tensor_parallel_size=1,data_parallel_size=4,max_model_len=4096", "fewshot": 5, "tasks": "mmlu"} \ No newline at end of file diff --git a/llama3_Meta-Llama-3-8B_fs5_mmlu_c36b3564_group_cache.npy b/llama3_Meta-Llama-3-8B_fs5_mmlu_c36b3564_group_cache.npy new file mode 100644 index 0000000000000000000000000000000000000000..368e571ae8cd8e10f5c2aeb7b3712a3f9c310433 --- /dev/null +++ b/llama3_Meta-Llama-3-8B_fs5_mmlu_c36b3564_group_cache.npy @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:afe71ec5b4d955664d554b87c9565db4abcfee2f58073e9b034988bdf6bda3e4 +size 16594546 diff --git a/llama3_Meta-Llama-3.1-8B-Instruct-AWQ-INT4_fs0_mmlu_6366e7b4.json b/llama3_Meta-Llama-3.1-8B-Instruct-AWQ-INT4_fs0_mmlu_6366e7b4.json new file mode 100644 index 0000000000000000000000000000000000000000..2b6b74d3c97b135916abb5e9782ea95e4ce5c92f --- /dev/null +++ b/llama3_Meta-Llama-3.1-8B-Instruct-AWQ-INT4_fs0_mmlu_6366e7b4.json @@ -0,0 +1 @@ +{"uid": "6366e7b4", "model": "hugging-quants/Meta-Llama-3.1-8B-Instruct-AWQ-INT4", "model_name": "Meta-Llama-3.1-8B-Instruct-AWQ-INT4", "group": "llama3", "model_args": "pretrained=hugging-quants/Meta-Llama-3.1-8B-Instruct-AWQ-INT4,dtype=auto,gpu_memory_utilization=0.7,tensor_parallel_size=1,data_parallel_size=1,max_model_len=4096", "fewshot": 0, "tasks": "mmlu"} \ No newline at end of file diff --git a/llama3_Meta-Llama-3.1-8B-Instruct-AWQ-INT4_fs0_mmlu_6366e7b4_group_cache.npy b/llama3_Meta-Llama-3.1-8B-Instruct-AWQ-INT4_fs0_mmlu_6366e7b4_group_cache.npy new file mode 100644 index 0000000000000000000000000000000000000000..939edc326a85a3f059e11f113a22b7826720aa13 --- /dev/null +++ b/llama3_Meta-Llama-3.1-8B-Instruct-AWQ-INT4_fs0_mmlu_6366e7b4_group_cache.npy @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:803e999a90bfdad7b2993eabf559d239ad3e2b32a46a0b656b2aa424bf14bf46 +size 921826 diff --git a/llama3_Meta-Llama-3.1-8B-Instruct-AWQ-INT4_fs1_mmlu_ed7a82c7.json b/llama3_Meta-Llama-3.1-8B-Instruct-AWQ-INT4_fs1_mmlu_ed7a82c7.json new file mode 100644 index 0000000000000000000000000000000000000000..e9f37d461d62f025634bc732f162979bac73f302 --- /dev/null +++ b/llama3_Meta-Llama-3.1-8B-Instruct-AWQ-INT4_fs1_mmlu_ed7a82c7.json @@ -0,0 +1 @@ +{"uid": "ed7a82c7", "model": "hugging-quants/Meta-Llama-3.1-8B-Instruct-AWQ-INT4", "model_name": "Meta-Llama-3.1-8B-Instruct-AWQ-INT4", "group": "llama3", "model_args": "pretrained=hugging-quants/Meta-Llama-3.1-8B-Instruct-AWQ-INT4,dtype=auto,gpu_memory_utilization=0.7,tensor_parallel_size=1,data_parallel_size=1,max_model_len=4096", "fewshot": 1, "tasks": "mmlu"} \ No newline at end of file diff --git a/llama3_Meta-Llama-3.1-8B-Instruct-AWQ-INT4_fs1_mmlu_ed7a82c7_group_cache.npy b/llama3_Meta-Llama-3.1-8B-Instruct-AWQ-INT4_fs1_mmlu_ed7a82c7_group_cache.npy new file mode 100644 index 0000000000000000000000000000000000000000..b36d8104adc8fc3101671eb93389bff345dd9769 --- /dev/null +++ b/llama3_Meta-Llama-3.1-8B-Instruct-AWQ-INT4_fs1_mmlu_ed7a82c7_group_cache.npy @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:8f205bc33e57e8794fcbfb91c42db1d4858424ed2bd7cf9c8759e36216d9cd04 +size 15672671 diff --git a/llama3_Meta-Llama-3.1-8B-Instruct-AWQ-INT4_fs2_mmlu_595f8a9d.json b/llama3_Meta-Llama-3.1-8B-Instruct-AWQ-INT4_fs2_mmlu_595f8a9d.json new file mode 100644 index 0000000000000000000000000000000000000000..73cdd657eae0292f0fde39d59b9829e0f934d5d6 --- /dev/null +++ b/llama3_Meta-Llama-3.1-8B-Instruct-AWQ-INT4_fs2_mmlu_595f8a9d.json @@ -0,0 +1 @@ +{"uid": "595f8a9d", "model": "hugging-quants/Meta-Llama-3.1-8B-Instruct-AWQ-INT4", "model_name": "Meta-Llama-3.1-8B-Instruct-AWQ-INT4", "group": "llama3", "model_args": "pretrained=hugging-quants/Meta-Llama-3.1-8B-Instruct-AWQ-INT4,dtype=auto,gpu_memory_utilization=0.7,tensor_parallel_size=1,data_parallel_size=1,max_model_len=4096", "fewshot": 2, "tasks": "mmlu"} \ No newline at end of file diff --git a/llama3_Meta-Llama-3.1-8B-Instruct-AWQ-INT4_fs2_mmlu_595f8a9d_group_cache.npy b/llama3_Meta-Llama-3.1-8B-Instruct-AWQ-INT4_fs2_mmlu_595f8a9d_group_cache.npy new file mode 100644 index 0000000000000000000000000000000000000000..b984a4565650df5eb122eebe238e39ad6278dcad --- /dev/null +++ b/llama3_Meta-Llama-3.1-8B-Instruct-AWQ-INT4_fs2_mmlu_595f8a9d_group_cache.npy @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:e9b2919ceeb4242f23148901e95dda8dd3221a9ef995cb749f3b98cefd9ceb9d +size 2304715 diff --git a/llama3_Meta-Llama-3.1-8B-Instruct-AWQ-INT4_fs3_mmlu_03373053.json b/llama3_Meta-Llama-3.1-8B-Instruct-AWQ-INT4_fs3_mmlu_03373053.json new file mode 100644 index 0000000000000000000000000000000000000000..0fc28b99af0e1ab7288540df5c137fa555fc6020 --- /dev/null +++ b/llama3_Meta-Llama-3.1-8B-Instruct-AWQ-INT4_fs3_mmlu_03373053.json @@ -0,0 +1 @@ +{"uid": "03373053", "model": "hugging-quants/Meta-Llama-3.1-8B-Instruct-AWQ-INT4", "model_name": "Meta-Llama-3.1-8B-Instruct-AWQ-INT4", "group": "llama3", "model_args": "pretrained=hugging-quants/Meta-Llama-3.1-8B-Instruct-AWQ-INT4,dtype=auto,gpu_memory_utilization=0.7,tensor_parallel_size=1,data_parallel_size=1,max_model_len=4096", "fewshot": 3, "tasks": "mmlu"} \ No newline at end of file diff --git a/llama3_Meta-Llama-3.1-8B-Instruct-AWQ-INT4_fs3_mmlu_03373053_group_cache.npy b/llama3_Meta-Llama-3.1-8B-Instruct-AWQ-INT4_fs3_mmlu_03373053_group_cache.npy new file mode 100644 index 0000000000000000000000000000000000000000..ba3b3483812dbb02258b2b7fdf2d6cf8253cec02 --- /dev/null +++ b/llama3_Meta-Llama-3.1-8B-Instruct-AWQ-INT4_fs3_mmlu_03373053_group_cache.npy @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:ccbea4ac514ab6f3a8e9aec759ee36e37b60146e760e0cbee76d10e7c08fd0ba +size 7375306 diff --git a/llama3_Meta-Llama-3.1-8B-Instruct-AWQ-INT4_fs4_mmlu_60ac1c71.json b/llama3_Meta-Llama-3.1-8B-Instruct-AWQ-INT4_fs4_mmlu_60ac1c71.json new file mode 100644 index 0000000000000000000000000000000000000000..b66ed149c6fa6d2a3c741a2b15e9f8bbaf284b9b --- /dev/null +++ b/llama3_Meta-Llama-3.1-8B-Instruct-AWQ-INT4_fs4_mmlu_60ac1c71.json @@ -0,0 +1 @@ +{"uid": "60ac1c71", "model": "hugging-quants/Meta-Llama-3.1-8B-Instruct-AWQ-INT4", "model_name": "Meta-Llama-3.1-8B-Instruct-AWQ-INT4", "group": "llama3", "model_args": "pretrained=hugging-quants/Meta-Llama-3.1-8B-Instruct-AWQ-INT4,dtype=auto,gpu_memory_utilization=0.7,tensor_parallel_size=1,data_parallel_size=1,max_model_len=4096", "fewshot": 4, "tasks": "mmlu"} \ No newline at end of file diff --git a/llama3_Meta-Llama-3.1-8B-Instruct-AWQ-INT4_fs4_mmlu_60ac1c71_group_cache.npy b/llama3_Meta-Llama-3.1-8B-Instruct-AWQ-INT4_fs4_mmlu_60ac1c71_group_cache.npy new file mode 100644 index 0000000000000000000000000000000000000000..bf13ec2effcc61254a8920718e8695309e717f82 --- /dev/null +++ b/llama3_Meta-Llama-3.1-8B-Instruct-AWQ-INT4_fs4_mmlu_60ac1c71_group_cache.npy @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:a1844590269d3cc949f976a7c0cd8d1f667baea666ecd96bebeec291fbe769e4 +size 10602061 diff --git a/llama3_Meta-Llama-3.1-8B-Instruct-AWQ-INT4_fs5_mmlu_dbc5e6a5.json b/llama3_Meta-Llama-3.1-8B-Instruct-AWQ-INT4_fs5_mmlu_dbc5e6a5.json new file mode 100644 index 0000000000000000000000000000000000000000..3c726f7054eba3fde3770fcb07340a9f0437fbd4 --- /dev/null +++ b/llama3_Meta-Llama-3.1-8B-Instruct-AWQ-INT4_fs5_mmlu_dbc5e6a5.json @@ -0,0 +1 @@ +{"uid": "dbc5e6a5", "model": "hugging-quants/Meta-Llama-3.1-8B-Instruct-AWQ-INT4", "model_name": "Meta-Llama-3.1-8B-Instruct-AWQ-INT4", "group": "llama3", "model_args": "pretrained=hugging-quants/Meta-Llama-3.1-8B-Instruct-AWQ-INT4,dtype=auto,gpu_memory_utilization=0.7,tensor_parallel_size=1,data_parallel_size=1,max_model_len=4096", "fewshot": 5, "tasks": "mmlu"} \ No newline at end of file diff --git a/llama3_Meta-Llama-3.1-8B-Instruct-AWQ-INT4_fs5_mmlu_dbc5e6a5_group_cache.npy b/llama3_Meta-Llama-3.1-8B-Instruct-AWQ-INT4_fs5_mmlu_dbc5e6a5_group_cache.npy new file mode 100644 index 0000000000000000000000000000000000000000..662c95644e1d3026e7f9e7501318a175a0bb229c --- /dev/null +++ b/llama3_Meta-Llama-3.1-8B-Instruct-AWQ-INT4_fs5_mmlu_dbc5e6a5_group_cache.npy @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:140ea1e6cc44aca99b490af9714c2ec34c5397fde095f59cbf467ad4bb40a9fe +size 10141074 diff --git a/llama3_Meta-Llama-3.1-8B-Instruct_fs0_mmlu_56bc86f0.json b/llama3_Meta-Llama-3.1-8B-Instruct_fs0_mmlu_56bc86f0.json new file mode 100644 index 0000000000000000000000000000000000000000..9a37262f41228c783cf3e04d6ab8c9fdc973b319 --- /dev/null +++ b/llama3_Meta-Llama-3.1-8B-Instruct_fs0_mmlu_56bc86f0.json @@ -0,0 +1 @@ +{"uid": "56bc86f0", "model": "NousResearch/Meta-Llama-3.1-8B-Instruct", "model_name": "Meta-Llama-3.1-8B-Instruct", "group": "llama3", "model_args": "pretrained=NousResearch/Meta-Llama-3.1-8B-Instruct,dtype=auto,gpu_memory_utilization=0.7,tensor_parallel_size=1,data_parallel_size=4,max_model_len=4096", "fewshot": 0, "tasks": "mmlu"} \ No newline at end of file diff --git a/llama3_Meta-Llama-3.1-8B-Instruct_fs0_mmlu_56bc86f0_group_cache.npy b/llama3_Meta-Llama-3.1-8B-Instruct_fs0_mmlu_56bc86f0_group_cache.npy new file mode 100644 index 0000000000000000000000000000000000000000..c41be4ccf090255127a033b5d24a0218491e4de3 --- /dev/null +++ b/llama3_Meta-Llama-3.1-8B-Instruct_fs0_mmlu_56bc86f0_group_cache.npy @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:73628756f24e52330a8b24579bdaaca5f0b5414f8f02190a5a532f50bb5b7ac6 +size 18438352 diff --git a/llama3_Meta-Llama-3.1-8B-Instruct_fs1_mmlu_6f66aeaa.json b/llama3_Meta-Llama-3.1-8B-Instruct_fs1_mmlu_6f66aeaa.json new file mode 100644 index 0000000000000000000000000000000000000000..17d2d27e4585efb3fadd85aa1efa6028b1e9b7e3 --- /dev/null +++ b/llama3_Meta-Llama-3.1-8B-Instruct_fs1_mmlu_6f66aeaa.json @@ -0,0 +1 @@ +{"uid": "6f66aeaa", "model": "NousResearch/Meta-Llama-3.1-8B-Instruct", "model_name": "Meta-Llama-3.1-8B-Instruct", "group": "llama3", "model_args": "pretrained=NousResearch/Meta-Llama-3.1-8B-Instruct,dtype=auto,gpu_memory_utilization=0.7,tensor_parallel_size=1,data_parallel_size=4,max_model_len=4096", "fewshot": 1, "tasks": "mmlu"} \ No newline at end of file diff --git a/llama3_Meta-Llama-3.1-8B-Instruct_fs1_mmlu_6f66aeaa_group_cache.npy b/llama3_Meta-Llama-3.1-8B-Instruct_fs1_mmlu_6f66aeaa_group_cache.npy new file mode 100644 index 0000000000000000000000000000000000000000..97575f78f7519010a84101bc9c07c82548c835b3 --- /dev/null +++ b/llama3_Meta-Llama-3.1-8B-Instruct_fs1_mmlu_6f66aeaa_group_cache.npy @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:792ebd4d16cc92aad6e8cc8d0bb6502c7d97f33989833bbea22916744d3c4a39 +size 16133627 diff --git a/llama3_Meta-Llama-3.1-8B-Instruct_fs2_mmlu_ad346594.json b/llama3_Meta-Llama-3.1-8B-Instruct_fs2_mmlu_ad346594.json new file mode 100644 index 0000000000000000000000000000000000000000..267f2efb7e55d9c4000d257a84280a09e585c384 --- /dev/null +++ b/llama3_Meta-Llama-3.1-8B-Instruct_fs2_mmlu_ad346594.json @@ -0,0 +1 @@ +{"uid": "ad346594", "model": "NousResearch/Meta-Llama-3.1-8B-Instruct", "model_name": "Meta-Llama-3.1-8B-Instruct", "group": "llama3", "model_args": "pretrained=NousResearch/Meta-Llama-3.1-8B-Instruct,dtype=auto,gpu_memory_utilization=0.7,tensor_parallel_size=1,data_parallel_size=4,max_model_len=4096", "fewshot": 2, "tasks": "mmlu"} \ No newline at end of file diff --git a/llama3_Meta-Llama-3.1-8B-Instruct_fs2_mmlu_ad346594_group_cache.npy b/llama3_Meta-Llama-3.1-8B-Instruct_fs2_mmlu_ad346594_group_cache.npy new file mode 100644 index 0000000000000000000000000000000000000000..02619c4dc346ed07153db00cf4512ea61b9539a9 --- /dev/null +++ b/llama3_Meta-Llama-3.1-8B-Instruct_fs2_mmlu_ad346594_group_cache.npy @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:deeb5aaa11d57f9442323edd0a2d64d54c73855c0cb986b7e39e53d6051a7c71 +size 13828817 diff --git a/llama3_Meta-Llama-3.1-8B-Instruct_fs3_mmlu_5227f586.json b/llama3_Meta-Llama-3.1-8B-Instruct_fs3_mmlu_5227f586.json new file mode 100644 index 0000000000000000000000000000000000000000..a6c9f95a01d2b4431cad6efc554ebed03c8aba6a --- /dev/null +++ b/llama3_Meta-Llama-3.1-8B-Instruct_fs3_mmlu_5227f586.json @@ -0,0 +1 @@ +{"uid": "5227f586", "model": "NousResearch/Meta-Llama-3.1-8B-Instruct", "model_name": "Meta-Llama-3.1-8B-Instruct", "group": "llama3", "model_args": "pretrained=NousResearch/Meta-Llama-3.1-8B-Instruct,dtype=auto,gpu_memory_utilization=0.7,tensor_parallel_size=1,data_parallel_size=4,max_model_len=4096", "fewshot": 3, "tasks": "mmlu"} \ No newline at end of file diff --git a/llama3_Meta-Llama-3.1-8B-Instruct_fs3_mmlu_5227f586_group_cache.npy b/llama3_Meta-Llama-3.1-8B-Instruct_fs3_mmlu_5227f586_group_cache.npy new file mode 100644 index 0000000000000000000000000000000000000000..4987bec09d4c264d5240df8645379841e3158a60 --- /dev/null +++ b/llama3_Meta-Llama-3.1-8B-Instruct_fs3_mmlu_5227f586_group_cache.npy @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:b9ea740ff5a5c42d134352ffcc68688fa795cb6f4a1ae7d14f6e9685b83b41ba +size 4609494 diff --git a/llama3_Meta-Llama-3.1-8B-Instruct_fs4_mmlu_1881efe7.json b/llama3_Meta-Llama-3.1-8B-Instruct_fs4_mmlu_1881efe7.json new file mode 100644 index 0000000000000000000000000000000000000000..79d898ce06250b20999f2f50dc2b3b4b8019e100 --- /dev/null +++ b/llama3_Meta-Llama-3.1-8B-Instruct_fs4_mmlu_1881efe7.json @@ -0,0 +1 @@ +{"uid": "1881efe7", "model": "NousResearch/Meta-Llama-3.1-8B-Instruct", "model_name": "Meta-Llama-3.1-8B-Instruct", "group": "llama3", "model_args": "pretrained=NousResearch/Meta-Llama-3.1-8B-Instruct,dtype=auto,gpu_memory_utilization=0.7,tensor_parallel_size=1,data_parallel_size=4,max_model_len=4096", "fewshot": 4, "tasks": "mmlu"} \ No newline at end of file diff --git a/llama3_Meta-Llama-3.1-8B-Instruct_fs4_mmlu_1881efe7_group_cache.npy b/llama3_Meta-Llama-3.1-8B-Instruct_fs4_mmlu_1881efe7_group_cache.npy new file mode 100644 index 0000000000000000000000000000000000000000..fec28ef7d38adbe6ba360299669e23f5a1ea3957 --- /dev/null +++ b/llama3_Meta-Llama-3.1-8B-Instruct_fs4_mmlu_1881efe7_group_cache.npy @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:3f6363fd7e87503774ac40c0059e6ee945094d2a6e35ad8d9a70b47db2154748 +size 13367861 diff --git a/llama3_Meta-Llama-3.1-8B-Instruct_fs5_mmlu_f94ef16d.json b/llama3_Meta-Llama-3.1-8B-Instruct_fs5_mmlu_f94ef16d.json new file mode 100644 index 0000000000000000000000000000000000000000..699e6140b433a486e727b99e9a03e9983549ee76 --- /dev/null +++ b/llama3_Meta-Llama-3.1-8B-Instruct_fs5_mmlu_f94ef16d.json @@ -0,0 +1 @@ +{"uid": "f94ef16d", "model": "NousResearch/Meta-Llama-3.1-8B-Instruct", "model_name": "Meta-Llama-3.1-8B-Instruct", "group": "llama3", "model_args": "pretrained=NousResearch/Meta-Llama-3.1-8B-Instruct,dtype=auto,gpu_memory_utilization=0.7,tensor_parallel_size=1,data_parallel_size=4,max_model_len=4096", "fewshot": 5, "tasks": "mmlu"} \ No newline at end of file diff --git a/llama3_Meta-Llama-3.1-8B-Instruct_fs5_mmlu_f94ef16d_group_cache.npy b/llama3_Meta-Llama-3.1-8B-Instruct_fs5_mmlu_f94ef16d_group_cache.npy new file mode 100644 index 0000000000000000000000000000000000000000..00b18f57c6e73b26835b53db92890ee2d4a4d899 --- /dev/null +++ b/llama3_Meta-Llama-3.1-8B-Instruct_fs5_mmlu_f94ef16d_group_cache.npy @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:dcbf96fcd8d9a8fef34ecac2e8f4489a53b5d824c0209712a112d2266fe9eb17 +size 1382782 diff --git a/llama3_Meta-Llama-3.1-8B_fs0_mmlu_10ba43b1.json b/llama3_Meta-Llama-3.1-8B_fs0_mmlu_10ba43b1.json new file mode 100644 index 0000000000000000000000000000000000000000..dbd26bf0190d5a334ca4f2dd4aa0d6e20d54ab15 --- /dev/null +++ b/llama3_Meta-Llama-3.1-8B_fs0_mmlu_10ba43b1.json @@ -0,0 +1 @@ +{"uid": "10ba43b1", "model": "NousResearch/Meta-Llama-3.1-8B", "model_name": "Meta-Llama-3.1-8B", "group": "llama3", "model_args": "pretrained=NousResearch/Meta-Llama-3.1-8B,dtype=auto,gpu_memory_utilization=0.7,tensor_parallel_size=1,data_parallel_size=4,max_model_len=4096", "fewshot": 0, "tasks": "mmlu"} \ No newline at end of file diff --git a/llama3_Meta-Llama-3.1-8B_fs0_mmlu_10ba43b1_group_cache.npy b/llama3_Meta-Llama-3.1-8B_fs0_mmlu_10ba43b1_group_cache.npy new file mode 100644 index 0000000000000000000000000000000000000000..4a9cecf2451e6644d26b43ec1f80f747c1f7a035 --- /dev/null +++ b/llama3_Meta-Llama-3.1-8B_fs0_mmlu_10ba43b1_group_cache.npy @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:ba70bd6e2ac90bec0d5c3a9e5878e3d8ef5084d2959d2ef8216efdbe110a9fcd +size 17516394 diff --git a/llama3_Meta-Llama-3.1-8B_fs1_mmlu_5d771df4.json b/llama3_Meta-Llama-3.1-8B_fs1_mmlu_5d771df4.json new file mode 100644 index 0000000000000000000000000000000000000000..10e1ccc71b8e4b0eb2f91d43f960775cf2598cf3 --- /dev/null +++ b/llama3_Meta-Llama-3.1-8B_fs1_mmlu_5d771df4.json @@ -0,0 +1 @@ +{"uid": "5d771df4", "model": "NousResearch/Meta-Llama-3.1-8B", "model_name": "Meta-Llama-3.1-8B", "group": "llama3", "model_args": "pretrained=NousResearch/Meta-Llama-3.1-8B,dtype=auto,gpu_memory_utilization=0.7,tensor_parallel_size=1,data_parallel_size=4,max_model_len=4096", "fewshot": 1, "tasks": "mmlu"} \ No newline at end of file diff --git a/llama3_Meta-Llama-3.1-8B_fs1_mmlu_5d771df4_group_cache.npy b/llama3_Meta-Llama-3.1-8B_fs1_mmlu_5d771df4_group_cache.npy new file mode 100644 index 0000000000000000000000000000000000000000..e3f95f79ff1da7cc33d140a60220f5e5dd9b6213 --- /dev/null +++ b/llama3_Meta-Llama-3.1-8B_fs1_mmlu_5d771df4_group_cache.npy @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:6499441772b18f39d8f04559410287ab03d4e8e519523164489a491eaadb17a3 +size 9219158 diff --git a/llama3_Meta-Llama-3.1-8B_fs2_mmlu_2588b8b6.json b/llama3_Meta-Llama-3.1-8B_fs2_mmlu_2588b8b6.json new file mode 100644 index 0000000000000000000000000000000000000000..cc6302a4d94e06f837a48cae398ebee190b566c9 --- /dev/null +++ b/llama3_Meta-Llama-3.1-8B_fs2_mmlu_2588b8b6.json @@ -0,0 +1 @@ +{"uid": "2588b8b6", "model": "NousResearch/Meta-Llama-3.1-8B", "model_name": "Meta-Llama-3.1-8B", "group": "llama3", "model_args": "pretrained=NousResearch/Meta-Llama-3.1-8B,dtype=auto,gpu_memory_utilization=0.7,tensor_parallel_size=1,data_parallel_size=4,max_model_len=4096", "fewshot": 2, "tasks": "mmlu"} \ No newline at end of file diff --git a/llama3_Meta-Llama-3.1-8B_fs2_mmlu_2588b8b6_group_cache.npy b/llama3_Meta-Llama-3.1-8B_fs2_mmlu_2588b8b6_group_cache.npy new file mode 100644 index 0000000000000000000000000000000000000000..c39f74b5598941c360357ff5b16ccdbe0e182d06 --- /dev/null +++ b/llama3_Meta-Llama-3.1-8B_fs2_mmlu_2588b8b6_group_cache.npy @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:0e143064b6b05997f859c24db5c1789123a821efa788d5faaf41d3783c9c48ea +size 2765644 diff --git a/llama3_Meta-Llama-3.1-8B_fs3_mmlu_820d306a.json b/llama3_Meta-Llama-3.1-8B_fs3_mmlu_820d306a.json new file mode 100644 index 0000000000000000000000000000000000000000..ca2d2d7d058e0dcae8d888c7f63c2953a78bacd7 --- /dev/null +++ b/llama3_Meta-Llama-3.1-8B_fs3_mmlu_820d306a.json @@ -0,0 +1 @@ +{"uid": "820d306a", "model": "NousResearch/Meta-Llama-3.1-8B", "model_name": "Meta-Llama-3.1-8B", "group": "llama3", "model_args": "pretrained=NousResearch/Meta-Llama-3.1-8B,dtype=auto,gpu_memory_utilization=0.7,tensor_parallel_size=1,data_parallel_size=4,max_model_len=4096", "fewshot": 3, "tasks": "mmlu"} \ No newline at end of file diff --git a/llama3_Meta-Llama-3.1-8B_fs3_mmlu_820d306a_group_cache.npy b/llama3_Meta-Llama-3.1-8B_fs3_mmlu_820d306a_group_cache.npy new file mode 100644 index 0000000000000000000000000000000000000000..1ad259de111f556c90ffd43108937e8ae578ff0e --- /dev/null +++ b/llama3_Meta-Llama-3.1-8B_fs3_mmlu_820d306a_group_cache.npy @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:e2d5eec8071eb2b1a4f8853b89348b21fdab6537e2848d51472ce0d4441483da +size 6453373 diff --git a/llama3_Meta-Llama-3.1-8B_fs4_mmlu_5d633d17.json b/llama3_Meta-Llama-3.1-8B_fs4_mmlu_5d633d17.json new file mode 100644 index 0000000000000000000000000000000000000000..51902849db071eb755accbd27622d12da14fcfc7 --- /dev/null +++ b/llama3_Meta-Llama-3.1-8B_fs4_mmlu_5d633d17.json @@ -0,0 +1 @@ +{"uid": "5d633d17", "model": "NousResearch/Meta-Llama-3.1-8B", "model_name": "Meta-Llama-3.1-8B", "group": "llama3", "model_args": "pretrained=NousResearch/Meta-Llama-3.1-8B,dtype=auto,gpu_memory_utilization=0.7,tensor_parallel_size=1,data_parallel_size=4,max_model_len=4096", "fewshot": 4, "tasks": "mmlu"} \ No newline at end of file diff --git a/llama3_Meta-Llama-3.1-8B_fs4_mmlu_5d633d17_group_cache.npy b/llama3_Meta-Llama-3.1-8B_fs4_mmlu_5d633d17_group_cache.npy new file mode 100644 index 0000000000000000000000000000000000000000..be179e172deb72f0ea6fdee95af9b11913cf64ca --- /dev/null +++ b/llama3_Meta-Llama-3.1-8B_fs4_mmlu_5d633d17_group_cache.npy @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:ecb66b758aa97a80ac6f39b0e8186819702833246b364101495a45dd10affd6c +size 18899281 diff --git a/llama3_Meta-Llama-3.1-8B_fs5_mmlu_91e5fc2c.json b/llama3_Meta-Llama-3.1-8B_fs5_mmlu_91e5fc2c.json new file mode 100644 index 0000000000000000000000000000000000000000..a8e39661a7b49ea4fda934ebebf41367ee1e8dfd --- /dev/null +++ b/llama3_Meta-Llama-3.1-8B_fs5_mmlu_91e5fc2c.json @@ -0,0 +1 @@ +{"uid": "91e5fc2c", "model": "NousResearch/Meta-Llama-3.1-8B", "model_name": "Meta-Llama-3.1-8B", "group": "llama3", "model_args": "pretrained=NousResearch/Meta-Llama-3.1-8B,dtype=auto,gpu_memory_utilization=0.7,tensor_parallel_size=1,data_parallel_size=4,max_model_len=4096", "fewshot": 5, "tasks": "mmlu"} \ No newline at end of file diff --git a/llama3_Meta-Llama-3.1-8B_fs5_mmlu_91e5fc2c_group_cache.npy b/llama3_Meta-Llama-3.1-8B_fs5_mmlu_91e5fc2c_group_cache.npy new file mode 100644 index 0000000000000000000000000000000000000000..a6299e9eef965b5e6ec4d43cd30d5e14869d08dd --- /dev/null +++ b/llama3_Meta-Llama-3.1-8B_fs5_mmlu_91e5fc2c_group_cache.npy @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:2bb1b8bab730a9157c66ccb9aac96cea851e450625bbe5c3dd2075ac905d0fee +size 9680087 diff --git a/tulu2.5_Llama-2-13b-hf_fs0_mmlu_bf3ec6f8.json b/tulu2.5_Llama-2-13b-hf_fs0_mmlu_bf3ec6f8.json new file mode 100644 index 0000000000000000000000000000000000000000..9506f65c6c5b0b125f9fa6009bd3e3ebce6fe397 --- /dev/null +++ b/tulu2.5_Llama-2-13b-hf_fs0_mmlu_bf3ec6f8.json @@ -0,0 +1 @@ +{"uid": "bf3ec6f8", "model": "meta-llama/Llama-2-13b-hf", "model_name": "Llama-2-13b-hf", "group": "tulu2.5", "model_args": "pretrained=meta-llama/Llama-2-13b-hf,dtype=auto,gpu_memory_utilization=0.7,tensor_parallel_size=1,data_parallel_size=4,max_model_len=4096", "fewshot": 0, "tasks": "mmlu"} \ No newline at end of file diff --git a/tulu2.5_Llama-2-13b-hf_fs0_mmlu_bf3ec6f8_group_cache.npy b/tulu2.5_Llama-2-13b-hf_fs0_mmlu_bf3ec6f8_group_cache.npy new file mode 100644 index 0000000000000000000000000000000000000000..8318cd0f52a2244b4c65cd0ac8a328adb7d0606b --- /dev/null +++ b/tulu2.5_Llama-2-13b-hf_fs0_mmlu_bf3ec6f8_group_cache.npy @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:f85d222d55feb771df6764b3ebb8f4ed053aafeb3727da455101428baeac8186 +size 3226241 diff --git a/tulu2.5_Llama-2-13b-hf_fs1_mmlu_3891fdd5.json b/tulu2.5_Llama-2-13b-hf_fs1_mmlu_3891fdd5.json new file mode 100644 index 0000000000000000000000000000000000000000..98c346eb1b60931f0e7bcd330aea7966801cae49 --- /dev/null +++ b/tulu2.5_Llama-2-13b-hf_fs1_mmlu_3891fdd5.json @@ -0,0 +1 @@ +{"uid": "3891fdd5", "model": "meta-llama/Llama-2-13b-hf", "model_name": "Llama-2-13b-hf", "group": "tulu2.5", "model_args": "pretrained=meta-llama/Llama-2-13b-hf,dtype=auto,gpu_memory_utilization=0.7,tensor_parallel_size=1,data_parallel_size=4,max_model_len=4096", "fewshot": 1, "tasks": "mmlu"} \ No newline at end of file diff --git a/tulu2.5_Llama-2-13b-hf_fs1_mmlu_3891fdd5_group_cache.npy b/tulu2.5_Llama-2-13b-hf_fs1_mmlu_3891fdd5_group_cache.npy new file mode 100644 index 0000000000000000000000000000000000000000..56d82b12362d91faf5588e90083b801a18d5cdf3 --- /dev/null +++ b/tulu2.5_Llama-2-13b-hf_fs1_mmlu_3891fdd5_group_cache.npy @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:4006277d4a2b70af5e8c16e31335037e733b20217ddf99c72a849eb83cf153b0 +size 5991908 diff --git a/tulu2.5_Llama-2-13b-hf_fs2_mmlu_e85838a0.json b/tulu2.5_Llama-2-13b-hf_fs2_mmlu_e85838a0.json new file mode 100644 index 0000000000000000000000000000000000000000..edf30d5cd85b9b5552433a11d0d937af482fff2e --- /dev/null +++ b/tulu2.5_Llama-2-13b-hf_fs2_mmlu_e85838a0.json @@ -0,0 +1 @@ +{"uid": "e85838a0", "model": "meta-llama/Llama-2-13b-hf", "model_name": "Llama-2-13b-hf", "group": "tulu2.5", "model_args": "pretrained=meta-llama/Llama-2-13b-hf,dtype=auto,gpu_memory_utilization=0.7,tensor_parallel_size=1,data_parallel_size=4,max_model_len=4096", "fewshot": 2, "tasks": "mmlu"} \ No newline at end of file diff --git a/tulu2.5_Llama-2-13b-hf_fs2_mmlu_e85838a0_group_cache.npy b/tulu2.5_Llama-2-13b-hf_fs2_mmlu_e85838a0_group_cache.npy new file mode 100644 index 0000000000000000000000000000000000000000..44b55772ebde0f5baa25ff48119892201782cc41 --- /dev/null +++ b/tulu2.5_Llama-2-13b-hf_fs2_mmlu_e85838a0_group_cache.npy @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:5ef7e06c1c8e199eb2c699ee5c21b5f67421466072235c00ae2b90ec803a843b +size 4609106 diff --git a/tulu2.5_Llama-2-13b-hf_fs3_mmlu_a0400f3e.json b/tulu2.5_Llama-2-13b-hf_fs3_mmlu_a0400f3e.json new file mode 100644 index 0000000000000000000000000000000000000000..83e7d555973a4ba8c58caa776f11d5727a96518f --- /dev/null +++ b/tulu2.5_Llama-2-13b-hf_fs3_mmlu_a0400f3e.json @@ -0,0 +1 @@ +{"uid": "a0400f3e", "model": "meta-llama/Llama-2-13b-hf", "model_name": "Llama-2-13b-hf", "group": "tulu2.5", "model_args": "pretrained=meta-llama/Llama-2-13b-hf,dtype=auto,gpu_memory_utilization=0.7,tensor_parallel_size=1,data_parallel_size=4,max_model_len=4096", "fewshot": 3, "tasks": "mmlu"} \ No newline at end of file diff --git a/tulu2.5_Llama-2-13b-hf_fs3_mmlu_a0400f3e_group_cache.npy b/tulu2.5_Llama-2-13b-hf_fs3_mmlu_a0400f3e_group_cache.npy new file mode 100644 index 0000000000000000000000000000000000000000..5c0dd04324bffb7c2c1e275dba2e7245782713d2 --- /dev/null +++ b/tulu2.5_Llama-2-13b-hf_fs3_mmlu_a0400f3e_group_cache.npy @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:f129568ef35297932107f59bb5dfe9713b6d535c98d0bf7d86cc7c8d39d9d5fc +size 10601144 diff --git a/tulu2.5_Llama-2-13b-hf_fs4_mmlu_c409e276.json b/tulu2.5_Llama-2-13b-hf_fs4_mmlu_c409e276.json new file mode 100644 index 0000000000000000000000000000000000000000..adb3eb7cd54a84d0dfce6964a9fc51c16b6d867a --- /dev/null +++ b/tulu2.5_Llama-2-13b-hf_fs4_mmlu_c409e276.json @@ -0,0 +1 @@ +{"uid": "c409e276", "model": "meta-llama/Llama-2-13b-hf", "model_name": "Llama-2-13b-hf", "group": "tulu2.5", "model_args": "pretrained=meta-llama/Llama-2-13b-hf,dtype=auto,gpu_memory_utilization=0.7,tensor_parallel_size=1,data_parallel_size=4,max_model_len=4096", "fewshot": 4, "tasks": "mmlu"} \ No newline at end of file diff --git a/tulu2.5_Llama-2-13b-hf_fs4_mmlu_c409e276_group_cache.npy b/tulu2.5_Llama-2-13b-hf_fs4_mmlu_c409e276_group_cache.npy new file mode 100644 index 0000000000000000000000000000000000000000..c089f29e6eb79fa2eb20c01d5fe0723ccd29f24e --- /dev/null +++ b/tulu2.5_Llama-2-13b-hf_fs4_mmlu_c409e276_group_cache.npy @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:f48561d73c89093f442c0e56d1189c6a2c26c5c7f4a10741828537a9db48148e +size 460753 diff --git a/tulu2.5_Llama-2-13b-hf_fs5_mmlu_3f54b48d.json b/tulu2.5_Llama-2-13b-hf_fs5_mmlu_3f54b48d.json new file mode 100644 index 0000000000000000000000000000000000000000..c2582cc05272d3eb2d08e4798e6cc2e128797142 --- /dev/null +++ b/tulu2.5_Llama-2-13b-hf_fs5_mmlu_3f54b48d.json @@ -0,0 +1 @@ +{"uid": "3f54b48d", "model": "meta-llama/Llama-2-13b-hf", "model_name": "Llama-2-13b-hf", "group": "tulu2.5", "model_args": "pretrained=meta-llama/Llama-2-13b-hf,dtype=auto,gpu_memory_utilization=0.7,tensor_parallel_size=1,data_parallel_size=4,max_model_len=4096", "fewshot": 5, "tasks": "mmlu"} \ No newline at end of file diff --git a/tulu2.5_Llama-2-13b-hf_fs5_mmlu_3f54b48d_group_cache.npy b/tulu2.5_Llama-2-13b-hf_fs5_mmlu_3f54b48d_group_cache.npy new file mode 100644 index 0000000000000000000000000000000000000000..e63990f8d5102eb38f2d92e3b0017cb55639fb3e --- /dev/null +++ b/tulu2.5_Llama-2-13b-hf_fs5_mmlu_3f54b48d_group_cache.npy @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:b6b58961c787eca96b95a4e853337ce8d2e3de093e27c418097f6f4c574a0c82 +size 9679316 diff --git a/tulu2.5_tulu-2-13b_fs0_mmlu_4103aaff.json b/tulu2.5_tulu-2-13b_fs0_mmlu_4103aaff.json new file mode 100644 index 0000000000000000000000000000000000000000..f58b04d434c4133cba16df58c378e432685a4f05 --- /dev/null +++ b/tulu2.5_tulu-2-13b_fs0_mmlu_4103aaff.json @@ -0,0 +1 @@ +{"uid": "4103aaff", "model": "allenai/tulu-2-13b", "model_name": "tulu-2-13b", "group": "tulu2.5", "model_args": "pretrained=allenai/tulu-2-13b,dtype=auto,gpu_memory_utilization=0.7,tensor_parallel_size=1,data_parallel_size=4,max_model_len=4096", "fewshot": 0, "tasks": "mmlu"} \ No newline at end of file diff --git a/tulu2.5_tulu-2-13b_fs0_mmlu_4103aaff_group_cache.npy b/tulu2.5_tulu-2-13b_fs0_mmlu_4103aaff_group_cache.npy new file mode 100644 index 0000000000000000000000000000000000000000..31318077a8fe48849491ecea0e4d2910500fb979 --- /dev/null +++ b/tulu2.5_tulu-2-13b_fs0_mmlu_4103aaff_group_cache.npy @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:1e177c8b515fdc20b44a8f3dd1bea17b4a94bbc4ccbb346ee2ef2d7b2d14cb60 +size 1843451 diff --git a/tulu2.5_tulu-2-13b_fs1_mmlu_9c7215a5.json b/tulu2.5_tulu-2-13b_fs1_mmlu_9c7215a5.json new file mode 100644 index 0000000000000000000000000000000000000000..7cbb5af1056edd187a3ee298c02b509243717bd2 --- /dev/null +++ b/tulu2.5_tulu-2-13b_fs1_mmlu_9c7215a5.json @@ -0,0 +1 @@ +{"uid": "9c7215a5", "model": "allenai/tulu-2-13b", "model_name": "tulu-2-13b", "group": "tulu2.5", "model_args": "pretrained=allenai/tulu-2-13b,dtype=auto,gpu_memory_utilization=0.7,tensor_parallel_size=1,data_parallel_size=4,max_model_len=4096", "fewshot": 1, "tasks": "mmlu"} \ No newline at end of file diff --git a/tulu2.5_tulu-2-13b_fs1_mmlu_9c7215a5_group_cache.npy b/tulu2.5_tulu-2-13b_fs1_mmlu_9c7215a5_group_cache.npy new file mode 100644 index 0000000000000000000000000000000000000000..e5efbaf495ce89f6c50549226417a69f28a62910 --- /dev/null +++ b/tulu2.5_tulu-2-13b_fs1_mmlu_9c7215a5_group_cache.npy @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:dd515d2249517f2fac410e6c08a66f3579e6475e293b39bc201605eb3f54ac7f +size 921653 diff --git a/tulu2.5_tulu-2-13b_fs2_mmlu_1a74c045.json b/tulu2.5_tulu-2-13b_fs2_mmlu_1a74c045.json new file mode 100644 index 0000000000000000000000000000000000000000..9c4e064d9c6d7f9233e7da59bc03801abccc8aea --- /dev/null +++ b/tulu2.5_tulu-2-13b_fs2_mmlu_1a74c045.json @@ -0,0 +1 @@ +{"uid": "1a74c045", "model": "allenai/tulu-2-13b", "model_name": "tulu-2-13b", "group": "tulu2.5", "model_args": "pretrained=allenai/tulu-2-13b,dtype=auto,gpu_memory_utilization=0.7,tensor_parallel_size=1,data_parallel_size=4,max_model_len=4096", "fewshot": 2, "tasks": "mmlu"} \ No newline at end of file diff --git a/tulu2.5_tulu-2-13b_fs2_mmlu_1a74c045_group_cache.npy b/tulu2.5_tulu-2-13b_fs2_mmlu_1a74c045_group_cache.npy new file mode 100644 index 0000000000000000000000000000000000000000..28d6a342555e3190f2164fe2bb0426e1e5479768 --- /dev/null +++ b/tulu2.5_tulu-2-13b_fs2_mmlu_1a74c045_group_cache.npy @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:24ae33cecfea005f0515d75aad87cba5c8b605cf1a1627d7c31d7b1b5ef16d7a +size 11062043 diff --git a/tulu2.5_tulu-2-13b_fs3_mmlu_f84b3c89.json b/tulu2.5_tulu-2-13b_fs3_mmlu_f84b3c89.json new file mode 100644 index 0000000000000000000000000000000000000000..f4270316e698bb39a27b52b1035e0ea1aa526d93 --- /dev/null +++ b/tulu2.5_tulu-2-13b_fs3_mmlu_f84b3c89.json @@ -0,0 +1 @@ +{"uid": "f84b3c89", "model": "allenai/tulu-2-13b", "model_name": "tulu-2-13b", "group": "tulu2.5", "model_args": "pretrained=allenai/tulu-2-13b,dtype=auto,gpu_memory_utilization=0.7,tensor_parallel_size=1,data_parallel_size=4,max_model_len=4096", "fewshot": 3, "tasks": "mmlu"} \ No newline at end of file diff --git a/tulu2.5_tulu-2-13b_fs3_mmlu_f84b3c89_group_cache.npy b/tulu2.5_tulu-2-13b_fs3_mmlu_f84b3c89_group_cache.npy new file mode 100644 index 0000000000000000000000000000000000000000..e677dc8b542ca61ab7579390c793f46c2175559a --- /dev/null +++ b/tulu2.5_tulu-2-13b_fs3_mmlu_f84b3c89_group_cache.npy @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:d6b3bfb94b6d4d1257636d6a21231e3d46cc338c030f2477d1f653960400cae6 +size 2765324 diff --git a/tulu2.5_tulu-2-13b_fs4_mmlu_4adf6584.json b/tulu2.5_tulu-2-13b_fs4_mmlu_4adf6584.json new file mode 100644 index 0000000000000000000000000000000000000000..295f7eac578c99c828450c9d64a0b11f11622e90 --- /dev/null +++ b/tulu2.5_tulu-2-13b_fs4_mmlu_4adf6584.json @@ -0,0 +1 @@ +{"uid": "4adf6584", "model": "allenai/tulu-2-13b", "model_name": "tulu-2-13b", "group": "tulu2.5", "model_args": "pretrained=allenai/tulu-2-13b,dtype=auto,gpu_memory_utilization=0.7,tensor_parallel_size=1,data_parallel_size=4,max_model_len=4096", "fewshot": 4, "tasks": "mmlu"} \ No newline at end of file diff --git a/tulu2.5_tulu-2-13b_fs4_mmlu_4adf6584_group_cache.npy b/tulu2.5_tulu-2-13b_fs4_mmlu_4adf6584_group_cache.npy new file mode 100644 index 0000000000000000000000000000000000000000..a865b3f7ebf20eb05f7b63d20e2711d38c652aaa --- /dev/null +++ b/tulu2.5_tulu-2-13b_fs4_mmlu_4adf6584_group_cache.npy @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:e3139c26fbd283bbad8564183d53d7c876f0222da2237aa02c4b73d1ecf9dffd +size 7835603 diff --git a/tulu2.5_tulu-2-13b_fs5_mmlu_c658e991.json b/tulu2.5_tulu-2-13b_fs5_mmlu_c658e991.json new file mode 100644 index 0000000000000000000000000000000000000000..649c8efb0ec21d555176ee5aa6c17a0a3a795844 --- /dev/null +++ b/tulu2.5_tulu-2-13b_fs5_mmlu_c658e991.json @@ -0,0 +1 @@ +{"uid": "c658e991", "model": "allenai/tulu-2-13b", "model_name": "tulu-2-13b", "group": "tulu2.5", "model_args": "pretrained=allenai/tulu-2-13b,dtype=auto,gpu_memory_utilization=0.7,tensor_parallel_size=1,data_parallel_size=4,max_model_len=4096", "fewshot": 5, "tasks": "mmlu"} \ No newline at end of file diff --git a/tulu2.5_tulu-2-13b_fs5_mmlu_c658e991_group_cache.npy b/tulu2.5_tulu-2-13b_fs5_mmlu_c658e991_group_cache.npy new file mode 100644 index 0000000000000000000000000000000000000000..d97e4f74cdb2adf415e27be30d34b7a6d11ce386 --- /dev/null +++ b/tulu2.5_tulu-2-13b_fs5_mmlu_c658e991_group_cache.npy @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:2ab87e9c8d3ca73b1aa25bc30a192a7589d3558d344add1c0c9b55030f4ea786 +size 1382552 diff --git a/tulu2.5_tulu-2-dpo-13b_fs0_mmlu_eb17b60e.json b/tulu2.5_tulu-2-dpo-13b_fs0_mmlu_eb17b60e.json new file mode 100644 index 0000000000000000000000000000000000000000..4d8ae49e6d3c9c911b30f7a7e040b7da4f3ca4c0 --- /dev/null +++ b/tulu2.5_tulu-2-dpo-13b_fs0_mmlu_eb17b60e.json @@ -0,0 +1 @@ +{"uid": "eb17b60e", "model": "allenai/tulu-2-dpo-13b", "model_name": "tulu-2-dpo-13b", "group": "tulu2.5", "model_args": "pretrained=allenai/tulu-2-dpo-13b,dtype=auto,gpu_memory_utilization=0.7,tensor_parallel_size=1,data_parallel_size=4,max_model_len=4096", "fewshot": 0, "tasks": "mmlu"} \ No newline at end of file diff --git a/tulu2.5_tulu-2-dpo-13b_fs0_mmlu_eb17b60e_group_cache.npy b/tulu2.5_tulu-2-dpo-13b_fs0_mmlu_eb17b60e_group_cache.npy new file mode 100644 index 0000000000000000000000000000000000000000..a90276de7e1c27f71669286f738ccb5df1404cff --- /dev/null +++ b/tulu2.5_tulu-2-dpo-13b_fs0_mmlu_eb17b60e_group_cache.npy @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:10a3d96dfbeb4b9c5afa6c203154c8756e66798d236958863f61c6403454998f +size 5070017 diff --git a/tulu2.5_tulu-2-dpo-13b_fs1_mmlu_495b335d.json b/tulu2.5_tulu-2-dpo-13b_fs1_mmlu_495b335d.json new file mode 100644 index 0000000000000000000000000000000000000000..90a14427763e8574e832ed20fac2554d3cbd4322 --- /dev/null +++ b/tulu2.5_tulu-2-dpo-13b_fs1_mmlu_495b335d.json @@ -0,0 +1 @@ +{"uid": "495b335d", "model": "allenai/tulu-2-dpo-13b", "model_name": "tulu-2-dpo-13b", "group": "tulu2.5", "model_args": "pretrained=allenai/tulu-2-dpo-13b,dtype=auto,gpu_memory_utilization=0.7,tensor_parallel_size=1,data_parallel_size=4,max_model_len=4096", "fewshot": 1, "tasks": "mmlu"} \ No newline at end of file diff --git a/tulu2.5_tulu-2-dpo-13b_fs1_mmlu_495b335d_group_cache.npy b/tulu2.5_tulu-2-dpo-13b_fs1_mmlu_495b335d_group_cache.npy new file mode 100644 index 0000000000000000000000000000000000000000..caac574d42ead786601818edfabfb764529453b2 --- /dev/null +++ b/tulu2.5_tulu-2-dpo-13b_fs1_mmlu_495b335d_group_cache.npy @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:14f26a58144d2230eec39a67f6b22a4ec53008aabc56a96e593d697d795873ae +size 8757488 diff --git a/tulu2.5_tulu-2-dpo-13b_fs2_mmlu_9d7a74d9.json b/tulu2.5_tulu-2-dpo-13b_fs2_mmlu_9d7a74d9.json new file mode 100644 index 0000000000000000000000000000000000000000..a86ca27f2de27154b9b30f5792c0c5403876b235 --- /dev/null +++ b/tulu2.5_tulu-2-dpo-13b_fs2_mmlu_9d7a74d9.json @@ -0,0 +1 @@ +{"uid": "9d7a74d9", "model": "allenai/tulu-2-dpo-13b", "model_name": "tulu-2-dpo-13b", "group": "tulu2.5", "model_args": "pretrained=allenai/tulu-2-dpo-13b,dtype=auto,gpu_memory_utilization=0.7,tensor_parallel_size=1,data_parallel_size=4,max_model_len=4096", "fewshot": 2, "tasks": "mmlu"} \ No newline at end of file diff --git a/tulu2.5_tulu-2-dpo-13b_fs2_mmlu_9d7a74d9_group_cache.npy b/tulu2.5_tulu-2-dpo-13b_fs2_mmlu_9d7a74d9_group_cache.npy new file mode 100644 index 0000000000000000000000000000000000000000..fa250af7b34b1b6e38541f411c0ef1d7189d137c --- /dev/null +++ b/tulu2.5_tulu-2-dpo-13b_fs2_mmlu_9d7a74d9_group_cache.npy @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:1bb3ee670157465ddf7672467b4d5758a44153a82266fd23f97086a12b37605a +size 6452819 diff --git a/tulu2.5_tulu-2-dpo-13b_fs3_mmlu_11e118a1.json b/tulu2.5_tulu-2-dpo-13b_fs3_mmlu_11e118a1.json new file mode 100644 index 0000000000000000000000000000000000000000..863753da0edfaa2fdd56ad97603d9870e49c2099 --- /dev/null +++ b/tulu2.5_tulu-2-dpo-13b_fs3_mmlu_11e118a1.json @@ -0,0 +1 @@ +{"uid": "11e118a1", "model": "allenai/tulu-2-dpo-13b", "model_name": "tulu-2-dpo-13b", "group": "tulu2.5", "model_args": "pretrained=allenai/tulu-2-dpo-13b,dtype=auto,gpu_memory_utilization=0.7,tensor_parallel_size=1,data_parallel_size=4,max_model_len=4096", "fewshot": 3, "tasks": "mmlu"} \ No newline at end of file diff --git a/tulu2.5_tulu-2-dpo-13b_fs3_mmlu_11e118a1_group_cache.npy b/tulu2.5_tulu-2-dpo-13b_fs3_mmlu_11e118a1_group_cache.npy new file mode 100644 index 0000000000000000000000000000000000000000..c92b439bdfc57e09c233103cc665a3d06d550074 --- /dev/null +++ b/tulu2.5_tulu-2-dpo-13b_fs3_mmlu_11e118a1_group_cache.npy @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:ea80fd5dd5e4da404d5778386b61cd8221b41082a0bc14ee9edf729b36bde23c +size 9218399 diff --git a/tulu2.5_tulu-2-dpo-13b_fs4_mmlu_0424aea5.json b/tulu2.5_tulu-2-dpo-13b_fs4_mmlu_0424aea5.json new file mode 100644 index 0000000000000000000000000000000000000000..bf7cc9263f91fa2b3d96fce9cecb3c128d672b07 --- /dev/null +++ b/tulu2.5_tulu-2-dpo-13b_fs4_mmlu_0424aea5.json @@ -0,0 +1 @@ +{"uid": "0424aea5", "model": "allenai/tulu-2-dpo-13b", "model_name": "tulu-2-dpo-13b", "group": "tulu2.5", "model_args": "pretrained=allenai/tulu-2-dpo-13b,dtype=auto,gpu_memory_utilization=0.7,tensor_parallel_size=1,data_parallel_size=4,max_model_len=4096", "fewshot": 4, "tasks": "mmlu"} \ No newline at end of file diff --git a/tulu2.5_tulu-2-dpo-13b_fs4_mmlu_0424aea5_group_cache.npy b/tulu2.5_tulu-2-dpo-13b_fs4_mmlu_0424aea5_group_cache.npy new file mode 100644 index 0000000000000000000000000000000000000000..1f7bf72ea4ba35590f354c4c0e11f1635ff145a5 --- /dev/null +++ b/tulu2.5_tulu-2-dpo-13b_fs4_mmlu_0424aea5_group_cache.npy @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:b641503906716aad5e04e15cf1d7a31d6f585a5434ccaefdcaa234fa06827379 +size 7374704 diff --git a/tulu2.5_tulu-2-dpo-13b_fs5_mmlu_78b648f7.json b/tulu2.5_tulu-2-dpo-13b_fs5_mmlu_78b648f7.json new file mode 100644 index 0000000000000000000000000000000000000000..ae06ad1e7bea150dcae5c64a13f12c7b8b9f5172 --- /dev/null +++ b/tulu2.5_tulu-2-dpo-13b_fs5_mmlu_78b648f7.json @@ -0,0 +1 @@ +{"uid": "78b648f7", "model": "allenai/tulu-2-dpo-13b", "model_name": "tulu-2-dpo-13b", "group": "tulu2.5", "model_args": "pretrained=allenai/tulu-2-dpo-13b,dtype=auto,gpu_memory_utilization=0.7,tensor_parallel_size=1,data_parallel_size=4,max_model_len=4096", "fewshot": 5, "tasks": "mmlu"} \ No newline at end of file diff --git a/tulu2.5_tulu-2-dpo-13b_fs5_mmlu_78b648f7_group_cache.npy b/tulu2.5_tulu-2-dpo-13b_fs5_mmlu_78b648f7_group_cache.npy new file mode 100644 index 0000000000000000000000000000000000000000..74060fcbe9fba2782436a0e23001ad8d80a2f741 --- /dev/null +++ b/tulu2.5_tulu-2-dpo-13b_fs5_mmlu_78b648f7_group_cache.npy @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:200242a2040a445416479c2670977bfe5ac993d16a66a0e95a6f4fbbda60ae95 +size 10140227 diff --git a/tulu2.5_tulu-v2.5-ppo-13b-uf-mean-70b-uf-rm_fs0_mmlu_025df173.json b/tulu2.5_tulu-v2.5-ppo-13b-uf-mean-70b-uf-rm_fs0_mmlu_025df173.json new file mode 100644 index 0000000000000000000000000000000000000000..e7319582cd93b7d99e942aa598eb643c5e221701 --- /dev/null +++ b/tulu2.5_tulu-v2.5-ppo-13b-uf-mean-70b-uf-rm_fs0_mmlu_025df173.json @@ -0,0 +1 @@ +{"uid": "025df173", "model": "allenai/tulu-v2.5-ppo-13b-uf-mean-70b-uf-rm", "model_name": "tulu-v2.5-ppo-13b-uf-mean-70b-uf-rm", "group": "tulu2.5", "model_args": "pretrained=allenai/tulu-v2.5-ppo-13b-uf-mean-70b-uf-rm,dtype=auto,gpu_memory_utilization=0.7,tensor_parallel_size=1,data_parallel_size=4,max_model_len=4096", "fewshot": 0, "tasks": "mmlu"} \ No newline at end of file diff --git a/tulu2.5_tulu-v2.5-ppo-13b-uf-mean-70b-uf-rm_fs0_mmlu_025df173_group_cache.npy b/tulu2.5_tulu-v2.5-ppo-13b-uf-mean-70b-uf-rm_fs0_mmlu_025df173_group_cache.npy new file mode 100644 index 0000000000000000000000000000000000000000..d15a32337a59d0b4687960eda33926d3e4b78256 --- /dev/null +++ b/tulu2.5_tulu-v2.5-ppo-13b-uf-mean-70b-uf-rm_fs0_mmlu_025df173_group_cache.npy @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:a8eafebe2f029ef96f0c887f9eecc06acc2809064cb120f5bb1e7465a51589fc +size 6913793 diff --git a/tulu2.5_tulu-v2.5-ppo-13b-uf-mean-70b-uf-rm_fs1_mmlu_fd1abc56.json b/tulu2.5_tulu-v2.5-ppo-13b-uf-mean-70b-uf-rm_fs1_mmlu_fd1abc56.json new file mode 100644 index 0000000000000000000000000000000000000000..6f76deca4ebe2ba7947c9cd272347107ff7f679d --- /dev/null +++ b/tulu2.5_tulu-v2.5-ppo-13b-uf-mean-70b-uf-rm_fs1_mmlu_fd1abc56.json @@ -0,0 +1 @@ +{"uid": "fd1abc56", "model": "allenai/tulu-v2.5-ppo-13b-uf-mean-70b-uf-rm", "model_name": "tulu-v2.5-ppo-13b-uf-mean-70b-uf-rm", "group": "tulu2.5", "model_args": "pretrained=allenai/tulu-v2.5-ppo-13b-uf-mean-70b-uf-rm,dtype=auto,gpu_memory_utilization=0.7,tensor_parallel_size=1,data_parallel_size=4,max_model_len=4096", "fewshot": 1, "tasks": "mmlu"} \ No newline at end of file diff --git a/tulu2.5_tulu-v2.5-ppo-13b-uf-mean-70b-uf-rm_fs1_mmlu_fd1abc56_group_cache.npy b/tulu2.5_tulu-v2.5-ppo-13b-uf-mean-70b-uf-rm_fs1_mmlu_fd1abc56_group_cache.npy new file mode 100644 index 0000000000000000000000000000000000000000..c8c1a4486b73b3adb5c7b6b1c9bc13eaf3907f85 --- /dev/null +++ b/tulu2.5_tulu-v2.5-ppo-13b-uf-mean-70b-uf-rm_fs1_mmlu_fd1abc56_group_cache.npy @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:9650d6f3a1cf6cfdfd6af1f0353e9cf863108e28d549b5195423385ebbd876e8 +size 4148189 diff --git a/tulu2.5_tulu-v2.5-ppo-13b-uf-mean-70b-uf-rm_fs2_mmlu_c2166e54.json b/tulu2.5_tulu-v2.5-ppo-13b-uf-mean-70b-uf-rm_fs2_mmlu_c2166e54.json new file mode 100644 index 0000000000000000000000000000000000000000..c4b2c38a3185d267fa683adc0745382980733fe2 --- /dev/null +++ b/tulu2.5_tulu-v2.5-ppo-13b-uf-mean-70b-uf-rm_fs2_mmlu_c2166e54.json @@ -0,0 +1 @@ +{"uid": "c2166e54", "model": "allenai/tulu-v2.5-ppo-13b-uf-mean-70b-uf-rm", "model_name": "tulu-v2.5-ppo-13b-uf-mean-70b-uf-rm", "group": "tulu2.5", "model_args": "pretrained=allenai/tulu-v2.5-ppo-13b-uf-mean-70b-uf-rm,dtype=auto,gpu_memory_utilization=0.7,tensor_parallel_size=1,data_parallel_size=4,max_model_len=4096", "fewshot": 2, "tasks": "mmlu"} \ No newline at end of file diff --git a/tulu2.5_tulu-v2.5-ppo-13b-uf-mean-70b-uf-rm_fs2_mmlu_c2166e54_group_cache.npy b/tulu2.5_tulu-v2.5-ppo-13b-uf-mean-70b-uf-rm_fs2_mmlu_c2166e54_group_cache.npy new file mode 100644 index 0000000000000000000000000000000000000000..a2807a17b4b4fedcb72380d4a74bf4eb67a6874a --- /dev/null +++ b/tulu2.5_tulu-v2.5-ppo-13b-uf-mean-70b-uf-rm_fs2_mmlu_c2166e54_group_cache.npy @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:ce88456db7d783491fd0001e036174fa68b304b7775433d9a1c64f56bafe9a43 +size 8296577 diff --git a/tulu2.5_tulu-v2.5-ppo-13b-uf-mean-70b-uf-rm_fs3_mmlu_7784c23d.json b/tulu2.5_tulu-v2.5-ppo-13b-uf-mean-70b-uf-rm_fs3_mmlu_7784c23d.json new file mode 100644 index 0000000000000000000000000000000000000000..aac16c39ebd8b80867278d8e81215627c4015f76 --- /dev/null +++ b/tulu2.5_tulu-v2.5-ppo-13b-uf-mean-70b-uf-rm_fs3_mmlu_7784c23d.json @@ -0,0 +1 @@ +{"uid": "7784c23d", "model": "allenai/tulu-v2.5-ppo-13b-uf-mean-70b-uf-rm", "model_name": "tulu-v2.5-ppo-13b-uf-mean-70b-uf-rm", "group": "tulu2.5", "model_args": "pretrained=allenai/tulu-v2.5-ppo-13b-uf-mean-70b-uf-rm,dtype=auto,gpu_memory_utilization=0.7,tensor_parallel_size=1,data_parallel_size=4,max_model_len=4096", "fewshot": 3, "tasks": "mmlu"} \ No newline at end of file diff --git a/tulu2.5_tulu-v2.5-ppo-13b-uf-mean-70b-uf-rm_fs3_mmlu_7784c23d_group_cache.npy b/tulu2.5_tulu-v2.5-ppo-13b-uf-mean-70b-uf-rm_fs3_mmlu_7784c23d_group_cache.npy new file mode 100644 index 0000000000000000000000000000000000000000..d3a91f2082556ab110e5fd0f1fcc56339092e8fc --- /dev/null +++ b/tulu2.5_tulu-v2.5-ppo-13b-uf-mean-70b-uf-rm_fs3_mmlu_7784c23d_group_cache.npy @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:e5ec172d4893279f3dd47f906bc4efb1f426862ae0e6e24a3207da7cfa15710f +size 3687215 diff --git a/tulu2.5_tulu-v2.5-ppo-13b-uf-mean-70b-uf-rm_fs4_mmlu_9f8276f3.json b/tulu2.5_tulu-v2.5-ppo-13b-uf-mean-70b-uf-rm_fs4_mmlu_9f8276f3.json new file mode 100644 index 0000000000000000000000000000000000000000..9820b1d49a11fa87617865fbe2ff595d1ecf3b6f --- /dev/null +++ b/tulu2.5_tulu-v2.5-ppo-13b-uf-mean-70b-uf-rm_fs4_mmlu_9f8276f3.json @@ -0,0 +1 @@ +{"uid": "9f8276f3", "model": "allenai/tulu-v2.5-ppo-13b-uf-mean-70b-uf-rm", "model_name": "tulu-v2.5-ppo-13b-uf-mean-70b-uf-rm", "group": "tulu2.5", "model_args": "pretrained=allenai/tulu-v2.5-ppo-13b-uf-mean-70b-uf-rm,dtype=auto,gpu_memory_utilization=0.7,tensor_parallel_size=1,data_parallel_size=4,max_model_len=4096", "fewshot": 4, "tasks": "mmlu"} \ No newline at end of file diff --git a/tulu2.5_tulu-v2.5-ppo-13b-uf-mean-70b-uf-rm_fs4_mmlu_9f8276f3_group_cache.npy b/tulu2.5_tulu-v2.5-ppo-13b-uf-mean-70b-uf-rm_fs4_mmlu_9f8276f3_group_cache.npy new file mode 100644 index 0000000000000000000000000000000000000000..0ad2791f646acbda6a0a6950e9cf9b31e93ab3c2 --- /dev/null +++ b/tulu2.5_tulu-v2.5-ppo-13b-uf-mean-70b-uf-rm_fs4_mmlu_9f8276f3_group_cache.npy @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:0584f31d6fb85ebdb2bddd791b7cba1c156aca4afc0a0bef5f11ce4a4908b257 +size 2304425 diff --git a/tulu2.5_tulu-v2.5-ppo-13b-uf-mean-70b-uf-rm_fs5_mmlu_1cb06d74.json b/tulu2.5_tulu-v2.5-ppo-13b-uf-mean-70b-uf-rm_fs5_mmlu_1cb06d74.json new file mode 100644 index 0000000000000000000000000000000000000000..b0c7b03ddaa3a48b57a8868f2a24db6352370187 --- /dev/null +++ b/tulu2.5_tulu-v2.5-ppo-13b-uf-mean-70b-uf-rm_fs5_mmlu_1cb06d74.json @@ -0,0 +1 @@ +{"uid": "1cb06d74", "model": "allenai/tulu-v2.5-ppo-13b-uf-mean-70b-uf-rm", "model_name": "tulu-v2.5-ppo-13b-uf-mean-70b-uf-rm", "group": "tulu2.5", "model_args": "pretrained=allenai/tulu-v2.5-ppo-13b-uf-mean-70b-uf-rm,dtype=auto,gpu_memory_utilization=0.7,tensor_parallel_size=1,data_parallel_size=4,max_model_len=4096", "fewshot": 5, "tasks": "mmlu"} \ No newline at end of file diff --git a/tulu2.5_tulu-v2.5-ppo-13b-uf-mean-70b-uf-rm_fs5_mmlu_1cb06d74_group_cache.npy b/tulu2.5_tulu-v2.5-ppo-13b-uf-mean-70b-uf-rm_fs5_mmlu_1cb06d74_group_cache.npy new file mode 100644 index 0000000000000000000000000000000000000000..446ebf23f7291b355cc1eb5ebd3ef78fd65e5684 --- /dev/null +++ b/tulu2.5_tulu-v2.5-ppo-13b-uf-mean-70b-uf-rm_fs5_mmlu_1cb06d74_group_cache.npy @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:3856c7aa0e3ef59add124d6b910ca8ea227ad4c0d3b75cb60ae24ce82ebcc0c6 +size 5530991