diff --git a/model-00001-of-00061.safetensors b/model-00001-of-00061.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..71a3cefa9b5cca7dac2acda58ae42b1b8a81941b --- /dev/null +++ b/model-00001-of-00061.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:6751e75dd5bc9925e212218bbc7d3aab88b0a04323da9becbeadb7e098175516 +size 1070099832 diff --git a/model-00002-of-00061.safetensors b/model-00002-of-00061.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..bb9089baab40e275c5e821d1a199aa5edc808e48 --- /dev/null +++ b/model-00002-of-00061.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:dc3a31be6245cf628aa77da916374d60c65ea9b9ecb176222893ebf3c0e7d6b9 +size 1070099832 diff --git a/model-00003-of-00061.safetensors b/model-00003-of-00061.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..dc13cfb7bf471ef37ea6d4112d41f291d3a027d3 --- /dev/null +++ b/model-00003-of-00061.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:500e93988212bb3189e9e85be4272cb49542d437c69663108f74b1776d39afc7 +size 1070099832 diff --git a/model-00004-of-00061.safetensors b/model-00004-of-00061.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..bd1ac9b02858315d1b8a341c0e5447883d0e28ba --- /dev/null +++ b/model-00004-of-00061.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:c306ead8519285260f137a2ef3c77f450b7cab8285a8561f9ad5efbb8acfe078 +size 1070099832 diff --git a/model-00005-of-00061.safetensors b/model-00005-of-00061.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..863a4e0e7aebf2647f8f496faca31d4fa532fa5e --- /dev/null +++ b/model-00005-of-00061.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:347bbea0140b57a5177c6caa51e40444041bb9b9d6599f1a6f8bbc5ab643cf86 +size 1070099832 diff --git a/model-00006-of-00061.safetensors b/model-00006-of-00061.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..ebf1f1b6a4b8b6b94a50360240926907f2098f1e --- /dev/null +++ b/model-00006-of-00061.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:85723c70ef9ae6d45c9384a1b92d78ad37a95e2279134749a4ffe0c29f33b3e8 +size 1070099832 diff --git a/model-00007-of-00061.safetensors b/model-00007-of-00061.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..c1a59d1f9872239ba6a9f690a5ec9232161d0562 --- /dev/null +++ b/model-00007-of-00061.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:08a22bc2a37c9974b58cdd93f9d0a4b34249a9362aea4e3169453dbde9d8f3ca +size 1070099832 diff --git a/model-00008-of-00061.safetensors b/model-00008-of-00061.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..ca5714a738fb7b49c83c80df962587f87a8f2614 --- /dev/null +++ b/model-00008-of-00061.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:a15de39305842c87019d316ce621869935d08a3a2de70d27567a02bad84f9bde +size 1070099832 diff --git a/model-00009-of-00061.safetensors b/model-00009-of-00061.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..63afb57935b73820a81395684b67ae840aebb024 --- /dev/null +++ b/model-00009-of-00061.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:4a7bfbf428d30d7b4196893f6fca990205944b8c8ae4997f09ac8afb3075b133 +size 1070099832 diff --git a/model-00010-of-00061.safetensors b/model-00010-of-00061.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..936c1e3f1894da430c198b64c4c3664e94aef5d6 --- /dev/null +++ b/model-00010-of-00061.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:956f1e6c5c85c558701bc8bb4c7167ee187c60ed952db1d6ded795ca762703e6 +size 1070099832 diff --git a/model-00011-of-00061.safetensors b/model-00011-of-00061.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..2d71ad7eb393c40b9de47133c833904f6330ff39 --- /dev/null +++ b/model-00011-of-00061.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:4359ffd762f0b54db08f9b71e5540d50dc3a638e061d072d850c7a90bc52b1f5 +size 1070099840 diff --git a/model-00012-of-00061.safetensors b/model-00012-of-00061.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..3b78294a50030e248189fdad09da618d7d42bd0b --- /dev/null +++ b/model-00012-of-00061.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:ecb547bc56f0f3d3528c9958a468fec60b8437768f0740843e1c8c8f588e7545 +size 1070099840 diff --git a/model-00013-of-00061.safetensors b/model-00013-of-00061.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..c5c552efe2b3d149e8d04c4ff60fa10e12146cf1 --- /dev/null +++ b/model-00013-of-00061.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:ea2443dadab8990610f3caf841a65e457ce30bc30d4e31d923b574d1bcd88134 +size 1070099840 diff --git a/model-00014-of-00061.safetensors b/model-00014-of-00061.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..91a90598fcb5940990a9e5e7602cc99a1a9a507e --- /dev/null +++ b/model-00014-of-00061.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:c1c40eeac56b3ca74caea8a26ee760e5ade920bbdb7b5f99f14f64dcc1f80def +size 1070099840 diff --git a/model-00015-of-00061.safetensors b/model-00015-of-00061.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..f3c05e4e72d7be593168de386c7c89e339d8100b --- /dev/null +++ b/model-00015-of-00061.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:43814df94c647d7700b752ca7e88847c0cc5c6dbb87f47ea520686c1c1cf5531 +size 1070099840 diff --git a/model-00016-of-00061.safetensors b/model-00016-of-00061.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..39d572fea2b83ea9cb78c77c397af14e5e1059d1 --- /dev/null +++ b/model-00016-of-00061.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:b3697edaa3b2390c30b5d38487395f589ee10416230f602f560424031395c8b6 +size 1070099840 diff --git a/model-00017-of-00061.safetensors b/model-00017-of-00061.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..3339c166680094042bb71b2eb2e8a36fae2e1b12 --- /dev/null +++ b/model-00017-of-00061.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:b71ba6f3ff5d5d354865a77c7c82cad750b92113e3b30fc8913fe392e8126ebb +size 1070099840 diff --git a/model-00018-of-00061.safetensors b/model-00018-of-00061.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..af3752092a6d862ebfdc3a714cb0f693bc725414 --- /dev/null +++ b/model-00018-of-00061.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:2a612e5006b04e438208f10f742d0b295f24cb8a5cb025291e411101ad70a7f1 +size 1070099840 diff --git a/model-00019-of-00061.safetensors b/model-00019-of-00061.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..1be4f3463792225d243ca390a6a9db54efd97dde --- /dev/null +++ b/model-00019-of-00061.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:d3773cecb779dca0889f9c84c2fb633ae57e23ac74446bd1db07293b5fcd8169 +size 1070099840 diff --git a/model-00020-of-00061.safetensors b/model-00020-of-00061.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..995ba3824b0f84e8476d5c4f113515ac4436ae1c --- /dev/null +++ b/model-00020-of-00061.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:5a38fa28b75a6a60083cc3cced4c249c40b99bd6c84cfe01729312563ee6d24c +size 1070099840 diff --git a/model-00021-of-00061.safetensors b/model-00021-of-00061.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..f519fe1db2b7a640e773d8949483afd9c070be10 --- /dev/null +++ b/model-00021-of-00061.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:ece35f1bbc0f1224f00d1940a013b08d09c32a7e53c4bf2e032ca5a3abd582b3 +size 1070099840 diff --git a/model-00022-of-00061.safetensors b/model-00022-of-00061.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..0a4b83d3c8ddfa587c75b39596292fbef7aed80a --- /dev/null +++ b/model-00022-of-00061.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:ef0315cd95f636975a70085efdc70f42e814140eb8345c63fe5e6df5e6f6673f +size 1070099840 diff --git a/model-00023-of-00061.safetensors b/model-00023-of-00061.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..9907dfa1c55701d0762090033379b86d00a51968 --- /dev/null +++ b/model-00023-of-00061.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:7012d0f5ff63d321248b4c94e7bb4533b12da5778fb387090fd5da18e633be21 +size 1070099840 diff --git a/model-00024-of-00061.safetensors b/model-00024-of-00061.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..710152404a889edd3662900c637946f4e78a5b88 --- /dev/null +++ b/model-00024-of-00061.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:edde81b0689a56b15c0eff5c07f7c02c053d3d1ebb3e03d1fa93f47f978538ae +size 1070099840 diff --git a/model-00025-of-00061.safetensors b/model-00025-of-00061.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..0368c454cfa3b3a8591cb43b78e0297981900182 --- /dev/null +++ b/model-00025-of-00061.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:389f18438bca71fbf97df49444664c49119947584f2308c5e259b6b352d7bc4a +size 1070099840 diff --git a/model-00026-of-00061.safetensors b/model-00026-of-00061.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..eb5ee9df43db43627b36e0154a54a25c4efffd84 --- /dev/null +++ b/model-00026-of-00061.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:dc8c6e9d9ec8fa25653e0cbc0016c977f28618836070ff63c0c3a25bcb2192e6 +size 1070099840 diff --git a/model-00027-of-00061.safetensors b/model-00027-of-00061.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..62fa79dd33a27fff65ef54571c4005bd27da1798 --- /dev/null +++ b/model-00027-of-00061.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:b62f12428cf01705e7d4f627dc56cb6d7806b9adfa0425678eea13c1d1a74622 +size 1070099840 diff --git a/model-00028-of-00061.safetensors b/model-00028-of-00061.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..614a04f6e199905f8035b190fab5b3957d8c3096 --- /dev/null +++ b/model-00028-of-00061.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:3d184d64f22b6ef2f67fbd4583be700fcebe33852bc1609766d9118a5534fbd7 +size 1070099840 diff --git a/model-00029-of-00061.safetensors b/model-00029-of-00061.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..30578e857fec4b05a8bbf024c5957b12949f5d87 --- /dev/null +++ b/model-00029-of-00061.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:602aa006e1e2210de68716cd3edaf0aba98fba0ae18af1642b6595bbf03c79d2 +size 1070099840 diff --git a/model-00030-of-00061.safetensors b/model-00030-of-00061.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..944710c884cff491a5090394fbe237e1440ca5ec --- /dev/null +++ b/model-00030-of-00061.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:0038166ac772e20417c19bbf52faa4c30e59f01beed4bef42cca3e02305bcfc5 +size 1070099840 diff --git a/model-00031-of-00061.safetensors b/model-00031-of-00061.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..0a361940f8cd695e6a9e94e00dc9ab6ee9a8318c --- /dev/null +++ b/model-00031-of-00061.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:c0ea5521486161b99ef811dd067dcfa9519f6a6b8e1efe5e0ce8c1a93637abca +size 1070099840 diff --git a/model-00032-of-00061.safetensors b/model-00032-of-00061.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..e6faaa51eadbc69b9c7db1348804a9fdac11795a --- /dev/null +++ b/model-00032-of-00061.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:14d312296eb93db833fa02f1986fb82992f8431b6e82ebdf4e5424b1ba4238de +size 1070099840 diff --git a/model-00033-of-00061.safetensors b/model-00033-of-00061.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..c24eceafa8c61e1277ab84b59649064d23fe7e35 --- /dev/null +++ b/model-00033-of-00061.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:dba431610183608aa8e609660dfc97eeeffe6843118ea9ceed58e43ae0a1b3a6 +size 1070099840 diff --git a/model-00034-of-00061.safetensors b/model-00034-of-00061.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..7d31ea826551612f946c8a69be18d964499f0aa9 --- /dev/null +++ b/model-00034-of-00061.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:803247d19ec9bb16428399c78dcb555f85a05635726f3ae9e14c83a9cb242a7c +size 1070099840 diff --git a/model-00035-of-00061.safetensors b/model-00035-of-00061.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..c07b573214a3db54f7f899fe8f698d7313b2f862 --- /dev/null +++ b/model-00035-of-00061.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:951fcc42dea876da2180160943eaef298a0db72ba8b79c1c45d43a124ca39266 +size 1070099840 diff --git a/model-00036-of-00061.safetensors b/model-00036-of-00061.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..232a0147be71d63786234a99952bb6f4880747c7 --- /dev/null +++ b/model-00036-of-00061.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:90cb0aaf74f1cd6879703faf7917ce62f4647fa9500729ba66a99c64ac330ef9 +size 1070099840 diff --git a/model-00037-of-00061.safetensors b/model-00037-of-00061.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..c839f20c2336ce1fc7592ac6d0598896af7d641c --- /dev/null +++ b/model-00037-of-00061.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:f96cc8c17ca90b893870da12a3e050216dbbd34a713d08de0bb9802219740bd6 +size 1070099840 diff --git a/model-00038-of-00061.safetensors b/model-00038-of-00061.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..9fbc718552cd360091f4d82164f15589db81ef29 --- /dev/null +++ b/model-00038-of-00061.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:ee3e0456cd58558b07eb00ec69ea0ea558974fa6dc1d5d80e942ea578bcbc92c +size 1070099840 diff --git a/model-00039-of-00061.safetensors b/model-00039-of-00061.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..7e96febcb526bab1583292bf1fcb30c06ff25f4c --- /dev/null +++ b/model-00039-of-00061.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:4feeacc25943bf67c9c16d8277341ee473d2dd131b3b12fa8f0c9dfb2e659759 +size 1070099840 diff --git a/model-00040-of-00061.safetensors b/model-00040-of-00061.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..b4d06ce86052e7a1a61ea9c062ba448cd30216d2 --- /dev/null +++ b/model-00040-of-00061.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:ab063fa496776ae6203cf177c0a75c6397936baa720bcd915a8ae7dd77af79ed +size 1070099840 diff --git a/model-00041-of-00061.safetensors b/model-00041-of-00061.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..379a5c3bc1cbf4fdf2a9c8b9a8c8dc10ab210bad --- /dev/null +++ b/model-00041-of-00061.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:66cc81eee107c7f1215f9bb48529276a8bb0e998af02a3a0853f2cd9b072283b +size 1070099840 diff --git a/model-00042-of-00061.safetensors b/model-00042-of-00061.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..246f43a8e1d47692373d043719ec4d70f54a3156 --- /dev/null +++ b/model-00042-of-00061.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:b26d51d86c2406244924b84613858173f830f8712e48daa5a5e41816ff91a36d +size 1070099840 diff --git a/model-00043-of-00061.safetensors b/model-00043-of-00061.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..305dcde8b04767bbede17262ec150dea6cb3f47e --- /dev/null +++ b/model-00043-of-00061.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:9740ea20f786ea225c47d88b041d56540bbab6803250447ee05b7ac28e3d4f59 +size 1070099840 diff --git a/model-00044-of-00061.safetensors b/model-00044-of-00061.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..e0e12cfd23268825220c0f5d3cc0f97d7b6e63e7 --- /dev/null +++ b/model-00044-of-00061.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:ecd018591744b596dd61163c5dceb0a185d607ef88664d53d8cb30deab1a4b27 +size 1070099840 diff --git a/model-00045-of-00061.safetensors b/model-00045-of-00061.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..2666798a1df0c39b92ac1d34fff415fa19fcb577 --- /dev/null +++ b/model-00045-of-00061.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:fa73e3632441d3bc0f8f1802a7ef2b5251cabc35c5030331a7c00b09658085b0 +size 1070099840 diff --git a/model-00046-of-00061.safetensors b/model-00046-of-00061.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..db46637a73aef895d9379212882cae79f4aff750 --- /dev/null +++ b/model-00046-of-00061.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:9c7475a372bbf8278f7e5f0313a5ed7b9f489a9dacbba9467e38bd3ccde38bba +size 1070099840 diff --git a/model-00047-of-00061.safetensors b/model-00047-of-00061.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..1f390500b6638320ea50d065ee99eea2706b4ce4 --- /dev/null +++ b/model-00047-of-00061.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:cdca88957974b92718467177e5985bf2b9ecc9aa7a68736ceafa97ced89f9149 +size 1070099840 diff --git a/model-00048-of-00061.safetensors b/model-00048-of-00061.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..27acd36234624eef014d52dc5214363dd590361c --- /dev/null +++ b/model-00048-of-00061.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:1ee9f59a2e8dce79005c258ea6851b3430607502393a972d60e87579dc54c4a6 +size 1070099840 diff --git a/model-00049-of-00061.safetensors b/model-00049-of-00061.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..92e1bc1bbc6e2c4efa5a8e688743345ec3a9702f --- /dev/null +++ b/model-00049-of-00061.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:90a8c72a13168fee704c0414ae2b98493766659ac9e3b7922b8b8e4b4a41816c +size 1070099840 diff --git a/model-00050-of-00061.safetensors b/model-00050-of-00061.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..70b5ce42590efc71b149c41a8b78b85b4f84279c --- /dev/null +++ b/model-00050-of-00061.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:a8ee83fe64ea55d23497ac5f33b2f3aa475398af0aa67364308397de821a0b2f +size 1070099840 diff --git a/model-00051-of-00061.safetensors b/model-00051-of-00061.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..a7a2aa59335b0b434101511436bd5ebc075563a0 --- /dev/null +++ b/model-00051-of-00061.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:598df079d427818bcb8a4ff78120af959cc795d859dfbd28defad396697a0892 +size 1070099840 diff --git a/model-00052-of-00061.safetensors b/model-00052-of-00061.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..fa9904a4ea62ce1b161b4488aa5352b6296c9868 --- /dev/null +++ b/model-00052-of-00061.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:a99630cba6405b5a86c63e636a6ae90a4fe70eb51f0c748f6dac7d5463a60c8f +size 1070099840 diff --git a/model-00053-of-00061.safetensors b/model-00053-of-00061.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..7d6e3e23b6224b8618c85484b63db4cf79b13aee --- /dev/null +++ b/model-00053-of-00061.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:d2f170f47f63476d1c876e6e1f5bceaeb9460983905f96d27900a13cc981392a +size 1070099840 diff --git a/model-00054-of-00061.safetensors b/model-00054-of-00061.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..2d6ecdb9ad464ae937ec78429caed7fc8c0c7d59 --- /dev/null +++ b/model-00054-of-00061.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:1b5fcc2e42cb082a0c06c7c9ffddfe87109e8ae0422217a08e3b52c388b4b0c7 +size 1070099840 diff --git a/model-00055-of-00061.safetensors b/model-00055-of-00061.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..b490edcbcb67bb7f460bb64b07d2385b7e24b6df --- /dev/null +++ b/model-00055-of-00061.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:e2590d1234bd8ebfe682ab16df4fc445a497bc112359cfd7545cb72678d12db5 +size 1070099840 diff --git a/model-00056-of-00061.safetensors b/model-00056-of-00061.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..1e9bf5708de277712e3adacdb3e0fa4dce5ce539 --- /dev/null +++ b/model-00056-of-00061.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:f5ff15ce1f011e2b8294154ef343b98aefdff7db02019da62df80f01b0f7cd31 +size 1070099840 diff --git a/model-00057-of-00061.safetensors b/model-00057-of-00061.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..c65811e7c62a1854bcaff4146fd10be86e40fdc0 --- /dev/null +++ b/model-00057-of-00061.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:3588a55ad2fb86d5bcbd92bd22236f2091790fe64ee471c8321487a41447106d +size 1070099840 diff --git a/model-00058-of-00061.safetensors b/model-00058-of-00061.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..12e808224c30bc1134275c1474be1b91fc5938e0 --- /dev/null +++ b/model-00058-of-00061.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:edb58cde7ec57e071d04e3ef6a32b431d24140e5ed282b012bc765cce896ce59 +size 1070099840 diff --git a/model-00059-of-00061.safetensors b/model-00059-of-00061.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..0c700cf519318c190fab20c6450c279488cb9c1b --- /dev/null +++ b/model-00059-of-00061.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:328eaa0a80fdb1e3c80f3edb5616ddf3410c9df4abcc146b5bb411164efe5918 +size 1070099840 diff --git a/model-00060-of-00061.safetensors b/model-00060-of-00061.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..8ee57389ee2eccdf9b339eb224c36478cff0b1b1 --- /dev/null +++ b/model-00060-of-00061.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:817241af9f62620d091bed6eaabe93aa48f4169d3f61f72722b2abda555ec73a +size 1070099840 diff --git a/model-00061-of-00061.safetensors b/model-00061-of-00061.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..e816c6d5e17e816d3190478bddc170ccfef7c329 --- /dev/null +++ b/model-00061-of-00061.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:fc87c8fdbe245e0597c05652faeefeffc180cc63cfe9fa8589eda8e5f5f96be1 +size 851981632 diff --git a/model.safetensors.index.json b/model.safetensors.index.json new file mode 100644 index 0000000000000000000000000000000000000000..4d999eeb1a1b5342bd808ecacf6e63922cc3b934 --- /dev/null +++ b/model.safetensors.index.json @@ -0,0 +1,610 @@ +{ + "weight_map": { + "model.layers.0.input_layernorm.weight": "model-00001-of-00061.safetensors", + "model.layers.0.post_attention_layernorm.weight": "model-00001-of-00061.safetensors", + "model.layers.0.self_attn.q_proj.weight": "model-00001-of-00061.safetensors", + "model.layers.0.self_attn.k_proj.weight": "model-00001-of-00061.safetensors", + "model.layers.0.self_attn.v_proj.weight": "model-00001-of-00061.safetensors", + "model.layers.0.self_attn.o_proj.weight": "model-00001-of-00061.safetensors", + "model.layers.0.mlp.gate_proj.weight": "model-00001-of-00061.safetensors", + "model.layers.0.mlp.down_proj.weight": "model-00001-of-00061.safetensors", + "model.layers.0.mlp.up_proj.weight": "model-00001-of-00061.safetensors", + "model.layers.0.self_attn.rotary_emb.inv_freq": "model-00001-of-00061.safetensors", + "model.layers.1.input_layernorm.weight": "model-00002-of-00061.safetensors", + "model.layers.1.post_attention_layernorm.weight": "model-00002-of-00061.safetensors", + "model.layers.1.self_attn.q_proj.weight": "model-00002-of-00061.safetensors", + "model.layers.1.self_attn.k_proj.weight": "model-00002-of-00061.safetensors", + "model.layers.1.self_attn.v_proj.weight": "model-00002-of-00061.safetensors", + "model.layers.1.self_attn.o_proj.weight": "model-00002-of-00061.safetensors", + "model.layers.1.mlp.gate_proj.weight": "model-00002-of-00061.safetensors", + "model.layers.1.mlp.down_proj.weight": "model-00002-of-00061.safetensors", + "model.layers.1.mlp.up_proj.weight": "model-00002-of-00061.safetensors", + "model.layers.1.self_attn.rotary_emb.inv_freq": "model-00002-of-00061.safetensors", + "model.layers.2.input_layernorm.weight": "model-00003-of-00061.safetensors", + "model.layers.2.post_attention_layernorm.weight": "model-00003-of-00061.safetensors", + "model.layers.2.self_attn.q_proj.weight": "model-00003-of-00061.safetensors", + "model.layers.2.self_attn.k_proj.weight": "model-00003-of-00061.safetensors", + "model.layers.2.self_attn.v_proj.weight": "model-00003-of-00061.safetensors", + "model.layers.2.self_attn.o_proj.weight": "model-00003-of-00061.safetensors", + "model.layers.2.mlp.gate_proj.weight": "model-00003-of-00061.safetensors", + "model.layers.2.mlp.down_proj.weight": "model-00003-of-00061.safetensors", + "model.layers.2.mlp.up_proj.weight": "model-00003-of-00061.safetensors", + "model.layers.2.self_attn.rotary_emb.inv_freq": "model-00003-of-00061.safetensors", + "model.layers.3.input_layernorm.weight": "model-00004-of-00061.safetensors", + "model.layers.3.post_attention_layernorm.weight": "model-00004-of-00061.safetensors", + "model.layers.3.self_attn.q_proj.weight": "model-00004-of-00061.safetensors", + "model.layers.3.self_attn.k_proj.weight": "model-00004-of-00061.safetensors", + "model.layers.3.self_attn.v_proj.weight": "model-00004-of-00061.safetensors", + "model.layers.3.self_attn.o_proj.weight": "model-00004-of-00061.safetensors", + "model.layers.3.mlp.gate_proj.weight": "model-00004-of-00061.safetensors", + "model.layers.3.mlp.down_proj.weight": "model-00004-of-00061.safetensors", + "model.layers.3.mlp.up_proj.weight": "model-00004-of-00061.safetensors", + "model.layers.3.self_attn.rotary_emb.inv_freq": "model-00004-of-00061.safetensors", + "model.layers.4.input_layernorm.weight": "model-00005-of-00061.safetensors", + "model.layers.4.post_attention_layernorm.weight": "model-00005-of-00061.safetensors", + "model.layers.4.self_attn.q_proj.weight": "model-00005-of-00061.safetensors", + "model.layers.4.self_attn.k_proj.weight": "model-00005-of-00061.safetensors", + "model.layers.4.self_attn.v_proj.weight": "model-00005-of-00061.safetensors", + "model.layers.4.self_attn.o_proj.weight": "model-00005-of-00061.safetensors", + "model.layers.4.mlp.gate_proj.weight": "model-00005-of-00061.safetensors", + "model.layers.4.mlp.down_proj.weight": "model-00005-of-00061.safetensors", + "model.layers.4.mlp.up_proj.weight": "model-00005-of-00061.safetensors", + "model.layers.4.self_attn.rotary_emb.inv_freq": "model-00005-of-00061.safetensors", + "model.layers.5.input_layernorm.weight": "model-00006-of-00061.safetensors", + "model.layers.5.post_attention_layernorm.weight": "model-00006-of-00061.safetensors", + "model.layers.5.self_attn.q_proj.weight": "model-00006-of-00061.safetensors", + "model.layers.5.self_attn.k_proj.weight": "model-00006-of-00061.safetensors", + "model.layers.5.self_attn.v_proj.weight": "model-00006-of-00061.safetensors", + "model.layers.5.self_attn.o_proj.weight": "model-00006-of-00061.safetensors", + "model.layers.5.mlp.gate_proj.weight": "model-00006-of-00061.safetensors", + "model.layers.5.mlp.down_proj.weight": "model-00006-of-00061.safetensors", + "model.layers.5.mlp.up_proj.weight": "model-00006-of-00061.safetensors", + "model.layers.5.self_attn.rotary_emb.inv_freq": "model-00006-of-00061.safetensors", + "model.layers.6.input_layernorm.weight": "model-00007-of-00061.safetensors", + "model.layers.6.post_attention_layernorm.weight": "model-00007-of-00061.safetensors", + "model.layers.6.self_attn.q_proj.weight": "model-00007-of-00061.safetensors", + "model.layers.6.self_attn.k_proj.weight": "model-00007-of-00061.safetensors", + "model.layers.6.self_attn.v_proj.weight": "model-00007-of-00061.safetensors", + "model.layers.6.self_attn.o_proj.weight": "model-00007-of-00061.safetensors", + "model.layers.6.mlp.gate_proj.weight": "model-00007-of-00061.safetensors", + "model.layers.6.mlp.down_proj.weight": "model-00007-of-00061.safetensors", + "model.layers.6.mlp.up_proj.weight": "model-00007-of-00061.safetensors", + "model.layers.6.self_attn.rotary_emb.inv_freq": "model-00007-of-00061.safetensors", + "model.layers.7.input_layernorm.weight": "model-00008-of-00061.safetensors", + "model.layers.7.post_attention_layernorm.weight": "model-00008-of-00061.safetensors", + "model.layers.7.self_attn.q_proj.weight": "model-00008-of-00061.safetensors", + "model.layers.7.self_attn.k_proj.weight": "model-00008-of-00061.safetensors", + "model.layers.7.self_attn.v_proj.weight": "model-00008-of-00061.safetensors", + "model.layers.7.self_attn.o_proj.weight": "model-00008-of-00061.safetensors", + "model.layers.7.mlp.gate_proj.weight": "model-00008-of-00061.safetensors", + "model.layers.7.mlp.down_proj.weight": "model-00008-of-00061.safetensors", + "model.layers.7.mlp.up_proj.weight": "model-00008-of-00061.safetensors", + "model.layers.7.self_attn.rotary_emb.inv_freq": "model-00008-of-00061.safetensors", + "model.layers.8.input_layernorm.weight": "model-00009-of-00061.safetensors", + "model.layers.8.post_attention_layernorm.weight": "model-00009-of-00061.safetensors", + "model.layers.8.self_attn.q_proj.weight": "model-00009-of-00061.safetensors", + "model.layers.8.self_attn.k_proj.weight": "model-00009-of-00061.safetensors", + "model.layers.8.self_attn.v_proj.weight": "model-00009-of-00061.safetensors", + "model.layers.8.self_attn.o_proj.weight": "model-00009-of-00061.safetensors", + "model.layers.8.mlp.gate_proj.weight": "model-00009-of-00061.safetensors", + "model.layers.8.mlp.down_proj.weight": "model-00009-of-00061.safetensors", + "model.layers.8.mlp.up_proj.weight": "model-00009-of-00061.safetensors", + "model.layers.8.self_attn.rotary_emb.inv_freq": "model-00009-of-00061.safetensors", + "model.layers.9.input_layernorm.weight": "model-00010-of-00061.safetensors", + "model.layers.9.post_attention_layernorm.weight": "model-00010-of-00061.safetensors", + "model.layers.9.self_attn.q_proj.weight": "model-00010-of-00061.safetensors", + "model.layers.9.self_attn.k_proj.weight": "model-00010-of-00061.safetensors", + "model.layers.9.self_attn.v_proj.weight": "model-00010-of-00061.safetensors", + "model.layers.9.self_attn.o_proj.weight": "model-00010-of-00061.safetensors", + "model.layers.9.mlp.gate_proj.weight": "model-00010-of-00061.safetensors", + "model.layers.9.mlp.down_proj.weight": "model-00010-of-00061.safetensors", + "model.layers.9.mlp.up_proj.weight": "model-00010-of-00061.safetensors", + "model.layers.9.self_attn.rotary_emb.inv_freq": "model-00010-of-00061.safetensors", + "model.layers.10.input_layernorm.weight": "model-00011-of-00061.safetensors", + "model.layers.10.post_attention_layernorm.weight": "model-00011-of-00061.safetensors", + "model.layers.10.self_attn.q_proj.weight": "model-00011-of-00061.safetensors", + "model.layers.10.self_attn.k_proj.weight": "model-00011-of-00061.safetensors", + "model.layers.10.self_attn.v_proj.weight": "model-00011-of-00061.safetensors", + "model.layers.10.self_attn.o_proj.weight": "model-00011-of-00061.safetensors", + "model.layers.10.mlp.gate_proj.weight": "model-00011-of-00061.safetensors", + "model.layers.10.mlp.down_proj.weight": "model-00011-of-00061.safetensors", + "model.layers.10.mlp.up_proj.weight": "model-00011-of-00061.safetensors", + "model.layers.10.self_attn.rotary_emb.inv_freq": "model-00011-of-00061.safetensors", + "model.layers.11.input_layernorm.weight": "model-00012-of-00061.safetensors", + "model.layers.11.post_attention_layernorm.weight": "model-00012-of-00061.safetensors", + "model.layers.11.self_attn.q_proj.weight": "model-00012-of-00061.safetensors", + "model.layers.11.self_attn.k_proj.weight": "model-00012-of-00061.safetensors", + "model.layers.11.self_attn.v_proj.weight": "model-00012-of-00061.safetensors", + "model.layers.11.self_attn.o_proj.weight": "model-00012-of-00061.safetensors", + "model.layers.11.mlp.gate_proj.weight": "model-00012-of-00061.safetensors", + "model.layers.11.mlp.down_proj.weight": "model-00012-of-00061.safetensors", + "model.layers.11.mlp.up_proj.weight": "model-00012-of-00061.safetensors", + "model.layers.11.self_attn.rotary_emb.inv_freq": "model-00012-of-00061.safetensors", + "model.layers.12.input_layernorm.weight": "model-00013-of-00061.safetensors", + "model.layers.12.post_attention_layernorm.weight": "model-00013-of-00061.safetensors", + "model.layers.12.self_attn.q_proj.weight": "model-00013-of-00061.safetensors", + "model.layers.12.self_attn.k_proj.weight": "model-00013-of-00061.safetensors", + "model.layers.12.self_attn.v_proj.weight": "model-00013-of-00061.safetensors", + "model.layers.12.self_attn.o_proj.weight": "model-00013-of-00061.safetensors", + "model.layers.12.mlp.gate_proj.weight": "model-00013-of-00061.safetensors", + "model.layers.12.mlp.down_proj.weight": "model-00013-of-00061.safetensors", + "model.layers.12.mlp.up_proj.weight": "model-00013-of-00061.safetensors", + "model.layers.12.self_attn.rotary_emb.inv_freq": "model-00013-of-00061.safetensors", + "model.layers.13.input_layernorm.weight": "model-00014-of-00061.safetensors", + "model.layers.13.post_attention_layernorm.weight": "model-00014-of-00061.safetensors", + "model.layers.13.self_attn.q_proj.weight": "model-00014-of-00061.safetensors", + "model.layers.13.self_attn.k_proj.weight": "model-00014-of-00061.safetensors", + "model.layers.13.self_attn.v_proj.weight": "model-00014-of-00061.safetensors", + "model.layers.13.self_attn.o_proj.weight": "model-00014-of-00061.safetensors", + "model.layers.13.mlp.gate_proj.weight": "model-00014-of-00061.safetensors", + "model.layers.13.mlp.down_proj.weight": "model-00014-of-00061.safetensors", + "model.layers.13.mlp.up_proj.weight": "model-00014-of-00061.safetensors", + "model.layers.13.self_attn.rotary_emb.inv_freq": "model-00014-of-00061.safetensors", + "model.layers.14.input_layernorm.weight": "model-00015-of-00061.safetensors", + "model.layers.14.post_attention_layernorm.weight": "model-00015-of-00061.safetensors", + "model.layers.14.self_attn.q_proj.weight": "model-00015-of-00061.safetensors", + "model.layers.14.self_attn.k_proj.weight": "model-00015-of-00061.safetensors", + "model.layers.14.self_attn.v_proj.weight": "model-00015-of-00061.safetensors", + "model.layers.14.self_attn.o_proj.weight": "model-00015-of-00061.safetensors", + "model.layers.14.mlp.gate_proj.weight": "model-00015-of-00061.safetensors", + "model.layers.14.mlp.down_proj.weight": "model-00015-of-00061.safetensors", + "model.layers.14.mlp.up_proj.weight": "model-00015-of-00061.safetensors", + "model.layers.14.self_attn.rotary_emb.inv_freq": "model-00015-of-00061.safetensors", + "model.layers.15.input_layernorm.weight": "model-00016-of-00061.safetensors", + "model.layers.15.post_attention_layernorm.weight": "model-00016-of-00061.safetensors", + "model.layers.15.self_attn.q_proj.weight": "model-00016-of-00061.safetensors", + "model.layers.15.self_attn.k_proj.weight": "model-00016-of-00061.safetensors", + "model.layers.15.self_attn.v_proj.weight": "model-00016-of-00061.safetensors", + "model.layers.15.self_attn.o_proj.weight": "model-00016-of-00061.safetensors", + "model.layers.15.mlp.gate_proj.weight": "model-00016-of-00061.safetensors", + "model.layers.15.mlp.down_proj.weight": "model-00016-of-00061.safetensors", + "model.layers.15.mlp.up_proj.weight": "model-00016-of-00061.safetensors", + "model.layers.15.self_attn.rotary_emb.inv_freq": "model-00016-of-00061.safetensors", + "model.layers.16.input_layernorm.weight": "model-00017-of-00061.safetensors", + "model.layers.16.post_attention_layernorm.weight": "model-00017-of-00061.safetensors", + "model.layers.16.self_attn.q_proj.weight": "model-00017-of-00061.safetensors", + "model.layers.16.self_attn.k_proj.weight": "model-00017-of-00061.safetensors", + "model.layers.16.self_attn.v_proj.weight": "model-00017-of-00061.safetensors", + "model.layers.16.self_attn.o_proj.weight": "model-00017-of-00061.safetensors", + "model.layers.16.mlp.gate_proj.weight": "model-00017-of-00061.safetensors", + "model.layers.16.mlp.down_proj.weight": "model-00017-of-00061.safetensors", + "model.layers.16.mlp.up_proj.weight": "model-00017-of-00061.safetensors", + "model.layers.16.self_attn.rotary_emb.inv_freq": "model-00017-of-00061.safetensors", + "model.layers.17.input_layernorm.weight": "model-00018-of-00061.safetensors", + "model.layers.17.post_attention_layernorm.weight": "model-00018-of-00061.safetensors", + "model.layers.17.self_attn.q_proj.weight": "model-00018-of-00061.safetensors", + "model.layers.17.self_attn.k_proj.weight": "model-00018-of-00061.safetensors", + "model.layers.17.self_attn.v_proj.weight": "model-00018-of-00061.safetensors", + "model.layers.17.self_attn.o_proj.weight": "model-00018-of-00061.safetensors", + "model.layers.17.mlp.gate_proj.weight": "model-00018-of-00061.safetensors", + "model.layers.17.mlp.down_proj.weight": "model-00018-of-00061.safetensors", + "model.layers.17.mlp.up_proj.weight": "model-00018-of-00061.safetensors", + "model.layers.17.self_attn.rotary_emb.inv_freq": "model-00018-of-00061.safetensors", + "model.layers.18.input_layernorm.weight": "model-00019-of-00061.safetensors", + "model.layers.18.post_attention_layernorm.weight": "model-00019-of-00061.safetensors", + "model.layers.18.self_attn.q_proj.weight": "model-00019-of-00061.safetensors", + "model.layers.18.self_attn.k_proj.weight": "model-00019-of-00061.safetensors", + "model.layers.18.self_attn.v_proj.weight": "model-00019-of-00061.safetensors", + "model.layers.18.self_attn.o_proj.weight": "model-00019-of-00061.safetensors", + "model.layers.18.mlp.gate_proj.weight": "model-00019-of-00061.safetensors", + "model.layers.18.mlp.down_proj.weight": "model-00019-of-00061.safetensors", + "model.layers.18.mlp.up_proj.weight": "model-00019-of-00061.safetensors", + "model.layers.18.self_attn.rotary_emb.inv_freq": "model-00019-of-00061.safetensors", + "model.layers.19.input_layernorm.weight": "model-00020-of-00061.safetensors", + "model.layers.19.post_attention_layernorm.weight": "model-00020-of-00061.safetensors", + "model.layers.19.self_attn.q_proj.weight": "model-00020-of-00061.safetensors", + "model.layers.19.self_attn.k_proj.weight": "model-00020-of-00061.safetensors", + "model.layers.19.self_attn.v_proj.weight": "model-00020-of-00061.safetensors", + "model.layers.19.self_attn.o_proj.weight": "model-00020-of-00061.safetensors", + "model.layers.19.mlp.gate_proj.weight": "model-00020-of-00061.safetensors", + "model.layers.19.mlp.down_proj.weight": "model-00020-of-00061.safetensors", + "model.layers.19.mlp.up_proj.weight": "model-00020-of-00061.safetensors", + "model.layers.19.self_attn.rotary_emb.inv_freq": "model-00020-of-00061.safetensors", + "model.layers.20.input_layernorm.weight": "model-00021-of-00061.safetensors", + "model.layers.20.post_attention_layernorm.weight": "model-00021-of-00061.safetensors", + "model.layers.20.self_attn.q_proj.weight": "model-00021-of-00061.safetensors", + "model.layers.20.self_attn.k_proj.weight": "model-00021-of-00061.safetensors", + "model.layers.20.self_attn.v_proj.weight": "model-00021-of-00061.safetensors", + "model.layers.20.self_attn.o_proj.weight": "model-00021-of-00061.safetensors", + "model.layers.20.mlp.gate_proj.weight": "model-00021-of-00061.safetensors", + "model.layers.20.mlp.down_proj.weight": "model-00021-of-00061.safetensors", + "model.layers.20.mlp.up_proj.weight": "model-00021-of-00061.safetensors", + "model.layers.20.self_attn.rotary_emb.inv_freq": "model-00021-of-00061.safetensors", + "model.layers.21.input_layernorm.weight": "model-00022-of-00061.safetensors", + "model.layers.21.post_attention_layernorm.weight": "model-00022-of-00061.safetensors", + "model.layers.21.self_attn.q_proj.weight": "model-00022-of-00061.safetensors", + "model.layers.21.self_attn.k_proj.weight": "model-00022-of-00061.safetensors", + "model.layers.21.self_attn.v_proj.weight": "model-00022-of-00061.safetensors", + "model.layers.21.self_attn.o_proj.weight": "model-00022-of-00061.safetensors", + "model.layers.21.mlp.gate_proj.weight": "model-00022-of-00061.safetensors", + "model.layers.21.mlp.down_proj.weight": "model-00022-of-00061.safetensors", + "model.layers.21.mlp.up_proj.weight": "model-00022-of-00061.safetensors", + "model.layers.21.self_attn.rotary_emb.inv_freq": "model-00022-of-00061.safetensors", + "model.layers.22.input_layernorm.weight": "model-00023-of-00061.safetensors", + "model.layers.22.post_attention_layernorm.weight": "model-00023-of-00061.safetensors", + "model.layers.22.self_attn.q_proj.weight": "model-00023-of-00061.safetensors", + "model.layers.22.self_attn.k_proj.weight": "model-00023-of-00061.safetensors", + "model.layers.22.self_attn.v_proj.weight": "model-00023-of-00061.safetensors", + "model.layers.22.self_attn.o_proj.weight": "model-00023-of-00061.safetensors", + "model.layers.22.mlp.gate_proj.weight": "model-00023-of-00061.safetensors", + "model.layers.22.mlp.down_proj.weight": "model-00023-of-00061.safetensors", + "model.layers.22.mlp.up_proj.weight": "model-00023-of-00061.safetensors", + "model.layers.22.self_attn.rotary_emb.inv_freq": "model-00023-of-00061.safetensors", + "model.layers.23.input_layernorm.weight": "model-00024-of-00061.safetensors", + "model.layers.23.post_attention_layernorm.weight": "model-00024-of-00061.safetensors", + "model.layers.23.self_attn.q_proj.weight": "model-00024-of-00061.safetensors", + "model.layers.23.self_attn.k_proj.weight": "model-00024-of-00061.safetensors", + "model.layers.23.self_attn.v_proj.weight": "model-00024-of-00061.safetensors", + "model.layers.23.self_attn.o_proj.weight": "model-00024-of-00061.safetensors", + "model.layers.23.mlp.gate_proj.weight": "model-00024-of-00061.safetensors", + "model.layers.23.mlp.down_proj.weight": "model-00024-of-00061.safetensors", + "model.layers.23.mlp.up_proj.weight": "model-00024-of-00061.safetensors", + "model.layers.23.self_attn.rotary_emb.inv_freq": "model-00024-of-00061.safetensors", + "model.layers.24.input_layernorm.weight": "model-00025-of-00061.safetensors", + "model.layers.24.post_attention_layernorm.weight": "model-00025-of-00061.safetensors", + "model.layers.24.self_attn.q_proj.weight": "model-00025-of-00061.safetensors", + "model.layers.24.self_attn.k_proj.weight": "model-00025-of-00061.safetensors", + "model.layers.24.self_attn.v_proj.weight": "model-00025-of-00061.safetensors", + "model.layers.24.self_attn.o_proj.weight": "model-00025-of-00061.safetensors", + "model.layers.24.mlp.gate_proj.weight": "model-00025-of-00061.safetensors", + "model.layers.24.mlp.down_proj.weight": "model-00025-of-00061.safetensors", + "model.layers.24.mlp.up_proj.weight": "model-00025-of-00061.safetensors", + "model.layers.24.self_attn.rotary_emb.inv_freq": "model-00025-of-00061.safetensors", + "model.layers.25.input_layernorm.weight": "model-00026-of-00061.safetensors", + "model.layers.25.post_attention_layernorm.weight": "model-00026-of-00061.safetensors", + "model.layers.25.self_attn.q_proj.weight": "model-00026-of-00061.safetensors", + "model.layers.25.self_attn.k_proj.weight": "model-00026-of-00061.safetensors", + "model.layers.25.self_attn.v_proj.weight": "model-00026-of-00061.safetensors", + "model.layers.25.self_attn.o_proj.weight": "model-00026-of-00061.safetensors", + "model.layers.25.mlp.gate_proj.weight": "model-00026-of-00061.safetensors", + "model.layers.25.mlp.down_proj.weight": "model-00026-of-00061.safetensors", + "model.layers.25.mlp.up_proj.weight": "model-00026-of-00061.safetensors", + "model.layers.25.self_attn.rotary_emb.inv_freq": "model-00026-of-00061.safetensors", + "model.layers.26.input_layernorm.weight": "model-00027-of-00061.safetensors", + "model.layers.26.post_attention_layernorm.weight": "model-00027-of-00061.safetensors", + "model.layers.26.self_attn.q_proj.weight": "model-00027-of-00061.safetensors", + "model.layers.26.self_attn.k_proj.weight": "model-00027-of-00061.safetensors", + "model.layers.26.self_attn.v_proj.weight": "model-00027-of-00061.safetensors", + "model.layers.26.self_attn.o_proj.weight": "model-00027-of-00061.safetensors", + "model.layers.26.mlp.gate_proj.weight": "model-00027-of-00061.safetensors", + "model.layers.26.mlp.down_proj.weight": "model-00027-of-00061.safetensors", + "model.layers.26.mlp.up_proj.weight": "model-00027-of-00061.safetensors", + "model.layers.26.self_attn.rotary_emb.inv_freq": "model-00027-of-00061.safetensors", + "model.layers.27.input_layernorm.weight": "model-00028-of-00061.safetensors", + "model.layers.27.post_attention_layernorm.weight": "model-00028-of-00061.safetensors", + "model.layers.27.self_attn.q_proj.weight": "model-00028-of-00061.safetensors", + "model.layers.27.self_attn.k_proj.weight": "model-00028-of-00061.safetensors", + "model.layers.27.self_attn.v_proj.weight": "model-00028-of-00061.safetensors", + "model.layers.27.self_attn.o_proj.weight": "model-00028-of-00061.safetensors", + "model.layers.27.mlp.gate_proj.weight": "model-00028-of-00061.safetensors", + "model.layers.27.mlp.down_proj.weight": "model-00028-of-00061.safetensors", + "model.layers.27.mlp.up_proj.weight": "model-00028-of-00061.safetensors", + "model.layers.27.self_attn.rotary_emb.inv_freq": "model-00028-of-00061.safetensors", + "model.layers.28.input_layernorm.weight": "model-00029-of-00061.safetensors", + "model.layers.28.post_attention_layernorm.weight": "model-00029-of-00061.safetensors", + "model.layers.28.self_attn.q_proj.weight": "model-00029-of-00061.safetensors", + "model.layers.28.self_attn.k_proj.weight": "model-00029-of-00061.safetensors", + "model.layers.28.self_attn.v_proj.weight": "model-00029-of-00061.safetensors", + "model.layers.28.self_attn.o_proj.weight": "model-00029-of-00061.safetensors", + "model.layers.28.mlp.gate_proj.weight": "model-00029-of-00061.safetensors", + "model.layers.28.mlp.down_proj.weight": "model-00029-of-00061.safetensors", + "model.layers.28.mlp.up_proj.weight": "model-00029-of-00061.safetensors", + "model.layers.28.self_attn.rotary_emb.inv_freq": "model-00029-of-00061.safetensors", + "model.layers.29.input_layernorm.weight": "model-00030-of-00061.safetensors", + "model.layers.29.post_attention_layernorm.weight": "model-00030-of-00061.safetensors", + "model.layers.29.self_attn.q_proj.weight": "model-00030-of-00061.safetensors", + "model.layers.29.self_attn.k_proj.weight": "model-00030-of-00061.safetensors", + "model.layers.29.self_attn.v_proj.weight": "model-00030-of-00061.safetensors", + "model.layers.29.self_attn.o_proj.weight": "model-00030-of-00061.safetensors", + "model.layers.29.mlp.gate_proj.weight": "model-00030-of-00061.safetensors", + "model.layers.29.mlp.down_proj.weight": "model-00030-of-00061.safetensors", + "model.layers.29.mlp.up_proj.weight": "model-00030-of-00061.safetensors", + "model.layers.29.self_attn.rotary_emb.inv_freq": "model-00030-of-00061.safetensors", + "model.layers.30.input_layernorm.weight": "model-00031-of-00061.safetensors", + "model.layers.30.post_attention_layernorm.weight": "model-00031-of-00061.safetensors", + "model.layers.30.self_attn.q_proj.weight": "model-00031-of-00061.safetensors", + "model.layers.30.self_attn.k_proj.weight": "model-00031-of-00061.safetensors", + "model.layers.30.self_attn.v_proj.weight": "model-00031-of-00061.safetensors", + "model.layers.30.self_attn.o_proj.weight": "model-00031-of-00061.safetensors", + "model.layers.30.mlp.gate_proj.weight": "model-00031-of-00061.safetensors", + "model.layers.30.mlp.down_proj.weight": "model-00031-of-00061.safetensors", + "model.layers.30.mlp.up_proj.weight": "model-00031-of-00061.safetensors", + "model.layers.30.self_attn.rotary_emb.inv_freq": "model-00031-of-00061.safetensors", + "model.layers.31.input_layernorm.weight": "model-00032-of-00061.safetensors", + "model.layers.31.post_attention_layernorm.weight": "model-00032-of-00061.safetensors", + "model.layers.31.self_attn.q_proj.weight": "model-00032-of-00061.safetensors", + "model.layers.31.self_attn.k_proj.weight": "model-00032-of-00061.safetensors", + "model.layers.31.self_attn.v_proj.weight": "model-00032-of-00061.safetensors", + "model.layers.31.self_attn.o_proj.weight": "model-00032-of-00061.safetensors", + "model.layers.31.mlp.gate_proj.weight": "model-00032-of-00061.safetensors", + "model.layers.31.mlp.down_proj.weight": "model-00032-of-00061.safetensors", + "model.layers.31.mlp.up_proj.weight": "model-00032-of-00061.safetensors", + "model.layers.31.self_attn.rotary_emb.inv_freq": "model-00032-of-00061.safetensors", + "model.layers.32.input_layernorm.weight": "model-00033-of-00061.safetensors", + "model.layers.32.post_attention_layernorm.weight": "model-00033-of-00061.safetensors", + "model.layers.32.self_attn.q_proj.weight": "model-00033-of-00061.safetensors", + "model.layers.32.self_attn.k_proj.weight": "model-00033-of-00061.safetensors", + "model.layers.32.self_attn.v_proj.weight": "model-00033-of-00061.safetensors", + "model.layers.32.self_attn.o_proj.weight": "model-00033-of-00061.safetensors", + "model.layers.32.mlp.gate_proj.weight": "model-00033-of-00061.safetensors", + "model.layers.32.mlp.down_proj.weight": "model-00033-of-00061.safetensors", + "model.layers.32.mlp.up_proj.weight": "model-00033-of-00061.safetensors", + "model.layers.32.self_attn.rotary_emb.inv_freq": "model-00033-of-00061.safetensors", + "model.layers.33.input_layernorm.weight": "model-00034-of-00061.safetensors", + "model.layers.33.post_attention_layernorm.weight": "model-00034-of-00061.safetensors", + "model.layers.33.self_attn.q_proj.weight": "model-00034-of-00061.safetensors", + "model.layers.33.self_attn.k_proj.weight": "model-00034-of-00061.safetensors", + "model.layers.33.self_attn.v_proj.weight": "model-00034-of-00061.safetensors", + "model.layers.33.self_attn.o_proj.weight": "model-00034-of-00061.safetensors", + "model.layers.33.mlp.gate_proj.weight": "model-00034-of-00061.safetensors", + "model.layers.33.mlp.down_proj.weight": "model-00034-of-00061.safetensors", + "model.layers.33.mlp.up_proj.weight": "model-00034-of-00061.safetensors", + "model.layers.33.self_attn.rotary_emb.inv_freq": "model-00034-of-00061.safetensors", + "model.layers.34.input_layernorm.weight": "model-00035-of-00061.safetensors", + "model.layers.34.post_attention_layernorm.weight": "model-00035-of-00061.safetensors", + "model.layers.34.self_attn.q_proj.weight": "model-00035-of-00061.safetensors", + "model.layers.34.self_attn.k_proj.weight": "model-00035-of-00061.safetensors", + "model.layers.34.self_attn.v_proj.weight": "model-00035-of-00061.safetensors", + "model.layers.34.self_attn.o_proj.weight": "model-00035-of-00061.safetensors", + "model.layers.34.mlp.gate_proj.weight": "model-00035-of-00061.safetensors", + "model.layers.34.mlp.down_proj.weight": "model-00035-of-00061.safetensors", + "model.layers.34.mlp.up_proj.weight": "model-00035-of-00061.safetensors", + "model.layers.34.self_attn.rotary_emb.inv_freq": "model-00035-of-00061.safetensors", + "model.layers.35.input_layernorm.weight": "model-00036-of-00061.safetensors", + "model.layers.35.post_attention_layernorm.weight": "model-00036-of-00061.safetensors", + "model.layers.35.self_attn.q_proj.weight": "model-00036-of-00061.safetensors", + "model.layers.35.self_attn.k_proj.weight": "model-00036-of-00061.safetensors", + "model.layers.35.self_attn.v_proj.weight": "model-00036-of-00061.safetensors", + "model.layers.35.self_attn.o_proj.weight": "model-00036-of-00061.safetensors", + "model.layers.35.mlp.gate_proj.weight": "model-00036-of-00061.safetensors", + "model.layers.35.mlp.down_proj.weight": "model-00036-of-00061.safetensors", + "model.layers.35.mlp.up_proj.weight": "model-00036-of-00061.safetensors", + "model.layers.35.self_attn.rotary_emb.inv_freq": "model-00036-of-00061.safetensors", + "model.layers.36.input_layernorm.weight": "model-00037-of-00061.safetensors", + "model.layers.36.post_attention_layernorm.weight": "model-00037-of-00061.safetensors", + "model.layers.36.self_attn.q_proj.weight": "model-00037-of-00061.safetensors", + "model.layers.36.self_attn.k_proj.weight": "model-00037-of-00061.safetensors", + "model.layers.36.self_attn.v_proj.weight": "model-00037-of-00061.safetensors", + "model.layers.36.self_attn.o_proj.weight": "model-00037-of-00061.safetensors", + "model.layers.36.mlp.gate_proj.weight": "model-00037-of-00061.safetensors", + "model.layers.36.mlp.down_proj.weight": "model-00037-of-00061.safetensors", + "model.layers.36.mlp.up_proj.weight": "model-00037-of-00061.safetensors", + "model.layers.36.self_attn.rotary_emb.inv_freq": "model-00037-of-00061.safetensors", + "model.layers.37.input_layernorm.weight": "model-00038-of-00061.safetensors", + "model.layers.37.post_attention_layernorm.weight": "model-00038-of-00061.safetensors", + "model.layers.37.self_attn.q_proj.weight": "model-00038-of-00061.safetensors", + "model.layers.37.self_attn.k_proj.weight": "model-00038-of-00061.safetensors", + "model.layers.37.self_attn.v_proj.weight": "model-00038-of-00061.safetensors", + "model.layers.37.self_attn.o_proj.weight": "model-00038-of-00061.safetensors", + "model.layers.37.mlp.gate_proj.weight": "model-00038-of-00061.safetensors", + "model.layers.37.mlp.down_proj.weight": "model-00038-of-00061.safetensors", + "model.layers.37.mlp.up_proj.weight": "model-00038-of-00061.safetensors", + "model.layers.37.self_attn.rotary_emb.inv_freq": "model-00038-of-00061.safetensors", + "model.layers.38.input_layernorm.weight": "model-00039-of-00061.safetensors", + "model.layers.38.post_attention_layernorm.weight": "model-00039-of-00061.safetensors", + "model.layers.38.self_attn.q_proj.weight": "model-00039-of-00061.safetensors", + "model.layers.38.self_attn.k_proj.weight": "model-00039-of-00061.safetensors", + "model.layers.38.self_attn.v_proj.weight": "model-00039-of-00061.safetensors", + "model.layers.38.self_attn.o_proj.weight": "model-00039-of-00061.safetensors", + "model.layers.38.mlp.gate_proj.weight": "model-00039-of-00061.safetensors", + "model.layers.38.mlp.down_proj.weight": "model-00039-of-00061.safetensors", + "model.layers.38.mlp.up_proj.weight": "model-00039-of-00061.safetensors", + "model.layers.38.self_attn.rotary_emb.inv_freq": "model-00039-of-00061.safetensors", + "model.layers.39.input_layernorm.weight": "model-00040-of-00061.safetensors", + "model.layers.39.post_attention_layernorm.weight": "model-00040-of-00061.safetensors", + "model.layers.39.self_attn.q_proj.weight": "model-00040-of-00061.safetensors", + "model.layers.39.self_attn.k_proj.weight": "model-00040-of-00061.safetensors", + "model.layers.39.self_attn.v_proj.weight": "model-00040-of-00061.safetensors", + "model.layers.39.self_attn.o_proj.weight": "model-00040-of-00061.safetensors", + "model.layers.39.mlp.gate_proj.weight": "model-00040-of-00061.safetensors", + "model.layers.39.mlp.down_proj.weight": "model-00040-of-00061.safetensors", + "model.layers.39.mlp.up_proj.weight": "model-00040-of-00061.safetensors", + "model.layers.39.self_attn.rotary_emb.inv_freq": "model-00040-of-00061.safetensors", + "model.layers.40.input_layernorm.weight": "model-00041-of-00061.safetensors", + "model.layers.40.post_attention_layernorm.weight": "model-00041-of-00061.safetensors", + "model.layers.40.self_attn.q_proj.weight": "model-00041-of-00061.safetensors", + "model.layers.40.self_attn.k_proj.weight": "model-00041-of-00061.safetensors", + "model.layers.40.self_attn.v_proj.weight": "model-00041-of-00061.safetensors", + "model.layers.40.self_attn.o_proj.weight": "model-00041-of-00061.safetensors", + "model.layers.40.mlp.gate_proj.weight": "model-00041-of-00061.safetensors", + "model.layers.40.mlp.down_proj.weight": "model-00041-of-00061.safetensors", + "model.layers.40.mlp.up_proj.weight": "model-00041-of-00061.safetensors", + "model.layers.40.self_attn.rotary_emb.inv_freq": "model-00041-of-00061.safetensors", + "model.layers.41.input_layernorm.weight": "model-00042-of-00061.safetensors", + "model.layers.41.post_attention_layernorm.weight": "model-00042-of-00061.safetensors", + "model.layers.41.self_attn.q_proj.weight": "model-00042-of-00061.safetensors", + "model.layers.41.self_attn.k_proj.weight": "model-00042-of-00061.safetensors", + "model.layers.41.self_attn.v_proj.weight": "model-00042-of-00061.safetensors", + "model.layers.41.self_attn.o_proj.weight": "model-00042-of-00061.safetensors", + "model.layers.41.mlp.gate_proj.weight": "model-00042-of-00061.safetensors", + "model.layers.41.mlp.down_proj.weight": "model-00042-of-00061.safetensors", + "model.layers.41.mlp.up_proj.weight": "model-00042-of-00061.safetensors", + "model.layers.41.self_attn.rotary_emb.inv_freq": "model-00042-of-00061.safetensors", + "model.layers.42.input_layernorm.weight": "model-00043-of-00061.safetensors", + "model.layers.42.post_attention_layernorm.weight": "model-00043-of-00061.safetensors", + "model.layers.42.self_attn.q_proj.weight": "model-00043-of-00061.safetensors", + "model.layers.42.self_attn.k_proj.weight": "model-00043-of-00061.safetensors", + "model.layers.42.self_attn.v_proj.weight": "model-00043-of-00061.safetensors", + "model.layers.42.self_attn.o_proj.weight": "model-00043-of-00061.safetensors", + "model.layers.42.mlp.gate_proj.weight": "model-00043-of-00061.safetensors", + "model.layers.42.mlp.down_proj.weight": "model-00043-of-00061.safetensors", + "model.layers.42.mlp.up_proj.weight": "model-00043-of-00061.safetensors", + "model.layers.42.self_attn.rotary_emb.inv_freq": "model-00043-of-00061.safetensors", + "model.layers.43.input_layernorm.weight": "model-00044-of-00061.safetensors", + "model.layers.43.post_attention_layernorm.weight": "model-00044-of-00061.safetensors", + "model.layers.43.self_attn.q_proj.weight": "model-00044-of-00061.safetensors", + "model.layers.43.self_attn.k_proj.weight": "model-00044-of-00061.safetensors", + "model.layers.43.self_attn.v_proj.weight": "model-00044-of-00061.safetensors", + "model.layers.43.self_attn.o_proj.weight": "model-00044-of-00061.safetensors", + "model.layers.43.mlp.gate_proj.weight": "model-00044-of-00061.safetensors", + "model.layers.43.mlp.down_proj.weight": "model-00044-of-00061.safetensors", + "model.layers.43.mlp.up_proj.weight": "model-00044-of-00061.safetensors", + "model.layers.43.self_attn.rotary_emb.inv_freq": "model-00044-of-00061.safetensors", + "model.layers.44.input_layernorm.weight": "model-00045-of-00061.safetensors", + "model.layers.44.post_attention_layernorm.weight": "model-00045-of-00061.safetensors", + "model.layers.44.self_attn.q_proj.weight": "model-00045-of-00061.safetensors", + "model.layers.44.self_attn.k_proj.weight": "model-00045-of-00061.safetensors", + "model.layers.44.self_attn.v_proj.weight": "model-00045-of-00061.safetensors", + "model.layers.44.self_attn.o_proj.weight": "model-00045-of-00061.safetensors", + "model.layers.44.mlp.gate_proj.weight": "model-00045-of-00061.safetensors", + "model.layers.44.mlp.down_proj.weight": "model-00045-of-00061.safetensors", + "model.layers.44.mlp.up_proj.weight": "model-00045-of-00061.safetensors", + "model.layers.44.self_attn.rotary_emb.inv_freq": "model-00045-of-00061.safetensors", + "model.layers.45.input_layernorm.weight": "model-00046-of-00061.safetensors", + "model.layers.45.post_attention_layernorm.weight": "model-00046-of-00061.safetensors", + "model.layers.45.self_attn.q_proj.weight": "model-00046-of-00061.safetensors", + "model.layers.45.self_attn.k_proj.weight": "model-00046-of-00061.safetensors", + "model.layers.45.self_attn.v_proj.weight": "model-00046-of-00061.safetensors", + "model.layers.45.self_attn.o_proj.weight": "model-00046-of-00061.safetensors", + "model.layers.45.mlp.gate_proj.weight": "model-00046-of-00061.safetensors", + "model.layers.45.mlp.down_proj.weight": "model-00046-of-00061.safetensors", + "model.layers.45.mlp.up_proj.weight": "model-00046-of-00061.safetensors", + "model.layers.45.self_attn.rotary_emb.inv_freq": "model-00046-of-00061.safetensors", + "model.layers.46.input_layernorm.weight": "model-00047-of-00061.safetensors", + "model.layers.46.post_attention_layernorm.weight": "model-00047-of-00061.safetensors", + "model.layers.46.self_attn.q_proj.weight": "model-00047-of-00061.safetensors", + "model.layers.46.self_attn.k_proj.weight": "model-00047-of-00061.safetensors", + "model.layers.46.self_attn.v_proj.weight": "model-00047-of-00061.safetensors", + "model.layers.46.self_attn.o_proj.weight": "model-00047-of-00061.safetensors", + "model.layers.46.mlp.gate_proj.weight": "model-00047-of-00061.safetensors", + "model.layers.46.mlp.down_proj.weight": "model-00047-of-00061.safetensors", + "model.layers.46.mlp.up_proj.weight": "model-00047-of-00061.safetensors", + "model.layers.46.self_attn.rotary_emb.inv_freq": "model-00047-of-00061.safetensors", + "model.layers.47.input_layernorm.weight": "model-00048-of-00061.safetensors", + "model.layers.47.post_attention_layernorm.weight": "model-00048-of-00061.safetensors", + "model.layers.47.self_attn.q_proj.weight": "model-00048-of-00061.safetensors", + "model.layers.47.self_attn.k_proj.weight": "model-00048-of-00061.safetensors", + "model.layers.47.self_attn.v_proj.weight": "model-00048-of-00061.safetensors", + "model.layers.47.self_attn.o_proj.weight": "model-00048-of-00061.safetensors", + "model.layers.47.mlp.gate_proj.weight": "model-00048-of-00061.safetensors", + "model.layers.47.mlp.down_proj.weight": "model-00048-of-00061.safetensors", + "model.layers.47.mlp.up_proj.weight": "model-00048-of-00061.safetensors", + "model.layers.47.self_attn.rotary_emb.inv_freq": "model-00048-of-00061.safetensors", + "model.layers.48.input_layernorm.weight": "model-00049-of-00061.safetensors", + "model.layers.48.post_attention_layernorm.weight": "model-00049-of-00061.safetensors", + "model.layers.48.self_attn.q_proj.weight": "model-00049-of-00061.safetensors", + "model.layers.48.self_attn.k_proj.weight": "model-00049-of-00061.safetensors", + "model.layers.48.self_attn.v_proj.weight": "model-00049-of-00061.safetensors", + "model.layers.48.self_attn.o_proj.weight": "model-00049-of-00061.safetensors", + "model.layers.48.mlp.gate_proj.weight": "model-00049-of-00061.safetensors", + "model.layers.48.mlp.down_proj.weight": "model-00049-of-00061.safetensors", + "model.layers.48.mlp.up_proj.weight": "model-00049-of-00061.safetensors", + "model.layers.48.self_attn.rotary_emb.inv_freq": "model-00049-of-00061.safetensors", + "model.layers.49.input_layernorm.weight": "model-00050-of-00061.safetensors", + "model.layers.49.post_attention_layernorm.weight": "model-00050-of-00061.safetensors", + "model.layers.49.self_attn.q_proj.weight": "model-00050-of-00061.safetensors", + "model.layers.49.self_attn.k_proj.weight": "model-00050-of-00061.safetensors", + "model.layers.49.self_attn.v_proj.weight": "model-00050-of-00061.safetensors", + "model.layers.49.self_attn.o_proj.weight": "model-00050-of-00061.safetensors", + "model.layers.49.mlp.gate_proj.weight": "model-00050-of-00061.safetensors", + "model.layers.49.mlp.down_proj.weight": "model-00050-of-00061.safetensors", + "model.layers.49.mlp.up_proj.weight": "model-00050-of-00061.safetensors", + "model.layers.49.self_attn.rotary_emb.inv_freq": "model-00050-of-00061.safetensors", + "model.layers.50.input_layernorm.weight": "model-00051-of-00061.safetensors", + "model.layers.50.post_attention_layernorm.weight": "model-00051-of-00061.safetensors", + "model.layers.50.self_attn.q_proj.weight": "model-00051-of-00061.safetensors", + "model.layers.50.self_attn.k_proj.weight": "model-00051-of-00061.safetensors", + "model.layers.50.self_attn.v_proj.weight": "model-00051-of-00061.safetensors", + "model.layers.50.self_attn.o_proj.weight": "model-00051-of-00061.safetensors", + "model.layers.50.mlp.gate_proj.weight": "model-00051-of-00061.safetensors", + "model.layers.50.mlp.down_proj.weight": "model-00051-of-00061.safetensors", + "model.layers.50.mlp.up_proj.weight": "model-00051-of-00061.safetensors", + "model.layers.50.self_attn.rotary_emb.inv_freq": "model-00051-of-00061.safetensors", + "model.layers.51.input_layernorm.weight": "model-00052-of-00061.safetensors", + "model.layers.51.post_attention_layernorm.weight": "model-00052-of-00061.safetensors", + "model.layers.51.self_attn.q_proj.weight": "model-00052-of-00061.safetensors", + "model.layers.51.self_attn.k_proj.weight": "model-00052-of-00061.safetensors", + "model.layers.51.self_attn.v_proj.weight": "model-00052-of-00061.safetensors", + "model.layers.51.self_attn.o_proj.weight": "model-00052-of-00061.safetensors", + "model.layers.51.mlp.gate_proj.weight": "model-00052-of-00061.safetensors", + "model.layers.51.mlp.down_proj.weight": "model-00052-of-00061.safetensors", + "model.layers.51.mlp.up_proj.weight": "model-00052-of-00061.safetensors", + "model.layers.51.self_attn.rotary_emb.inv_freq": "model-00052-of-00061.safetensors", + "model.layers.52.input_layernorm.weight": "model-00053-of-00061.safetensors", + "model.layers.52.post_attention_layernorm.weight": "model-00053-of-00061.safetensors", + "model.layers.52.self_attn.q_proj.weight": "model-00053-of-00061.safetensors", + "model.layers.52.self_attn.k_proj.weight": "model-00053-of-00061.safetensors", + "model.layers.52.self_attn.v_proj.weight": "model-00053-of-00061.safetensors", + "model.layers.52.self_attn.o_proj.weight": "model-00053-of-00061.safetensors", + "model.layers.52.mlp.gate_proj.weight": "model-00053-of-00061.safetensors", + "model.layers.52.mlp.down_proj.weight": "model-00053-of-00061.safetensors", + "model.layers.52.mlp.up_proj.weight": "model-00053-of-00061.safetensors", + "model.layers.52.self_attn.rotary_emb.inv_freq": "model-00053-of-00061.safetensors", + "model.layers.53.input_layernorm.weight": "model-00054-of-00061.safetensors", + "model.layers.53.post_attention_layernorm.weight": "model-00054-of-00061.safetensors", + "model.layers.53.self_attn.q_proj.weight": "model-00054-of-00061.safetensors", + "model.layers.53.self_attn.k_proj.weight": "model-00054-of-00061.safetensors", + "model.layers.53.self_attn.v_proj.weight": "model-00054-of-00061.safetensors", + "model.layers.53.self_attn.o_proj.weight": "model-00054-of-00061.safetensors", + "model.layers.53.mlp.gate_proj.weight": "model-00054-of-00061.safetensors", + "model.layers.53.mlp.down_proj.weight": "model-00054-of-00061.safetensors", + "model.layers.53.mlp.up_proj.weight": "model-00054-of-00061.safetensors", + "model.layers.53.self_attn.rotary_emb.inv_freq": "model-00054-of-00061.safetensors", + "model.layers.54.input_layernorm.weight": "model-00055-of-00061.safetensors", + "model.layers.54.post_attention_layernorm.weight": "model-00055-of-00061.safetensors", + "model.layers.54.self_attn.q_proj.weight": "model-00055-of-00061.safetensors", + "model.layers.54.self_attn.k_proj.weight": "model-00055-of-00061.safetensors", + "model.layers.54.self_attn.v_proj.weight": "model-00055-of-00061.safetensors", + "model.layers.54.self_attn.o_proj.weight": "model-00055-of-00061.safetensors", + "model.layers.54.mlp.gate_proj.weight": "model-00055-of-00061.safetensors", + "model.layers.54.mlp.down_proj.weight": "model-00055-of-00061.safetensors", + "model.layers.54.mlp.up_proj.weight": "model-00055-of-00061.safetensors", + "model.layers.54.self_attn.rotary_emb.inv_freq": "model-00055-of-00061.safetensors", + "model.layers.55.input_layernorm.weight": "model-00056-of-00061.safetensors", + "model.layers.55.post_attention_layernorm.weight": "model-00056-of-00061.safetensors", + "model.layers.55.self_attn.q_proj.weight": "model-00056-of-00061.safetensors", + "model.layers.55.self_attn.k_proj.weight": "model-00056-of-00061.safetensors", + "model.layers.55.self_attn.v_proj.weight": "model-00056-of-00061.safetensors", + "model.layers.55.self_attn.o_proj.weight": "model-00056-of-00061.safetensors", + "model.layers.55.mlp.gate_proj.weight": "model-00056-of-00061.safetensors", + "model.layers.55.mlp.down_proj.weight": "model-00056-of-00061.safetensors", + "model.layers.55.mlp.up_proj.weight": "model-00056-of-00061.safetensors", + "model.layers.55.self_attn.rotary_emb.inv_freq": "model-00056-of-00061.safetensors", + "model.layers.56.input_layernorm.weight": "model-00057-of-00061.safetensors", + "model.layers.56.post_attention_layernorm.weight": "model-00057-of-00061.safetensors", + "model.layers.56.self_attn.q_proj.weight": "model-00057-of-00061.safetensors", + "model.layers.56.self_attn.k_proj.weight": "model-00057-of-00061.safetensors", + "model.layers.56.self_attn.v_proj.weight": "model-00057-of-00061.safetensors", + "model.layers.56.self_attn.o_proj.weight": "model-00057-of-00061.safetensors", + "model.layers.56.mlp.gate_proj.weight": "model-00057-of-00061.safetensors", + "model.layers.56.mlp.down_proj.weight": "model-00057-of-00061.safetensors", + "model.layers.56.mlp.up_proj.weight": "model-00057-of-00061.safetensors", + "model.layers.56.self_attn.rotary_emb.inv_freq": "model-00057-of-00061.safetensors", + "model.layers.57.input_layernorm.weight": "model-00058-of-00061.safetensors", + "model.layers.57.post_attention_layernorm.weight": "model-00058-of-00061.safetensors", + "model.layers.57.self_attn.q_proj.weight": "model-00058-of-00061.safetensors", + "model.layers.57.self_attn.k_proj.weight": "model-00058-of-00061.safetensors", + "model.layers.57.self_attn.v_proj.weight": "model-00058-of-00061.safetensors", + "model.layers.57.self_attn.o_proj.weight": "model-00058-of-00061.safetensors", + "model.layers.57.mlp.gate_proj.weight": "model-00058-of-00061.safetensors", + "model.layers.57.mlp.down_proj.weight": "model-00058-of-00061.safetensors", + "model.layers.57.mlp.up_proj.weight": "model-00058-of-00061.safetensors", + "model.layers.57.self_attn.rotary_emb.inv_freq": "model-00058-of-00061.safetensors", + "model.layers.58.input_layernorm.weight": "model-00059-of-00061.safetensors", + "model.layers.58.post_attention_layernorm.weight": "model-00059-of-00061.safetensors", + "model.layers.58.self_attn.q_proj.weight": "model-00059-of-00061.safetensors", + "model.layers.58.self_attn.k_proj.weight": "model-00059-of-00061.safetensors", + "model.layers.58.self_attn.v_proj.weight": "model-00059-of-00061.safetensors", + "model.layers.58.self_attn.o_proj.weight": "model-00059-of-00061.safetensors", + "model.layers.58.mlp.gate_proj.weight": "model-00059-of-00061.safetensors", + "model.layers.58.mlp.down_proj.weight": "model-00059-of-00061.safetensors", + "model.layers.58.mlp.up_proj.weight": "model-00059-of-00061.safetensors", + "model.layers.58.self_attn.rotary_emb.inv_freq": "model-00059-of-00061.safetensors", + "model.layers.59.input_layernorm.weight": "model-00060-of-00061.safetensors", + "model.layers.59.post_attention_layernorm.weight": "model-00060-of-00061.safetensors", + "model.layers.59.self_attn.q_proj.weight": "model-00060-of-00061.safetensors", + "model.layers.59.self_attn.k_proj.weight": "model-00060-of-00061.safetensors", + "model.layers.59.self_attn.v_proj.weight": "model-00060-of-00061.safetensors", + "model.layers.59.self_attn.o_proj.weight": "model-00060-of-00061.safetensors", + "model.layers.59.mlp.gate_proj.weight": "model-00060-of-00061.safetensors", + "model.layers.59.mlp.down_proj.weight": "model-00060-of-00061.safetensors", + "model.layers.59.mlp.up_proj.weight": "model-00060-of-00061.safetensors", + "model.layers.59.self_attn.rotary_emb.inv_freq": "model-00060-of-00061.safetensors", + "model.norm.weight": "model-00061-of-00061.safetensors", + "model.embed_tokens.weight": "model-00061-of-00061.safetensors", + "lm_head.weight": "model-00061-of-00061.safetensors" + }, + "metadata": { + "total_size": 65057894912 + } +} \ No newline at end of file