diff --git a/model-00001-of-00051.safetensors b/model-00001-of-00051.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..4ee5fa5f3c52bbb34a148c5117fc2179c5b1280a --- /dev/null +++ b/model-00001-of-00051.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:dacaebfbec943c266d87d97f5b88355cb223b4d97122cef595ea152da6abecc6 +size 4932552432 diff --git a/model-00002-of-00051.safetensors b/model-00002-of-00051.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..75a7586371459a9bc161200b99a9d10ca2793265 --- /dev/null +++ b/model-00002-of-00051.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:65d25ab077405e37a2a6f752c577084ca66346efdbe3d8917e9e4c94e146c4f5 +size 4831938528 diff --git a/model-00003-of-00051.safetensors b/model-00003-of-00051.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..76c48e9d3a0cc79e71b1d7ff7206392f93ec3311 --- /dev/null +++ b/model-00003-of-00051.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:cd98eb2714215176d9c65ae5546216bf31ef857a486fbc92899cfd43f6b8d371 +size 4882269816 diff --git a/model-00004-of-00051.safetensors b/model-00004-of-00051.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..823d2d13adda54b273447b2869111fbe8843a1d5 --- /dev/null +++ b/model-00004-of-00051.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:0429d8c8808627747ee842384df07e659d96507fbc72903c5a582fb0c8cf3390 +size 4831889136 diff --git a/model-00005-of-00051.safetensors b/model-00005-of-00051.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..2a4d1ff394e7253cae5f6f84382cdcf31048134b --- /dev/null +++ b/model-00005-of-00051.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:d5828666e7c4752a9134955d878fdd5e6e8d7bfc920cc9ed85dab65cb4ba2739 +size 4831938520 diff --git a/model-00006-of-00051.safetensors b/model-00006-of-00051.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..b52fb8dc09f2415b294857c70e87b0e37f070a56 --- /dev/null +++ b/model-00006-of-00051.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:db76a8adf7c4bc4a713467ef5aa7e196b9669cfa1ca995e4e21d7d1f5320c751 +size 4831938528 diff --git a/model-00007-of-00051.safetensors b/model-00007-of-00051.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..84be89d3d9d79c13cdc2205531fd0b209f6ac57e --- /dev/null +++ b/model-00007-of-00051.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:dccbc2e6de3234ccc49bf708041e47479c87782cd5c914ed9ac3a5522d397c4b +size 4882269832 diff --git a/model-00008-of-00051.safetensors b/model-00008-of-00051.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..0ad4b004664fd7730f73b4dbc283d6dd6b53b520 --- /dev/null +++ b/model-00008-of-00051.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:d60bea67a1204b2aa22541f6a75d83daa559d0bb90757ecd286a5dabd71dda4f +size 4831889152 diff --git a/model-00009-of-00051.safetensors b/model-00009-of-00051.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..c422c397d8fb540ecaa1e07868988daf07b5df9a --- /dev/null +++ b/model-00009-of-00051.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:3e7b0499073b1090bc63d3004c21206075efa17fdb261644ff6f8264307e8734 +size 4831938536 diff --git a/model-00010-of-00051.safetensors b/model-00010-of-00051.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..eb1279d94914124f5a9cc5555187562a684c5d3b --- /dev/null +++ b/model-00010-of-00051.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:d8781b3efb12420bd672071fcff977443d6cdc6b01f87078a6eec685d332e7f9 +size 4831938544 diff --git a/model-00011-of-00051.safetensors b/model-00011-of-00051.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..f477339dda73f51e1210c1df2fb74a377b7a6be3 --- /dev/null +++ b/model-00011-of-00051.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:ec93f1d8589f246ed6da6d65174b88779c35517e0073e45f7a0fb7ff37ae72fa +size 4882269832 diff --git a/model-00012-of-00051.safetensors b/model-00012-of-00051.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..7c25542b4dee63f54880b070ded926eb2c5862ca --- /dev/null +++ b/model-00012-of-00051.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:dfde546e9e8f27e935d29c13ce3f3aa20778372db34e21fb20c1dcf866c2fed5 +size 4831889152 diff --git a/model-00013-of-00051.safetensors b/model-00013-of-00051.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..cfabd5710f11c3c82f3141e27280acd274422d6c --- /dev/null +++ b/model-00013-of-00051.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:f2c0727117d66337635dc197cad68befdae7b2114c015e4d2b5680dae2702645 +size 4831938536 diff --git a/model-00014-of-00051.safetensors b/model-00014-of-00051.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..fbaab305f1ea1cdfb9c57634793d020b15726061 --- /dev/null +++ b/model-00014-of-00051.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:96e3f3fdc53c19963c2e4f9ba20be3c67ecbd73a11f683e85022a04c514f5a3c +size 4831938544 diff --git a/model-00015-of-00051.safetensors b/model-00015-of-00051.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..2ce73450e5f152d62f5b5391955b3d11db3fb56b --- /dev/null +++ b/model-00015-of-00051.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:e4ae9ddc880c6981d13f9a7b10c0f92baf4ad8f3f95cbd5c329e155b5e4b9565 +size 4882269832 diff --git a/model-00016-of-00051.safetensors b/model-00016-of-00051.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..bfdd0a469e8a5a81134c33e59120ef4306b66497 --- /dev/null +++ b/model-00016-of-00051.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:74dd469610b90e8daefd4cfe985702cbbb7b377df99970fd194ba289830eb68b +size 4831889152 diff --git a/model-00017-of-00051.safetensors b/model-00017-of-00051.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..fe0c341ce5b8ba6bf3f4ef3c5a2bfbbd5def7d59 --- /dev/null +++ b/model-00017-of-00051.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:e1c7ca5bd3046bd08b93c96bf22d792316ed3a4058cfcf071b4c11163b146187 +size 4831938536 diff --git a/model-00018-of-00051.safetensors b/model-00018-of-00051.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..ac094c79713c32ef910ac0b61c68e99c143ce0c7 --- /dev/null +++ b/model-00018-of-00051.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:30bf3d94624d63e1776888608460efcb428771d81e925748487562a2e9f3e6ac +size 4831938544 diff --git a/model-00019-of-00051.safetensors b/model-00019-of-00051.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..3f657df1b32bbf4e0b1d0d852440895bb0930ecb --- /dev/null +++ b/model-00019-of-00051.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:490d6e252ec85f1a3d8e017c37dfd71f80ea6c83494ee86b35043bbfef8d2bc4 +size 4882269832 diff --git a/model-00020-of-00051.safetensors b/model-00020-of-00051.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..aa4d93daddc6da0f1553da9af17d68e05296ed02 --- /dev/null +++ b/model-00020-of-00051.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:bc47e09a350386b302ae6b69a5dd3427bf3562d4090218a852f84374c838f54b +size 4831889152 diff --git a/model-00021-of-00051.safetensors b/model-00021-of-00051.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..b2ff1dd12b5d327783dbf3e7b6f30ca37902d1e2 --- /dev/null +++ b/model-00021-of-00051.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:5fb914963d54c0aeed8613c9145ded8a2880bc8a442e2e45d08a9225262e07c1 +size 4831938536 diff --git a/model-00022-of-00051.safetensors b/model-00022-of-00051.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..2d894082a96eb2c055ce2b5853fb3cd909d35e0f --- /dev/null +++ b/model-00022-of-00051.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:9a9a38c9c5f50f1a3b550018866b5d7559c9b01836b3b9518eef6972729b9770 +size 4831938544 diff --git a/model-00023-of-00051.safetensors b/model-00023-of-00051.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..8ddefbe4c4475b2c4940b29f7f327b8582286064 --- /dev/null +++ b/model-00023-of-00051.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:fddc5dbc01efea9f101f65891bc03f0deeb32795935235b8559a7bb9d3a296fa +size 4882269832 diff --git a/model-00024-of-00051.safetensors b/model-00024-of-00051.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..67dadbed2a7f3fbd9d0ec811cb154bd56dc16abd --- /dev/null +++ b/model-00024-of-00051.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:392fbae3cd2d9394a97ac42a0598c8d0899dfd078d72e99a7aaad93b9cbf9757 +size 4831889152 diff --git a/model-00025-of-00051.safetensors b/model-00025-of-00051.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..5acf47dcd56ee6fb9ebaebf90a8ffedd0ee1387c --- /dev/null +++ b/model-00025-of-00051.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:2e089cefb7827de7a5d1b1214bf7f2a3b1ea77272701122829ce4ca857ca2113 +size 4831938536 diff --git a/model-00026-of-00051.safetensors b/model-00026-of-00051.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..e73f7622d28783b1e67b40e409181f4e2570e167 --- /dev/null +++ b/model-00026-of-00051.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:1dcedb1a76e825d23951b77fd7969cc8282b5d2081b9389f65dce911d5b34560 +size 4831938544 diff --git a/model-00027-of-00051.safetensors b/model-00027-of-00051.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..5a345e6053334bb9e698b2c0f656c8a847dbedfa --- /dev/null +++ b/model-00027-of-00051.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:838737d590f37cf1acac9b66dde3bbdac3a62aef12cee9535ae951f8abc20021 +size 4882269832 diff --git a/model-00028-of-00051.safetensors b/model-00028-of-00051.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..342b0bdefaf92b2a41d585c446f7c16fddddf3ff --- /dev/null +++ b/model-00028-of-00051.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:256a7f413eee5414fa5a1ee186fd56f578344e3835700f4c7d6ea31afeb833e8 +size 4831889152 diff --git a/model-00029-of-00051.safetensors b/model-00029-of-00051.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..939918c0636d7a3e0d3e36c1697f08d101f25ac8 --- /dev/null +++ b/model-00029-of-00051.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:e32793990d7950023c52c9126d2cc3b6a773a3d7a96c08c335cee4771a669fc3 +size 4831938536 diff --git a/model-00030-of-00051.safetensors b/model-00030-of-00051.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..c4392a1dcb649e101ba0d7d291a109cc274a7836 --- /dev/null +++ b/model-00030-of-00051.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:da13f230e8999ff2200ec7c50f235fefac1bcaf6c20a8b9713c8b7089ce844b1 +size 4831938544 diff --git a/model-00031-of-00051.safetensors b/model-00031-of-00051.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..e8680aebb3973f5fc90757b9c8ed3d9fdca19723 --- /dev/null +++ b/model-00031-of-00051.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:c754e2336209118a4b562af938603ab2295ac26917022bd0111939ed885cd0af +size 4882269832 diff --git a/model-00032-of-00051.safetensors b/model-00032-of-00051.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..d6e78da0658e1348c98bc79a2e98556a46705556 --- /dev/null +++ b/model-00032-of-00051.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:6dc9cbd1a92d703f3924996cb16cea91ed86485fe8ce3183f7651bc6e97a23dc +size 4831889152 diff --git a/model-00033-of-00051.safetensors b/model-00033-of-00051.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..b99a5f2fbfac6ac874b41f2849fe088a4ec110e0 --- /dev/null +++ b/model-00033-of-00051.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:a3ea5ad14364966a79b7134feb2297176e522d79007924e630263d35efefa70d +size 4831938536 diff --git a/model-00034-of-00051.safetensors b/model-00034-of-00051.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..e6b4612c22766a0a55c0d9f924ffe8231a87c253 --- /dev/null +++ b/model-00034-of-00051.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:90702857e9e9f5db28a72908b20c3b5cfe7253c4ace0e160233617cfd5215c8d +size 4831938544 diff --git a/model-00035-of-00051.safetensors b/model-00035-of-00051.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..6eb061c0fb8df7e4c2e51dcd7c803dab2ab9d761 --- /dev/null +++ b/model-00035-of-00051.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:bf4179ca1c1445780e44a68b4f7b70984ff921982dafba727d1e669b63310fbf +size 4882269832 diff --git a/model-00036-of-00051.safetensors b/model-00036-of-00051.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..c2716d42195e7c5ff446b73329a62e211b4058d6 --- /dev/null +++ b/model-00036-of-00051.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:fd6c0f4241749dfdea9142038cfe5f3c770a02cd1edf77d197245c3cdc695730 +size 4831889152 diff --git a/model-00037-of-00051.safetensors b/model-00037-of-00051.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..1a48daf47f4fb5498697bc2a9cb9595656575bbe --- /dev/null +++ b/model-00037-of-00051.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:15f709510753847f72cfe473a39e5fe32517b52aacf4b7b20dbc97f96f37186b +size 4831938536 diff --git a/model-00038-of-00051.safetensors b/model-00038-of-00051.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..8fe89fe82c20e789c9fb8cb5870ee9fb84953116 --- /dev/null +++ b/model-00038-of-00051.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:c659230731b53ed2665f5916ac88c5b19ea49fb3769e663910ab8bca22373ead +size 4831938544 diff --git a/model-00039-of-00051.safetensors b/model-00039-of-00051.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..45d6c5cd197c7238f2e8e14b9648c545a675f3f1 --- /dev/null +++ b/model-00039-of-00051.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:a166d01af29d5ead30509b27ee30eede2fe67ce67c031a93c7f6b1a9a13bfe54 +size 4882269832 diff --git a/model-00040-of-00051.safetensors b/model-00040-of-00051.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..db25ddd3b13b35bf7243a5f1ddc19178086827c9 --- /dev/null +++ b/model-00040-of-00051.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:89a9a1132215070b30aa9f8889d94d24aa94c6cf9893465baf64de718f6ecfd8 +size 4831889152 diff --git a/model-00041-of-00051.safetensors b/model-00041-of-00051.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..49065be5d6034b082b1cd83ccaf4ac498a657c61 --- /dev/null +++ b/model-00041-of-00051.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:7571db93a5a9ee2337b211e6e333784b3f2808a53e0a1a261cdc78ec8dd58f51 +size 4831938536 diff --git a/model-00042-of-00051.safetensors b/model-00042-of-00051.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..92b15c7a52e86db66d4a52823497ab83bb4d99c7 --- /dev/null +++ b/model-00042-of-00051.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:39e2b7467ee49f7494c1c44fc96b57cb679d34d907d1ef031f2bf102a2b9c764 +size 4831938544 diff --git a/model-00043-of-00051.safetensors b/model-00043-of-00051.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..1af2cb87145761257a82c78b89e62582185ba1bd --- /dev/null +++ b/model-00043-of-00051.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:3fa9a842d76aa4c3ace327df9e6999810262d3cbccb95bee3159f855ed29d1ac +size 4882269832 diff --git a/model-00044-of-00051.safetensors b/model-00044-of-00051.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..2ae45f0d1428afd0d14410705ece36a093e009c3 --- /dev/null +++ b/model-00044-of-00051.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:6d702936501327f2471154c4e8424be5ea874a216753baafe176c837a870a42e +size 4831889152 diff --git a/model-00045-of-00051.safetensors b/model-00045-of-00051.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..fa2099510662ffc362bef3dba80067fca6986a50 --- /dev/null +++ b/model-00045-of-00051.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:188edadebd028bf411244f2f974249ada06f3f7f5124700e5cd5157e15586f53 +size 4831938536 diff --git a/model-00046-of-00051.safetensors b/model-00046-of-00051.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..ab5a1e56ffdcd75613b56cddfafccb23e3d1955e --- /dev/null +++ b/model-00046-of-00051.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:26cf099661fa5aef71a0534f8f4c32a8aa14aa55912dd53ea309b723309c2caf +size 4831938544 diff --git a/model-00047-of-00051.safetensors b/model-00047-of-00051.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..2a3607eb79ae508383358aa5042d4a1bcc3661e7 --- /dev/null +++ b/model-00047-of-00051.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:da73147f5990ed6ef215227f7fcf93f10a270db517b7a11d79bd2c989fe9e14f +size 4882269832 diff --git a/model-00048-of-00051.safetensors b/model-00048-of-00051.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..8747327b414cd32f53e802f12a8d43786f9f9b14 --- /dev/null +++ b/model-00048-of-00051.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:0074a9cdcbf75592fe6b8c078149772e5bb745ddc35c765ecf8a4bea00ce7191 +size 4831889152 diff --git a/model-00049-of-00051.safetensors b/model-00049-of-00051.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..18c54eab2144ba478fc186acbb3d55bd28ae81c6 --- /dev/null +++ b/model-00049-of-00051.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:798ab0bee1814fc37fd132a919607526ac089b8b2f631addbbbae947e17eeeb1 +size 4831938536 diff --git a/model-00050-of-00051.safetensors b/model-00050-of-00051.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..d7711e3807e4278222e68416477942eaf67b1bdd --- /dev/null +++ b/model-00050-of-00051.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:3851308356b375fc6b4beb27099bc0c37d2dcaaeb6f1c8c5fe87d85702baba8f +size 4831938544 diff --git a/model-00051-of-00051.safetensors b/model-00051-of-00051.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..babdd7675ce788e03a07d401d75873d046ee5422 --- /dev/null +++ b/model-00051-of-00051.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:1967a7fd269cd25064a91adc0d64473d6ea21f7a66d7b1d4e096d6613fff00f1 +size 2919310112 diff --git a/model.safetensors.index.json b/model.safetensors.index.json new file mode 100644 index 0000000000000000000000000000000000000000..0cad9f6df5cc1b2f35f3264e713d486e87972d8b --- /dev/null +++ b/model.safetensors.index.json @@ -0,0 +1,802 @@ +{ + "metadata": { + "total_size": 245220139008 + }, + "weight_map": { + "lm_head.weight": "model-00051-of-00051.safetensors", + "model.embed_tokens.weight": "model-00001-of-00051.safetensors", + "model.layers.0.input_layernorm.weight": "model-00001-of-00051.safetensors", + "model.layers.0.mlp.down_proj.weight": "model-00001-of-00051.safetensors", + "model.layers.0.mlp.gate_proj.weight": "model-00001-of-00051.safetensors", + "model.layers.0.mlp.up_proj.weight": "model-00001-of-00051.safetensors", + "model.layers.0.post_attention_layernorm.weight": "model-00001-of-00051.safetensors", + "model.layers.0.self_attn.k_proj.weight": "model-00001-of-00051.safetensors", + "model.layers.0.self_attn.o_proj.weight": "model-00001-of-00051.safetensors", + "model.layers.0.self_attn.q_proj.weight": "model-00001-of-00051.safetensors", + "model.layers.0.self_attn.v_proj.weight": "model-00001-of-00051.safetensors", + "model.layers.1.input_layernorm.weight": "model-00002-of-00051.safetensors", + "model.layers.1.mlp.down_proj.weight": "model-00002-of-00051.safetensors", + "model.layers.1.mlp.gate_proj.weight": "model-00001-of-00051.safetensors", + "model.layers.1.mlp.up_proj.weight": "model-00002-of-00051.safetensors", + "model.layers.1.post_attention_layernorm.weight": "model-00002-of-00051.safetensors", + "model.layers.1.self_attn.k_proj.weight": "model-00001-of-00051.safetensors", + "model.layers.1.self_attn.o_proj.weight": "model-00001-of-00051.safetensors", + "model.layers.1.self_attn.q_proj.weight": "model-00001-of-00051.safetensors", + "model.layers.1.self_attn.v_proj.weight": "model-00001-of-00051.safetensors", + "model.layers.10.input_layernorm.weight": "model-00007-of-00051.safetensors", + "model.layers.10.mlp.down_proj.weight": "model-00007-of-00051.safetensors", + "model.layers.10.mlp.gate_proj.weight": "model-00007-of-00051.safetensors", + "model.layers.10.mlp.up_proj.weight": "model-00007-of-00051.safetensors", + "model.layers.10.post_attention_layernorm.weight": "model-00007-of-00051.safetensors", + "model.layers.10.self_attn.k_proj.weight": "model-00006-of-00051.safetensors", + "model.layers.10.self_attn.o_proj.weight": "model-00006-of-00051.safetensors", + "model.layers.10.self_attn.q_proj.weight": "model-00006-of-00051.safetensors", + "model.layers.10.self_attn.v_proj.weight": "model-00006-of-00051.safetensors", + "model.layers.11.input_layernorm.weight": "model-00007-of-00051.safetensors", + "model.layers.11.mlp.down_proj.weight": "model-00007-of-00051.safetensors", + "model.layers.11.mlp.gate_proj.weight": "model-00007-of-00051.safetensors", + "model.layers.11.mlp.up_proj.weight": "model-00007-of-00051.safetensors", + "model.layers.11.post_attention_layernorm.weight": "model-00007-of-00051.safetensors", + "model.layers.11.self_attn.k_proj.weight": "model-00007-of-00051.safetensors", + "model.layers.11.self_attn.o_proj.weight": "model-00007-of-00051.safetensors", + "model.layers.11.self_attn.q_proj.weight": "model-00007-of-00051.safetensors", + "model.layers.11.self_attn.v_proj.weight": "model-00007-of-00051.safetensors", + "model.layers.12.input_layernorm.weight": "model-00008-of-00051.safetensors", + "model.layers.12.mlp.down_proj.weight": "model-00008-of-00051.safetensors", + "model.layers.12.mlp.gate_proj.weight": "model-00008-of-00051.safetensors", + "model.layers.12.mlp.up_proj.weight": "model-00008-of-00051.safetensors", + "model.layers.12.post_attention_layernorm.weight": "model-00008-of-00051.safetensors", + "model.layers.12.self_attn.k_proj.weight": "model-00008-of-00051.safetensors", + "model.layers.12.self_attn.o_proj.weight": "model-00008-of-00051.safetensors", + "model.layers.12.self_attn.q_proj.weight": "model-00008-of-00051.safetensors", + "model.layers.12.self_attn.v_proj.weight": "model-00008-of-00051.safetensors", + "model.layers.13.input_layernorm.weight": "model-00009-of-00051.safetensors", + "model.layers.13.mlp.down_proj.weight": "model-00009-of-00051.safetensors", + "model.layers.13.mlp.gate_proj.weight": "model-00008-of-00051.safetensors", + "model.layers.13.mlp.up_proj.weight": "model-00008-of-00051.safetensors", + "model.layers.13.post_attention_layernorm.weight": "model-00009-of-00051.safetensors", + "model.layers.13.self_attn.k_proj.weight": "model-00008-of-00051.safetensors", + "model.layers.13.self_attn.o_proj.weight": "model-00008-of-00051.safetensors", + "model.layers.13.self_attn.q_proj.weight": "model-00008-of-00051.safetensors", + "model.layers.13.self_attn.v_proj.weight": "model-00008-of-00051.safetensors", + "model.layers.14.input_layernorm.weight": "model-00009-of-00051.safetensors", + "model.layers.14.mlp.down_proj.weight": "model-00009-of-00051.safetensors", + "model.layers.14.mlp.gate_proj.weight": "model-00009-of-00051.safetensors", + "model.layers.14.mlp.up_proj.weight": "model-00009-of-00051.safetensors", + "model.layers.14.post_attention_layernorm.weight": "model-00009-of-00051.safetensors", + "model.layers.14.self_attn.k_proj.weight": "model-00009-of-00051.safetensors", + "model.layers.14.self_attn.o_proj.weight": "model-00009-of-00051.safetensors", + "model.layers.14.self_attn.q_proj.weight": "model-00009-of-00051.safetensors", + "model.layers.14.self_attn.v_proj.weight": "model-00009-of-00051.safetensors", + "model.layers.15.input_layernorm.weight": "model-00010-of-00051.safetensors", + "model.layers.15.mlp.down_proj.weight": "model-00010-of-00051.safetensors", + "model.layers.15.mlp.gate_proj.weight": "model-00009-of-00051.safetensors", + "model.layers.15.mlp.up_proj.weight": "model-00010-of-00051.safetensors", + "model.layers.15.post_attention_layernorm.weight": "model-00010-of-00051.safetensors", + "model.layers.15.self_attn.k_proj.weight": "model-00009-of-00051.safetensors", + "model.layers.15.self_attn.o_proj.weight": "model-00009-of-00051.safetensors", + "model.layers.15.self_attn.q_proj.weight": "model-00009-of-00051.safetensors", + "model.layers.15.self_attn.v_proj.weight": "model-00009-of-00051.safetensors", + "model.layers.16.input_layernorm.weight": "model-00010-of-00051.safetensors", + "model.layers.16.mlp.down_proj.weight": "model-00010-of-00051.safetensors", + "model.layers.16.mlp.gate_proj.weight": "model-00010-of-00051.safetensors", + "model.layers.16.mlp.up_proj.weight": "model-00010-of-00051.safetensors", + "model.layers.16.post_attention_layernorm.weight": "model-00010-of-00051.safetensors", + "model.layers.16.self_attn.k_proj.weight": "model-00010-of-00051.safetensors", + "model.layers.16.self_attn.o_proj.weight": "model-00010-of-00051.safetensors", + "model.layers.16.self_attn.q_proj.weight": "model-00010-of-00051.safetensors", + "model.layers.16.self_attn.v_proj.weight": "model-00010-of-00051.safetensors", + "model.layers.17.input_layernorm.weight": "model-00011-of-00051.safetensors", + "model.layers.17.mlp.down_proj.weight": "model-00011-of-00051.safetensors", + "model.layers.17.mlp.gate_proj.weight": "model-00011-of-00051.safetensors", + "model.layers.17.mlp.up_proj.weight": "model-00011-of-00051.safetensors", + "model.layers.17.post_attention_layernorm.weight": "model-00011-of-00051.safetensors", + "model.layers.17.self_attn.k_proj.weight": "model-00010-of-00051.safetensors", + "model.layers.17.self_attn.o_proj.weight": "model-00010-of-00051.safetensors", + "model.layers.17.self_attn.q_proj.weight": "model-00010-of-00051.safetensors", + "model.layers.17.self_attn.v_proj.weight": "model-00010-of-00051.safetensors", + "model.layers.18.input_layernorm.weight": "model-00011-of-00051.safetensors", + "model.layers.18.mlp.down_proj.weight": "model-00011-of-00051.safetensors", + "model.layers.18.mlp.gate_proj.weight": "model-00011-of-00051.safetensors", + "model.layers.18.mlp.up_proj.weight": "model-00011-of-00051.safetensors", + "model.layers.18.post_attention_layernorm.weight": "model-00011-of-00051.safetensors", + "model.layers.18.self_attn.k_proj.weight": "model-00011-of-00051.safetensors", + "model.layers.18.self_attn.o_proj.weight": "model-00011-of-00051.safetensors", + "model.layers.18.self_attn.q_proj.weight": "model-00011-of-00051.safetensors", + "model.layers.18.self_attn.v_proj.weight": "model-00011-of-00051.safetensors", + "model.layers.19.input_layernorm.weight": "model-00012-of-00051.safetensors", + "model.layers.19.mlp.down_proj.weight": "model-00012-of-00051.safetensors", + "model.layers.19.mlp.gate_proj.weight": "model-00012-of-00051.safetensors", + "model.layers.19.mlp.up_proj.weight": "model-00012-of-00051.safetensors", + "model.layers.19.post_attention_layernorm.weight": "model-00012-of-00051.safetensors", + "model.layers.19.self_attn.k_proj.weight": "model-00012-of-00051.safetensors", + "model.layers.19.self_attn.o_proj.weight": "model-00012-of-00051.safetensors", + "model.layers.19.self_attn.q_proj.weight": "model-00012-of-00051.safetensors", + "model.layers.19.self_attn.v_proj.weight": "model-00012-of-00051.safetensors", + "model.layers.2.input_layernorm.weight": "model-00002-of-00051.safetensors", + "model.layers.2.mlp.down_proj.weight": "model-00002-of-00051.safetensors", + "model.layers.2.mlp.gate_proj.weight": "model-00002-of-00051.safetensors", + "model.layers.2.mlp.up_proj.weight": "model-00002-of-00051.safetensors", + "model.layers.2.post_attention_layernorm.weight": "model-00002-of-00051.safetensors", + "model.layers.2.self_attn.k_proj.weight": "model-00002-of-00051.safetensors", + "model.layers.2.self_attn.o_proj.weight": "model-00002-of-00051.safetensors", + "model.layers.2.self_attn.q_proj.weight": "model-00002-of-00051.safetensors", + "model.layers.2.self_attn.v_proj.weight": "model-00002-of-00051.safetensors", + "model.layers.20.input_layernorm.weight": "model-00013-of-00051.safetensors", + "model.layers.20.mlp.down_proj.weight": "model-00013-of-00051.safetensors", + "model.layers.20.mlp.gate_proj.weight": "model-00012-of-00051.safetensors", + "model.layers.20.mlp.up_proj.weight": "model-00012-of-00051.safetensors", + "model.layers.20.post_attention_layernorm.weight": "model-00013-of-00051.safetensors", + "model.layers.20.self_attn.k_proj.weight": "model-00012-of-00051.safetensors", + "model.layers.20.self_attn.o_proj.weight": "model-00012-of-00051.safetensors", + "model.layers.20.self_attn.q_proj.weight": "model-00012-of-00051.safetensors", + "model.layers.20.self_attn.v_proj.weight": "model-00012-of-00051.safetensors", + "model.layers.21.input_layernorm.weight": "model-00013-of-00051.safetensors", + "model.layers.21.mlp.down_proj.weight": "model-00013-of-00051.safetensors", + "model.layers.21.mlp.gate_proj.weight": "model-00013-of-00051.safetensors", + "model.layers.21.mlp.up_proj.weight": "model-00013-of-00051.safetensors", + "model.layers.21.post_attention_layernorm.weight": "model-00013-of-00051.safetensors", + "model.layers.21.self_attn.k_proj.weight": "model-00013-of-00051.safetensors", + "model.layers.21.self_attn.o_proj.weight": "model-00013-of-00051.safetensors", + "model.layers.21.self_attn.q_proj.weight": "model-00013-of-00051.safetensors", + "model.layers.21.self_attn.v_proj.weight": "model-00013-of-00051.safetensors", + "model.layers.22.input_layernorm.weight": "model-00014-of-00051.safetensors", + "model.layers.22.mlp.down_proj.weight": "model-00014-of-00051.safetensors", + "model.layers.22.mlp.gate_proj.weight": "model-00013-of-00051.safetensors", + "model.layers.22.mlp.up_proj.weight": "model-00014-of-00051.safetensors", + "model.layers.22.post_attention_layernorm.weight": "model-00014-of-00051.safetensors", + "model.layers.22.self_attn.k_proj.weight": "model-00013-of-00051.safetensors", + "model.layers.22.self_attn.o_proj.weight": "model-00013-of-00051.safetensors", + "model.layers.22.self_attn.q_proj.weight": "model-00013-of-00051.safetensors", + "model.layers.22.self_attn.v_proj.weight": "model-00013-of-00051.safetensors", + "model.layers.23.input_layernorm.weight": "model-00014-of-00051.safetensors", + "model.layers.23.mlp.down_proj.weight": "model-00014-of-00051.safetensors", + "model.layers.23.mlp.gate_proj.weight": "model-00014-of-00051.safetensors", + "model.layers.23.mlp.up_proj.weight": "model-00014-of-00051.safetensors", + "model.layers.23.post_attention_layernorm.weight": "model-00014-of-00051.safetensors", + "model.layers.23.self_attn.k_proj.weight": "model-00014-of-00051.safetensors", + "model.layers.23.self_attn.o_proj.weight": "model-00014-of-00051.safetensors", + "model.layers.23.self_attn.q_proj.weight": "model-00014-of-00051.safetensors", + "model.layers.23.self_attn.v_proj.weight": "model-00014-of-00051.safetensors", + "model.layers.24.input_layernorm.weight": "model-00015-of-00051.safetensors", + "model.layers.24.mlp.down_proj.weight": "model-00015-of-00051.safetensors", + "model.layers.24.mlp.gate_proj.weight": "model-00015-of-00051.safetensors", + "model.layers.24.mlp.up_proj.weight": "model-00015-of-00051.safetensors", + "model.layers.24.post_attention_layernorm.weight": "model-00015-of-00051.safetensors", + "model.layers.24.self_attn.k_proj.weight": "model-00014-of-00051.safetensors", + "model.layers.24.self_attn.o_proj.weight": "model-00014-of-00051.safetensors", + "model.layers.24.self_attn.q_proj.weight": "model-00014-of-00051.safetensors", + "model.layers.24.self_attn.v_proj.weight": "model-00014-of-00051.safetensors", + "model.layers.25.input_layernorm.weight": "model-00015-of-00051.safetensors", + "model.layers.25.mlp.down_proj.weight": "model-00015-of-00051.safetensors", + "model.layers.25.mlp.gate_proj.weight": "model-00015-of-00051.safetensors", + "model.layers.25.mlp.up_proj.weight": "model-00015-of-00051.safetensors", + "model.layers.25.post_attention_layernorm.weight": "model-00015-of-00051.safetensors", + "model.layers.25.self_attn.k_proj.weight": "model-00015-of-00051.safetensors", + "model.layers.25.self_attn.o_proj.weight": "model-00015-of-00051.safetensors", + "model.layers.25.self_attn.q_proj.weight": "model-00015-of-00051.safetensors", + "model.layers.25.self_attn.v_proj.weight": "model-00015-of-00051.safetensors", + "model.layers.26.input_layernorm.weight": "model-00016-of-00051.safetensors", + "model.layers.26.mlp.down_proj.weight": "model-00016-of-00051.safetensors", + "model.layers.26.mlp.gate_proj.weight": "model-00016-of-00051.safetensors", + "model.layers.26.mlp.up_proj.weight": "model-00016-of-00051.safetensors", + "model.layers.26.post_attention_layernorm.weight": "model-00016-of-00051.safetensors", + "model.layers.26.self_attn.k_proj.weight": "model-00016-of-00051.safetensors", + "model.layers.26.self_attn.o_proj.weight": "model-00016-of-00051.safetensors", + "model.layers.26.self_attn.q_proj.weight": "model-00016-of-00051.safetensors", + "model.layers.26.self_attn.v_proj.weight": "model-00016-of-00051.safetensors", + "model.layers.27.input_layernorm.weight": "model-00017-of-00051.safetensors", + "model.layers.27.mlp.down_proj.weight": "model-00017-of-00051.safetensors", + "model.layers.27.mlp.gate_proj.weight": "model-00016-of-00051.safetensors", + "model.layers.27.mlp.up_proj.weight": "model-00016-of-00051.safetensors", + "model.layers.27.post_attention_layernorm.weight": "model-00017-of-00051.safetensors", + "model.layers.27.self_attn.k_proj.weight": "model-00016-of-00051.safetensors", + "model.layers.27.self_attn.o_proj.weight": "model-00016-of-00051.safetensors", + "model.layers.27.self_attn.q_proj.weight": "model-00016-of-00051.safetensors", + "model.layers.27.self_attn.v_proj.weight": "model-00016-of-00051.safetensors", + "model.layers.28.input_layernorm.weight": "model-00017-of-00051.safetensors", + "model.layers.28.mlp.down_proj.weight": "model-00017-of-00051.safetensors", + "model.layers.28.mlp.gate_proj.weight": "model-00017-of-00051.safetensors", + "model.layers.28.mlp.up_proj.weight": "model-00017-of-00051.safetensors", + "model.layers.28.post_attention_layernorm.weight": "model-00017-of-00051.safetensors", + "model.layers.28.self_attn.k_proj.weight": "model-00017-of-00051.safetensors", + "model.layers.28.self_attn.o_proj.weight": "model-00017-of-00051.safetensors", + "model.layers.28.self_attn.q_proj.weight": "model-00017-of-00051.safetensors", + "model.layers.28.self_attn.v_proj.weight": "model-00017-of-00051.safetensors", + "model.layers.29.input_layernorm.weight": "model-00018-of-00051.safetensors", + "model.layers.29.mlp.down_proj.weight": "model-00018-of-00051.safetensors", + "model.layers.29.mlp.gate_proj.weight": "model-00017-of-00051.safetensors", + "model.layers.29.mlp.up_proj.weight": "model-00018-of-00051.safetensors", + "model.layers.29.post_attention_layernorm.weight": "model-00018-of-00051.safetensors", + "model.layers.29.self_attn.k_proj.weight": "model-00017-of-00051.safetensors", + "model.layers.29.self_attn.o_proj.weight": "model-00017-of-00051.safetensors", + "model.layers.29.self_attn.q_proj.weight": "model-00017-of-00051.safetensors", + "model.layers.29.self_attn.v_proj.weight": "model-00017-of-00051.safetensors", + "model.layers.3.input_layernorm.weight": "model-00003-of-00051.safetensors", + "model.layers.3.mlp.down_proj.weight": "model-00003-of-00051.safetensors", + "model.layers.3.mlp.gate_proj.weight": "model-00003-of-00051.safetensors", + "model.layers.3.mlp.up_proj.weight": "model-00003-of-00051.safetensors", + "model.layers.3.post_attention_layernorm.weight": "model-00003-of-00051.safetensors", + "model.layers.3.self_attn.k_proj.weight": "model-00002-of-00051.safetensors", + "model.layers.3.self_attn.o_proj.weight": "model-00002-of-00051.safetensors", + "model.layers.3.self_attn.q_proj.weight": "model-00002-of-00051.safetensors", + "model.layers.3.self_attn.v_proj.weight": "model-00002-of-00051.safetensors", + "model.layers.30.input_layernorm.weight": "model-00018-of-00051.safetensors", + "model.layers.30.mlp.down_proj.weight": "model-00018-of-00051.safetensors", + "model.layers.30.mlp.gate_proj.weight": "model-00018-of-00051.safetensors", + "model.layers.30.mlp.up_proj.weight": "model-00018-of-00051.safetensors", + "model.layers.30.post_attention_layernorm.weight": "model-00018-of-00051.safetensors", + "model.layers.30.self_attn.k_proj.weight": "model-00018-of-00051.safetensors", + "model.layers.30.self_attn.o_proj.weight": "model-00018-of-00051.safetensors", + "model.layers.30.self_attn.q_proj.weight": "model-00018-of-00051.safetensors", + "model.layers.30.self_attn.v_proj.weight": "model-00018-of-00051.safetensors", + "model.layers.31.input_layernorm.weight": "model-00019-of-00051.safetensors", + "model.layers.31.mlp.down_proj.weight": "model-00019-of-00051.safetensors", + "model.layers.31.mlp.gate_proj.weight": "model-00019-of-00051.safetensors", + "model.layers.31.mlp.up_proj.weight": "model-00019-of-00051.safetensors", + "model.layers.31.post_attention_layernorm.weight": "model-00019-of-00051.safetensors", + "model.layers.31.self_attn.k_proj.weight": "model-00018-of-00051.safetensors", + "model.layers.31.self_attn.o_proj.weight": "model-00018-of-00051.safetensors", + "model.layers.31.self_attn.q_proj.weight": "model-00018-of-00051.safetensors", + "model.layers.31.self_attn.v_proj.weight": "model-00018-of-00051.safetensors", + "model.layers.32.input_layernorm.weight": "model-00019-of-00051.safetensors", + "model.layers.32.mlp.down_proj.weight": "model-00019-of-00051.safetensors", + "model.layers.32.mlp.gate_proj.weight": "model-00019-of-00051.safetensors", + "model.layers.32.mlp.up_proj.weight": "model-00019-of-00051.safetensors", + "model.layers.32.post_attention_layernorm.weight": "model-00019-of-00051.safetensors", + "model.layers.32.self_attn.k_proj.weight": "model-00019-of-00051.safetensors", + "model.layers.32.self_attn.o_proj.weight": "model-00019-of-00051.safetensors", + "model.layers.32.self_attn.q_proj.weight": "model-00019-of-00051.safetensors", + "model.layers.32.self_attn.v_proj.weight": "model-00019-of-00051.safetensors", + "model.layers.33.input_layernorm.weight": "model-00020-of-00051.safetensors", + "model.layers.33.mlp.down_proj.weight": "model-00020-of-00051.safetensors", + "model.layers.33.mlp.gate_proj.weight": "model-00020-of-00051.safetensors", + "model.layers.33.mlp.up_proj.weight": "model-00020-of-00051.safetensors", + "model.layers.33.post_attention_layernorm.weight": "model-00020-of-00051.safetensors", + "model.layers.33.self_attn.k_proj.weight": "model-00020-of-00051.safetensors", + "model.layers.33.self_attn.o_proj.weight": "model-00020-of-00051.safetensors", + "model.layers.33.self_attn.q_proj.weight": "model-00020-of-00051.safetensors", + "model.layers.33.self_attn.v_proj.weight": "model-00020-of-00051.safetensors", + "model.layers.34.input_layernorm.weight": "model-00021-of-00051.safetensors", + "model.layers.34.mlp.down_proj.weight": "model-00021-of-00051.safetensors", + "model.layers.34.mlp.gate_proj.weight": "model-00020-of-00051.safetensors", + "model.layers.34.mlp.up_proj.weight": "model-00020-of-00051.safetensors", + "model.layers.34.post_attention_layernorm.weight": "model-00021-of-00051.safetensors", + "model.layers.34.self_attn.k_proj.weight": "model-00020-of-00051.safetensors", + "model.layers.34.self_attn.o_proj.weight": "model-00020-of-00051.safetensors", + "model.layers.34.self_attn.q_proj.weight": "model-00020-of-00051.safetensors", + "model.layers.34.self_attn.v_proj.weight": "model-00020-of-00051.safetensors", + "model.layers.35.input_layernorm.weight": "model-00021-of-00051.safetensors", + "model.layers.35.mlp.down_proj.weight": "model-00021-of-00051.safetensors", + "model.layers.35.mlp.gate_proj.weight": "model-00021-of-00051.safetensors", + "model.layers.35.mlp.up_proj.weight": "model-00021-of-00051.safetensors", + "model.layers.35.post_attention_layernorm.weight": "model-00021-of-00051.safetensors", + "model.layers.35.self_attn.k_proj.weight": "model-00021-of-00051.safetensors", + "model.layers.35.self_attn.o_proj.weight": "model-00021-of-00051.safetensors", + "model.layers.35.self_attn.q_proj.weight": "model-00021-of-00051.safetensors", + "model.layers.35.self_attn.v_proj.weight": "model-00021-of-00051.safetensors", + "model.layers.36.input_layernorm.weight": "model-00022-of-00051.safetensors", + "model.layers.36.mlp.down_proj.weight": "model-00022-of-00051.safetensors", + "model.layers.36.mlp.gate_proj.weight": "model-00021-of-00051.safetensors", + "model.layers.36.mlp.up_proj.weight": "model-00022-of-00051.safetensors", + "model.layers.36.post_attention_layernorm.weight": "model-00022-of-00051.safetensors", + "model.layers.36.self_attn.k_proj.weight": "model-00021-of-00051.safetensors", + "model.layers.36.self_attn.o_proj.weight": "model-00021-of-00051.safetensors", + "model.layers.36.self_attn.q_proj.weight": "model-00021-of-00051.safetensors", + "model.layers.36.self_attn.v_proj.weight": "model-00021-of-00051.safetensors", + "model.layers.37.input_layernorm.weight": "model-00022-of-00051.safetensors", + "model.layers.37.mlp.down_proj.weight": "model-00022-of-00051.safetensors", + "model.layers.37.mlp.gate_proj.weight": "model-00022-of-00051.safetensors", + "model.layers.37.mlp.up_proj.weight": "model-00022-of-00051.safetensors", + "model.layers.37.post_attention_layernorm.weight": "model-00022-of-00051.safetensors", + "model.layers.37.self_attn.k_proj.weight": "model-00022-of-00051.safetensors", + "model.layers.37.self_attn.o_proj.weight": "model-00022-of-00051.safetensors", + "model.layers.37.self_attn.q_proj.weight": "model-00022-of-00051.safetensors", + "model.layers.37.self_attn.v_proj.weight": "model-00022-of-00051.safetensors", + "model.layers.38.input_layernorm.weight": "model-00023-of-00051.safetensors", + "model.layers.38.mlp.down_proj.weight": "model-00023-of-00051.safetensors", + "model.layers.38.mlp.gate_proj.weight": "model-00023-of-00051.safetensors", + "model.layers.38.mlp.up_proj.weight": "model-00023-of-00051.safetensors", + "model.layers.38.post_attention_layernorm.weight": "model-00023-of-00051.safetensors", + "model.layers.38.self_attn.k_proj.weight": "model-00022-of-00051.safetensors", + "model.layers.38.self_attn.o_proj.weight": "model-00022-of-00051.safetensors", + "model.layers.38.self_attn.q_proj.weight": "model-00022-of-00051.safetensors", + "model.layers.38.self_attn.v_proj.weight": "model-00022-of-00051.safetensors", + "model.layers.39.input_layernorm.weight": "model-00023-of-00051.safetensors", + "model.layers.39.mlp.down_proj.weight": "model-00023-of-00051.safetensors", + "model.layers.39.mlp.gate_proj.weight": "model-00023-of-00051.safetensors", + "model.layers.39.mlp.up_proj.weight": "model-00023-of-00051.safetensors", + "model.layers.39.post_attention_layernorm.weight": "model-00023-of-00051.safetensors", + "model.layers.39.self_attn.k_proj.weight": "model-00023-of-00051.safetensors", + "model.layers.39.self_attn.o_proj.weight": "model-00023-of-00051.safetensors", + "model.layers.39.self_attn.q_proj.weight": "model-00023-of-00051.safetensors", + "model.layers.39.self_attn.v_proj.weight": "model-00023-of-00051.safetensors", + "model.layers.4.input_layernorm.weight": "model-00003-of-00051.safetensors", + "model.layers.4.mlp.down_proj.weight": "model-00003-of-00051.safetensors", + "model.layers.4.mlp.gate_proj.weight": "model-00003-of-00051.safetensors", + "model.layers.4.mlp.up_proj.weight": "model-00003-of-00051.safetensors", + "model.layers.4.post_attention_layernorm.weight": "model-00003-of-00051.safetensors", + "model.layers.4.self_attn.k_proj.weight": "model-00003-of-00051.safetensors", + "model.layers.4.self_attn.o_proj.weight": "model-00003-of-00051.safetensors", + "model.layers.4.self_attn.q_proj.weight": "model-00003-of-00051.safetensors", + "model.layers.4.self_attn.v_proj.weight": "model-00003-of-00051.safetensors", + "model.layers.40.input_layernorm.weight": "model-00024-of-00051.safetensors", + "model.layers.40.mlp.down_proj.weight": "model-00024-of-00051.safetensors", + "model.layers.40.mlp.gate_proj.weight": "model-00024-of-00051.safetensors", + "model.layers.40.mlp.up_proj.weight": "model-00024-of-00051.safetensors", + "model.layers.40.post_attention_layernorm.weight": "model-00024-of-00051.safetensors", + "model.layers.40.self_attn.k_proj.weight": "model-00024-of-00051.safetensors", + "model.layers.40.self_attn.o_proj.weight": "model-00024-of-00051.safetensors", + "model.layers.40.self_attn.q_proj.weight": "model-00024-of-00051.safetensors", + "model.layers.40.self_attn.v_proj.weight": "model-00024-of-00051.safetensors", + "model.layers.41.input_layernorm.weight": "model-00025-of-00051.safetensors", + "model.layers.41.mlp.down_proj.weight": "model-00025-of-00051.safetensors", + "model.layers.41.mlp.gate_proj.weight": "model-00024-of-00051.safetensors", + "model.layers.41.mlp.up_proj.weight": "model-00024-of-00051.safetensors", + "model.layers.41.post_attention_layernorm.weight": "model-00025-of-00051.safetensors", + "model.layers.41.self_attn.k_proj.weight": "model-00024-of-00051.safetensors", + "model.layers.41.self_attn.o_proj.weight": "model-00024-of-00051.safetensors", + "model.layers.41.self_attn.q_proj.weight": "model-00024-of-00051.safetensors", + "model.layers.41.self_attn.v_proj.weight": "model-00024-of-00051.safetensors", + "model.layers.42.input_layernorm.weight": "model-00025-of-00051.safetensors", + "model.layers.42.mlp.down_proj.weight": "model-00025-of-00051.safetensors", + "model.layers.42.mlp.gate_proj.weight": "model-00025-of-00051.safetensors", + "model.layers.42.mlp.up_proj.weight": "model-00025-of-00051.safetensors", + "model.layers.42.post_attention_layernorm.weight": "model-00025-of-00051.safetensors", + "model.layers.42.self_attn.k_proj.weight": "model-00025-of-00051.safetensors", + "model.layers.42.self_attn.o_proj.weight": "model-00025-of-00051.safetensors", + "model.layers.42.self_attn.q_proj.weight": "model-00025-of-00051.safetensors", + "model.layers.42.self_attn.v_proj.weight": "model-00025-of-00051.safetensors", + "model.layers.43.input_layernorm.weight": "model-00026-of-00051.safetensors", + "model.layers.43.mlp.down_proj.weight": "model-00026-of-00051.safetensors", + "model.layers.43.mlp.gate_proj.weight": "model-00025-of-00051.safetensors", + "model.layers.43.mlp.up_proj.weight": "model-00026-of-00051.safetensors", + "model.layers.43.post_attention_layernorm.weight": "model-00026-of-00051.safetensors", + "model.layers.43.self_attn.k_proj.weight": "model-00025-of-00051.safetensors", + "model.layers.43.self_attn.o_proj.weight": "model-00025-of-00051.safetensors", + "model.layers.43.self_attn.q_proj.weight": "model-00025-of-00051.safetensors", + "model.layers.43.self_attn.v_proj.weight": "model-00025-of-00051.safetensors", + "model.layers.44.input_layernorm.weight": "model-00026-of-00051.safetensors", + "model.layers.44.mlp.down_proj.weight": "model-00026-of-00051.safetensors", + "model.layers.44.mlp.gate_proj.weight": "model-00026-of-00051.safetensors", + "model.layers.44.mlp.up_proj.weight": "model-00026-of-00051.safetensors", + "model.layers.44.post_attention_layernorm.weight": "model-00026-of-00051.safetensors", + "model.layers.44.self_attn.k_proj.weight": "model-00026-of-00051.safetensors", + "model.layers.44.self_attn.o_proj.weight": "model-00026-of-00051.safetensors", + "model.layers.44.self_attn.q_proj.weight": "model-00026-of-00051.safetensors", + "model.layers.44.self_attn.v_proj.weight": "model-00026-of-00051.safetensors", + "model.layers.45.input_layernorm.weight": "model-00027-of-00051.safetensors", + "model.layers.45.mlp.down_proj.weight": "model-00027-of-00051.safetensors", + "model.layers.45.mlp.gate_proj.weight": "model-00027-of-00051.safetensors", + "model.layers.45.mlp.up_proj.weight": "model-00027-of-00051.safetensors", + "model.layers.45.post_attention_layernorm.weight": "model-00027-of-00051.safetensors", + "model.layers.45.self_attn.k_proj.weight": "model-00026-of-00051.safetensors", + "model.layers.45.self_attn.o_proj.weight": "model-00026-of-00051.safetensors", + "model.layers.45.self_attn.q_proj.weight": "model-00026-of-00051.safetensors", + "model.layers.45.self_attn.v_proj.weight": "model-00026-of-00051.safetensors", + "model.layers.46.input_layernorm.weight": "model-00027-of-00051.safetensors", + "model.layers.46.mlp.down_proj.weight": "model-00027-of-00051.safetensors", + "model.layers.46.mlp.gate_proj.weight": "model-00027-of-00051.safetensors", + "model.layers.46.mlp.up_proj.weight": "model-00027-of-00051.safetensors", + "model.layers.46.post_attention_layernorm.weight": "model-00027-of-00051.safetensors", + "model.layers.46.self_attn.k_proj.weight": "model-00027-of-00051.safetensors", + "model.layers.46.self_attn.o_proj.weight": "model-00027-of-00051.safetensors", + "model.layers.46.self_attn.q_proj.weight": "model-00027-of-00051.safetensors", + "model.layers.46.self_attn.v_proj.weight": "model-00027-of-00051.safetensors", + "model.layers.47.input_layernorm.weight": "model-00028-of-00051.safetensors", + "model.layers.47.mlp.down_proj.weight": "model-00028-of-00051.safetensors", + "model.layers.47.mlp.gate_proj.weight": "model-00028-of-00051.safetensors", + "model.layers.47.mlp.up_proj.weight": "model-00028-of-00051.safetensors", + "model.layers.47.post_attention_layernorm.weight": "model-00028-of-00051.safetensors", + "model.layers.47.self_attn.k_proj.weight": "model-00028-of-00051.safetensors", + "model.layers.47.self_attn.o_proj.weight": "model-00028-of-00051.safetensors", + "model.layers.47.self_attn.q_proj.weight": "model-00028-of-00051.safetensors", + "model.layers.47.self_attn.v_proj.weight": "model-00028-of-00051.safetensors", + "model.layers.48.input_layernorm.weight": "model-00029-of-00051.safetensors", + "model.layers.48.mlp.down_proj.weight": "model-00029-of-00051.safetensors", + "model.layers.48.mlp.gate_proj.weight": "model-00028-of-00051.safetensors", + "model.layers.48.mlp.up_proj.weight": "model-00028-of-00051.safetensors", + "model.layers.48.post_attention_layernorm.weight": "model-00029-of-00051.safetensors", + "model.layers.48.self_attn.k_proj.weight": "model-00028-of-00051.safetensors", + "model.layers.48.self_attn.o_proj.weight": "model-00028-of-00051.safetensors", + "model.layers.48.self_attn.q_proj.weight": "model-00028-of-00051.safetensors", + "model.layers.48.self_attn.v_proj.weight": "model-00028-of-00051.safetensors", + "model.layers.49.input_layernorm.weight": "model-00029-of-00051.safetensors", + "model.layers.49.mlp.down_proj.weight": "model-00029-of-00051.safetensors", + "model.layers.49.mlp.gate_proj.weight": "model-00029-of-00051.safetensors", + "model.layers.49.mlp.up_proj.weight": "model-00029-of-00051.safetensors", + "model.layers.49.post_attention_layernorm.weight": "model-00029-of-00051.safetensors", + "model.layers.49.self_attn.k_proj.weight": "model-00029-of-00051.safetensors", + "model.layers.49.self_attn.o_proj.weight": "model-00029-of-00051.safetensors", + "model.layers.49.self_attn.q_proj.weight": "model-00029-of-00051.safetensors", + "model.layers.49.self_attn.v_proj.weight": "model-00029-of-00051.safetensors", + "model.layers.5.input_layernorm.weight": "model-00004-of-00051.safetensors", + "model.layers.5.mlp.down_proj.weight": "model-00004-of-00051.safetensors", + "model.layers.5.mlp.gate_proj.weight": "model-00004-of-00051.safetensors", + "model.layers.5.mlp.up_proj.weight": "model-00004-of-00051.safetensors", + "model.layers.5.post_attention_layernorm.weight": "model-00004-of-00051.safetensors", + "model.layers.5.self_attn.k_proj.weight": "model-00004-of-00051.safetensors", + "model.layers.5.self_attn.o_proj.weight": "model-00004-of-00051.safetensors", + "model.layers.5.self_attn.q_proj.weight": "model-00004-of-00051.safetensors", + "model.layers.5.self_attn.v_proj.weight": "model-00004-of-00051.safetensors", + "model.layers.50.input_layernorm.weight": "model-00030-of-00051.safetensors", + "model.layers.50.mlp.down_proj.weight": "model-00030-of-00051.safetensors", + "model.layers.50.mlp.gate_proj.weight": "model-00029-of-00051.safetensors", + "model.layers.50.mlp.up_proj.weight": "model-00030-of-00051.safetensors", + "model.layers.50.post_attention_layernorm.weight": "model-00030-of-00051.safetensors", + "model.layers.50.self_attn.k_proj.weight": "model-00029-of-00051.safetensors", + "model.layers.50.self_attn.o_proj.weight": "model-00029-of-00051.safetensors", + "model.layers.50.self_attn.q_proj.weight": "model-00029-of-00051.safetensors", + "model.layers.50.self_attn.v_proj.weight": "model-00029-of-00051.safetensors", + "model.layers.51.input_layernorm.weight": "model-00030-of-00051.safetensors", + "model.layers.51.mlp.down_proj.weight": "model-00030-of-00051.safetensors", + "model.layers.51.mlp.gate_proj.weight": "model-00030-of-00051.safetensors", + "model.layers.51.mlp.up_proj.weight": "model-00030-of-00051.safetensors", + "model.layers.51.post_attention_layernorm.weight": "model-00030-of-00051.safetensors", + "model.layers.51.self_attn.k_proj.weight": "model-00030-of-00051.safetensors", + "model.layers.51.self_attn.o_proj.weight": "model-00030-of-00051.safetensors", + "model.layers.51.self_attn.q_proj.weight": "model-00030-of-00051.safetensors", + "model.layers.51.self_attn.v_proj.weight": "model-00030-of-00051.safetensors", + "model.layers.52.input_layernorm.weight": "model-00031-of-00051.safetensors", + "model.layers.52.mlp.down_proj.weight": "model-00031-of-00051.safetensors", + "model.layers.52.mlp.gate_proj.weight": "model-00031-of-00051.safetensors", + "model.layers.52.mlp.up_proj.weight": "model-00031-of-00051.safetensors", + "model.layers.52.post_attention_layernorm.weight": "model-00031-of-00051.safetensors", + "model.layers.52.self_attn.k_proj.weight": "model-00030-of-00051.safetensors", + "model.layers.52.self_attn.o_proj.weight": "model-00030-of-00051.safetensors", + "model.layers.52.self_attn.q_proj.weight": "model-00030-of-00051.safetensors", + "model.layers.52.self_attn.v_proj.weight": "model-00030-of-00051.safetensors", + "model.layers.53.input_layernorm.weight": "model-00031-of-00051.safetensors", + "model.layers.53.mlp.down_proj.weight": "model-00031-of-00051.safetensors", + "model.layers.53.mlp.gate_proj.weight": "model-00031-of-00051.safetensors", + "model.layers.53.mlp.up_proj.weight": "model-00031-of-00051.safetensors", + "model.layers.53.post_attention_layernorm.weight": "model-00031-of-00051.safetensors", + "model.layers.53.self_attn.k_proj.weight": "model-00031-of-00051.safetensors", + "model.layers.53.self_attn.o_proj.weight": "model-00031-of-00051.safetensors", + "model.layers.53.self_attn.q_proj.weight": "model-00031-of-00051.safetensors", + "model.layers.53.self_attn.v_proj.weight": "model-00031-of-00051.safetensors", + "model.layers.54.input_layernorm.weight": "model-00032-of-00051.safetensors", + "model.layers.54.mlp.down_proj.weight": "model-00032-of-00051.safetensors", + "model.layers.54.mlp.gate_proj.weight": "model-00032-of-00051.safetensors", + "model.layers.54.mlp.up_proj.weight": "model-00032-of-00051.safetensors", + "model.layers.54.post_attention_layernorm.weight": "model-00032-of-00051.safetensors", + "model.layers.54.self_attn.k_proj.weight": "model-00032-of-00051.safetensors", + "model.layers.54.self_attn.o_proj.weight": "model-00032-of-00051.safetensors", + "model.layers.54.self_attn.q_proj.weight": "model-00032-of-00051.safetensors", + "model.layers.54.self_attn.v_proj.weight": "model-00032-of-00051.safetensors", + "model.layers.55.input_layernorm.weight": "model-00033-of-00051.safetensors", + "model.layers.55.mlp.down_proj.weight": "model-00033-of-00051.safetensors", + "model.layers.55.mlp.gate_proj.weight": "model-00032-of-00051.safetensors", + "model.layers.55.mlp.up_proj.weight": "model-00032-of-00051.safetensors", + "model.layers.55.post_attention_layernorm.weight": "model-00033-of-00051.safetensors", + "model.layers.55.self_attn.k_proj.weight": "model-00032-of-00051.safetensors", + "model.layers.55.self_attn.o_proj.weight": "model-00032-of-00051.safetensors", + "model.layers.55.self_attn.q_proj.weight": "model-00032-of-00051.safetensors", + "model.layers.55.self_attn.v_proj.weight": "model-00032-of-00051.safetensors", + "model.layers.56.input_layernorm.weight": "model-00033-of-00051.safetensors", + "model.layers.56.mlp.down_proj.weight": "model-00033-of-00051.safetensors", + "model.layers.56.mlp.gate_proj.weight": "model-00033-of-00051.safetensors", + "model.layers.56.mlp.up_proj.weight": "model-00033-of-00051.safetensors", + "model.layers.56.post_attention_layernorm.weight": "model-00033-of-00051.safetensors", + "model.layers.56.self_attn.k_proj.weight": "model-00033-of-00051.safetensors", + "model.layers.56.self_attn.o_proj.weight": "model-00033-of-00051.safetensors", + "model.layers.56.self_attn.q_proj.weight": "model-00033-of-00051.safetensors", + "model.layers.56.self_attn.v_proj.weight": "model-00033-of-00051.safetensors", + "model.layers.57.input_layernorm.weight": "model-00034-of-00051.safetensors", + "model.layers.57.mlp.down_proj.weight": "model-00034-of-00051.safetensors", + "model.layers.57.mlp.gate_proj.weight": "model-00033-of-00051.safetensors", + "model.layers.57.mlp.up_proj.weight": "model-00034-of-00051.safetensors", + "model.layers.57.post_attention_layernorm.weight": "model-00034-of-00051.safetensors", + "model.layers.57.self_attn.k_proj.weight": "model-00033-of-00051.safetensors", + "model.layers.57.self_attn.o_proj.weight": "model-00033-of-00051.safetensors", + "model.layers.57.self_attn.q_proj.weight": "model-00033-of-00051.safetensors", + "model.layers.57.self_attn.v_proj.weight": "model-00033-of-00051.safetensors", + "model.layers.58.input_layernorm.weight": "model-00034-of-00051.safetensors", + "model.layers.58.mlp.down_proj.weight": "model-00034-of-00051.safetensors", + "model.layers.58.mlp.gate_proj.weight": "model-00034-of-00051.safetensors", + "model.layers.58.mlp.up_proj.weight": "model-00034-of-00051.safetensors", + "model.layers.58.post_attention_layernorm.weight": "model-00034-of-00051.safetensors", + "model.layers.58.self_attn.k_proj.weight": "model-00034-of-00051.safetensors", + "model.layers.58.self_attn.o_proj.weight": "model-00034-of-00051.safetensors", + "model.layers.58.self_attn.q_proj.weight": "model-00034-of-00051.safetensors", + "model.layers.58.self_attn.v_proj.weight": "model-00034-of-00051.safetensors", + "model.layers.59.input_layernorm.weight": "model-00035-of-00051.safetensors", + "model.layers.59.mlp.down_proj.weight": "model-00035-of-00051.safetensors", + "model.layers.59.mlp.gate_proj.weight": "model-00035-of-00051.safetensors", + "model.layers.59.mlp.up_proj.weight": "model-00035-of-00051.safetensors", + "model.layers.59.post_attention_layernorm.weight": "model-00035-of-00051.safetensors", + "model.layers.59.self_attn.k_proj.weight": "model-00034-of-00051.safetensors", + "model.layers.59.self_attn.o_proj.weight": "model-00034-of-00051.safetensors", + "model.layers.59.self_attn.q_proj.weight": "model-00034-of-00051.safetensors", + "model.layers.59.self_attn.v_proj.weight": "model-00034-of-00051.safetensors", + "model.layers.6.input_layernorm.weight": "model-00005-of-00051.safetensors", + "model.layers.6.mlp.down_proj.weight": "model-00005-of-00051.safetensors", + "model.layers.6.mlp.gate_proj.weight": "model-00004-of-00051.safetensors", + "model.layers.6.mlp.up_proj.weight": "model-00004-of-00051.safetensors", + "model.layers.6.post_attention_layernorm.weight": "model-00005-of-00051.safetensors", + "model.layers.6.self_attn.k_proj.weight": "model-00004-of-00051.safetensors", + "model.layers.6.self_attn.o_proj.weight": "model-00004-of-00051.safetensors", + "model.layers.6.self_attn.q_proj.weight": "model-00004-of-00051.safetensors", + "model.layers.6.self_attn.v_proj.weight": "model-00004-of-00051.safetensors", + "model.layers.60.input_layernorm.weight": "model-00035-of-00051.safetensors", + "model.layers.60.mlp.down_proj.weight": "model-00035-of-00051.safetensors", + "model.layers.60.mlp.gate_proj.weight": "model-00035-of-00051.safetensors", + "model.layers.60.mlp.up_proj.weight": "model-00035-of-00051.safetensors", + "model.layers.60.post_attention_layernorm.weight": "model-00035-of-00051.safetensors", + "model.layers.60.self_attn.k_proj.weight": "model-00035-of-00051.safetensors", + "model.layers.60.self_attn.o_proj.weight": "model-00035-of-00051.safetensors", + "model.layers.60.self_attn.q_proj.weight": "model-00035-of-00051.safetensors", + "model.layers.60.self_attn.v_proj.weight": "model-00035-of-00051.safetensors", + "model.layers.61.input_layernorm.weight": "model-00036-of-00051.safetensors", + "model.layers.61.mlp.down_proj.weight": "model-00036-of-00051.safetensors", + "model.layers.61.mlp.gate_proj.weight": "model-00036-of-00051.safetensors", + "model.layers.61.mlp.up_proj.weight": "model-00036-of-00051.safetensors", + "model.layers.61.post_attention_layernorm.weight": "model-00036-of-00051.safetensors", + "model.layers.61.self_attn.k_proj.weight": "model-00036-of-00051.safetensors", + "model.layers.61.self_attn.o_proj.weight": "model-00036-of-00051.safetensors", + "model.layers.61.self_attn.q_proj.weight": "model-00036-of-00051.safetensors", + "model.layers.61.self_attn.v_proj.weight": "model-00036-of-00051.safetensors", + "model.layers.62.input_layernorm.weight": "model-00037-of-00051.safetensors", + "model.layers.62.mlp.down_proj.weight": "model-00037-of-00051.safetensors", + "model.layers.62.mlp.gate_proj.weight": "model-00036-of-00051.safetensors", + "model.layers.62.mlp.up_proj.weight": "model-00036-of-00051.safetensors", + "model.layers.62.post_attention_layernorm.weight": "model-00037-of-00051.safetensors", + "model.layers.62.self_attn.k_proj.weight": "model-00036-of-00051.safetensors", + "model.layers.62.self_attn.o_proj.weight": "model-00036-of-00051.safetensors", + "model.layers.62.self_attn.q_proj.weight": "model-00036-of-00051.safetensors", + "model.layers.62.self_attn.v_proj.weight": "model-00036-of-00051.safetensors", + "model.layers.63.input_layernorm.weight": "model-00037-of-00051.safetensors", + "model.layers.63.mlp.down_proj.weight": "model-00037-of-00051.safetensors", + "model.layers.63.mlp.gate_proj.weight": "model-00037-of-00051.safetensors", + "model.layers.63.mlp.up_proj.weight": "model-00037-of-00051.safetensors", + "model.layers.63.post_attention_layernorm.weight": "model-00037-of-00051.safetensors", + "model.layers.63.self_attn.k_proj.weight": "model-00037-of-00051.safetensors", + "model.layers.63.self_attn.o_proj.weight": "model-00037-of-00051.safetensors", + "model.layers.63.self_attn.q_proj.weight": "model-00037-of-00051.safetensors", + "model.layers.63.self_attn.v_proj.weight": "model-00037-of-00051.safetensors", + "model.layers.64.input_layernorm.weight": "model-00038-of-00051.safetensors", + "model.layers.64.mlp.down_proj.weight": "model-00038-of-00051.safetensors", + "model.layers.64.mlp.gate_proj.weight": "model-00037-of-00051.safetensors", + "model.layers.64.mlp.up_proj.weight": "model-00038-of-00051.safetensors", + "model.layers.64.post_attention_layernorm.weight": "model-00038-of-00051.safetensors", + "model.layers.64.self_attn.k_proj.weight": "model-00037-of-00051.safetensors", + "model.layers.64.self_attn.o_proj.weight": "model-00037-of-00051.safetensors", + "model.layers.64.self_attn.q_proj.weight": "model-00037-of-00051.safetensors", + "model.layers.64.self_attn.v_proj.weight": "model-00037-of-00051.safetensors", + "model.layers.65.input_layernorm.weight": "model-00038-of-00051.safetensors", + "model.layers.65.mlp.down_proj.weight": "model-00038-of-00051.safetensors", + "model.layers.65.mlp.gate_proj.weight": "model-00038-of-00051.safetensors", + "model.layers.65.mlp.up_proj.weight": "model-00038-of-00051.safetensors", + "model.layers.65.post_attention_layernorm.weight": "model-00038-of-00051.safetensors", + "model.layers.65.self_attn.k_proj.weight": "model-00038-of-00051.safetensors", + "model.layers.65.self_attn.o_proj.weight": "model-00038-of-00051.safetensors", + "model.layers.65.self_attn.q_proj.weight": "model-00038-of-00051.safetensors", + "model.layers.65.self_attn.v_proj.weight": "model-00038-of-00051.safetensors", + "model.layers.66.input_layernorm.weight": "model-00039-of-00051.safetensors", + "model.layers.66.mlp.down_proj.weight": "model-00039-of-00051.safetensors", + "model.layers.66.mlp.gate_proj.weight": "model-00039-of-00051.safetensors", + "model.layers.66.mlp.up_proj.weight": "model-00039-of-00051.safetensors", + "model.layers.66.post_attention_layernorm.weight": "model-00039-of-00051.safetensors", + "model.layers.66.self_attn.k_proj.weight": "model-00038-of-00051.safetensors", + "model.layers.66.self_attn.o_proj.weight": "model-00038-of-00051.safetensors", + "model.layers.66.self_attn.q_proj.weight": "model-00038-of-00051.safetensors", + "model.layers.66.self_attn.v_proj.weight": "model-00038-of-00051.safetensors", + "model.layers.67.input_layernorm.weight": "model-00039-of-00051.safetensors", + "model.layers.67.mlp.down_proj.weight": "model-00039-of-00051.safetensors", + "model.layers.67.mlp.gate_proj.weight": "model-00039-of-00051.safetensors", + "model.layers.67.mlp.up_proj.weight": "model-00039-of-00051.safetensors", + "model.layers.67.post_attention_layernorm.weight": "model-00039-of-00051.safetensors", + "model.layers.67.self_attn.k_proj.weight": "model-00039-of-00051.safetensors", + "model.layers.67.self_attn.o_proj.weight": "model-00039-of-00051.safetensors", + "model.layers.67.self_attn.q_proj.weight": "model-00039-of-00051.safetensors", + "model.layers.67.self_attn.v_proj.weight": "model-00039-of-00051.safetensors", + "model.layers.68.input_layernorm.weight": "model-00040-of-00051.safetensors", + "model.layers.68.mlp.down_proj.weight": "model-00040-of-00051.safetensors", + "model.layers.68.mlp.gate_proj.weight": "model-00040-of-00051.safetensors", + "model.layers.68.mlp.up_proj.weight": "model-00040-of-00051.safetensors", + "model.layers.68.post_attention_layernorm.weight": "model-00040-of-00051.safetensors", + "model.layers.68.self_attn.k_proj.weight": "model-00040-of-00051.safetensors", + "model.layers.68.self_attn.o_proj.weight": "model-00040-of-00051.safetensors", + "model.layers.68.self_attn.q_proj.weight": "model-00040-of-00051.safetensors", + "model.layers.68.self_attn.v_proj.weight": "model-00040-of-00051.safetensors", + "model.layers.69.input_layernorm.weight": "model-00041-of-00051.safetensors", + "model.layers.69.mlp.down_proj.weight": "model-00041-of-00051.safetensors", + "model.layers.69.mlp.gate_proj.weight": "model-00040-of-00051.safetensors", + "model.layers.69.mlp.up_proj.weight": "model-00040-of-00051.safetensors", + "model.layers.69.post_attention_layernorm.weight": "model-00041-of-00051.safetensors", + "model.layers.69.self_attn.k_proj.weight": "model-00040-of-00051.safetensors", + "model.layers.69.self_attn.o_proj.weight": "model-00040-of-00051.safetensors", + "model.layers.69.self_attn.q_proj.weight": "model-00040-of-00051.safetensors", + "model.layers.69.self_attn.v_proj.weight": "model-00040-of-00051.safetensors", + "model.layers.7.input_layernorm.weight": "model-00005-of-00051.safetensors", + "model.layers.7.mlp.down_proj.weight": "model-00005-of-00051.safetensors", + "model.layers.7.mlp.gate_proj.weight": "model-00005-of-00051.safetensors", + "model.layers.7.mlp.up_proj.weight": "model-00005-of-00051.safetensors", + "model.layers.7.post_attention_layernorm.weight": "model-00005-of-00051.safetensors", + "model.layers.7.self_attn.k_proj.weight": "model-00005-of-00051.safetensors", + "model.layers.7.self_attn.o_proj.weight": "model-00005-of-00051.safetensors", + "model.layers.7.self_attn.q_proj.weight": "model-00005-of-00051.safetensors", + "model.layers.7.self_attn.v_proj.weight": "model-00005-of-00051.safetensors", + "model.layers.70.input_layernorm.weight": "model-00041-of-00051.safetensors", + "model.layers.70.mlp.down_proj.weight": "model-00041-of-00051.safetensors", + "model.layers.70.mlp.gate_proj.weight": "model-00041-of-00051.safetensors", + "model.layers.70.mlp.up_proj.weight": "model-00041-of-00051.safetensors", + "model.layers.70.post_attention_layernorm.weight": "model-00041-of-00051.safetensors", + "model.layers.70.self_attn.k_proj.weight": "model-00041-of-00051.safetensors", + "model.layers.70.self_attn.o_proj.weight": "model-00041-of-00051.safetensors", + "model.layers.70.self_attn.q_proj.weight": "model-00041-of-00051.safetensors", + "model.layers.70.self_attn.v_proj.weight": "model-00041-of-00051.safetensors", + "model.layers.71.input_layernorm.weight": "model-00042-of-00051.safetensors", + "model.layers.71.mlp.down_proj.weight": "model-00042-of-00051.safetensors", + "model.layers.71.mlp.gate_proj.weight": "model-00041-of-00051.safetensors", + "model.layers.71.mlp.up_proj.weight": "model-00042-of-00051.safetensors", + "model.layers.71.post_attention_layernorm.weight": "model-00042-of-00051.safetensors", + "model.layers.71.self_attn.k_proj.weight": "model-00041-of-00051.safetensors", + "model.layers.71.self_attn.o_proj.weight": "model-00041-of-00051.safetensors", + "model.layers.71.self_attn.q_proj.weight": "model-00041-of-00051.safetensors", + "model.layers.71.self_attn.v_proj.weight": "model-00041-of-00051.safetensors", + "model.layers.72.input_layernorm.weight": "model-00042-of-00051.safetensors", + "model.layers.72.mlp.down_proj.weight": "model-00042-of-00051.safetensors", + "model.layers.72.mlp.gate_proj.weight": "model-00042-of-00051.safetensors", + "model.layers.72.mlp.up_proj.weight": "model-00042-of-00051.safetensors", + "model.layers.72.post_attention_layernorm.weight": "model-00042-of-00051.safetensors", + "model.layers.72.self_attn.k_proj.weight": "model-00042-of-00051.safetensors", + "model.layers.72.self_attn.o_proj.weight": "model-00042-of-00051.safetensors", + "model.layers.72.self_attn.q_proj.weight": "model-00042-of-00051.safetensors", + "model.layers.72.self_attn.v_proj.weight": "model-00042-of-00051.safetensors", + "model.layers.73.input_layernorm.weight": "model-00043-of-00051.safetensors", + "model.layers.73.mlp.down_proj.weight": "model-00043-of-00051.safetensors", + "model.layers.73.mlp.gate_proj.weight": "model-00043-of-00051.safetensors", + "model.layers.73.mlp.up_proj.weight": "model-00043-of-00051.safetensors", + "model.layers.73.post_attention_layernorm.weight": "model-00043-of-00051.safetensors", + "model.layers.73.self_attn.k_proj.weight": "model-00042-of-00051.safetensors", + "model.layers.73.self_attn.o_proj.weight": "model-00042-of-00051.safetensors", + "model.layers.73.self_attn.q_proj.weight": "model-00042-of-00051.safetensors", + "model.layers.73.self_attn.v_proj.weight": "model-00042-of-00051.safetensors", + "model.layers.74.input_layernorm.weight": "model-00043-of-00051.safetensors", + "model.layers.74.mlp.down_proj.weight": "model-00043-of-00051.safetensors", + "model.layers.74.mlp.gate_proj.weight": "model-00043-of-00051.safetensors", + "model.layers.74.mlp.up_proj.weight": "model-00043-of-00051.safetensors", + "model.layers.74.post_attention_layernorm.weight": "model-00043-of-00051.safetensors", + "model.layers.74.self_attn.k_proj.weight": "model-00043-of-00051.safetensors", + "model.layers.74.self_attn.o_proj.weight": "model-00043-of-00051.safetensors", + "model.layers.74.self_attn.q_proj.weight": "model-00043-of-00051.safetensors", + "model.layers.74.self_attn.v_proj.weight": "model-00043-of-00051.safetensors", + "model.layers.75.input_layernorm.weight": "model-00044-of-00051.safetensors", + "model.layers.75.mlp.down_proj.weight": "model-00044-of-00051.safetensors", + "model.layers.75.mlp.gate_proj.weight": "model-00044-of-00051.safetensors", + "model.layers.75.mlp.up_proj.weight": "model-00044-of-00051.safetensors", + "model.layers.75.post_attention_layernorm.weight": "model-00044-of-00051.safetensors", + "model.layers.75.self_attn.k_proj.weight": "model-00044-of-00051.safetensors", + "model.layers.75.self_attn.o_proj.weight": "model-00044-of-00051.safetensors", + "model.layers.75.self_attn.q_proj.weight": "model-00044-of-00051.safetensors", + "model.layers.75.self_attn.v_proj.weight": "model-00044-of-00051.safetensors", + "model.layers.76.input_layernorm.weight": "model-00045-of-00051.safetensors", + "model.layers.76.mlp.down_proj.weight": "model-00045-of-00051.safetensors", + "model.layers.76.mlp.gate_proj.weight": "model-00044-of-00051.safetensors", + "model.layers.76.mlp.up_proj.weight": "model-00044-of-00051.safetensors", + "model.layers.76.post_attention_layernorm.weight": "model-00045-of-00051.safetensors", + "model.layers.76.self_attn.k_proj.weight": "model-00044-of-00051.safetensors", + "model.layers.76.self_attn.o_proj.weight": "model-00044-of-00051.safetensors", + "model.layers.76.self_attn.q_proj.weight": "model-00044-of-00051.safetensors", + "model.layers.76.self_attn.v_proj.weight": "model-00044-of-00051.safetensors", + "model.layers.77.input_layernorm.weight": "model-00045-of-00051.safetensors", + "model.layers.77.mlp.down_proj.weight": "model-00045-of-00051.safetensors", + "model.layers.77.mlp.gate_proj.weight": "model-00045-of-00051.safetensors", + "model.layers.77.mlp.up_proj.weight": "model-00045-of-00051.safetensors", + "model.layers.77.post_attention_layernorm.weight": "model-00045-of-00051.safetensors", + "model.layers.77.self_attn.k_proj.weight": "model-00045-of-00051.safetensors", + "model.layers.77.self_attn.o_proj.weight": "model-00045-of-00051.safetensors", + "model.layers.77.self_attn.q_proj.weight": "model-00045-of-00051.safetensors", + "model.layers.77.self_attn.v_proj.weight": "model-00045-of-00051.safetensors", + "model.layers.78.input_layernorm.weight": "model-00046-of-00051.safetensors", + "model.layers.78.mlp.down_proj.weight": "model-00046-of-00051.safetensors", + "model.layers.78.mlp.gate_proj.weight": "model-00045-of-00051.safetensors", + "model.layers.78.mlp.up_proj.weight": "model-00046-of-00051.safetensors", + "model.layers.78.post_attention_layernorm.weight": "model-00046-of-00051.safetensors", + "model.layers.78.self_attn.k_proj.weight": "model-00045-of-00051.safetensors", + "model.layers.78.self_attn.o_proj.weight": "model-00045-of-00051.safetensors", + "model.layers.78.self_attn.q_proj.weight": "model-00045-of-00051.safetensors", + "model.layers.78.self_attn.v_proj.weight": "model-00045-of-00051.safetensors", + "model.layers.79.input_layernorm.weight": "model-00046-of-00051.safetensors", + "model.layers.79.mlp.down_proj.weight": "model-00046-of-00051.safetensors", + "model.layers.79.mlp.gate_proj.weight": "model-00046-of-00051.safetensors", + "model.layers.79.mlp.up_proj.weight": "model-00046-of-00051.safetensors", + "model.layers.79.post_attention_layernorm.weight": "model-00046-of-00051.safetensors", + "model.layers.79.self_attn.k_proj.weight": "model-00046-of-00051.safetensors", + "model.layers.79.self_attn.o_proj.weight": "model-00046-of-00051.safetensors", + "model.layers.79.self_attn.q_proj.weight": "model-00046-of-00051.safetensors", + "model.layers.79.self_attn.v_proj.weight": "model-00046-of-00051.safetensors", + "model.layers.8.input_layernorm.weight": "model-00006-of-00051.safetensors", + "model.layers.8.mlp.down_proj.weight": "model-00006-of-00051.safetensors", + "model.layers.8.mlp.gate_proj.weight": "model-00005-of-00051.safetensors", + "model.layers.8.mlp.up_proj.weight": "model-00006-of-00051.safetensors", + "model.layers.8.post_attention_layernorm.weight": "model-00006-of-00051.safetensors", + "model.layers.8.self_attn.k_proj.weight": "model-00005-of-00051.safetensors", + "model.layers.8.self_attn.o_proj.weight": "model-00005-of-00051.safetensors", + "model.layers.8.self_attn.q_proj.weight": "model-00005-of-00051.safetensors", + "model.layers.8.self_attn.v_proj.weight": "model-00005-of-00051.safetensors", + "model.layers.80.input_layernorm.weight": "model-00047-of-00051.safetensors", + "model.layers.80.mlp.down_proj.weight": "model-00047-of-00051.safetensors", + "model.layers.80.mlp.gate_proj.weight": "model-00047-of-00051.safetensors", + "model.layers.80.mlp.up_proj.weight": "model-00047-of-00051.safetensors", + "model.layers.80.post_attention_layernorm.weight": "model-00047-of-00051.safetensors", + "model.layers.80.self_attn.k_proj.weight": "model-00046-of-00051.safetensors", + "model.layers.80.self_attn.o_proj.weight": "model-00046-of-00051.safetensors", + "model.layers.80.self_attn.q_proj.weight": "model-00046-of-00051.safetensors", + "model.layers.80.self_attn.v_proj.weight": "model-00046-of-00051.safetensors", + "model.layers.81.input_layernorm.weight": "model-00047-of-00051.safetensors", + "model.layers.81.mlp.down_proj.weight": "model-00047-of-00051.safetensors", + "model.layers.81.mlp.gate_proj.weight": "model-00047-of-00051.safetensors", + "model.layers.81.mlp.up_proj.weight": "model-00047-of-00051.safetensors", + "model.layers.81.post_attention_layernorm.weight": "model-00047-of-00051.safetensors", + "model.layers.81.self_attn.k_proj.weight": "model-00047-of-00051.safetensors", + "model.layers.81.self_attn.o_proj.weight": "model-00047-of-00051.safetensors", + "model.layers.81.self_attn.q_proj.weight": "model-00047-of-00051.safetensors", + "model.layers.81.self_attn.v_proj.weight": "model-00047-of-00051.safetensors", + "model.layers.82.input_layernorm.weight": "model-00048-of-00051.safetensors", + "model.layers.82.mlp.down_proj.weight": "model-00048-of-00051.safetensors", + "model.layers.82.mlp.gate_proj.weight": "model-00048-of-00051.safetensors", + "model.layers.82.mlp.up_proj.weight": "model-00048-of-00051.safetensors", + "model.layers.82.post_attention_layernorm.weight": "model-00048-of-00051.safetensors", + "model.layers.82.self_attn.k_proj.weight": "model-00048-of-00051.safetensors", + "model.layers.82.self_attn.o_proj.weight": "model-00048-of-00051.safetensors", + "model.layers.82.self_attn.q_proj.weight": "model-00048-of-00051.safetensors", + "model.layers.82.self_attn.v_proj.weight": "model-00048-of-00051.safetensors", + "model.layers.83.input_layernorm.weight": "model-00049-of-00051.safetensors", + "model.layers.83.mlp.down_proj.weight": "model-00049-of-00051.safetensors", + "model.layers.83.mlp.gate_proj.weight": "model-00048-of-00051.safetensors", + "model.layers.83.mlp.up_proj.weight": "model-00048-of-00051.safetensors", + "model.layers.83.post_attention_layernorm.weight": "model-00049-of-00051.safetensors", + "model.layers.83.self_attn.k_proj.weight": "model-00048-of-00051.safetensors", + "model.layers.83.self_attn.o_proj.weight": "model-00048-of-00051.safetensors", + "model.layers.83.self_attn.q_proj.weight": "model-00048-of-00051.safetensors", + "model.layers.83.self_attn.v_proj.weight": "model-00048-of-00051.safetensors", + "model.layers.84.input_layernorm.weight": "model-00049-of-00051.safetensors", + "model.layers.84.mlp.down_proj.weight": "model-00049-of-00051.safetensors", + "model.layers.84.mlp.gate_proj.weight": "model-00049-of-00051.safetensors", + "model.layers.84.mlp.up_proj.weight": "model-00049-of-00051.safetensors", + "model.layers.84.post_attention_layernorm.weight": "model-00049-of-00051.safetensors", + "model.layers.84.self_attn.k_proj.weight": "model-00049-of-00051.safetensors", + "model.layers.84.self_attn.o_proj.weight": "model-00049-of-00051.safetensors", + "model.layers.84.self_attn.q_proj.weight": "model-00049-of-00051.safetensors", + "model.layers.84.self_attn.v_proj.weight": "model-00049-of-00051.safetensors", + "model.layers.85.input_layernorm.weight": "model-00050-of-00051.safetensors", + "model.layers.85.mlp.down_proj.weight": "model-00050-of-00051.safetensors", + "model.layers.85.mlp.gate_proj.weight": "model-00049-of-00051.safetensors", + "model.layers.85.mlp.up_proj.weight": "model-00050-of-00051.safetensors", + "model.layers.85.post_attention_layernorm.weight": "model-00050-of-00051.safetensors", + "model.layers.85.self_attn.k_proj.weight": "model-00049-of-00051.safetensors", + "model.layers.85.self_attn.o_proj.weight": "model-00049-of-00051.safetensors", + "model.layers.85.self_attn.q_proj.weight": "model-00049-of-00051.safetensors", + "model.layers.85.self_attn.v_proj.weight": "model-00049-of-00051.safetensors", + "model.layers.86.input_layernorm.weight": "model-00050-of-00051.safetensors", + "model.layers.86.mlp.down_proj.weight": "model-00050-of-00051.safetensors", + "model.layers.86.mlp.gate_proj.weight": "model-00050-of-00051.safetensors", + "model.layers.86.mlp.up_proj.weight": "model-00050-of-00051.safetensors", + "model.layers.86.post_attention_layernorm.weight": "model-00050-of-00051.safetensors", + "model.layers.86.self_attn.k_proj.weight": "model-00050-of-00051.safetensors", + "model.layers.86.self_attn.o_proj.weight": "model-00050-of-00051.safetensors", + "model.layers.86.self_attn.q_proj.weight": "model-00050-of-00051.safetensors", + "model.layers.86.self_attn.v_proj.weight": "model-00050-of-00051.safetensors", + "model.layers.87.input_layernorm.weight": "model-00051-of-00051.safetensors", + "model.layers.87.mlp.down_proj.weight": "model-00051-of-00051.safetensors", + "model.layers.87.mlp.gate_proj.weight": "model-00051-of-00051.safetensors", + "model.layers.87.mlp.up_proj.weight": "model-00051-of-00051.safetensors", + "model.layers.87.post_attention_layernorm.weight": "model-00051-of-00051.safetensors", + "model.layers.87.self_attn.k_proj.weight": "model-00050-of-00051.safetensors", + "model.layers.87.self_attn.o_proj.weight": "model-00050-of-00051.safetensors", + "model.layers.87.self_attn.q_proj.weight": "model-00050-of-00051.safetensors", + "model.layers.87.self_attn.v_proj.weight": "model-00050-of-00051.safetensors", + "model.layers.9.input_layernorm.weight": "model-00006-of-00051.safetensors", + "model.layers.9.mlp.down_proj.weight": "model-00006-of-00051.safetensors", + "model.layers.9.mlp.gate_proj.weight": "model-00006-of-00051.safetensors", + "model.layers.9.mlp.up_proj.weight": "model-00006-of-00051.safetensors", + "model.layers.9.post_attention_layernorm.weight": "model-00006-of-00051.safetensors", + "model.layers.9.self_attn.k_proj.weight": "model-00006-of-00051.safetensors", + "model.layers.9.self_attn.o_proj.weight": "model-00006-of-00051.safetensors", + "model.layers.9.self_attn.q_proj.weight": "model-00006-of-00051.safetensors", + "model.layers.9.self_attn.v_proj.weight": "model-00006-of-00051.safetensors", + "model.norm.weight": "model-00051-of-00051.safetensors" + } +} \ No newline at end of file