diff --git a/config.json b/config.json new file mode 100644 index 0000000000000000000000000000000000000000..5172a0904a7368236c5e50cd1f89117714e25ced --- /dev/null +++ b/config.json @@ -0,0 +1,39 @@ +{ + "architectures": [ + "LlamaForCausalLM" + ], + "attention_bias": false, + "attention_dropout": 0.0, + "bos_token_id": 128000, + "eos_token_id": [ + 128001, + 128008, + 128009 + ], + "head_dim": 128, + "hidden_act": "silu", + "hidden_size": 8192, + "initializer_range": 0.02, + "intermediate_size": 28672, + "max_position_embeddings": 131072, + "mlp_bias": false, + "model_type": "llama", + "num_attention_heads": 64, + "num_hidden_layers": 80, + "num_key_value_heads": 8, + "pretraining_tp": 1, + "rms_norm_eps": 1e-05, + "rope_scaling": { + "factor": 8.0, + "high_freq_factor": 4.0, + "low_freq_factor": 1.0, + "original_max_position_embeddings": 8192, + "rope_type": "llama3" + }, + "rope_theta": 500000.0, + "tie_word_embeddings": false, + "torch_dtype": "float32", + "transformers_version": "4.55.4", + "use_cache": true, + "vocab_size": 128256 +} diff --git a/generation_config.json b/generation_config.json new file mode 100644 index 0000000000000000000000000000000000000000..7bbf54087ff336b2346fed8703e5d147a20a025d --- /dev/null +++ b/generation_config.json @@ -0,0 +1,12 @@ +{ + "bos_token_id": 128000, + "do_sample": true, + "eos_token_id": [ + 128001, + 128008, + 128009 + ], + "temperature": 0.6, + "top_p": 0.9, + "transformers_version": "4.55.4" +} diff --git a/model-00001-of-00062.safetensors b/model-00001-of-00062.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..78c297cf1353c7eefd28b0089eea30ae613e8102 --- /dev/null +++ b/model-00001-of-00062.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:c0508c741f14a81d48939a8914390e6c81fb7baf12c64a1229000b2fd86c4c1c +size 4806672984 diff --git a/model-00002-of-00062.safetensors b/model-00002-of-00062.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..1d4439a4fd37e3365b776c76e4a1e6ae4335251e --- /dev/null +++ b/model-00002-of-00062.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:4c91c948b01faa5b782d99e80ea1e83b5be28370e52f9b40b9b1369b9771187f +size 4362142864 diff --git a/model-00003-of-00062.safetensors b/model-00003-of-00062.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..4c70500681902835f2d10ca0471b431d91db0d59 --- /dev/null +++ b/model-00003-of-00062.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:1de0ce3843e4269829108067a9847bc3ad64d067700de0ad825f9cd5db52dc09 +size 4362142864 diff --git a/model-00004-of-00062.safetensors b/model-00004-of-00062.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..f8927d6815a6d904d369bd72cca81a294fb846d9 --- /dev/null +++ b/model-00004-of-00062.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:5ba0f0fec9a0de39a7f9b052dfa22d788b297c420e8afefdd92317beb1263687 +size 4966188864 diff --git a/model-00005-of-00062.safetensors b/model-00005-of-00062.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..1417375a13d596d6e1141ee8626a1f27f8c13077 --- /dev/null +++ b/model-00005-of-00062.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:5304375b5f6e589972dcd2f1968bf2169b85a559d7fb914691e0880f5600dfb6 +size 4362142864 diff --git a/model-00006-of-00062.safetensors b/model-00006-of-00062.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..7b4ec0662952c66900ba95fe5ed9dad12e73dee3 --- /dev/null +++ b/model-00006-of-00062.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:7265f7f523e64f52478da5d7d4a3f6244d0f192b5ea3a22832b5441d03ea0cf1 +size 4362142864 diff --git a/model-00007-of-00062.safetensors b/model-00007-of-00062.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..803ce089aa15d1d7f063996852db2387bd3c06ad --- /dev/null +++ b/model-00007-of-00062.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:dd08393d2b258f739be730b0b7434eb4ace6a7b787629df6ea785df90335f7be +size 4966188864 diff --git a/model-00008-of-00062.safetensors b/model-00008-of-00062.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..da50509eec7ea1787f646da208657b1143378b24 --- /dev/null +++ b/model-00008-of-00062.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:84d7a8db4ea21f9d9aa7ae1556c1027a67110e49f46ecd69c299a1eb040623aa +size 4362142864 diff --git a/model-00009-of-00062.safetensors b/model-00009-of-00062.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..e85e1c1dfaaed414c43b6751c7ca982a1fae9485 --- /dev/null +++ b/model-00009-of-00062.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:75720b6d95ed5476f553492fd7aa03f40c4c326c84887e8b7a1b2b764cdada90 +size 4362142880 diff --git a/model-00010-of-00062.safetensors b/model-00010-of-00062.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..d0e08533b1bd4cd04dfb2597bbab37db0038a083 --- /dev/null +++ b/model-00010-of-00062.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:0e3f2571c0537b4376dd758886f96d64073790fc38341412af1e9db628dc9b0f +size 4966188880 diff --git a/model-00011-of-00062.safetensors b/model-00011-of-00062.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..06335a7a026e9360f04d146b277a494e2f4bbb03 --- /dev/null +++ b/model-00011-of-00062.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:c256c8b0cb99d86ad83d70532b373ddfde33e1454e2f02e1fafffd25773ea766 +size 4362142872 diff --git a/model-00012-of-00062.safetensors b/model-00012-of-00062.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..b340be83e21029ea3e440c9951dc507b531f23c4 --- /dev/null +++ b/model-00012-of-00062.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:e900554b94df3e317eb03d3e98cff39a1fffab83152f21511e928b568f4443b4 +size 4362142872 diff --git a/model-00013-of-00062.safetensors b/model-00013-of-00062.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..d08a57b96f866597be0af6207db43698407bad82 --- /dev/null +++ b/model-00013-of-00062.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:84a4ca0995c0f03143cbd2441b7da58231562b5ba82456e72d4e65ca50f60cc6 +size 4966188880 diff --git a/model-00014-of-00062.safetensors b/model-00014-of-00062.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..9abfdbbcec82fca3462a3cea04272c1be1b5603f --- /dev/null +++ b/model-00014-of-00062.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:e2ed247b06df6a41fa836b749c4f05c069b9c8d761ff9719fb38ea22b8187bea +size 4362142872 diff --git a/model-00015-of-00062.safetensors b/model-00015-of-00062.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..6e0526c42438a29061a3e8076e72d0480810a752 --- /dev/null +++ b/model-00015-of-00062.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:1fdef788283d72177a9cc756244e4f3109b93f6176dfd25203188a26d32f38c4 +size 4362142872 diff --git a/model-00016-of-00062.safetensors b/model-00016-of-00062.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..222122cff5dc0ca606034eccb7025bac65c77cf3 --- /dev/null +++ b/model-00016-of-00062.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:e01309a1d15eab168df74e3f45898247bba72bd37b121967b92e2ce78fef1d57 +size 4966188880 diff --git a/model-00017-of-00062.safetensors b/model-00017-of-00062.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..b22748f5173d69801862ff706f5aaccbda32b10b --- /dev/null +++ b/model-00017-of-00062.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:89eeb359fed2a5a5464842279e38e40aeb1e7f0c35b89b4d215b23856605f2ba +size 4362142872 diff --git a/model-00018-of-00062.safetensors b/model-00018-of-00062.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..54939c5ad52ce89d39944cdeecc6742cfad77b9b --- /dev/null +++ b/model-00018-of-00062.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:48768615c36034b987ecea279d798f7a193fce8b7abf74d81ed592536be4c505 +size 4362142872 diff --git a/model-00019-of-00062.safetensors b/model-00019-of-00062.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..e6e91ed4e0a3d74c22d60da77a90a0efbba7ceeb --- /dev/null +++ b/model-00019-of-00062.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:50fc90357fce91f42ce6cf43eb39c7d3b75fc41cd07b7a98d453137582d81b10 +size 4966188880 diff --git a/model-00020-of-00062.safetensors b/model-00020-of-00062.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..d5b7f056b1338084f558f2b5fbea81c3f716a433 --- /dev/null +++ b/model-00020-of-00062.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:f2044e0d55ecff82f80915d775e0446090a6b2a80b96ddc066cb0f370638c248 +size 4362142872 diff --git a/model-00021-of-00062.safetensors b/model-00021-of-00062.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..ff4721a557bd5d790950d36c1bad8915afc6dd7a --- /dev/null +++ b/model-00021-of-00062.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:66b5ae0b5c3eb3571bb800f15e30461dee9070fb10172d0ecda7c6847082a970 +size 4362142872 diff --git a/model-00022-of-00062.safetensors b/model-00022-of-00062.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..625d51d26ff092a594cc5fb9725c4c7a5d8d72eb --- /dev/null +++ b/model-00022-of-00062.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:42a41a23d9fbe7e9f1e6638a9143713d04fbd876ef328acc26af5d33dd4f6ff3 +size 4966188880 diff --git a/model-00023-of-00062.safetensors b/model-00023-of-00062.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..e89c6845245762548830d1ba1b2d3dc16963c445 --- /dev/null +++ b/model-00023-of-00062.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:4000eb1e81a12540366dc1187cc4ad6390432e97db3829fe6f761be5b9473211 +size 4362142872 diff --git a/model-00024-of-00062.safetensors b/model-00024-of-00062.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..728c51d5caff5e5aaabecf5633c3ea5a8a4a2aed --- /dev/null +++ b/model-00024-of-00062.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:9294f226e52ee97d21b1900366ac337704655af87e52deace45e30897da39b35 +size 4362142872 diff --git a/model-00025-of-00062.safetensors b/model-00025-of-00062.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..53f4a675ed0c135cf73db0be8914101796d49f11 --- /dev/null +++ b/model-00025-of-00062.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:a2ee0fa5bdd45cefdd442b4e8b1135947d6b9541ff1d9459d1fc17a9bbb46a24 +size 4966188880 diff --git a/model-00026-of-00062.safetensors b/model-00026-of-00062.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..cc74725cd4479ca3c0382f026a39e8c848c1b7cf --- /dev/null +++ b/model-00026-of-00062.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:ba79140600211832f84d2244b4795390721f5cf5aba11aec38983df4a54147d3 +size 4362142872 diff --git a/model-00027-of-00062.safetensors b/model-00027-of-00062.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..c6f987fcc9803da81459f27ea7e0bf489cbff393 --- /dev/null +++ b/model-00027-of-00062.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:a6a4419b08856470bb4be66bddc136807b4418c0807af6e9d6362b390ff66c84 +size 4362142872 diff --git a/model-00028-of-00062.safetensors b/model-00028-of-00062.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..5a0325b6f1f6f6b07143501a9501355b2894acb4 --- /dev/null +++ b/model-00028-of-00062.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:ca0e46def37f60864b8a801d16bfea5b7476140168995a8911c5bf49b812d502 +size 4966188880 diff --git a/model-00029-of-00062.safetensors b/model-00029-of-00062.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..c4bea02925b3042f866705c044d5635947fa8e25 --- /dev/null +++ b/model-00029-of-00062.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:81faa2c4402dabfd9041e1fffb8480dcdf3e0337bc6fed8f8bd0be82240c7dd9 +size 4362142872 diff --git a/model-00030-of-00062.safetensors b/model-00030-of-00062.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..5a589db603810fea31face8d0be006f908aa4a0d --- /dev/null +++ b/model-00030-of-00062.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:1d6cb8470bf5cb73bcdd0fafe57f26f1cca6fae07fdacb9b173ba9784d91f6d0 +size 4362142872 diff --git a/model-00031-of-00062.safetensors b/model-00031-of-00062.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..65dd6d0352640e738c0bbef3ec180bf3ddc1c24d --- /dev/null +++ b/model-00031-of-00062.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:9fbda00c24f7c9a128c6f0a9f551c6a4b8668829fd68f5f2459786dc4ce420a6 +size 4966188880 diff --git a/model-00032-of-00062.safetensors b/model-00032-of-00062.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..3e852204a11975a29f62c0b55ea63898c68496c9 --- /dev/null +++ b/model-00032-of-00062.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:5ae31c5bd1a7ce2c81b1a818c998da53cd5edf1d66dfd1673409b957344e2f93 +size 4362142872 diff --git a/model-00033-of-00062.safetensors b/model-00033-of-00062.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..e0fc9be509c9039b0298db9b668f310cbc41ca96 --- /dev/null +++ b/model-00033-of-00062.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:5f05d6f3186cd7a43f6d69584c6279a98a5063c31cdeddb3265c3950debf86ce +size 4362142872 diff --git a/model-00034-of-00062.safetensors b/model-00034-of-00062.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..18d6338a8ddf71486eb1810f44cece421250a3f5 --- /dev/null +++ b/model-00034-of-00062.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:be840476ed2a376f0ae48f610596a45bff750f68713725b2edccfea4fb4f700d +size 4966188880 diff --git a/model-00035-of-00062.safetensors b/model-00035-of-00062.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..489a2d5794bb7ad0de4637d1a795a472f989fb88 --- /dev/null +++ b/model-00035-of-00062.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:5a6304c20e12f45834a7914cfb7d8ee4bed83e90e2554891cf87f4c1c6bf0619 +size 4362142872 diff --git a/model-00036-of-00062.safetensors b/model-00036-of-00062.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..cddc3e5af1a76e682bf5a2f3889a879c9b24276e --- /dev/null +++ b/model-00036-of-00062.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:988f6d21cbade0a8cdf9bb9c087fc2ef5697605f6a99c6afc1a5fa24cb0b2983 +size 4362142872 diff --git a/model-00037-of-00062.safetensors b/model-00037-of-00062.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..4b2309e93104a071d79738ef77a847f31630a6b2 --- /dev/null +++ b/model-00037-of-00062.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:55c4bbb4b88a6350462849866f01f8127536ffe3686a6c91be0995f6e6f288c4 +size 4966188880 diff --git a/model-00038-of-00062.safetensors b/model-00038-of-00062.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..5296d4adede3e36803022f5cdefdfabc7c7629ef --- /dev/null +++ b/model-00038-of-00062.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:17b2a65b2d44cbf5c232c0e6cd6e0af546e2eca0c93dbabc77aec363e31ffcd0 +size 4362142872 diff --git a/model-00039-of-00062.safetensors b/model-00039-of-00062.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..59966d793029d201a87592e78fa0c3ce03e1fa9a --- /dev/null +++ b/model-00039-of-00062.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:bad18c25dde63d2ca5b11f11532a03dcb1cdeef7648391f60e9289798d7b7dbe +size 4362142872 diff --git a/model-00040-of-00062.safetensors b/model-00040-of-00062.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..667696d6daf6991ad24ef1031024a67621b2a1fa --- /dev/null +++ b/model-00040-of-00062.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:43b91b70638c7448d29ec353315140abbc0d55f6ac32f72136b4dac9712a3e08 +size 4966188880 diff --git a/model-00041-of-00062.safetensors b/model-00041-of-00062.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..b604b54d95981065415e30c3032b8e1472696ccf --- /dev/null +++ b/model-00041-of-00062.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:affb7608ebab90b7271143468f4ea3b1a514106e998924cb77b8c27b1752fd19 +size 4362142872 diff --git a/model-00042-of-00062.safetensors b/model-00042-of-00062.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..634373db6611e6a09f2945d62d08b270c3aceb8d --- /dev/null +++ b/model-00042-of-00062.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:9c8f173eac47d7d5d29eaf6078aa80e746087bca5ca27bcd9e8685af4ec8ec4f +size 4362142872 diff --git a/model-00043-of-00062.safetensors b/model-00043-of-00062.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..cff2cf3bb0cdd9a7f77fb3381ba9e8d5a4426de5 --- /dev/null +++ b/model-00043-of-00062.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:3a69220773e0b4165658dee2d9aa8afab1908de70f79463de637d90f77e74a6d +size 4966188880 diff --git a/model-00044-of-00062.safetensors b/model-00044-of-00062.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..39ae66f71d87ccceaa6e59a49c8fab754ded4ef2 --- /dev/null +++ b/model-00044-of-00062.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:089153f4bb3a180467cdbbfdf647dda2d5cef97ebc97564eca809d61c942a54a +size 4362142872 diff --git a/model-00045-of-00062.safetensors b/model-00045-of-00062.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..55d476e5307afb750f2c6effa3019cf2195792ca --- /dev/null +++ b/model-00045-of-00062.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:ab179e919b081cde43c215833bd920b83873e525244a77ec496efdad5bfb2648 +size 4362142872 diff --git a/model-00046-of-00062.safetensors b/model-00046-of-00062.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..6da283fa3df62aae98973a3942bdb49dfc8a5b2c --- /dev/null +++ b/model-00046-of-00062.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:ae114ecdfe01644eb47203fdd79a1fc0cecf8e426aa2442df04d034f6681b301 +size 4966188880 diff --git a/model-00047-of-00062.safetensors b/model-00047-of-00062.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..aa2d108dab1cfedbccc1ee5cd7aea7fa6bc3add3 --- /dev/null +++ b/model-00047-of-00062.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:945af301f5b977c538039fba1edd3cda0fdba49ba6f78bbb70556594b2667a5c +size 4362142872 diff --git a/model-00048-of-00062.safetensors b/model-00048-of-00062.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..5164d72b48e0a047761f42a0333f94d741d697ac --- /dev/null +++ b/model-00048-of-00062.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:25780374dc10114e61793bdf221a1319d39a7e2336e3cacf6385a3d8e65c5d44 +size 4362142872 diff --git a/model-00049-of-00062.safetensors b/model-00049-of-00062.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..b278e87ee3f2ec8178b41a5427d97e229a52ed06 --- /dev/null +++ b/model-00049-of-00062.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:4b2b3bf61130e8b75d685fedcb67dbddd91761ea7bb01979554f9a2e44e67e1d +size 4966188880 diff --git a/model-00050-of-00062.safetensors b/model-00050-of-00062.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..b4d195590a5fec1136ce4b7cb18813ce259dafb6 --- /dev/null +++ b/model-00050-of-00062.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:459546ff1af6ba026a83b6975c8c7ad4721f311b6c357923e386cd649d71c44b +size 4362142872 diff --git a/model-00051-of-00062.safetensors b/model-00051-of-00062.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..7e0b71008b0fca2fd7cbf602c156d0c8d18a666e --- /dev/null +++ b/model-00051-of-00062.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:2ae172c262a34ae21da0b6f39d3bdad451cc80f2394086ace75a8a3e6908f6dd +size 4362142872 diff --git a/model-00052-of-00062.safetensors b/model-00052-of-00062.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..cfcd6af14354163a625799962766c619ed7f32b5 --- /dev/null +++ b/model-00052-of-00062.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:41ec08cc568fe08bce7d3f129c5928f9f57770ea7f7c0e21fa3780a4a34d2492 +size 4966188880 diff --git a/model-00053-of-00062.safetensors b/model-00053-of-00062.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..d3bbff08ad8cd40b61b8ca1f46a30644116c4ea6 --- /dev/null +++ b/model-00053-of-00062.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:8555220d6e8fb72e35817b9629fa193f32a337f4e23a5abb192dec946fc8a7d7 +size 4362142872 diff --git a/model-00054-of-00062.safetensors b/model-00054-of-00062.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..e147235385bf6c2fd05110a754e403472ff7f118 --- /dev/null +++ b/model-00054-of-00062.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:884b8e6a71b2004bd166f26b637778b82a79f5e2645f9154a861e9eace0f8bf3 +size 4362142872 diff --git a/model-00055-of-00062.safetensors b/model-00055-of-00062.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..fd18b9c50dda2ab0a5a8458252925c87010b3326 --- /dev/null +++ b/model-00055-of-00062.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:762e5c141231033a50dbe27ec69af1a2253381835c78eac892801a2a2ee81082 +size 4966188880 diff --git a/model-00056-of-00062.safetensors b/model-00056-of-00062.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..e8317377591c48a33eb01a77a84162aed9fa0dcc --- /dev/null +++ b/model-00056-of-00062.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:b2ee59ea29b48bc997ec7c98eacf15a5d4561fce3988f250800a8265b17116e4 +size 4362142872 diff --git a/model-00057-of-00062.safetensors b/model-00057-of-00062.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..50dfb78ca82cfbd33dbd90c1df5ca2af0ebc0e6c --- /dev/null +++ b/model-00057-of-00062.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:c86781f48e4a7717d41491501696f99f20cd28e1dd99a2240e81446fba9c5117 +size 4362142872 diff --git a/model-00058-of-00062.safetensors b/model-00058-of-00062.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..ad18140ff87da4009db232f516497bd3e1faf5bb --- /dev/null +++ b/model-00058-of-00062.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:3ef144fcbfea898c053e5058c197c4f091236f49a30584dc534136dafab4d16c +size 4966188880 diff --git a/model-00059-of-00062.safetensors b/model-00059-of-00062.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..58f36ac60d849afec888c408a12eba64937a25ed --- /dev/null +++ b/model-00059-of-00062.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:b396e6376726d4a424bd92da6e96070b3b73a8aaa5032136563acdf571fb338f +size 4362142872 diff --git a/model-00060-of-00062.safetensors b/model-00060-of-00062.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..1b1fcee6d4a2e4d34679b9e591951a8347edcb68 --- /dev/null +++ b/model-00060-of-00062.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:aab1d35152d9a6d16d5f46dc22a95f4fb72af04925a3841784c18fcdcd526e5c +size 4362142872 diff --git a/model-00061-of-00062.safetensors b/model-00061-of-00062.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..0db9461331bcf1eb8013f6c3b73cfd8d427e14aa --- /dev/null +++ b/model-00061-of-00062.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:76350a615caf9f6396c672f31abc43064133102b611b7d847aabd55c0cd94afa +size 4362241496 diff --git a/model-00062-of-00062.safetensors b/model-00062-of-00062.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..908a6bd1df871d81d23f41bd765c1d5c2abf525a --- /dev/null +++ b/model-00062-of-00062.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:37adba49b605e4b59b81ef2930475973ff37b20bd4ddd0fbc20c302fc3c112d5 +size 4202692736 diff --git a/model.safetensors.index.json b/model.safetensors.index.json new file mode 100644 index 0000000000000000000000000000000000000000..ef0c642b145f0661c97c573578ea5f5c868b5c19 --- /dev/null +++ b/model.safetensors.index.json @@ -0,0 +1,731 @@ +{ + "metadata": { + "total_parameters": 70553706496, + "total_size": 282214825984 + }, + "weight_map": { + "lm_head.weight": "model-00062-of-00062.safetensors", + "model.embed_tokens.weight": "model-00001-of-00062.safetensors", + "model.layers.0.input_layernorm.weight": "model-00002-of-00062.safetensors", + "model.layers.0.mlp.down_proj.weight": "model-00002-of-00062.safetensors", + "model.layers.0.mlp.gate_proj.weight": "model-00002-of-00062.safetensors", + "model.layers.0.mlp.up_proj.weight": "model-00002-of-00062.safetensors", + "model.layers.0.post_attention_layernorm.weight": "model-00002-of-00062.safetensors", + "model.layers.0.self_attn.k_proj.weight": "model-00001-of-00062.safetensors", + "model.layers.0.self_attn.o_proj.weight": "model-00001-of-00062.safetensors", + "model.layers.0.self_attn.q_proj.weight": "model-00001-of-00062.safetensors", + "model.layers.0.self_attn.v_proj.weight": "model-00001-of-00062.safetensors", + "model.layers.1.input_layernorm.weight": "model-00003-of-00062.safetensors", + "model.layers.1.mlp.down_proj.weight": "model-00003-of-00062.safetensors", + "model.layers.1.mlp.gate_proj.weight": "model-00002-of-00062.safetensors", + "model.layers.1.mlp.up_proj.weight": "model-00003-of-00062.safetensors", + "model.layers.1.post_attention_layernorm.weight": "model-00003-of-00062.safetensors", + "model.layers.1.self_attn.k_proj.weight": "model-00002-of-00062.safetensors", + "model.layers.1.self_attn.o_proj.weight": "model-00002-of-00062.safetensors", + "model.layers.1.self_attn.q_proj.weight": "model-00002-of-00062.safetensors", + "model.layers.1.self_attn.v_proj.weight": "model-00002-of-00062.safetensors", + "model.layers.10.input_layernorm.weight": "model-00010-of-00062.safetensors", + "model.layers.10.mlp.down_proj.weight": "model-00010-of-00062.safetensors", + "model.layers.10.mlp.gate_proj.weight": "model-00009-of-00062.safetensors", + "model.layers.10.mlp.up_proj.weight": "model-00009-of-00062.safetensors", + "model.layers.10.post_attention_layernorm.weight": "model-00010-of-00062.safetensors", + "model.layers.10.self_attn.k_proj.weight": "model-00009-of-00062.safetensors", + "model.layers.10.self_attn.o_proj.weight": "model-00009-of-00062.safetensors", + "model.layers.10.self_attn.q_proj.weight": "model-00009-of-00062.safetensors", + "model.layers.10.self_attn.v_proj.weight": "model-00009-of-00062.safetensors", + "model.layers.11.input_layernorm.weight": "model-00010-of-00062.safetensors", + "model.layers.11.mlp.down_proj.weight": "model-00010-of-00062.safetensors", + "model.layers.11.mlp.gate_proj.weight": "model-00010-of-00062.safetensors", + "model.layers.11.mlp.up_proj.weight": "model-00010-of-00062.safetensors", + "model.layers.11.post_attention_layernorm.weight": "model-00010-of-00062.safetensors", + "model.layers.11.self_attn.k_proj.weight": "model-00010-of-00062.safetensors", + "model.layers.11.self_attn.o_proj.weight": "model-00010-of-00062.safetensors", + "model.layers.11.self_attn.q_proj.weight": "model-00010-of-00062.safetensors", + "model.layers.11.self_attn.v_proj.weight": "model-00010-of-00062.safetensors", + "model.layers.12.input_layernorm.weight": "model-00011-of-00062.safetensors", + "model.layers.12.mlp.down_proj.weight": "model-00011-of-00062.safetensors", + "model.layers.12.mlp.gate_proj.weight": "model-00011-of-00062.safetensors", + "model.layers.12.mlp.up_proj.weight": "model-00011-of-00062.safetensors", + "model.layers.12.post_attention_layernorm.weight": "model-00011-of-00062.safetensors", + "model.layers.12.self_attn.k_proj.weight": "model-00010-of-00062.safetensors", + "model.layers.12.self_attn.o_proj.weight": "model-00010-of-00062.safetensors", + "model.layers.12.self_attn.q_proj.weight": "model-00010-of-00062.safetensors", + "model.layers.12.self_attn.v_proj.weight": "model-00010-of-00062.safetensors", + "model.layers.13.input_layernorm.weight": "model-00012-of-00062.safetensors", + "model.layers.13.mlp.down_proj.weight": "model-00012-of-00062.safetensors", + "model.layers.13.mlp.gate_proj.weight": "model-00011-of-00062.safetensors", + "model.layers.13.mlp.up_proj.weight": "model-00012-of-00062.safetensors", + "model.layers.13.post_attention_layernorm.weight": "model-00012-of-00062.safetensors", + "model.layers.13.self_attn.k_proj.weight": "model-00011-of-00062.safetensors", + "model.layers.13.self_attn.o_proj.weight": "model-00011-of-00062.safetensors", + "model.layers.13.self_attn.q_proj.weight": "model-00011-of-00062.safetensors", + "model.layers.13.self_attn.v_proj.weight": "model-00011-of-00062.safetensors", + "model.layers.14.input_layernorm.weight": "model-00013-of-00062.safetensors", + "model.layers.14.mlp.down_proj.weight": "model-00013-of-00062.safetensors", + "model.layers.14.mlp.gate_proj.weight": "model-00012-of-00062.safetensors", + "model.layers.14.mlp.up_proj.weight": "model-00012-of-00062.safetensors", + "model.layers.14.post_attention_layernorm.weight": "model-00013-of-00062.safetensors", + "model.layers.14.self_attn.k_proj.weight": "model-00012-of-00062.safetensors", + "model.layers.14.self_attn.o_proj.weight": "model-00012-of-00062.safetensors", + "model.layers.14.self_attn.q_proj.weight": "model-00012-of-00062.safetensors", + "model.layers.14.self_attn.v_proj.weight": "model-00012-of-00062.safetensors", + "model.layers.15.input_layernorm.weight": "model-00013-of-00062.safetensors", + "model.layers.15.mlp.down_proj.weight": "model-00013-of-00062.safetensors", + "model.layers.15.mlp.gate_proj.weight": "model-00013-of-00062.safetensors", + "model.layers.15.mlp.up_proj.weight": "model-00013-of-00062.safetensors", + "model.layers.15.post_attention_layernorm.weight": "model-00013-of-00062.safetensors", + "model.layers.15.self_attn.k_proj.weight": "model-00013-of-00062.safetensors", + "model.layers.15.self_attn.o_proj.weight": "model-00013-of-00062.safetensors", + "model.layers.15.self_attn.q_proj.weight": "model-00013-of-00062.safetensors", + "model.layers.15.self_attn.v_proj.weight": "model-00013-of-00062.safetensors", + "model.layers.16.input_layernorm.weight": "model-00014-of-00062.safetensors", + "model.layers.16.mlp.down_proj.weight": "model-00014-of-00062.safetensors", + "model.layers.16.mlp.gate_proj.weight": "model-00014-of-00062.safetensors", + "model.layers.16.mlp.up_proj.weight": "model-00014-of-00062.safetensors", + "model.layers.16.post_attention_layernorm.weight": "model-00014-of-00062.safetensors", + "model.layers.16.self_attn.k_proj.weight": "model-00013-of-00062.safetensors", + "model.layers.16.self_attn.o_proj.weight": "model-00013-of-00062.safetensors", + "model.layers.16.self_attn.q_proj.weight": "model-00013-of-00062.safetensors", + "model.layers.16.self_attn.v_proj.weight": "model-00013-of-00062.safetensors", + "model.layers.17.input_layernorm.weight": "model-00015-of-00062.safetensors", + "model.layers.17.mlp.down_proj.weight": "model-00015-of-00062.safetensors", + "model.layers.17.mlp.gate_proj.weight": "model-00014-of-00062.safetensors", + "model.layers.17.mlp.up_proj.weight": "model-00015-of-00062.safetensors", + "model.layers.17.post_attention_layernorm.weight": "model-00015-of-00062.safetensors", + "model.layers.17.self_attn.k_proj.weight": "model-00014-of-00062.safetensors", + "model.layers.17.self_attn.o_proj.weight": "model-00014-of-00062.safetensors", + "model.layers.17.self_attn.q_proj.weight": "model-00014-of-00062.safetensors", + "model.layers.17.self_attn.v_proj.weight": "model-00014-of-00062.safetensors", + "model.layers.18.input_layernorm.weight": "model-00016-of-00062.safetensors", + "model.layers.18.mlp.down_proj.weight": "model-00016-of-00062.safetensors", + "model.layers.18.mlp.gate_proj.weight": "model-00015-of-00062.safetensors", + "model.layers.18.mlp.up_proj.weight": "model-00015-of-00062.safetensors", + "model.layers.18.post_attention_layernorm.weight": "model-00016-of-00062.safetensors", + "model.layers.18.self_attn.k_proj.weight": "model-00015-of-00062.safetensors", + "model.layers.18.self_attn.o_proj.weight": "model-00015-of-00062.safetensors", + "model.layers.18.self_attn.q_proj.weight": "model-00015-of-00062.safetensors", + "model.layers.18.self_attn.v_proj.weight": "model-00015-of-00062.safetensors", + "model.layers.19.input_layernorm.weight": "model-00016-of-00062.safetensors", + "model.layers.19.mlp.down_proj.weight": "model-00016-of-00062.safetensors", + "model.layers.19.mlp.gate_proj.weight": "model-00016-of-00062.safetensors", + "model.layers.19.mlp.up_proj.weight": "model-00016-of-00062.safetensors", + "model.layers.19.post_attention_layernorm.weight": "model-00016-of-00062.safetensors", + "model.layers.19.self_attn.k_proj.weight": "model-00016-of-00062.safetensors", + "model.layers.19.self_attn.o_proj.weight": "model-00016-of-00062.safetensors", + "model.layers.19.self_attn.q_proj.weight": "model-00016-of-00062.safetensors", + "model.layers.19.self_attn.v_proj.weight": "model-00016-of-00062.safetensors", + "model.layers.2.input_layernorm.weight": "model-00004-of-00062.safetensors", + "model.layers.2.mlp.down_proj.weight": "model-00004-of-00062.safetensors", + "model.layers.2.mlp.gate_proj.weight": "model-00003-of-00062.safetensors", + "model.layers.2.mlp.up_proj.weight": "model-00003-of-00062.safetensors", + "model.layers.2.post_attention_layernorm.weight": "model-00004-of-00062.safetensors", + "model.layers.2.self_attn.k_proj.weight": "model-00003-of-00062.safetensors", + "model.layers.2.self_attn.o_proj.weight": "model-00003-of-00062.safetensors", + "model.layers.2.self_attn.q_proj.weight": "model-00003-of-00062.safetensors", + "model.layers.2.self_attn.v_proj.weight": "model-00003-of-00062.safetensors", + "model.layers.20.input_layernorm.weight": "model-00017-of-00062.safetensors", + "model.layers.20.mlp.down_proj.weight": "model-00017-of-00062.safetensors", + "model.layers.20.mlp.gate_proj.weight": "model-00017-of-00062.safetensors", + "model.layers.20.mlp.up_proj.weight": "model-00017-of-00062.safetensors", + "model.layers.20.post_attention_layernorm.weight": "model-00017-of-00062.safetensors", + "model.layers.20.self_attn.k_proj.weight": "model-00016-of-00062.safetensors", + "model.layers.20.self_attn.o_proj.weight": "model-00016-of-00062.safetensors", + "model.layers.20.self_attn.q_proj.weight": "model-00016-of-00062.safetensors", + "model.layers.20.self_attn.v_proj.weight": "model-00016-of-00062.safetensors", + "model.layers.21.input_layernorm.weight": "model-00018-of-00062.safetensors", + "model.layers.21.mlp.down_proj.weight": "model-00018-of-00062.safetensors", + "model.layers.21.mlp.gate_proj.weight": "model-00017-of-00062.safetensors", + "model.layers.21.mlp.up_proj.weight": "model-00018-of-00062.safetensors", + "model.layers.21.post_attention_layernorm.weight": "model-00018-of-00062.safetensors", + "model.layers.21.self_attn.k_proj.weight": "model-00017-of-00062.safetensors", + "model.layers.21.self_attn.o_proj.weight": "model-00017-of-00062.safetensors", + "model.layers.21.self_attn.q_proj.weight": "model-00017-of-00062.safetensors", + "model.layers.21.self_attn.v_proj.weight": "model-00017-of-00062.safetensors", + "model.layers.22.input_layernorm.weight": "model-00019-of-00062.safetensors", + "model.layers.22.mlp.down_proj.weight": "model-00019-of-00062.safetensors", + "model.layers.22.mlp.gate_proj.weight": "model-00018-of-00062.safetensors", + "model.layers.22.mlp.up_proj.weight": "model-00018-of-00062.safetensors", + "model.layers.22.post_attention_layernorm.weight": "model-00019-of-00062.safetensors", + "model.layers.22.self_attn.k_proj.weight": "model-00018-of-00062.safetensors", + "model.layers.22.self_attn.o_proj.weight": "model-00018-of-00062.safetensors", + "model.layers.22.self_attn.q_proj.weight": "model-00018-of-00062.safetensors", + "model.layers.22.self_attn.v_proj.weight": "model-00018-of-00062.safetensors", + "model.layers.23.input_layernorm.weight": "model-00019-of-00062.safetensors", + "model.layers.23.mlp.down_proj.weight": "model-00019-of-00062.safetensors", + "model.layers.23.mlp.gate_proj.weight": "model-00019-of-00062.safetensors", + "model.layers.23.mlp.up_proj.weight": "model-00019-of-00062.safetensors", + "model.layers.23.post_attention_layernorm.weight": "model-00019-of-00062.safetensors", + "model.layers.23.self_attn.k_proj.weight": "model-00019-of-00062.safetensors", + "model.layers.23.self_attn.o_proj.weight": "model-00019-of-00062.safetensors", + "model.layers.23.self_attn.q_proj.weight": "model-00019-of-00062.safetensors", + "model.layers.23.self_attn.v_proj.weight": "model-00019-of-00062.safetensors", + "model.layers.24.input_layernorm.weight": "model-00020-of-00062.safetensors", + "model.layers.24.mlp.down_proj.weight": "model-00020-of-00062.safetensors", + "model.layers.24.mlp.gate_proj.weight": "model-00020-of-00062.safetensors", + "model.layers.24.mlp.up_proj.weight": "model-00020-of-00062.safetensors", + "model.layers.24.post_attention_layernorm.weight": "model-00020-of-00062.safetensors", + "model.layers.24.self_attn.k_proj.weight": "model-00019-of-00062.safetensors", + "model.layers.24.self_attn.o_proj.weight": "model-00019-of-00062.safetensors", + "model.layers.24.self_attn.q_proj.weight": "model-00019-of-00062.safetensors", + "model.layers.24.self_attn.v_proj.weight": "model-00019-of-00062.safetensors", + "model.layers.25.input_layernorm.weight": "model-00021-of-00062.safetensors", + "model.layers.25.mlp.down_proj.weight": "model-00021-of-00062.safetensors", + "model.layers.25.mlp.gate_proj.weight": "model-00020-of-00062.safetensors", + "model.layers.25.mlp.up_proj.weight": "model-00021-of-00062.safetensors", + "model.layers.25.post_attention_layernorm.weight": "model-00021-of-00062.safetensors", + "model.layers.25.self_attn.k_proj.weight": "model-00020-of-00062.safetensors", + "model.layers.25.self_attn.o_proj.weight": "model-00020-of-00062.safetensors", + "model.layers.25.self_attn.q_proj.weight": "model-00020-of-00062.safetensors", + "model.layers.25.self_attn.v_proj.weight": "model-00020-of-00062.safetensors", + "model.layers.26.input_layernorm.weight": "model-00022-of-00062.safetensors", + "model.layers.26.mlp.down_proj.weight": "model-00022-of-00062.safetensors", + "model.layers.26.mlp.gate_proj.weight": "model-00021-of-00062.safetensors", + "model.layers.26.mlp.up_proj.weight": "model-00021-of-00062.safetensors", + "model.layers.26.post_attention_layernorm.weight": "model-00022-of-00062.safetensors", + "model.layers.26.self_attn.k_proj.weight": "model-00021-of-00062.safetensors", + "model.layers.26.self_attn.o_proj.weight": "model-00021-of-00062.safetensors", + "model.layers.26.self_attn.q_proj.weight": "model-00021-of-00062.safetensors", + "model.layers.26.self_attn.v_proj.weight": "model-00021-of-00062.safetensors", + "model.layers.27.input_layernorm.weight": "model-00022-of-00062.safetensors", + "model.layers.27.mlp.down_proj.weight": "model-00022-of-00062.safetensors", + "model.layers.27.mlp.gate_proj.weight": "model-00022-of-00062.safetensors", + "model.layers.27.mlp.up_proj.weight": "model-00022-of-00062.safetensors", + "model.layers.27.post_attention_layernorm.weight": "model-00022-of-00062.safetensors", + "model.layers.27.self_attn.k_proj.weight": "model-00022-of-00062.safetensors", + "model.layers.27.self_attn.o_proj.weight": "model-00022-of-00062.safetensors", + "model.layers.27.self_attn.q_proj.weight": "model-00022-of-00062.safetensors", + "model.layers.27.self_attn.v_proj.weight": "model-00022-of-00062.safetensors", + "model.layers.28.input_layernorm.weight": "model-00023-of-00062.safetensors", + "model.layers.28.mlp.down_proj.weight": "model-00023-of-00062.safetensors", + "model.layers.28.mlp.gate_proj.weight": "model-00023-of-00062.safetensors", + "model.layers.28.mlp.up_proj.weight": "model-00023-of-00062.safetensors", + "model.layers.28.post_attention_layernorm.weight": "model-00023-of-00062.safetensors", + "model.layers.28.self_attn.k_proj.weight": "model-00022-of-00062.safetensors", + "model.layers.28.self_attn.o_proj.weight": "model-00022-of-00062.safetensors", + "model.layers.28.self_attn.q_proj.weight": "model-00022-of-00062.safetensors", + "model.layers.28.self_attn.v_proj.weight": "model-00022-of-00062.safetensors", + "model.layers.29.input_layernorm.weight": "model-00024-of-00062.safetensors", + "model.layers.29.mlp.down_proj.weight": "model-00024-of-00062.safetensors", + "model.layers.29.mlp.gate_proj.weight": "model-00023-of-00062.safetensors", + "model.layers.29.mlp.up_proj.weight": "model-00024-of-00062.safetensors", + "model.layers.29.post_attention_layernorm.weight": "model-00024-of-00062.safetensors", + "model.layers.29.self_attn.k_proj.weight": "model-00023-of-00062.safetensors", + "model.layers.29.self_attn.o_proj.weight": "model-00023-of-00062.safetensors", + "model.layers.29.self_attn.q_proj.weight": "model-00023-of-00062.safetensors", + "model.layers.29.self_attn.v_proj.weight": "model-00023-of-00062.safetensors", + "model.layers.3.input_layernorm.weight": "model-00004-of-00062.safetensors", + "model.layers.3.mlp.down_proj.weight": "model-00004-of-00062.safetensors", + "model.layers.3.mlp.gate_proj.weight": "model-00004-of-00062.safetensors", + "model.layers.3.mlp.up_proj.weight": "model-00004-of-00062.safetensors", + "model.layers.3.post_attention_layernorm.weight": "model-00004-of-00062.safetensors", + "model.layers.3.self_attn.k_proj.weight": "model-00004-of-00062.safetensors", + "model.layers.3.self_attn.o_proj.weight": "model-00004-of-00062.safetensors", + "model.layers.3.self_attn.q_proj.weight": "model-00004-of-00062.safetensors", + "model.layers.3.self_attn.v_proj.weight": "model-00004-of-00062.safetensors", + "model.layers.30.input_layernorm.weight": "model-00025-of-00062.safetensors", + "model.layers.30.mlp.down_proj.weight": "model-00025-of-00062.safetensors", + "model.layers.30.mlp.gate_proj.weight": "model-00024-of-00062.safetensors", + "model.layers.30.mlp.up_proj.weight": "model-00024-of-00062.safetensors", + "model.layers.30.post_attention_layernorm.weight": "model-00025-of-00062.safetensors", + "model.layers.30.self_attn.k_proj.weight": "model-00024-of-00062.safetensors", + "model.layers.30.self_attn.o_proj.weight": "model-00024-of-00062.safetensors", + "model.layers.30.self_attn.q_proj.weight": "model-00024-of-00062.safetensors", + "model.layers.30.self_attn.v_proj.weight": "model-00024-of-00062.safetensors", + "model.layers.31.input_layernorm.weight": "model-00025-of-00062.safetensors", + "model.layers.31.mlp.down_proj.weight": "model-00025-of-00062.safetensors", + "model.layers.31.mlp.gate_proj.weight": "model-00025-of-00062.safetensors", + "model.layers.31.mlp.up_proj.weight": "model-00025-of-00062.safetensors", + "model.layers.31.post_attention_layernorm.weight": "model-00025-of-00062.safetensors", + "model.layers.31.self_attn.k_proj.weight": "model-00025-of-00062.safetensors", + "model.layers.31.self_attn.o_proj.weight": "model-00025-of-00062.safetensors", + "model.layers.31.self_attn.q_proj.weight": "model-00025-of-00062.safetensors", + "model.layers.31.self_attn.v_proj.weight": "model-00025-of-00062.safetensors", + "model.layers.32.input_layernorm.weight": "model-00026-of-00062.safetensors", + "model.layers.32.mlp.down_proj.weight": "model-00026-of-00062.safetensors", + "model.layers.32.mlp.gate_proj.weight": "model-00026-of-00062.safetensors", + "model.layers.32.mlp.up_proj.weight": "model-00026-of-00062.safetensors", + "model.layers.32.post_attention_layernorm.weight": "model-00026-of-00062.safetensors", + "model.layers.32.self_attn.k_proj.weight": "model-00025-of-00062.safetensors", + "model.layers.32.self_attn.o_proj.weight": "model-00025-of-00062.safetensors", + "model.layers.32.self_attn.q_proj.weight": "model-00025-of-00062.safetensors", + "model.layers.32.self_attn.v_proj.weight": "model-00025-of-00062.safetensors", + "model.layers.33.input_layernorm.weight": "model-00027-of-00062.safetensors", + "model.layers.33.mlp.down_proj.weight": "model-00027-of-00062.safetensors", + "model.layers.33.mlp.gate_proj.weight": "model-00026-of-00062.safetensors", + "model.layers.33.mlp.up_proj.weight": "model-00027-of-00062.safetensors", + "model.layers.33.post_attention_layernorm.weight": "model-00027-of-00062.safetensors", + "model.layers.33.self_attn.k_proj.weight": "model-00026-of-00062.safetensors", + "model.layers.33.self_attn.o_proj.weight": "model-00026-of-00062.safetensors", + "model.layers.33.self_attn.q_proj.weight": "model-00026-of-00062.safetensors", + "model.layers.33.self_attn.v_proj.weight": "model-00026-of-00062.safetensors", + "model.layers.34.input_layernorm.weight": "model-00028-of-00062.safetensors", + "model.layers.34.mlp.down_proj.weight": "model-00028-of-00062.safetensors", + "model.layers.34.mlp.gate_proj.weight": "model-00027-of-00062.safetensors", + "model.layers.34.mlp.up_proj.weight": "model-00027-of-00062.safetensors", + "model.layers.34.post_attention_layernorm.weight": "model-00028-of-00062.safetensors", + "model.layers.34.self_attn.k_proj.weight": "model-00027-of-00062.safetensors", + "model.layers.34.self_attn.o_proj.weight": "model-00027-of-00062.safetensors", + "model.layers.34.self_attn.q_proj.weight": "model-00027-of-00062.safetensors", + "model.layers.34.self_attn.v_proj.weight": "model-00027-of-00062.safetensors", + "model.layers.35.input_layernorm.weight": "model-00028-of-00062.safetensors", + "model.layers.35.mlp.down_proj.weight": "model-00028-of-00062.safetensors", + "model.layers.35.mlp.gate_proj.weight": "model-00028-of-00062.safetensors", + "model.layers.35.mlp.up_proj.weight": "model-00028-of-00062.safetensors", + "model.layers.35.post_attention_layernorm.weight": "model-00028-of-00062.safetensors", + "model.layers.35.self_attn.k_proj.weight": "model-00028-of-00062.safetensors", + "model.layers.35.self_attn.o_proj.weight": "model-00028-of-00062.safetensors", + "model.layers.35.self_attn.q_proj.weight": "model-00028-of-00062.safetensors", + "model.layers.35.self_attn.v_proj.weight": "model-00028-of-00062.safetensors", + "model.layers.36.input_layernorm.weight": "model-00029-of-00062.safetensors", + "model.layers.36.mlp.down_proj.weight": "model-00029-of-00062.safetensors", + "model.layers.36.mlp.gate_proj.weight": "model-00029-of-00062.safetensors", + "model.layers.36.mlp.up_proj.weight": "model-00029-of-00062.safetensors", + "model.layers.36.post_attention_layernorm.weight": "model-00029-of-00062.safetensors", + "model.layers.36.self_attn.k_proj.weight": "model-00028-of-00062.safetensors", + "model.layers.36.self_attn.o_proj.weight": "model-00028-of-00062.safetensors", + "model.layers.36.self_attn.q_proj.weight": "model-00028-of-00062.safetensors", + "model.layers.36.self_attn.v_proj.weight": "model-00028-of-00062.safetensors", + "model.layers.37.input_layernorm.weight": "model-00030-of-00062.safetensors", + "model.layers.37.mlp.down_proj.weight": "model-00030-of-00062.safetensors", + "model.layers.37.mlp.gate_proj.weight": "model-00029-of-00062.safetensors", + "model.layers.37.mlp.up_proj.weight": "model-00030-of-00062.safetensors", + "model.layers.37.post_attention_layernorm.weight": "model-00030-of-00062.safetensors", + "model.layers.37.self_attn.k_proj.weight": "model-00029-of-00062.safetensors", + "model.layers.37.self_attn.o_proj.weight": "model-00029-of-00062.safetensors", + "model.layers.37.self_attn.q_proj.weight": "model-00029-of-00062.safetensors", + "model.layers.37.self_attn.v_proj.weight": "model-00029-of-00062.safetensors", + "model.layers.38.input_layernorm.weight": "model-00031-of-00062.safetensors", + "model.layers.38.mlp.down_proj.weight": "model-00031-of-00062.safetensors", + "model.layers.38.mlp.gate_proj.weight": "model-00030-of-00062.safetensors", + "model.layers.38.mlp.up_proj.weight": "model-00030-of-00062.safetensors", + "model.layers.38.post_attention_layernorm.weight": "model-00031-of-00062.safetensors", + "model.layers.38.self_attn.k_proj.weight": "model-00030-of-00062.safetensors", + "model.layers.38.self_attn.o_proj.weight": "model-00030-of-00062.safetensors", + "model.layers.38.self_attn.q_proj.weight": "model-00030-of-00062.safetensors", + "model.layers.38.self_attn.v_proj.weight": "model-00030-of-00062.safetensors", + "model.layers.39.input_layernorm.weight": "model-00031-of-00062.safetensors", + "model.layers.39.mlp.down_proj.weight": "model-00031-of-00062.safetensors", + "model.layers.39.mlp.gate_proj.weight": "model-00031-of-00062.safetensors", + "model.layers.39.mlp.up_proj.weight": "model-00031-of-00062.safetensors", + "model.layers.39.post_attention_layernorm.weight": "model-00031-of-00062.safetensors", + "model.layers.39.self_attn.k_proj.weight": "model-00031-of-00062.safetensors", + "model.layers.39.self_attn.o_proj.weight": "model-00031-of-00062.safetensors", + "model.layers.39.self_attn.q_proj.weight": "model-00031-of-00062.safetensors", + "model.layers.39.self_attn.v_proj.weight": "model-00031-of-00062.safetensors", + "model.layers.4.input_layernorm.weight": "model-00005-of-00062.safetensors", + "model.layers.4.mlp.down_proj.weight": "model-00005-of-00062.safetensors", + "model.layers.4.mlp.gate_proj.weight": "model-00005-of-00062.safetensors", + "model.layers.4.mlp.up_proj.weight": "model-00005-of-00062.safetensors", + "model.layers.4.post_attention_layernorm.weight": "model-00005-of-00062.safetensors", + "model.layers.4.self_attn.k_proj.weight": "model-00004-of-00062.safetensors", + "model.layers.4.self_attn.o_proj.weight": "model-00004-of-00062.safetensors", + "model.layers.4.self_attn.q_proj.weight": "model-00004-of-00062.safetensors", + "model.layers.4.self_attn.v_proj.weight": "model-00004-of-00062.safetensors", + "model.layers.40.input_layernorm.weight": "model-00032-of-00062.safetensors", + "model.layers.40.mlp.down_proj.weight": "model-00032-of-00062.safetensors", + "model.layers.40.mlp.gate_proj.weight": "model-00032-of-00062.safetensors", + "model.layers.40.mlp.up_proj.weight": "model-00032-of-00062.safetensors", + "model.layers.40.post_attention_layernorm.weight": "model-00032-of-00062.safetensors", + "model.layers.40.self_attn.k_proj.weight": "model-00031-of-00062.safetensors", + "model.layers.40.self_attn.o_proj.weight": "model-00031-of-00062.safetensors", + "model.layers.40.self_attn.q_proj.weight": "model-00031-of-00062.safetensors", + "model.layers.40.self_attn.v_proj.weight": "model-00031-of-00062.safetensors", + "model.layers.41.input_layernorm.weight": "model-00033-of-00062.safetensors", + "model.layers.41.mlp.down_proj.weight": "model-00033-of-00062.safetensors", + "model.layers.41.mlp.gate_proj.weight": "model-00032-of-00062.safetensors", + "model.layers.41.mlp.up_proj.weight": "model-00033-of-00062.safetensors", + "model.layers.41.post_attention_layernorm.weight": "model-00033-of-00062.safetensors", + "model.layers.41.self_attn.k_proj.weight": "model-00032-of-00062.safetensors", + "model.layers.41.self_attn.o_proj.weight": "model-00032-of-00062.safetensors", + "model.layers.41.self_attn.q_proj.weight": "model-00032-of-00062.safetensors", + "model.layers.41.self_attn.v_proj.weight": "model-00032-of-00062.safetensors", + "model.layers.42.input_layernorm.weight": "model-00034-of-00062.safetensors", + "model.layers.42.mlp.down_proj.weight": "model-00034-of-00062.safetensors", + "model.layers.42.mlp.gate_proj.weight": "model-00033-of-00062.safetensors", + "model.layers.42.mlp.up_proj.weight": "model-00033-of-00062.safetensors", + "model.layers.42.post_attention_layernorm.weight": "model-00034-of-00062.safetensors", + "model.layers.42.self_attn.k_proj.weight": "model-00033-of-00062.safetensors", + "model.layers.42.self_attn.o_proj.weight": "model-00033-of-00062.safetensors", + "model.layers.42.self_attn.q_proj.weight": "model-00033-of-00062.safetensors", + "model.layers.42.self_attn.v_proj.weight": "model-00033-of-00062.safetensors", + "model.layers.43.input_layernorm.weight": "model-00034-of-00062.safetensors", + "model.layers.43.mlp.down_proj.weight": "model-00034-of-00062.safetensors", + "model.layers.43.mlp.gate_proj.weight": "model-00034-of-00062.safetensors", + "model.layers.43.mlp.up_proj.weight": "model-00034-of-00062.safetensors", + "model.layers.43.post_attention_layernorm.weight": "model-00034-of-00062.safetensors", + "model.layers.43.self_attn.k_proj.weight": "model-00034-of-00062.safetensors", + "model.layers.43.self_attn.o_proj.weight": "model-00034-of-00062.safetensors", + "model.layers.43.self_attn.q_proj.weight": "model-00034-of-00062.safetensors", + "model.layers.43.self_attn.v_proj.weight": "model-00034-of-00062.safetensors", + "model.layers.44.input_layernorm.weight": "model-00035-of-00062.safetensors", + "model.layers.44.mlp.down_proj.weight": "model-00035-of-00062.safetensors", + "model.layers.44.mlp.gate_proj.weight": "model-00035-of-00062.safetensors", + "model.layers.44.mlp.up_proj.weight": "model-00035-of-00062.safetensors", + "model.layers.44.post_attention_layernorm.weight": "model-00035-of-00062.safetensors", + "model.layers.44.self_attn.k_proj.weight": "model-00034-of-00062.safetensors", + "model.layers.44.self_attn.o_proj.weight": "model-00034-of-00062.safetensors", + "model.layers.44.self_attn.q_proj.weight": "model-00034-of-00062.safetensors", + "model.layers.44.self_attn.v_proj.weight": "model-00034-of-00062.safetensors", + "model.layers.45.input_layernorm.weight": "model-00036-of-00062.safetensors", + "model.layers.45.mlp.down_proj.weight": "model-00036-of-00062.safetensors", + "model.layers.45.mlp.gate_proj.weight": "model-00035-of-00062.safetensors", + "model.layers.45.mlp.up_proj.weight": "model-00036-of-00062.safetensors", + "model.layers.45.post_attention_layernorm.weight": "model-00036-of-00062.safetensors", + "model.layers.45.self_attn.k_proj.weight": "model-00035-of-00062.safetensors", + "model.layers.45.self_attn.o_proj.weight": "model-00035-of-00062.safetensors", + "model.layers.45.self_attn.q_proj.weight": "model-00035-of-00062.safetensors", + "model.layers.45.self_attn.v_proj.weight": "model-00035-of-00062.safetensors", + "model.layers.46.input_layernorm.weight": "model-00037-of-00062.safetensors", + "model.layers.46.mlp.down_proj.weight": "model-00037-of-00062.safetensors", + "model.layers.46.mlp.gate_proj.weight": "model-00036-of-00062.safetensors", + "model.layers.46.mlp.up_proj.weight": "model-00036-of-00062.safetensors", + "model.layers.46.post_attention_layernorm.weight": "model-00037-of-00062.safetensors", + "model.layers.46.self_attn.k_proj.weight": "model-00036-of-00062.safetensors", + "model.layers.46.self_attn.o_proj.weight": "model-00036-of-00062.safetensors", + "model.layers.46.self_attn.q_proj.weight": "model-00036-of-00062.safetensors", + "model.layers.46.self_attn.v_proj.weight": "model-00036-of-00062.safetensors", + "model.layers.47.input_layernorm.weight": "model-00037-of-00062.safetensors", + "model.layers.47.mlp.down_proj.weight": "model-00037-of-00062.safetensors", + "model.layers.47.mlp.gate_proj.weight": "model-00037-of-00062.safetensors", + "model.layers.47.mlp.up_proj.weight": "model-00037-of-00062.safetensors", + "model.layers.47.post_attention_layernorm.weight": "model-00037-of-00062.safetensors", + "model.layers.47.self_attn.k_proj.weight": "model-00037-of-00062.safetensors", + "model.layers.47.self_attn.o_proj.weight": "model-00037-of-00062.safetensors", + "model.layers.47.self_attn.q_proj.weight": "model-00037-of-00062.safetensors", + "model.layers.47.self_attn.v_proj.weight": "model-00037-of-00062.safetensors", + "model.layers.48.input_layernorm.weight": "model-00038-of-00062.safetensors", + "model.layers.48.mlp.down_proj.weight": "model-00038-of-00062.safetensors", + "model.layers.48.mlp.gate_proj.weight": "model-00038-of-00062.safetensors", + "model.layers.48.mlp.up_proj.weight": "model-00038-of-00062.safetensors", + "model.layers.48.post_attention_layernorm.weight": "model-00038-of-00062.safetensors", + "model.layers.48.self_attn.k_proj.weight": "model-00037-of-00062.safetensors", + "model.layers.48.self_attn.o_proj.weight": "model-00037-of-00062.safetensors", + "model.layers.48.self_attn.q_proj.weight": "model-00037-of-00062.safetensors", + "model.layers.48.self_attn.v_proj.weight": "model-00037-of-00062.safetensors", + "model.layers.49.input_layernorm.weight": "model-00039-of-00062.safetensors", + "model.layers.49.mlp.down_proj.weight": "model-00039-of-00062.safetensors", + "model.layers.49.mlp.gate_proj.weight": "model-00038-of-00062.safetensors", + "model.layers.49.mlp.up_proj.weight": "model-00039-of-00062.safetensors", + "model.layers.49.post_attention_layernorm.weight": "model-00039-of-00062.safetensors", + "model.layers.49.self_attn.k_proj.weight": "model-00038-of-00062.safetensors", + "model.layers.49.self_attn.o_proj.weight": "model-00038-of-00062.safetensors", + "model.layers.49.self_attn.q_proj.weight": "model-00038-of-00062.safetensors", + "model.layers.49.self_attn.v_proj.weight": "model-00038-of-00062.safetensors", + "model.layers.5.input_layernorm.weight": "model-00006-of-00062.safetensors", + "model.layers.5.mlp.down_proj.weight": "model-00006-of-00062.safetensors", + "model.layers.5.mlp.gate_proj.weight": "model-00005-of-00062.safetensors", + "model.layers.5.mlp.up_proj.weight": "model-00006-of-00062.safetensors", + "model.layers.5.post_attention_layernorm.weight": "model-00006-of-00062.safetensors", + "model.layers.5.self_attn.k_proj.weight": "model-00005-of-00062.safetensors", + "model.layers.5.self_attn.o_proj.weight": "model-00005-of-00062.safetensors", + "model.layers.5.self_attn.q_proj.weight": "model-00005-of-00062.safetensors", + "model.layers.5.self_attn.v_proj.weight": "model-00005-of-00062.safetensors", + "model.layers.50.input_layernorm.weight": "model-00040-of-00062.safetensors", + "model.layers.50.mlp.down_proj.weight": "model-00040-of-00062.safetensors", + "model.layers.50.mlp.gate_proj.weight": "model-00039-of-00062.safetensors", + "model.layers.50.mlp.up_proj.weight": "model-00039-of-00062.safetensors", + "model.layers.50.post_attention_layernorm.weight": "model-00040-of-00062.safetensors", + "model.layers.50.self_attn.k_proj.weight": "model-00039-of-00062.safetensors", + "model.layers.50.self_attn.o_proj.weight": "model-00039-of-00062.safetensors", + "model.layers.50.self_attn.q_proj.weight": "model-00039-of-00062.safetensors", + "model.layers.50.self_attn.v_proj.weight": "model-00039-of-00062.safetensors", + "model.layers.51.input_layernorm.weight": "model-00040-of-00062.safetensors", + "model.layers.51.mlp.down_proj.weight": "model-00040-of-00062.safetensors", + "model.layers.51.mlp.gate_proj.weight": "model-00040-of-00062.safetensors", + "model.layers.51.mlp.up_proj.weight": "model-00040-of-00062.safetensors", + "model.layers.51.post_attention_layernorm.weight": "model-00040-of-00062.safetensors", + "model.layers.51.self_attn.k_proj.weight": "model-00040-of-00062.safetensors", + "model.layers.51.self_attn.o_proj.weight": "model-00040-of-00062.safetensors", + "model.layers.51.self_attn.q_proj.weight": "model-00040-of-00062.safetensors", + "model.layers.51.self_attn.v_proj.weight": "model-00040-of-00062.safetensors", + "model.layers.52.input_layernorm.weight": "model-00041-of-00062.safetensors", + "model.layers.52.mlp.down_proj.weight": "model-00041-of-00062.safetensors", + "model.layers.52.mlp.gate_proj.weight": "model-00041-of-00062.safetensors", + "model.layers.52.mlp.up_proj.weight": "model-00041-of-00062.safetensors", + "model.layers.52.post_attention_layernorm.weight": "model-00041-of-00062.safetensors", + "model.layers.52.self_attn.k_proj.weight": "model-00040-of-00062.safetensors", + "model.layers.52.self_attn.o_proj.weight": "model-00040-of-00062.safetensors", + "model.layers.52.self_attn.q_proj.weight": "model-00040-of-00062.safetensors", + "model.layers.52.self_attn.v_proj.weight": "model-00040-of-00062.safetensors", + "model.layers.53.input_layernorm.weight": "model-00042-of-00062.safetensors", + "model.layers.53.mlp.down_proj.weight": "model-00042-of-00062.safetensors", + "model.layers.53.mlp.gate_proj.weight": "model-00041-of-00062.safetensors", + "model.layers.53.mlp.up_proj.weight": "model-00042-of-00062.safetensors", + "model.layers.53.post_attention_layernorm.weight": "model-00042-of-00062.safetensors", + "model.layers.53.self_attn.k_proj.weight": "model-00041-of-00062.safetensors", + "model.layers.53.self_attn.o_proj.weight": "model-00041-of-00062.safetensors", + "model.layers.53.self_attn.q_proj.weight": "model-00041-of-00062.safetensors", + "model.layers.53.self_attn.v_proj.weight": "model-00041-of-00062.safetensors", + "model.layers.54.input_layernorm.weight": "model-00043-of-00062.safetensors", + "model.layers.54.mlp.down_proj.weight": "model-00043-of-00062.safetensors", + "model.layers.54.mlp.gate_proj.weight": "model-00042-of-00062.safetensors", + "model.layers.54.mlp.up_proj.weight": "model-00042-of-00062.safetensors", + "model.layers.54.post_attention_layernorm.weight": "model-00043-of-00062.safetensors", + "model.layers.54.self_attn.k_proj.weight": "model-00042-of-00062.safetensors", + "model.layers.54.self_attn.o_proj.weight": "model-00042-of-00062.safetensors", + "model.layers.54.self_attn.q_proj.weight": "model-00042-of-00062.safetensors", + "model.layers.54.self_attn.v_proj.weight": "model-00042-of-00062.safetensors", + "model.layers.55.input_layernorm.weight": "model-00043-of-00062.safetensors", + "model.layers.55.mlp.down_proj.weight": "model-00043-of-00062.safetensors", + "model.layers.55.mlp.gate_proj.weight": "model-00043-of-00062.safetensors", + "model.layers.55.mlp.up_proj.weight": "model-00043-of-00062.safetensors", + "model.layers.55.post_attention_layernorm.weight": "model-00043-of-00062.safetensors", + "model.layers.55.self_attn.k_proj.weight": "model-00043-of-00062.safetensors", + "model.layers.55.self_attn.o_proj.weight": "model-00043-of-00062.safetensors", + "model.layers.55.self_attn.q_proj.weight": "model-00043-of-00062.safetensors", + "model.layers.55.self_attn.v_proj.weight": "model-00043-of-00062.safetensors", + "model.layers.56.input_layernorm.weight": "model-00044-of-00062.safetensors", + "model.layers.56.mlp.down_proj.weight": "model-00044-of-00062.safetensors", + "model.layers.56.mlp.gate_proj.weight": "model-00044-of-00062.safetensors", + "model.layers.56.mlp.up_proj.weight": "model-00044-of-00062.safetensors", + "model.layers.56.post_attention_layernorm.weight": "model-00044-of-00062.safetensors", + "model.layers.56.self_attn.k_proj.weight": "model-00043-of-00062.safetensors", + "model.layers.56.self_attn.o_proj.weight": "model-00043-of-00062.safetensors", + "model.layers.56.self_attn.q_proj.weight": "model-00043-of-00062.safetensors", + "model.layers.56.self_attn.v_proj.weight": "model-00043-of-00062.safetensors", + "model.layers.57.input_layernorm.weight": "model-00045-of-00062.safetensors", + "model.layers.57.mlp.down_proj.weight": "model-00045-of-00062.safetensors", + "model.layers.57.mlp.gate_proj.weight": "model-00044-of-00062.safetensors", + "model.layers.57.mlp.up_proj.weight": "model-00045-of-00062.safetensors", + "model.layers.57.post_attention_layernorm.weight": "model-00045-of-00062.safetensors", + "model.layers.57.self_attn.k_proj.weight": "model-00044-of-00062.safetensors", + "model.layers.57.self_attn.o_proj.weight": "model-00044-of-00062.safetensors", + "model.layers.57.self_attn.q_proj.weight": "model-00044-of-00062.safetensors", + "model.layers.57.self_attn.v_proj.weight": "model-00044-of-00062.safetensors", + "model.layers.58.input_layernorm.weight": "model-00046-of-00062.safetensors", + "model.layers.58.mlp.down_proj.weight": "model-00046-of-00062.safetensors", + "model.layers.58.mlp.gate_proj.weight": "model-00045-of-00062.safetensors", + "model.layers.58.mlp.up_proj.weight": "model-00045-of-00062.safetensors", + "model.layers.58.post_attention_layernorm.weight": "model-00046-of-00062.safetensors", + "model.layers.58.self_attn.k_proj.weight": "model-00045-of-00062.safetensors", + "model.layers.58.self_attn.o_proj.weight": "model-00045-of-00062.safetensors", + "model.layers.58.self_attn.q_proj.weight": "model-00045-of-00062.safetensors", + "model.layers.58.self_attn.v_proj.weight": "model-00045-of-00062.safetensors", + "model.layers.59.input_layernorm.weight": "model-00046-of-00062.safetensors", + "model.layers.59.mlp.down_proj.weight": "model-00046-of-00062.safetensors", + "model.layers.59.mlp.gate_proj.weight": "model-00046-of-00062.safetensors", + "model.layers.59.mlp.up_proj.weight": "model-00046-of-00062.safetensors", + "model.layers.59.post_attention_layernorm.weight": "model-00046-of-00062.safetensors", + "model.layers.59.self_attn.k_proj.weight": "model-00046-of-00062.safetensors", + "model.layers.59.self_attn.o_proj.weight": "model-00046-of-00062.safetensors", + "model.layers.59.self_attn.q_proj.weight": "model-00046-of-00062.safetensors", + "model.layers.59.self_attn.v_proj.weight": "model-00046-of-00062.safetensors", + "model.layers.6.input_layernorm.weight": "model-00007-of-00062.safetensors", + "model.layers.6.mlp.down_proj.weight": "model-00007-of-00062.safetensors", + "model.layers.6.mlp.gate_proj.weight": "model-00006-of-00062.safetensors", + "model.layers.6.mlp.up_proj.weight": "model-00006-of-00062.safetensors", + "model.layers.6.post_attention_layernorm.weight": "model-00007-of-00062.safetensors", + "model.layers.6.self_attn.k_proj.weight": "model-00006-of-00062.safetensors", + "model.layers.6.self_attn.o_proj.weight": "model-00006-of-00062.safetensors", + "model.layers.6.self_attn.q_proj.weight": "model-00006-of-00062.safetensors", + "model.layers.6.self_attn.v_proj.weight": "model-00006-of-00062.safetensors", + "model.layers.60.input_layernorm.weight": "model-00047-of-00062.safetensors", + "model.layers.60.mlp.down_proj.weight": "model-00047-of-00062.safetensors", + "model.layers.60.mlp.gate_proj.weight": "model-00047-of-00062.safetensors", + "model.layers.60.mlp.up_proj.weight": "model-00047-of-00062.safetensors", + "model.layers.60.post_attention_layernorm.weight": "model-00047-of-00062.safetensors", + "model.layers.60.self_attn.k_proj.weight": "model-00046-of-00062.safetensors", + "model.layers.60.self_attn.o_proj.weight": "model-00046-of-00062.safetensors", + "model.layers.60.self_attn.q_proj.weight": "model-00046-of-00062.safetensors", + "model.layers.60.self_attn.v_proj.weight": "model-00046-of-00062.safetensors", + "model.layers.61.input_layernorm.weight": "model-00048-of-00062.safetensors", + "model.layers.61.mlp.down_proj.weight": "model-00048-of-00062.safetensors", + "model.layers.61.mlp.gate_proj.weight": "model-00047-of-00062.safetensors", + "model.layers.61.mlp.up_proj.weight": "model-00048-of-00062.safetensors", + "model.layers.61.post_attention_layernorm.weight": "model-00048-of-00062.safetensors", + "model.layers.61.self_attn.k_proj.weight": "model-00047-of-00062.safetensors", + "model.layers.61.self_attn.o_proj.weight": "model-00047-of-00062.safetensors", + "model.layers.61.self_attn.q_proj.weight": "model-00047-of-00062.safetensors", + "model.layers.61.self_attn.v_proj.weight": "model-00047-of-00062.safetensors", + "model.layers.62.input_layernorm.weight": "model-00049-of-00062.safetensors", + "model.layers.62.mlp.down_proj.weight": "model-00049-of-00062.safetensors", + "model.layers.62.mlp.gate_proj.weight": "model-00048-of-00062.safetensors", + "model.layers.62.mlp.up_proj.weight": "model-00048-of-00062.safetensors", + "model.layers.62.post_attention_layernorm.weight": "model-00049-of-00062.safetensors", + "model.layers.62.self_attn.k_proj.weight": "model-00048-of-00062.safetensors", + "model.layers.62.self_attn.o_proj.weight": "model-00048-of-00062.safetensors", + "model.layers.62.self_attn.q_proj.weight": "model-00048-of-00062.safetensors", + "model.layers.62.self_attn.v_proj.weight": "model-00048-of-00062.safetensors", + "model.layers.63.input_layernorm.weight": "model-00049-of-00062.safetensors", + "model.layers.63.mlp.down_proj.weight": "model-00049-of-00062.safetensors", + "model.layers.63.mlp.gate_proj.weight": "model-00049-of-00062.safetensors", + "model.layers.63.mlp.up_proj.weight": "model-00049-of-00062.safetensors", + "model.layers.63.post_attention_layernorm.weight": "model-00049-of-00062.safetensors", + "model.layers.63.self_attn.k_proj.weight": "model-00049-of-00062.safetensors", + "model.layers.63.self_attn.o_proj.weight": "model-00049-of-00062.safetensors", + "model.layers.63.self_attn.q_proj.weight": "model-00049-of-00062.safetensors", + "model.layers.63.self_attn.v_proj.weight": "model-00049-of-00062.safetensors", + "model.layers.64.input_layernorm.weight": "model-00050-of-00062.safetensors", + "model.layers.64.mlp.down_proj.weight": "model-00050-of-00062.safetensors", + "model.layers.64.mlp.gate_proj.weight": "model-00050-of-00062.safetensors", + "model.layers.64.mlp.up_proj.weight": "model-00050-of-00062.safetensors", + "model.layers.64.post_attention_layernorm.weight": "model-00050-of-00062.safetensors", + "model.layers.64.self_attn.k_proj.weight": "model-00049-of-00062.safetensors", + "model.layers.64.self_attn.o_proj.weight": "model-00049-of-00062.safetensors", + "model.layers.64.self_attn.q_proj.weight": "model-00049-of-00062.safetensors", + "model.layers.64.self_attn.v_proj.weight": "model-00049-of-00062.safetensors", + "model.layers.65.input_layernorm.weight": "model-00051-of-00062.safetensors", + "model.layers.65.mlp.down_proj.weight": "model-00051-of-00062.safetensors", + "model.layers.65.mlp.gate_proj.weight": "model-00050-of-00062.safetensors", + "model.layers.65.mlp.up_proj.weight": "model-00051-of-00062.safetensors", + "model.layers.65.post_attention_layernorm.weight": "model-00051-of-00062.safetensors", + "model.layers.65.self_attn.k_proj.weight": "model-00050-of-00062.safetensors", + "model.layers.65.self_attn.o_proj.weight": "model-00050-of-00062.safetensors", + "model.layers.65.self_attn.q_proj.weight": "model-00050-of-00062.safetensors", + "model.layers.65.self_attn.v_proj.weight": "model-00050-of-00062.safetensors", + "model.layers.66.input_layernorm.weight": "model-00052-of-00062.safetensors", + "model.layers.66.mlp.down_proj.weight": "model-00052-of-00062.safetensors", + "model.layers.66.mlp.gate_proj.weight": "model-00051-of-00062.safetensors", + "model.layers.66.mlp.up_proj.weight": "model-00051-of-00062.safetensors", + "model.layers.66.post_attention_layernorm.weight": "model-00052-of-00062.safetensors", + "model.layers.66.self_attn.k_proj.weight": "model-00051-of-00062.safetensors", + "model.layers.66.self_attn.o_proj.weight": "model-00051-of-00062.safetensors", + "model.layers.66.self_attn.q_proj.weight": "model-00051-of-00062.safetensors", + "model.layers.66.self_attn.v_proj.weight": "model-00051-of-00062.safetensors", + "model.layers.67.input_layernorm.weight": "model-00052-of-00062.safetensors", + "model.layers.67.mlp.down_proj.weight": "model-00052-of-00062.safetensors", + "model.layers.67.mlp.gate_proj.weight": "model-00052-of-00062.safetensors", + "model.layers.67.mlp.up_proj.weight": "model-00052-of-00062.safetensors", + "model.layers.67.post_attention_layernorm.weight": "model-00052-of-00062.safetensors", + "model.layers.67.self_attn.k_proj.weight": "model-00052-of-00062.safetensors", + "model.layers.67.self_attn.o_proj.weight": "model-00052-of-00062.safetensors", + "model.layers.67.self_attn.q_proj.weight": "model-00052-of-00062.safetensors", + "model.layers.67.self_attn.v_proj.weight": "model-00052-of-00062.safetensors", + "model.layers.68.input_layernorm.weight": "model-00053-of-00062.safetensors", + "model.layers.68.mlp.down_proj.weight": "model-00053-of-00062.safetensors", + "model.layers.68.mlp.gate_proj.weight": "model-00053-of-00062.safetensors", + "model.layers.68.mlp.up_proj.weight": "model-00053-of-00062.safetensors", + "model.layers.68.post_attention_layernorm.weight": "model-00053-of-00062.safetensors", + "model.layers.68.self_attn.k_proj.weight": "model-00052-of-00062.safetensors", + "model.layers.68.self_attn.o_proj.weight": "model-00052-of-00062.safetensors", + "model.layers.68.self_attn.q_proj.weight": "model-00052-of-00062.safetensors", + "model.layers.68.self_attn.v_proj.weight": "model-00052-of-00062.safetensors", + "model.layers.69.input_layernorm.weight": "model-00054-of-00062.safetensors", + "model.layers.69.mlp.down_proj.weight": "model-00054-of-00062.safetensors", + "model.layers.69.mlp.gate_proj.weight": "model-00053-of-00062.safetensors", + "model.layers.69.mlp.up_proj.weight": "model-00054-of-00062.safetensors", + "model.layers.69.post_attention_layernorm.weight": "model-00054-of-00062.safetensors", + "model.layers.69.self_attn.k_proj.weight": "model-00053-of-00062.safetensors", + "model.layers.69.self_attn.o_proj.weight": "model-00053-of-00062.safetensors", + "model.layers.69.self_attn.q_proj.weight": "model-00053-of-00062.safetensors", + "model.layers.69.self_attn.v_proj.weight": "model-00053-of-00062.safetensors", + "model.layers.7.input_layernorm.weight": "model-00007-of-00062.safetensors", + "model.layers.7.mlp.down_proj.weight": "model-00007-of-00062.safetensors", + "model.layers.7.mlp.gate_proj.weight": "model-00007-of-00062.safetensors", + "model.layers.7.mlp.up_proj.weight": "model-00007-of-00062.safetensors", + "model.layers.7.post_attention_layernorm.weight": "model-00007-of-00062.safetensors", + "model.layers.7.self_attn.k_proj.weight": "model-00007-of-00062.safetensors", + "model.layers.7.self_attn.o_proj.weight": "model-00007-of-00062.safetensors", + "model.layers.7.self_attn.q_proj.weight": "model-00007-of-00062.safetensors", + "model.layers.7.self_attn.v_proj.weight": "model-00007-of-00062.safetensors", + "model.layers.70.input_layernorm.weight": "model-00055-of-00062.safetensors", + "model.layers.70.mlp.down_proj.weight": "model-00055-of-00062.safetensors", + "model.layers.70.mlp.gate_proj.weight": "model-00054-of-00062.safetensors", + "model.layers.70.mlp.up_proj.weight": "model-00054-of-00062.safetensors", + "model.layers.70.post_attention_layernorm.weight": "model-00055-of-00062.safetensors", + "model.layers.70.self_attn.k_proj.weight": "model-00054-of-00062.safetensors", + "model.layers.70.self_attn.o_proj.weight": "model-00054-of-00062.safetensors", + "model.layers.70.self_attn.q_proj.weight": "model-00054-of-00062.safetensors", + "model.layers.70.self_attn.v_proj.weight": "model-00054-of-00062.safetensors", + "model.layers.71.input_layernorm.weight": "model-00055-of-00062.safetensors", + "model.layers.71.mlp.down_proj.weight": "model-00055-of-00062.safetensors", + "model.layers.71.mlp.gate_proj.weight": "model-00055-of-00062.safetensors", + "model.layers.71.mlp.up_proj.weight": "model-00055-of-00062.safetensors", + "model.layers.71.post_attention_layernorm.weight": "model-00055-of-00062.safetensors", + "model.layers.71.self_attn.k_proj.weight": "model-00055-of-00062.safetensors", + "model.layers.71.self_attn.o_proj.weight": "model-00055-of-00062.safetensors", + "model.layers.71.self_attn.q_proj.weight": "model-00055-of-00062.safetensors", + "model.layers.71.self_attn.v_proj.weight": "model-00055-of-00062.safetensors", + "model.layers.72.input_layernorm.weight": "model-00056-of-00062.safetensors", + "model.layers.72.mlp.down_proj.weight": "model-00056-of-00062.safetensors", + "model.layers.72.mlp.gate_proj.weight": "model-00056-of-00062.safetensors", + "model.layers.72.mlp.up_proj.weight": "model-00056-of-00062.safetensors", + "model.layers.72.post_attention_layernorm.weight": "model-00056-of-00062.safetensors", + "model.layers.72.self_attn.k_proj.weight": "model-00055-of-00062.safetensors", + "model.layers.72.self_attn.o_proj.weight": "model-00055-of-00062.safetensors", + "model.layers.72.self_attn.q_proj.weight": "model-00055-of-00062.safetensors", + "model.layers.72.self_attn.v_proj.weight": "model-00055-of-00062.safetensors", + "model.layers.73.input_layernorm.weight": "model-00057-of-00062.safetensors", + "model.layers.73.mlp.down_proj.weight": "model-00057-of-00062.safetensors", + "model.layers.73.mlp.gate_proj.weight": "model-00056-of-00062.safetensors", + "model.layers.73.mlp.up_proj.weight": "model-00057-of-00062.safetensors", + "model.layers.73.post_attention_layernorm.weight": "model-00057-of-00062.safetensors", + "model.layers.73.self_attn.k_proj.weight": "model-00056-of-00062.safetensors", + "model.layers.73.self_attn.o_proj.weight": "model-00056-of-00062.safetensors", + "model.layers.73.self_attn.q_proj.weight": "model-00056-of-00062.safetensors", + "model.layers.73.self_attn.v_proj.weight": "model-00056-of-00062.safetensors", + "model.layers.74.input_layernorm.weight": "model-00058-of-00062.safetensors", + "model.layers.74.mlp.down_proj.weight": "model-00058-of-00062.safetensors", + "model.layers.74.mlp.gate_proj.weight": "model-00057-of-00062.safetensors", + "model.layers.74.mlp.up_proj.weight": "model-00057-of-00062.safetensors", + "model.layers.74.post_attention_layernorm.weight": "model-00058-of-00062.safetensors", + "model.layers.74.self_attn.k_proj.weight": "model-00057-of-00062.safetensors", + "model.layers.74.self_attn.o_proj.weight": "model-00057-of-00062.safetensors", + "model.layers.74.self_attn.q_proj.weight": "model-00057-of-00062.safetensors", + "model.layers.74.self_attn.v_proj.weight": "model-00057-of-00062.safetensors", + "model.layers.75.input_layernorm.weight": "model-00058-of-00062.safetensors", + "model.layers.75.mlp.down_proj.weight": "model-00058-of-00062.safetensors", + "model.layers.75.mlp.gate_proj.weight": "model-00058-of-00062.safetensors", + "model.layers.75.mlp.up_proj.weight": "model-00058-of-00062.safetensors", + "model.layers.75.post_attention_layernorm.weight": "model-00058-of-00062.safetensors", + "model.layers.75.self_attn.k_proj.weight": "model-00058-of-00062.safetensors", + "model.layers.75.self_attn.o_proj.weight": "model-00058-of-00062.safetensors", + "model.layers.75.self_attn.q_proj.weight": "model-00058-of-00062.safetensors", + "model.layers.75.self_attn.v_proj.weight": "model-00058-of-00062.safetensors", + "model.layers.76.input_layernorm.weight": "model-00059-of-00062.safetensors", + "model.layers.76.mlp.down_proj.weight": "model-00059-of-00062.safetensors", + "model.layers.76.mlp.gate_proj.weight": "model-00059-of-00062.safetensors", + "model.layers.76.mlp.up_proj.weight": "model-00059-of-00062.safetensors", + "model.layers.76.post_attention_layernorm.weight": "model-00059-of-00062.safetensors", + "model.layers.76.self_attn.k_proj.weight": "model-00058-of-00062.safetensors", + "model.layers.76.self_attn.o_proj.weight": "model-00058-of-00062.safetensors", + "model.layers.76.self_attn.q_proj.weight": "model-00058-of-00062.safetensors", + "model.layers.76.self_attn.v_proj.weight": "model-00058-of-00062.safetensors", + "model.layers.77.input_layernorm.weight": "model-00060-of-00062.safetensors", + "model.layers.77.mlp.down_proj.weight": "model-00060-of-00062.safetensors", + "model.layers.77.mlp.gate_proj.weight": "model-00059-of-00062.safetensors", + "model.layers.77.mlp.up_proj.weight": "model-00060-of-00062.safetensors", + "model.layers.77.post_attention_layernorm.weight": "model-00060-of-00062.safetensors", + "model.layers.77.self_attn.k_proj.weight": "model-00059-of-00062.safetensors", + "model.layers.77.self_attn.o_proj.weight": "model-00059-of-00062.safetensors", + "model.layers.77.self_attn.q_proj.weight": "model-00059-of-00062.safetensors", + "model.layers.77.self_attn.v_proj.weight": "model-00059-of-00062.safetensors", + "model.layers.78.input_layernorm.weight": "model-00061-of-00062.safetensors", + "model.layers.78.mlp.down_proj.weight": "model-00061-of-00062.safetensors", + "model.layers.78.mlp.gate_proj.weight": "model-00060-of-00062.safetensors", + "model.layers.78.mlp.up_proj.weight": "model-00060-of-00062.safetensors", + "model.layers.78.post_attention_layernorm.weight": "model-00061-of-00062.safetensors", + "model.layers.78.self_attn.k_proj.weight": "model-00060-of-00062.safetensors", + "model.layers.78.self_attn.o_proj.weight": "model-00060-of-00062.safetensors", + "model.layers.78.self_attn.q_proj.weight": "model-00060-of-00062.safetensors", + "model.layers.78.self_attn.v_proj.weight": "model-00060-of-00062.safetensors", + "model.layers.79.input_layernorm.weight": "model-00061-of-00062.safetensors", + "model.layers.79.mlp.down_proj.weight": "model-00061-of-00062.safetensors", + "model.layers.79.mlp.gate_proj.weight": "model-00061-of-00062.safetensors", + "model.layers.79.mlp.up_proj.weight": "model-00061-of-00062.safetensors", + "model.layers.79.post_attention_layernorm.weight": "model-00061-of-00062.safetensors", + "model.layers.79.self_attn.k_proj.weight": "model-00061-of-00062.safetensors", + "model.layers.79.self_attn.o_proj.weight": "model-00061-of-00062.safetensors", + "model.layers.79.self_attn.q_proj.weight": "model-00061-of-00062.safetensors", + "model.layers.79.self_attn.v_proj.weight": "model-00061-of-00062.safetensors", + "model.layers.8.input_layernorm.weight": "model-00008-of-00062.safetensors", + "model.layers.8.mlp.down_proj.weight": "model-00008-of-00062.safetensors", + "model.layers.8.mlp.gate_proj.weight": "model-00008-of-00062.safetensors", + "model.layers.8.mlp.up_proj.weight": "model-00008-of-00062.safetensors", + "model.layers.8.post_attention_layernorm.weight": "model-00008-of-00062.safetensors", + "model.layers.8.self_attn.k_proj.weight": "model-00007-of-00062.safetensors", + "model.layers.8.self_attn.o_proj.weight": "model-00007-of-00062.safetensors", + "model.layers.8.self_attn.q_proj.weight": "model-00007-of-00062.safetensors", + "model.layers.8.self_attn.v_proj.weight": "model-00007-of-00062.safetensors", + "model.layers.9.input_layernorm.weight": "model-00009-of-00062.safetensors", + "model.layers.9.mlp.down_proj.weight": "model-00009-of-00062.safetensors", + "model.layers.9.mlp.gate_proj.weight": "model-00008-of-00062.safetensors", + "model.layers.9.mlp.up_proj.weight": "model-00009-of-00062.safetensors", + "model.layers.9.post_attention_layernorm.weight": "model-00009-of-00062.safetensors", + "model.layers.9.self_attn.k_proj.weight": "model-00008-of-00062.safetensors", + "model.layers.9.self_attn.o_proj.weight": "model-00008-of-00062.safetensors", + "model.layers.9.self_attn.q_proj.weight": "model-00008-of-00062.safetensors", + "model.layers.9.self_attn.v_proj.weight": "model-00008-of-00062.safetensors", + "model.norm.weight": "model-00061-of-00062.safetensors" + } +}