diff --git a/README.md b/README.md new file mode 100644 index 0000000000000000000000000000000000000000..bc5f30d6632ac0efdc7be2e9095e9e9579af2e33 --- /dev/null +++ b/README.md @@ -0,0 +1,199 @@ +--- +library_name: transformers +tags: [] +--- + +# Model Card for Model ID + + + + + +## Model Details + +### Model Description + + + +This is the model card of a 🤗 transformers model that has been pushed on the Hub. This model card has been automatically generated. + +- **Developed by:** [More Information Needed] +- **Funded by [optional]:** [More Information Needed] +- **Shared by [optional]:** [More Information Needed] +- **Model type:** [More Information Needed] +- **Language(s) (NLP):** [More Information Needed] +- **License:** [More Information Needed] +- **Finetuned from model [optional]:** [More Information Needed] + +### Model Sources [optional] + + + +- **Repository:** [More Information Needed] +- **Paper [optional]:** [More Information Needed] +- **Demo [optional]:** [More Information Needed] + +## Uses + + + +### Direct Use + + + +[More Information Needed] + +### Downstream Use [optional] + + + +[More Information Needed] + +### Out-of-Scope Use + + + +[More Information Needed] + +## Bias, Risks, and Limitations + + + +[More Information Needed] + +### Recommendations + + + +Users (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations. + +## How to Get Started with the Model + +Use the code below to get started with the model. + +[More Information Needed] + +## Training Details + +### Training Data + + + +[More Information Needed] + +### Training Procedure + + + +#### Preprocessing [optional] + +[More Information Needed] + + +#### Training Hyperparameters + +- **Training regime:** [More Information Needed] + +#### Speeds, Sizes, Times [optional] + + + +[More Information Needed] + +## Evaluation + + + +### Testing Data, Factors & Metrics + +#### Testing Data + + + +[More Information Needed] + +#### Factors + + + +[More Information Needed] + +#### Metrics + + + +[More Information Needed] + +### Results + +[More Information Needed] + +#### Summary + + + +## Model Examination [optional] + + + +[More Information Needed] + +## Environmental Impact + + + +Carbon emissions can be estimated using the [Machine Learning Impact calculator](https://mlco2.github.io/impact#compute) presented in [Lacoste et al. (2019)](https://arxiv.org/abs/1910.09700). + +- **Hardware Type:** [More Information Needed] +- **Hours used:** [More Information Needed] +- **Cloud Provider:** [More Information Needed] +- **Compute Region:** [More Information Needed] +- **Carbon Emitted:** [More Information Needed] + +## Technical Specifications [optional] + +### Model Architecture and Objective + +[More Information Needed] + +### Compute Infrastructure + +[More Information Needed] + +#### Hardware + +[More Information Needed] + +#### Software + +[More Information Needed] + +## Citation [optional] + + + +**BibTeX:** + +[More Information Needed] + +**APA:** + +[More Information Needed] + +## Glossary [optional] + + + +[More Information Needed] + +## More Information [optional] + +[More Information Needed] + +## Model Card Authors [optional] + +[More Information Needed] + +## Model Card Contact + +[More Information Needed] \ No newline at end of file diff --git a/config.json b/config.json new file mode 100644 index 0000000000000000000000000000000000000000..0069012e61e54cdd8edcf1232ec3a15a62e92360 --- /dev/null +++ b/config.json @@ -0,0 +1,39 @@ +{ + "architectures": [ + "LlamaForCausalLM" + ], + "attention_bias": false, + "attention_dropout": 0.0, + "bos_token_id": 128000, + "eos_token_id": [ + 128001, + 128008, + 128009 + ], + "head_dim": 128, + "hidden_act": "silu", + "hidden_size": 8192, + "initializer_range": 0.02, + "intermediate_size": 28672, + "max_position_embeddings": 131072, + "mlp_bias": false, + "model_type": "llama", + "num_attention_heads": 64, + "num_hidden_layers": 80, + "num_key_value_heads": 8, + "pretraining_tp": 1, + "rms_norm_eps": 1e-05, + "rope_scaling": { + "factor": 8.0, + "high_freq_factor": 4.0, + "low_freq_factor": 1.0, + "original_max_position_embeddings": 8192, + "rope_type": "llama3" + }, + "rope_theta": 500000.0, + "tie_word_embeddings": false, + "torch_dtype": "float32", + "transformers_version": "4.54.1", + "use_cache": true, + "vocab_size": 128256 +} diff --git a/generation_config.json b/generation_config.json new file mode 100644 index 0000000000000000000000000000000000000000..fb922075a9bf1e542ba674d5b5ed52a4c17fe46b --- /dev/null +++ b/generation_config.json @@ -0,0 +1,12 @@ +{ + "bos_token_id": 128000, + "do_sample": true, + "eos_token_id": [ + 128001, + 128008, + 128009 + ], + "temperature": 0.6, + "top_p": 0.9, + "transformers_version": "4.54.1" +} diff --git a/model-00001-of-00062.safetensors b/model-00001-of-00062.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..fe0d97a059483d2f8e944f143777883e508b22c6 --- /dev/null +++ b/model-00001-of-00062.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:50be45abf21b9ae565bae8f3db79acf001894dfd691b61e53d8abd47cdf3030f +size 4806672984 diff --git a/model-00002-of-00062.safetensors b/model-00002-of-00062.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..5d243e61da88e83fe500d10473b7653327f3e7c6 --- /dev/null +++ b/model-00002-of-00062.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:08476ac6ba1e5e6dfef781e7dc0aca8a8fd951b65f06b380d282c4221b585785 +size 4362142864 diff --git a/model-00003-of-00062.safetensors b/model-00003-of-00062.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..b6a17d73d0c5b4733319210c09a6d1fdd94ab379 --- /dev/null +++ b/model-00003-of-00062.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:c01a61a3fa7e65fcee4dfc15b8b470272dc52dd637befb73c763cdfa7cbe3a50 +size 4362142864 diff --git a/model-00004-of-00062.safetensors b/model-00004-of-00062.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..937a550eabb178d9b62e8dc9c998372d496037ac --- /dev/null +++ b/model-00004-of-00062.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:496e4b221a808a512a159b220ddc43761ca5ef5f9e2b0633394e497069c5e6e0 +size 4966188864 diff --git a/model-00005-of-00062.safetensors b/model-00005-of-00062.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..5a0bb9130ffef48de436baea02791cd68c17a05e --- /dev/null +++ b/model-00005-of-00062.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:b3fbc625afea94e4e11ee2be3ae097544635a6dcd557d16578003789bd50adcd +size 4362142864 diff --git a/model-00006-of-00062.safetensors b/model-00006-of-00062.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..aee72701b3bcaa508fbc4a87e9c293ed1a8f5b73 --- /dev/null +++ b/model-00006-of-00062.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:55594b4b6b46561cfab9517aed711b908c147c8e9d1ec1ad6cd478137ed581a5 +size 4362142864 diff --git a/model-00007-of-00062.safetensors b/model-00007-of-00062.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..c087a7b631b77a007f704daa4f7e02936cae7525 --- /dev/null +++ b/model-00007-of-00062.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:c4e38db4c99b7620e057ab2e54e3d527c3306ae3929f7fcd8b3ea5066b5464fd +size 4966188864 diff --git a/model-00008-of-00062.safetensors b/model-00008-of-00062.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..a478ddebcbb6e41bd66980a71b237028f46c031b --- /dev/null +++ b/model-00008-of-00062.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:e1119d3e364c392d54f8558db90621a4d49b99420ff25c23a1f3566b993470e6 +size 4362142864 diff --git a/model-00009-of-00062.safetensors b/model-00009-of-00062.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..9a96a18a5282feb54c8969c2c4d84d8c903f40bc --- /dev/null +++ b/model-00009-of-00062.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:c2f2525057363326d5f0ecc892485cfe998c7303ec3d4b6ab8d7775572c838ef +size 4362142880 diff --git a/model-00010-of-00062.safetensors b/model-00010-of-00062.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..bb23cb785dab5818450c39e6eaeb931fe1ce346b --- /dev/null +++ b/model-00010-of-00062.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:c4f9d930eeb358368b3b9f419b4bb7b8fae4148ef318532d4a509b6ed9d6aa01 +size 4966188880 diff --git a/model-00011-of-00062.safetensors b/model-00011-of-00062.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..88468eca101b4fa72d2c79bb9600375937995f03 --- /dev/null +++ b/model-00011-of-00062.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:4724c884fc7c778ea3b84b00c845e4097e11d96e524dcafbbe93c77f5e6e7fa0 +size 4362142872 diff --git a/model-00012-of-00062.safetensors b/model-00012-of-00062.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..a317055f298b802f624af800d7fba041ec7f833e --- /dev/null +++ b/model-00012-of-00062.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:39a14c55f32920a95185fb0b325f7f9896cc7272b67acaf3355cbf6ed5f0a9c4 +size 4362142872 diff --git a/model-00013-of-00062.safetensors b/model-00013-of-00062.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..42cf03394404797f869cd611d510de6b878f9040 --- /dev/null +++ b/model-00013-of-00062.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:1557d4fa5cde2cfad8e87a1d3ef2084f8291649f9e71af7a63b1cf37908284c5 +size 4966188880 diff --git a/model-00014-of-00062.safetensors b/model-00014-of-00062.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..2628b4a52fb877e9d78272b928c3c664d84c6a09 --- /dev/null +++ b/model-00014-of-00062.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:625dffb1957636f1e6f8b1a763264dae740bd6e6dc61e936cf239000ee2d66b9 +size 4362142872 diff --git a/model-00015-of-00062.safetensors b/model-00015-of-00062.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..35db7d101a87160558222f5e5d495df5aceaf116 --- /dev/null +++ b/model-00015-of-00062.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:db868145679816faaac3673e1f61ef118558d0139b702adc6da6c635af22dca5 +size 4362142872 diff --git a/model-00016-of-00062.safetensors b/model-00016-of-00062.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..bf12b1973fe4359339b4a87ead171d56604b73a6 --- /dev/null +++ b/model-00016-of-00062.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:34c9f7fb77ddc0924321718b0288efbe576785e85fb923cb9638ed0a8f8b9ddc +size 4966188880 diff --git a/model-00017-of-00062.safetensors b/model-00017-of-00062.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..3c5975b93ce9f5970e74a2e9d29b7f0ad3eed4d5 --- /dev/null +++ b/model-00017-of-00062.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:46affa010e7a43c99e9f950b4a168d7097fcc8c2f60543fed79516e6ed199fde +size 4362142872 diff --git a/model-00018-of-00062.safetensors b/model-00018-of-00062.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..ebfb74316abe99555cf045c51fbefdebe4897e79 --- /dev/null +++ b/model-00018-of-00062.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:4e9d5317c6333872dca4092f839a9e3fdc93a8847f08d68e38a006d130b6f75f +size 4362142872 diff --git a/model-00019-of-00062.safetensors b/model-00019-of-00062.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..c15596444f4805ec137820cefded61ef8cd3dac7 --- /dev/null +++ b/model-00019-of-00062.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:bcff59c10d5601f35765970b552c0f56a94e67352cd725dd4de6ab5139792ead +size 4966188880 diff --git a/model-00020-of-00062.safetensors b/model-00020-of-00062.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..1475e486c0194ca91089259e11d28e890de39c90 --- /dev/null +++ b/model-00020-of-00062.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:c37ebf4309b58ba8f2e7322de62eeec1f8f2da0784abd2e021619b203cc74184 +size 4362142872 diff --git a/model-00021-of-00062.safetensors b/model-00021-of-00062.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..ee0bad94f0bdd387fe6dae985001ba2afaed98bc --- /dev/null +++ b/model-00021-of-00062.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:c1c190cce86354c955978fa67fc6b354b822c35237d3cee2fdf5016327992a80 +size 4362142872 diff --git a/model-00022-of-00062.safetensors b/model-00022-of-00062.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..4b82d5b832a4bf252a3f0213a27ec47a3a900fa8 --- /dev/null +++ b/model-00022-of-00062.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:47aa3a0c0989dfe590814ef7bf6a736c834dd3cfbab6bc7a1275114a45f7b2c5 +size 4966188880 diff --git a/model-00023-of-00062.safetensors b/model-00023-of-00062.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..a8c29b3bebd5d26e0ebdcd3aa361add94df2a771 --- /dev/null +++ b/model-00023-of-00062.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:46fb2872c53d919f4bad44971c452acee088da380aeaa0287300891d676d486f +size 4362142872 diff --git a/model-00024-of-00062.safetensors b/model-00024-of-00062.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..641e0971cec6e2d0dc3fdfdcf80066ba8800b45f --- /dev/null +++ b/model-00024-of-00062.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:c3ab5bf7b35acd4dd475e1b8bcaa26a3a91fd48e9491d45d423c8fa2f7e1550c +size 4362142872 diff --git a/model-00025-of-00062.safetensors b/model-00025-of-00062.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..8c660a295f0ae39a7e7ea37e5d768679238a1f47 --- /dev/null +++ b/model-00025-of-00062.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:9cdb536be5008c56d1260ff191d2158f58297ca5abb9cc657048ff0897a50a84 +size 4966188880 diff --git a/model-00026-of-00062.safetensors b/model-00026-of-00062.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..1de57f842c8ee8a5a6a9f981bd34ff6bfaeea9db --- /dev/null +++ b/model-00026-of-00062.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:e6c46a2790c45afd9d27af60be2c5ae9868a095c5bad34865e302b848fcee9b4 +size 4362142872 diff --git a/model-00027-of-00062.safetensors b/model-00027-of-00062.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..e645bc17287ac8cc462af2528472dd64bbe304d5 --- /dev/null +++ b/model-00027-of-00062.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:0306fdd13a897f53b3d006cf0aaaf18e54ca8e527d2000ec773b38a529cf077a +size 4362142872 diff --git a/model-00028-of-00062.safetensors b/model-00028-of-00062.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..d0b5927daa298b6765df1400b59184ecd4f853ef --- /dev/null +++ b/model-00028-of-00062.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:05a8ab6761233e673cebf64e00ffe352ec7f2318345d3096b042315c35db2c0b +size 4966188880 diff --git a/model-00029-of-00062.safetensors b/model-00029-of-00062.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..048ef0073b09dc3ee9e1c62991ff7c1ba5096a69 --- /dev/null +++ b/model-00029-of-00062.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:72c44e73c861b13e0a6c1a9c2b4fdb6b2ac877f6f37fca3a7eee703c7a56a28e +size 4362142872 diff --git a/model-00030-of-00062.safetensors b/model-00030-of-00062.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..f09f2bbd7ba55cec63f9872413e01041e4d83e9c --- /dev/null +++ b/model-00030-of-00062.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:068fa6d15a86661fc4987da3dcbd746423d1ec21dd1a72dd619067ad6b332370 +size 4362142872 diff --git a/model-00031-of-00062.safetensors b/model-00031-of-00062.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..bc1d349d49600c7706f4770c6ffdcd83ac0afc01 --- /dev/null +++ b/model-00031-of-00062.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:ea96521b13f13bae4ef91893db0d6a4afa9000c4eab38f4f361ef0dbd1273f72 +size 4966188880 diff --git a/model-00032-of-00062.safetensors b/model-00032-of-00062.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..6194697d89ec0566b43207f774d56f7efc00f2d2 --- /dev/null +++ b/model-00032-of-00062.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:ecaf0bd5094a46e6fe52411cfb32c146c357f2a4d18b678324d1511b140b4e38 +size 4362142872 diff --git a/model-00033-of-00062.safetensors b/model-00033-of-00062.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..9ff2323fba6e9c55f85bbcd7a96682e472aa7b60 --- /dev/null +++ b/model-00033-of-00062.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:633160abe541852fba1b3b930ad493b02178e02769d9cb95b02d30179395a29b +size 4362142872 diff --git a/model-00034-of-00062.safetensors b/model-00034-of-00062.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..a791d0a8d50555ef3a8faf0a121d9d16a1755608 --- /dev/null +++ b/model-00034-of-00062.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:7791fe732d64d2bb595aff3b6075fbd1d6b7298a9ffc591b1018b4cd7dfb4612 +size 4966188880 diff --git a/model-00035-of-00062.safetensors b/model-00035-of-00062.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..009991496b3428bc0aa787094f4a080fb7ae92a8 --- /dev/null +++ b/model-00035-of-00062.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:ed7d061d84963f52e1aae97b987c80d847a6811f79106d4d2e9ad7564b582820 +size 4362142872 diff --git a/model-00036-of-00062.safetensors b/model-00036-of-00062.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..28464dae6c99f05f9ef79f8de69947afc2f9ce34 --- /dev/null +++ b/model-00036-of-00062.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:4bc489514798c83d5da1befc02decc1459a06b64a23c631ef9f53f999b48d555 +size 4362142872 diff --git a/model-00037-of-00062.safetensors b/model-00037-of-00062.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..2379d5a392375f66b29086110fe1ef0eda6776e4 --- /dev/null +++ b/model-00037-of-00062.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:3b5d88a3fa1145937304a4a10d92c58b84f497b68e0645366a35e41d7455ed6f +size 4966188880 diff --git a/model-00038-of-00062.safetensors b/model-00038-of-00062.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..44962775c688e16b142a7e11a7aeb1ed3e366365 --- /dev/null +++ b/model-00038-of-00062.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:d79c8612b7d5d43a8fd37ac06a64482ad6712c901090b1297181449d44659ede +size 4362142872 diff --git a/model-00039-of-00062.safetensors b/model-00039-of-00062.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..d70898edb52669f98c60f288033d07d5a3857017 --- /dev/null +++ b/model-00039-of-00062.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:6d895f25fb9880517389eb9457941eef8c1c7e71d14ecee71d00200e8fe117ea +size 4362142872 diff --git a/model-00040-of-00062.safetensors b/model-00040-of-00062.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..a9db8d1f38f90ccb66ce6502651394133778a70f --- /dev/null +++ b/model-00040-of-00062.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:99a90ab4bcb5195a08b79f4daa4e4d0d12d245d2ffc712f8e43a912ec55b6f81 +size 4966188880 diff --git a/model-00041-of-00062.safetensors b/model-00041-of-00062.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..8aefacc700a5f15d6a7518a67175d5a65ff7913e --- /dev/null +++ b/model-00041-of-00062.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:503805016e5f095b0bbc6cb024a85e58cc9f780f3a9e3c3819880c4889bc1a0a +size 4362142872 diff --git a/model-00042-of-00062.safetensors b/model-00042-of-00062.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..a318d9c0be2974b1f55a586def2de775237cfea9 --- /dev/null +++ b/model-00042-of-00062.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:7f53d909b1ff655069e792f1b4909d6375640eb8f95e8f5271864de49fbc56d2 +size 4362142872 diff --git a/model-00043-of-00062.safetensors b/model-00043-of-00062.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..79689010c502169a257e5471677030dc130b6375 --- /dev/null +++ b/model-00043-of-00062.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:f6b101cec1b6220c5e8456980af69a21a20a6880e74c739e8ff9d2ec51b8a71a +size 4966188880 diff --git a/model-00044-of-00062.safetensors b/model-00044-of-00062.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..e20666b5ca99071bcd418a3cf8c7228c46b8ae91 --- /dev/null +++ b/model-00044-of-00062.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:fbfe2386b8d5771dd21adcc05724e7bc08ecefa87876f7cf9f8d762393cf56de +size 4362142872 diff --git a/model-00045-of-00062.safetensors b/model-00045-of-00062.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..f0819a84bba8dee202a587f834ac148e19287c95 --- /dev/null +++ b/model-00045-of-00062.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:24f493134662d373ff6f3ef438e7aa9d34b89af0c1ab3a351b9befb619ee332e +size 4362142872 diff --git a/model-00046-of-00062.safetensors b/model-00046-of-00062.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..120ac9e1f2e43390cd51068bf7b25183b9bec3c9 --- /dev/null +++ b/model-00046-of-00062.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:4fdb5e551d905a566ac3484f08444f03aea5646ed7c5012c8a4f580c067adf75 +size 4966188880 diff --git a/model-00047-of-00062.safetensors b/model-00047-of-00062.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..f6beeade1380d35722884a68e86523d199862ceb --- /dev/null +++ b/model-00047-of-00062.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:60b4b267dd55f3c9ce5db0b183c0f4be93fbeef81d60757a40cf4e5880fa4581 +size 4362142872 diff --git a/model-00048-of-00062.safetensors b/model-00048-of-00062.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..26631ece6d9d93cb91122377f625b1cd72700c00 --- /dev/null +++ b/model-00048-of-00062.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:1dd821c75bfc5823bcc927170b2a67f8142a492761d95411b137cd384ca06a83 +size 4362142872 diff --git a/model-00049-of-00062.safetensors b/model-00049-of-00062.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..5c34e3d1ba0402e918fde084c3c9c098c4b71101 --- /dev/null +++ b/model-00049-of-00062.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:0dcdf80ba8030624d2b6c7969093dea0fceb6dcd98f91ec6dcb1995292413964 +size 4966188880 diff --git a/model-00050-of-00062.safetensors b/model-00050-of-00062.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..008ca5834df35c98318441ae526079097ea2a1bc --- /dev/null +++ b/model-00050-of-00062.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:00eaff4b80a8e18c514cbcd397c7d99ff57714620ec986b08642bab07bb0160b +size 4362142872 diff --git a/model-00051-of-00062.safetensors b/model-00051-of-00062.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..dc0549a377d5029bf1d2332029835cae5bbc80eb --- /dev/null +++ b/model-00051-of-00062.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:7b53dd688a418fb220223bbd2147f265a0a47c7f0b13f7686614073d7b46e3e6 +size 4362142872 diff --git a/model-00052-of-00062.safetensors b/model-00052-of-00062.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..46dd20bf6da7e0f250769a7c53219c24725ff3d7 --- /dev/null +++ b/model-00052-of-00062.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:97d288729490c3794d43e03ac5a861cdb32a20abf06d450900f490e0508bbddd +size 4966188880 diff --git a/model-00053-of-00062.safetensors b/model-00053-of-00062.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..b76b1dec7cbc6babae97f58766e75581bc10bb27 --- /dev/null +++ b/model-00053-of-00062.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:2415f93f8957534db5a4643b9bc432ac85ddb6aab71867dd1fd03e8f16f38614 +size 4362142872 diff --git a/model-00054-of-00062.safetensors b/model-00054-of-00062.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..e10841cd57f23f25a5a8c555a9a4f7b4624902d7 --- /dev/null +++ b/model-00054-of-00062.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:a5ceef61380cb8b35813021c6c48c1c8cfe67d27269d841046246e8b5d8bf350 +size 4362142872 diff --git a/model-00055-of-00062.safetensors b/model-00055-of-00062.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..384b71a76bb7003203f1f32287f8fee30fb46e09 --- /dev/null +++ b/model-00055-of-00062.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:dcd45edb83221c0a3548f700484e6df81e5835d21e6de292a0d2c88a869a0986 +size 4966188880 diff --git a/model-00056-of-00062.safetensors b/model-00056-of-00062.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..f48df330afda5df2a3fa8dda3001614ac060fccc --- /dev/null +++ b/model-00056-of-00062.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:99338b60fb087972f77d4cb19c870f177e34e325b84679bd2d932bde57d89c55 +size 4362142872 diff --git a/model-00057-of-00062.safetensors b/model-00057-of-00062.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..ad7c80ce50dfe8147668a4c72b8b2486b5ea5490 --- /dev/null +++ b/model-00057-of-00062.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:059036a75a336b74588bd2e7bf9303e8dfc70b051985ce79741f0c45680d75a0 +size 4362142872 diff --git a/model-00058-of-00062.safetensors b/model-00058-of-00062.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..4a84db2e457d9cfb58d57b88ebbc203006c86111 --- /dev/null +++ b/model-00058-of-00062.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:7ac5c50bcd0039d537a6311229c7cfc08ffd9f63141a795c796f4f8db66fac4d +size 4966188880 diff --git a/model-00059-of-00062.safetensors b/model-00059-of-00062.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..73d0c853c917641a745e86ccdd31691dd59ce755 --- /dev/null +++ b/model-00059-of-00062.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:2db07e44c127b8f8f12a3dad50ba35207fad4ce6030ee94b9cdfd78491175193 +size 4362142872 diff --git a/model-00060-of-00062.safetensors b/model-00060-of-00062.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..e42b15aca72bf8d387b6f6ecaf5ab19758200ffe --- /dev/null +++ b/model-00060-of-00062.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:d16d9316cce8d427b456f2bbec353f4cf1b70d369c6fc716991274fe9490db48 +size 4362142872 diff --git a/model-00061-of-00062.safetensors b/model-00061-of-00062.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..ed8c33fc3f7a2eeb61828f44b68b9d26ff7f0639 --- /dev/null +++ b/model-00061-of-00062.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:f2a353bdd7d9505080ca32dec56ebd5e310f151fde49c7eb87814aff755ce1de +size 4362241496 diff --git a/model-00062-of-00062.safetensors b/model-00062-of-00062.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..908a6bd1df871d81d23f41bd765c1d5c2abf525a --- /dev/null +++ b/model-00062-of-00062.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:37adba49b605e4b59b81ef2930475973ff37b20bd4ddd0fbc20c302fc3c112d5 +size 4202692736 diff --git a/model.safetensors.index.json b/model.safetensors.index.json new file mode 100644 index 0000000000000000000000000000000000000000..ef0c642b145f0661c97c573578ea5f5c868b5c19 --- /dev/null +++ b/model.safetensors.index.json @@ -0,0 +1,731 @@ +{ + "metadata": { + "total_parameters": 70553706496, + "total_size": 282214825984 + }, + "weight_map": { + "lm_head.weight": "model-00062-of-00062.safetensors", + "model.embed_tokens.weight": "model-00001-of-00062.safetensors", + "model.layers.0.input_layernorm.weight": "model-00002-of-00062.safetensors", + "model.layers.0.mlp.down_proj.weight": "model-00002-of-00062.safetensors", + "model.layers.0.mlp.gate_proj.weight": "model-00002-of-00062.safetensors", + "model.layers.0.mlp.up_proj.weight": "model-00002-of-00062.safetensors", + "model.layers.0.post_attention_layernorm.weight": "model-00002-of-00062.safetensors", + "model.layers.0.self_attn.k_proj.weight": "model-00001-of-00062.safetensors", + "model.layers.0.self_attn.o_proj.weight": "model-00001-of-00062.safetensors", + "model.layers.0.self_attn.q_proj.weight": "model-00001-of-00062.safetensors", + "model.layers.0.self_attn.v_proj.weight": "model-00001-of-00062.safetensors", + "model.layers.1.input_layernorm.weight": "model-00003-of-00062.safetensors", + "model.layers.1.mlp.down_proj.weight": "model-00003-of-00062.safetensors", + "model.layers.1.mlp.gate_proj.weight": "model-00002-of-00062.safetensors", + "model.layers.1.mlp.up_proj.weight": "model-00003-of-00062.safetensors", + "model.layers.1.post_attention_layernorm.weight": "model-00003-of-00062.safetensors", + "model.layers.1.self_attn.k_proj.weight": "model-00002-of-00062.safetensors", + "model.layers.1.self_attn.o_proj.weight": "model-00002-of-00062.safetensors", + "model.layers.1.self_attn.q_proj.weight": "model-00002-of-00062.safetensors", + "model.layers.1.self_attn.v_proj.weight": "model-00002-of-00062.safetensors", + "model.layers.10.input_layernorm.weight": "model-00010-of-00062.safetensors", + "model.layers.10.mlp.down_proj.weight": "model-00010-of-00062.safetensors", + "model.layers.10.mlp.gate_proj.weight": "model-00009-of-00062.safetensors", + "model.layers.10.mlp.up_proj.weight": "model-00009-of-00062.safetensors", + "model.layers.10.post_attention_layernorm.weight": "model-00010-of-00062.safetensors", + "model.layers.10.self_attn.k_proj.weight": "model-00009-of-00062.safetensors", + "model.layers.10.self_attn.o_proj.weight": "model-00009-of-00062.safetensors", + "model.layers.10.self_attn.q_proj.weight": "model-00009-of-00062.safetensors", + "model.layers.10.self_attn.v_proj.weight": "model-00009-of-00062.safetensors", + "model.layers.11.input_layernorm.weight": "model-00010-of-00062.safetensors", + "model.layers.11.mlp.down_proj.weight": "model-00010-of-00062.safetensors", + "model.layers.11.mlp.gate_proj.weight": "model-00010-of-00062.safetensors", + "model.layers.11.mlp.up_proj.weight": "model-00010-of-00062.safetensors", + "model.layers.11.post_attention_layernorm.weight": "model-00010-of-00062.safetensors", + "model.layers.11.self_attn.k_proj.weight": "model-00010-of-00062.safetensors", + "model.layers.11.self_attn.o_proj.weight": "model-00010-of-00062.safetensors", + "model.layers.11.self_attn.q_proj.weight": "model-00010-of-00062.safetensors", + "model.layers.11.self_attn.v_proj.weight": "model-00010-of-00062.safetensors", + "model.layers.12.input_layernorm.weight": "model-00011-of-00062.safetensors", + "model.layers.12.mlp.down_proj.weight": "model-00011-of-00062.safetensors", + "model.layers.12.mlp.gate_proj.weight": "model-00011-of-00062.safetensors", + "model.layers.12.mlp.up_proj.weight": "model-00011-of-00062.safetensors", + "model.layers.12.post_attention_layernorm.weight": "model-00011-of-00062.safetensors", + "model.layers.12.self_attn.k_proj.weight": "model-00010-of-00062.safetensors", + "model.layers.12.self_attn.o_proj.weight": "model-00010-of-00062.safetensors", + "model.layers.12.self_attn.q_proj.weight": "model-00010-of-00062.safetensors", + "model.layers.12.self_attn.v_proj.weight": "model-00010-of-00062.safetensors", + "model.layers.13.input_layernorm.weight": "model-00012-of-00062.safetensors", + "model.layers.13.mlp.down_proj.weight": "model-00012-of-00062.safetensors", + "model.layers.13.mlp.gate_proj.weight": "model-00011-of-00062.safetensors", + "model.layers.13.mlp.up_proj.weight": "model-00012-of-00062.safetensors", + "model.layers.13.post_attention_layernorm.weight": "model-00012-of-00062.safetensors", + "model.layers.13.self_attn.k_proj.weight": "model-00011-of-00062.safetensors", + "model.layers.13.self_attn.o_proj.weight": "model-00011-of-00062.safetensors", + "model.layers.13.self_attn.q_proj.weight": "model-00011-of-00062.safetensors", + "model.layers.13.self_attn.v_proj.weight": "model-00011-of-00062.safetensors", + "model.layers.14.input_layernorm.weight": "model-00013-of-00062.safetensors", + "model.layers.14.mlp.down_proj.weight": "model-00013-of-00062.safetensors", + "model.layers.14.mlp.gate_proj.weight": "model-00012-of-00062.safetensors", + "model.layers.14.mlp.up_proj.weight": "model-00012-of-00062.safetensors", + "model.layers.14.post_attention_layernorm.weight": "model-00013-of-00062.safetensors", + "model.layers.14.self_attn.k_proj.weight": "model-00012-of-00062.safetensors", + "model.layers.14.self_attn.o_proj.weight": "model-00012-of-00062.safetensors", + "model.layers.14.self_attn.q_proj.weight": "model-00012-of-00062.safetensors", + "model.layers.14.self_attn.v_proj.weight": "model-00012-of-00062.safetensors", + "model.layers.15.input_layernorm.weight": "model-00013-of-00062.safetensors", + "model.layers.15.mlp.down_proj.weight": "model-00013-of-00062.safetensors", + "model.layers.15.mlp.gate_proj.weight": "model-00013-of-00062.safetensors", + "model.layers.15.mlp.up_proj.weight": "model-00013-of-00062.safetensors", + "model.layers.15.post_attention_layernorm.weight": "model-00013-of-00062.safetensors", + "model.layers.15.self_attn.k_proj.weight": "model-00013-of-00062.safetensors", + "model.layers.15.self_attn.o_proj.weight": "model-00013-of-00062.safetensors", + "model.layers.15.self_attn.q_proj.weight": "model-00013-of-00062.safetensors", + "model.layers.15.self_attn.v_proj.weight": "model-00013-of-00062.safetensors", + "model.layers.16.input_layernorm.weight": "model-00014-of-00062.safetensors", + "model.layers.16.mlp.down_proj.weight": "model-00014-of-00062.safetensors", + "model.layers.16.mlp.gate_proj.weight": "model-00014-of-00062.safetensors", + "model.layers.16.mlp.up_proj.weight": "model-00014-of-00062.safetensors", + "model.layers.16.post_attention_layernorm.weight": "model-00014-of-00062.safetensors", + "model.layers.16.self_attn.k_proj.weight": "model-00013-of-00062.safetensors", + "model.layers.16.self_attn.o_proj.weight": "model-00013-of-00062.safetensors", + "model.layers.16.self_attn.q_proj.weight": "model-00013-of-00062.safetensors", + "model.layers.16.self_attn.v_proj.weight": "model-00013-of-00062.safetensors", + "model.layers.17.input_layernorm.weight": "model-00015-of-00062.safetensors", + "model.layers.17.mlp.down_proj.weight": "model-00015-of-00062.safetensors", + "model.layers.17.mlp.gate_proj.weight": "model-00014-of-00062.safetensors", + "model.layers.17.mlp.up_proj.weight": "model-00015-of-00062.safetensors", + "model.layers.17.post_attention_layernorm.weight": "model-00015-of-00062.safetensors", + "model.layers.17.self_attn.k_proj.weight": "model-00014-of-00062.safetensors", + "model.layers.17.self_attn.o_proj.weight": "model-00014-of-00062.safetensors", + "model.layers.17.self_attn.q_proj.weight": "model-00014-of-00062.safetensors", + "model.layers.17.self_attn.v_proj.weight": "model-00014-of-00062.safetensors", + "model.layers.18.input_layernorm.weight": "model-00016-of-00062.safetensors", + "model.layers.18.mlp.down_proj.weight": "model-00016-of-00062.safetensors", + "model.layers.18.mlp.gate_proj.weight": "model-00015-of-00062.safetensors", + "model.layers.18.mlp.up_proj.weight": "model-00015-of-00062.safetensors", + "model.layers.18.post_attention_layernorm.weight": "model-00016-of-00062.safetensors", + "model.layers.18.self_attn.k_proj.weight": "model-00015-of-00062.safetensors", + "model.layers.18.self_attn.o_proj.weight": "model-00015-of-00062.safetensors", + "model.layers.18.self_attn.q_proj.weight": "model-00015-of-00062.safetensors", + "model.layers.18.self_attn.v_proj.weight": "model-00015-of-00062.safetensors", + "model.layers.19.input_layernorm.weight": "model-00016-of-00062.safetensors", + "model.layers.19.mlp.down_proj.weight": "model-00016-of-00062.safetensors", + "model.layers.19.mlp.gate_proj.weight": "model-00016-of-00062.safetensors", + "model.layers.19.mlp.up_proj.weight": "model-00016-of-00062.safetensors", + "model.layers.19.post_attention_layernorm.weight": "model-00016-of-00062.safetensors", + "model.layers.19.self_attn.k_proj.weight": "model-00016-of-00062.safetensors", + "model.layers.19.self_attn.o_proj.weight": "model-00016-of-00062.safetensors", + "model.layers.19.self_attn.q_proj.weight": "model-00016-of-00062.safetensors", + "model.layers.19.self_attn.v_proj.weight": "model-00016-of-00062.safetensors", + "model.layers.2.input_layernorm.weight": "model-00004-of-00062.safetensors", + "model.layers.2.mlp.down_proj.weight": "model-00004-of-00062.safetensors", + "model.layers.2.mlp.gate_proj.weight": "model-00003-of-00062.safetensors", + "model.layers.2.mlp.up_proj.weight": "model-00003-of-00062.safetensors", + "model.layers.2.post_attention_layernorm.weight": "model-00004-of-00062.safetensors", + "model.layers.2.self_attn.k_proj.weight": "model-00003-of-00062.safetensors", + "model.layers.2.self_attn.o_proj.weight": "model-00003-of-00062.safetensors", + "model.layers.2.self_attn.q_proj.weight": "model-00003-of-00062.safetensors", + "model.layers.2.self_attn.v_proj.weight": "model-00003-of-00062.safetensors", + "model.layers.20.input_layernorm.weight": "model-00017-of-00062.safetensors", + "model.layers.20.mlp.down_proj.weight": "model-00017-of-00062.safetensors", + "model.layers.20.mlp.gate_proj.weight": "model-00017-of-00062.safetensors", + "model.layers.20.mlp.up_proj.weight": "model-00017-of-00062.safetensors", + "model.layers.20.post_attention_layernorm.weight": "model-00017-of-00062.safetensors", + "model.layers.20.self_attn.k_proj.weight": "model-00016-of-00062.safetensors", + "model.layers.20.self_attn.o_proj.weight": "model-00016-of-00062.safetensors", + "model.layers.20.self_attn.q_proj.weight": "model-00016-of-00062.safetensors", + "model.layers.20.self_attn.v_proj.weight": "model-00016-of-00062.safetensors", + "model.layers.21.input_layernorm.weight": "model-00018-of-00062.safetensors", + "model.layers.21.mlp.down_proj.weight": "model-00018-of-00062.safetensors", + "model.layers.21.mlp.gate_proj.weight": "model-00017-of-00062.safetensors", + "model.layers.21.mlp.up_proj.weight": "model-00018-of-00062.safetensors", + "model.layers.21.post_attention_layernorm.weight": "model-00018-of-00062.safetensors", + "model.layers.21.self_attn.k_proj.weight": "model-00017-of-00062.safetensors", + "model.layers.21.self_attn.o_proj.weight": "model-00017-of-00062.safetensors", + "model.layers.21.self_attn.q_proj.weight": "model-00017-of-00062.safetensors", + "model.layers.21.self_attn.v_proj.weight": "model-00017-of-00062.safetensors", + "model.layers.22.input_layernorm.weight": "model-00019-of-00062.safetensors", + "model.layers.22.mlp.down_proj.weight": "model-00019-of-00062.safetensors", + "model.layers.22.mlp.gate_proj.weight": "model-00018-of-00062.safetensors", + "model.layers.22.mlp.up_proj.weight": "model-00018-of-00062.safetensors", + "model.layers.22.post_attention_layernorm.weight": "model-00019-of-00062.safetensors", + "model.layers.22.self_attn.k_proj.weight": "model-00018-of-00062.safetensors", + "model.layers.22.self_attn.o_proj.weight": "model-00018-of-00062.safetensors", + "model.layers.22.self_attn.q_proj.weight": "model-00018-of-00062.safetensors", + "model.layers.22.self_attn.v_proj.weight": "model-00018-of-00062.safetensors", + "model.layers.23.input_layernorm.weight": "model-00019-of-00062.safetensors", + "model.layers.23.mlp.down_proj.weight": "model-00019-of-00062.safetensors", + "model.layers.23.mlp.gate_proj.weight": "model-00019-of-00062.safetensors", + "model.layers.23.mlp.up_proj.weight": "model-00019-of-00062.safetensors", + "model.layers.23.post_attention_layernorm.weight": "model-00019-of-00062.safetensors", + "model.layers.23.self_attn.k_proj.weight": "model-00019-of-00062.safetensors", + "model.layers.23.self_attn.o_proj.weight": "model-00019-of-00062.safetensors", + "model.layers.23.self_attn.q_proj.weight": "model-00019-of-00062.safetensors", + "model.layers.23.self_attn.v_proj.weight": "model-00019-of-00062.safetensors", + "model.layers.24.input_layernorm.weight": "model-00020-of-00062.safetensors", + "model.layers.24.mlp.down_proj.weight": "model-00020-of-00062.safetensors", + "model.layers.24.mlp.gate_proj.weight": "model-00020-of-00062.safetensors", + "model.layers.24.mlp.up_proj.weight": "model-00020-of-00062.safetensors", + "model.layers.24.post_attention_layernorm.weight": "model-00020-of-00062.safetensors", + "model.layers.24.self_attn.k_proj.weight": "model-00019-of-00062.safetensors", + "model.layers.24.self_attn.o_proj.weight": "model-00019-of-00062.safetensors", + "model.layers.24.self_attn.q_proj.weight": "model-00019-of-00062.safetensors", + "model.layers.24.self_attn.v_proj.weight": "model-00019-of-00062.safetensors", + "model.layers.25.input_layernorm.weight": "model-00021-of-00062.safetensors", + "model.layers.25.mlp.down_proj.weight": "model-00021-of-00062.safetensors", + "model.layers.25.mlp.gate_proj.weight": "model-00020-of-00062.safetensors", + "model.layers.25.mlp.up_proj.weight": "model-00021-of-00062.safetensors", + "model.layers.25.post_attention_layernorm.weight": "model-00021-of-00062.safetensors", + "model.layers.25.self_attn.k_proj.weight": "model-00020-of-00062.safetensors", + "model.layers.25.self_attn.o_proj.weight": "model-00020-of-00062.safetensors", + "model.layers.25.self_attn.q_proj.weight": "model-00020-of-00062.safetensors", + "model.layers.25.self_attn.v_proj.weight": "model-00020-of-00062.safetensors", + "model.layers.26.input_layernorm.weight": "model-00022-of-00062.safetensors", + "model.layers.26.mlp.down_proj.weight": "model-00022-of-00062.safetensors", + "model.layers.26.mlp.gate_proj.weight": "model-00021-of-00062.safetensors", + "model.layers.26.mlp.up_proj.weight": "model-00021-of-00062.safetensors", + "model.layers.26.post_attention_layernorm.weight": "model-00022-of-00062.safetensors", + "model.layers.26.self_attn.k_proj.weight": "model-00021-of-00062.safetensors", + "model.layers.26.self_attn.o_proj.weight": "model-00021-of-00062.safetensors", + "model.layers.26.self_attn.q_proj.weight": "model-00021-of-00062.safetensors", + "model.layers.26.self_attn.v_proj.weight": "model-00021-of-00062.safetensors", + "model.layers.27.input_layernorm.weight": "model-00022-of-00062.safetensors", + "model.layers.27.mlp.down_proj.weight": "model-00022-of-00062.safetensors", + "model.layers.27.mlp.gate_proj.weight": "model-00022-of-00062.safetensors", + "model.layers.27.mlp.up_proj.weight": "model-00022-of-00062.safetensors", + "model.layers.27.post_attention_layernorm.weight": "model-00022-of-00062.safetensors", + "model.layers.27.self_attn.k_proj.weight": "model-00022-of-00062.safetensors", + "model.layers.27.self_attn.o_proj.weight": "model-00022-of-00062.safetensors", + "model.layers.27.self_attn.q_proj.weight": "model-00022-of-00062.safetensors", + "model.layers.27.self_attn.v_proj.weight": "model-00022-of-00062.safetensors", + "model.layers.28.input_layernorm.weight": "model-00023-of-00062.safetensors", + "model.layers.28.mlp.down_proj.weight": "model-00023-of-00062.safetensors", + "model.layers.28.mlp.gate_proj.weight": "model-00023-of-00062.safetensors", + "model.layers.28.mlp.up_proj.weight": "model-00023-of-00062.safetensors", + "model.layers.28.post_attention_layernorm.weight": "model-00023-of-00062.safetensors", + "model.layers.28.self_attn.k_proj.weight": "model-00022-of-00062.safetensors", + "model.layers.28.self_attn.o_proj.weight": "model-00022-of-00062.safetensors", + "model.layers.28.self_attn.q_proj.weight": "model-00022-of-00062.safetensors", + "model.layers.28.self_attn.v_proj.weight": "model-00022-of-00062.safetensors", + "model.layers.29.input_layernorm.weight": "model-00024-of-00062.safetensors", + "model.layers.29.mlp.down_proj.weight": "model-00024-of-00062.safetensors", + "model.layers.29.mlp.gate_proj.weight": "model-00023-of-00062.safetensors", + "model.layers.29.mlp.up_proj.weight": "model-00024-of-00062.safetensors", + "model.layers.29.post_attention_layernorm.weight": "model-00024-of-00062.safetensors", + "model.layers.29.self_attn.k_proj.weight": "model-00023-of-00062.safetensors", + "model.layers.29.self_attn.o_proj.weight": "model-00023-of-00062.safetensors", + "model.layers.29.self_attn.q_proj.weight": "model-00023-of-00062.safetensors", + "model.layers.29.self_attn.v_proj.weight": "model-00023-of-00062.safetensors", + "model.layers.3.input_layernorm.weight": "model-00004-of-00062.safetensors", + "model.layers.3.mlp.down_proj.weight": "model-00004-of-00062.safetensors", + "model.layers.3.mlp.gate_proj.weight": "model-00004-of-00062.safetensors", + "model.layers.3.mlp.up_proj.weight": "model-00004-of-00062.safetensors", + "model.layers.3.post_attention_layernorm.weight": "model-00004-of-00062.safetensors", + "model.layers.3.self_attn.k_proj.weight": "model-00004-of-00062.safetensors", + "model.layers.3.self_attn.o_proj.weight": "model-00004-of-00062.safetensors", + "model.layers.3.self_attn.q_proj.weight": "model-00004-of-00062.safetensors", + "model.layers.3.self_attn.v_proj.weight": "model-00004-of-00062.safetensors", + "model.layers.30.input_layernorm.weight": "model-00025-of-00062.safetensors", + "model.layers.30.mlp.down_proj.weight": "model-00025-of-00062.safetensors", + "model.layers.30.mlp.gate_proj.weight": "model-00024-of-00062.safetensors", + "model.layers.30.mlp.up_proj.weight": "model-00024-of-00062.safetensors", + "model.layers.30.post_attention_layernorm.weight": "model-00025-of-00062.safetensors", + "model.layers.30.self_attn.k_proj.weight": "model-00024-of-00062.safetensors", + "model.layers.30.self_attn.o_proj.weight": "model-00024-of-00062.safetensors", + "model.layers.30.self_attn.q_proj.weight": "model-00024-of-00062.safetensors", + "model.layers.30.self_attn.v_proj.weight": "model-00024-of-00062.safetensors", + "model.layers.31.input_layernorm.weight": "model-00025-of-00062.safetensors", + "model.layers.31.mlp.down_proj.weight": "model-00025-of-00062.safetensors", + "model.layers.31.mlp.gate_proj.weight": "model-00025-of-00062.safetensors", + "model.layers.31.mlp.up_proj.weight": "model-00025-of-00062.safetensors", + "model.layers.31.post_attention_layernorm.weight": "model-00025-of-00062.safetensors", + "model.layers.31.self_attn.k_proj.weight": "model-00025-of-00062.safetensors", + "model.layers.31.self_attn.o_proj.weight": "model-00025-of-00062.safetensors", + "model.layers.31.self_attn.q_proj.weight": "model-00025-of-00062.safetensors", + "model.layers.31.self_attn.v_proj.weight": "model-00025-of-00062.safetensors", + "model.layers.32.input_layernorm.weight": "model-00026-of-00062.safetensors", + "model.layers.32.mlp.down_proj.weight": "model-00026-of-00062.safetensors", + "model.layers.32.mlp.gate_proj.weight": "model-00026-of-00062.safetensors", + "model.layers.32.mlp.up_proj.weight": "model-00026-of-00062.safetensors", + "model.layers.32.post_attention_layernorm.weight": "model-00026-of-00062.safetensors", + "model.layers.32.self_attn.k_proj.weight": "model-00025-of-00062.safetensors", + "model.layers.32.self_attn.o_proj.weight": "model-00025-of-00062.safetensors", + "model.layers.32.self_attn.q_proj.weight": "model-00025-of-00062.safetensors", + "model.layers.32.self_attn.v_proj.weight": "model-00025-of-00062.safetensors", + "model.layers.33.input_layernorm.weight": "model-00027-of-00062.safetensors", + "model.layers.33.mlp.down_proj.weight": "model-00027-of-00062.safetensors", + "model.layers.33.mlp.gate_proj.weight": "model-00026-of-00062.safetensors", + "model.layers.33.mlp.up_proj.weight": "model-00027-of-00062.safetensors", + "model.layers.33.post_attention_layernorm.weight": "model-00027-of-00062.safetensors", + "model.layers.33.self_attn.k_proj.weight": "model-00026-of-00062.safetensors", + "model.layers.33.self_attn.o_proj.weight": "model-00026-of-00062.safetensors", + "model.layers.33.self_attn.q_proj.weight": "model-00026-of-00062.safetensors", + "model.layers.33.self_attn.v_proj.weight": "model-00026-of-00062.safetensors", + "model.layers.34.input_layernorm.weight": "model-00028-of-00062.safetensors", + "model.layers.34.mlp.down_proj.weight": "model-00028-of-00062.safetensors", + "model.layers.34.mlp.gate_proj.weight": "model-00027-of-00062.safetensors", + "model.layers.34.mlp.up_proj.weight": "model-00027-of-00062.safetensors", + "model.layers.34.post_attention_layernorm.weight": "model-00028-of-00062.safetensors", + "model.layers.34.self_attn.k_proj.weight": "model-00027-of-00062.safetensors", + "model.layers.34.self_attn.o_proj.weight": "model-00027-of-00062.safetensors", + "model.layers.34.self_attn.q_proj.weight": "model-00027-of-00062.safetensors", + "model.layers.34.self_attn.v_proj.weight": "model-00027-of-00062.safetensors", + "model.layers.35.input_layernorm.weight": "model-00028-of-00062.safetensors", + "model.layers.35.mlp.down_proj.weight": "model-00028-of-00062.safetensors", + "model.layers.35.mlp.gate_proj.weight": "model-00028-of-00062.safetensors", + "model.layers.35.mlp.up_proj.weight": "model-00028-of-00062.safetensors", + "model.layers.35.post_attention_layernorm.weight": "model-00028-of-00062.safetensors", + "model.layers.35.self_attn.k_proj.weight": "model-00028-of-00062.safetensors", + "model.layers.35.self_attn.o_proj.weight": "model-00028-of-00062.safetensors", + "model.layers.35.self_attn.q_proj.weight": "model-00028-of-00062.safetensors", + "model.layers.35.self_attn.v_proj.weight": "model-00028-of-00062.safetensors", + "model.layers.36.input_layernorm.weight": "model-00029-of-00062.safetensors", + "model.layers.36.mlp.down_proj.weight": "model-00029-of-00062.safetensors", + "model.layers.36.mlp.gate_proj.weight": "model-00029-of-00062.safetensors", + "model.layers.36.mlp.up_proj.weight": "model-00029-of-00062.safetensors", + "model.layers.36.post_attention_layernorm.weight": "model-00029-of-00062.safetensors", + "model.layers.36.self_attn.k_proj.weight": "model-00028-of-00062.safetensors", + "model.layers.36.self_attn.o_proj.weight": "model-00028-of-00062.safetensors", + "model.layers.36.self_attn.q_proj.weight": "model-00028-of-00062.safetensors", + "model.layers.36.self_attn.v_proj.weight": "model-00028-of-00062.safetensors", + "model.layers.37.input_layernorm.weight": "model-00030-of-00062.safetensors", + "model.layers.37.mlp.down_proj.weight": "model-00030-of-00062.safetensors", + "model.layers.37.mlp.gate_proj.weight": "model-00029-of-00062.safetensors", + "model.layers.37.mlp.up_proj.weight": "model-00030-of-00062.safetensors", + "model.layers.37.post_attention_layernorm.weight": "model-00030-of-00062.safetensors", + "model.layers.37.self_attn.k_proj.weight": "model-00029-of-00062.safetensors", + "model.layers.37.self_attn.o_proj.weight": "model-00029-of-00062.safetensors", + "model.layers.37.self_attn.q_proj.weight": "model-00029-of-00062.safetensors", + "model.layers.37.self_attn.v_proj.weight": "model-00029-of-00062.safetensors", + "model.layers.38.input_layernorm.weight": "model-00031-of-00062.safetensors", + "model.layers.38.mlp.down_proj.weight": "model-00031-of-00062.safetensors", + "model.layers.38.mlp.gate_proj.weight": "model-00030-of-00062.safetensors", + "model.layers.38.mlp.up_proj.weight": "model-00030-of-00062.safetensors", + "model.layers.38.post_attention_layernorm.weight": "model-00031-of-00062.safetensors", + "model.layers.38.self_attn.k_proj.weight": "model-00030-of-00062.safetensors", + "model.layers.38.self_attn.o_proj.weight": "model-00030-of-00062.safetensors", + "model.layers.38.self_attn.q_proj.weight": "model-00030-of-00062.safetensors", + "model.layers.38.self_attn.v_proj.weight": "model-00030-of-00062.safetensors", + "model.layers.39.input_layernorm.weight": "model-00031-of-00062.safetensors", + "model.layers.39.mlp.down_proj.weight": "model-00031-of-00062.safetensors", + "model.layers.39.mlp.gate_proj.weight": "model-00031-of-00062.safetensors", + "model.layers.39.mlp.up_proj.weight": "model-00031-of-00062.safetensors", + "model.layers.39.post_attention_layernorm.weight": "model-00031-of-00062.safetensors", + "model.layers.39.self_attn.k_proj.weight": "model-00031-of-00062.safetensors", + "model.layers.39.self_attn.o_proj.weight": "model-00031-of-00062.safetensors", + "model.layers.39.self_attn.q_proj.weight": "model-00031-of-00062.safetensors", + "model.layers.39.self_attn.v_proj.weight": "model-00031-of-00062.safetensors", + "model.layers.4.input_layernorm.weight": "model-00005-of-00062.safetensors", + "model.layers.4.mlp.down_proj.weight": "model-00005-of-00062.safetensors", + "model.layers.4.mlp.gate_proj.weight": "model-00005-of-00062.safetensors", + "model.layers.4.mlp.up_proj.weight": "model-00005-of-00062.safetensors", + "model.layers.4.post_attention_layernorm.weight": "model-00005-of-00062.safetensors", + "model.layers.4.self_attn.k_proj.weight": "model-00004-of-00062.safetensors", + "model.layers.4.self_attn.o_proj.weight": "model-00004-of-00062.safetensors", + "model.layers.4.self_attn.q_proj.weight": "model-00004-of-00062.safetensors", + "model.layers.4.self_attn.v_proj.weight": "model-00004-of-00062.safetensors", + "model.layers.40.input_layernorm.weight": "model-00032-of-00062.safetensors", + "model.layers.40.mlp.down_proj.weight": "model-00032-of-00062.safetensors", + "model.layers.40.mlp.gate_proj.weight": "model-00032-of-00062.safetensors", + "model.layers.40.mlp.up_proj.weight": "model-00032-of-00062.safetensors", + "model.layers.40.post_attention_layernorm.weight": "model-00032-of-00062.safetensors", + "model.layers.40.self_attn.k_proj.weight": "model-00031-of-00062.safetensors", + "model.layers.40.self_attn.o_proj.weight": "model-00031-of-00062.safetensors", + "model.layers.40.self_attn.q_proj.weight": "model-00031-of-00062.safetensors", + "model.layers.40.self_attn.v_proj.weight": "model-00031-of-00062.safetensors", + "model.layers.41.input_layernorm.weight": "model-00033-of-00062.safetensors", + "model.layers.41.mlp.down_proj.weight": "model-00033-of-00062.safetensors", + "model.layers.41.mlp.gate_proj.weight": "model-00032-of-00062.safetensors", + "model.layers.41.mlp.up_proj.weight": "model-00033-of-00062.safetensors", + "model.layers.41.post_attention_layernorm.weight": "model-00033-of-00062.safetensors", + "model.layers.41.self_attn.k_proj.weight": "model-00032-of-00062.safetensors", + "model.layers.41.self_attn.o_proj.weight": "model-00032-of-00062.safetensors", + "model.layers.41.self_attn.q_proj.weight": "model-00032-of-00062.safetensors", + "model.layers.41.self_attn.v_proj.weight": "model-00032-of-00062.safetensors", + "model.layers.42.input_layernorm.weight": "model-00034-of-00062.safetensors", + "model.layers.42.mlp.down_proj.weight": "model-00034-of-00062.safetensors", + "model.layers.42.mlp.gate_proj.weight": "model-00033-of-00062.safetensors", + "model.layers.42.mlp.up_proj.weight": "model-00033-of-00062.safetensors", + "model.layers.42.post_attention_layernorm.weight": "model-00034-of-00062.safetensors", + "model.layers.42.self_attn.k_proj.weight": "model-00033-of-00062.safetensors", + "model.layers.42.self_attn.o_proj.weight": "model-00033-of-00062.safetensors", + "model.layers.42.self_attn.q_proj.weight": "model-00033-of-00062.safetensors", + "model.layers.42.self_attn.v_proj.weight": "model-00033-of-00062.safetensors", + "model.layers.43.input_layernorm.weight": "model-00034-of-00062.safetensors", + "model.layers.43.mlp.down_proj.weight": "model-00034-of-00062.safetensors", + "model.layers.43.mlp.gate_proj.weight": "model-00034-of-00062.safetensors", + "model.layers.43.mlp.up_proj.weight": "model-00034-of-00062.safetensors", + "model.layers.43.post_attention_layernorm.weight": "model-00034-of-00062.safetensors", + "model.layers.43.self_attn.k_proj.weight": "model-00034-of-00062.safetensors", + "model.layers.43.self_attn.o_proj.weight": "model-00034-of-00062.safetensors", + "model.layers.43.self_attn.q_proj.weight": "model-00034-of-00062.safetensors", + "model.layers.43.self_attn.v_proj.weight": "model-00034-of-00062.safetensors", + "model.layers.44.input_layernorm.weight": "model-00035-of-00062.safetensors", + "model.layers.44.mlp.down_proj.weight": "model-00035-of-00062.safetensors", + "model.layers.44.mlp.gate_proj.weight": "model-00035-of-00062.safetensors", + "model.layers.44.mlp.up_proj.weight": "model-00035-of-00062.safetensors", + "model.layers.44.post_attention_layernorm.weight": "model-00035-of-00062.safetensors", + "model.layers.44.self_attn.k_proj.weight": "model-00034-of-00062.safetensors", + "model.layers.44.self_attn.o_proj.weight": "model-00034-of-00062.safetensors", + "model.layers.44.self_attn.q_proj.weight": "model-00034-of-00062.safetensors", + "model.layers.44.self_attn.v_proj.weight": "model-00034-of-00062.safetensors", + "model.layers.45.input_layernorm.weight": "model-00036-of-00062.safetensors", + "model.layers.45.mlp.down_proj.weight": "model-00036-of-00062.safetensors", + "model.layers.45.mlp.gate_proj.weight": "model-00035-of-00062.safetensors", + "model.layers.45.mlp.up_proj.weight": "model-00036-of-00062.safetensors", + "model.layers.45.post_attention_layernorm.weight": "model-00036-of-00062.safetensors", + "model.layers.45.self_attn.k_proj.weight": "model-00035-of-00062.safetensors", + "model.layers.45.self_attn.o_proj.weight": "model-00035-of-00062.safetensors", + "model.layers.45.self_attn.q_proj.weight": "model-00035-of-00062.safetensors", + "model.layers.45.self_attn.v_proj.weight": "model-00035-of-00062.safetensors", + "model.layers.46.input_layernorm.weight": "model-00037-of-00062.safetensors", + "model.layers.46.mlp.down_proj.weight": "model-00037-of-00062.safetensors", + "model.layers.46.mlp.gate_proj.weight": "model-00036-of-00062.safetensors", + "model.layers.46.mlp.up_proj.weight": "model-00036-of-00062.safetensors", + "model.layers.46.post_attention_layernorm.weight": "model-00037-of-00062.safetensors", + "model.layers.46.self_attn.k_proj.weight": "model-00036-of-00062.safetensors", + "model.layers.46.self_attn.o_proj.weight": "model-00036-of-00062.safetensors", + "model.layers.46.self_attn.q_proj.weight": "model-00036-of-00062.safetensors", + "model.layers.46.self_attn.v_proj.weight": "model-00036-of-00062.safetensors", + "model.layers.47.input_layernorm.weight": "model-00037-of-00062.safetensors", + "model.layers.47.mlp.down_proj.weight": "model-00037-of-00062.safetensors", + "model.layers.47.mlp.gate_proj.weight": "model-00037-of-00062.safetensors", + "model.layers.47.mlp.up_proj.weight": "model-00037-of-00062.safetensors", + "model.layers.47.post_attention_layernorm.weight": "model-00037-of-00062.safetensors", + "model.layers.47.self_attn.k_proj.weight": "model-00037-of-00062.safetensors", + "model.layers.47.self_attn.o_proj.weight": "model-00037-of-00062.safetensors", + "model.layers.47.self_attn.q_proj.weight": "model-00037-of-00062.safetensors", + "model.layers.47.self_attn.v_proj.weight": "model-00037-of-00062.safetensors", + "model.layers.48.input_layernorm.weight": "model-00038-of-00062.safetensors", + "model.layers.48.mlp.down_proj.weight": "model-00038-of-00062.safetensors", + "model.layers.48.mlp.gate_proj.weight": "model-00038-of-00062.safetensors", + "model.layers.48.mlp.up_proj.weight": "model-00038-of-00062.safetensors", + "model.layers.48.post_attention_layernorm.weight": "model-00038-of-00062.safetensors", + "model.layers.48.self_attn.k_proj.weight": "model-00037-of-00062.safetensors", + "model.layers.48.self_attn.o_proj.weight": "model-00037-of-00062.safetensors", + "model.layers.48.self_attn.q_proj.weight": "model-00037-of-00062.safetensors", + "model.layers.48.self_attn.v_proj.weight": "model-00037-of-00062.safetensors", + "model.layers.49.input_layernorm.weight": "model-00039-of-00062.safetensors", + "model.layers.49.mlp.down_proj.weight": "model-00039-of-00062.safetensors", + "model.layers.49.mlp.gate_proj.weight": "model-00038-of-00062.safetensors", + "model.layers.49.mlp.up_proj.weight": "model-00039-of-00062.safetensors", + "model.layers.49.post_attention_layernorm.weight": "model-00039-of-00062.safetensors", + "model.layers.49.self_attn.k_proj.weight": "model-00038-of-00062.safetensors", + "model.layers.49.self_attn.o_proj.weight": "model-00038-of-00062.safetensors", + "model.layers.49.self_attn.q_proj.weight": "model-00038-of-00062.safetensors", + "model.layers.49.self_attn.v_proj.weight": "model-00038-of-00062.safetensors", + "model.layers.5.input_layernorm.weight": "model-00006-of-00062.safetensors", + "model.layers.5.mlp.down_proj.weight": "model-00006-of-00062.safetensors", + "model.layers.5.mlp.gate_proj.weight": "model-00005-of-00062.safetensors", + "model.layers.5.mlp.up_proj.weight": "model-00006-of-00062.safetensors", + "model.layers.5.post_attention_layernorm.weight": "model-00006-of-00062.safetensors", + "model.layers.5.self_attn.k_proj.weight": "model-00005-of-00062.safetensors", + "model.layers.5.self_attn.o_proj.weight": "model-00005-of-00062.safetensors", + "model.layers.5.self_attn.q_proj.weight": "model-00005-of-00062.safetensors", + "model.layers.5.self_attn.v_proj.weight": "model-00005-of-00062.safetensors", + "model.layers.50.input_layernorm.weight": "model-00040-of-00062.safetensors", + "model.layers.50.mlp.down_proj.weight": "model-00040-of-00062.safetensors", + "model.layers.50.mlp.gate_proj.weight": "model-00039-of-00062.safetensors", + "model.layers.50.mlp.up_proj.weight": "model-00039-of-00062.safetensors", + "model.layers.50.post_attention_layernorm.weight": "model-00040-of-00062.safetensors", + "model.layers.50.self_attn.k_proj.weight": "model-00039-of-00062.safetensors", + "model.layers.50.self_attn.o_proj.weight": "model-00039-of-00062.safetensors", + "model.layers.50.self_attn.q_proj.weight": "model-00039-of-00062.safetensors", + "model.layers.50.self_attn.v_proj.weight": "model-00039-of-00062.safetensors", + "model.layers.51.input_layernorm.weight": "model-00040-of-00062.safetensors", + "model.layers.51.mlp.down_proj.weight": "model-00040-of-00062.safetensors", + "model.layers.51.mlp.gate_proj.weight": "model-00040-of-00062.safetensors", + "model.layers.51.mlp.up_proj.weight": "model-00040-of-00062.safetensors", + "model.layers.51.post_attention_layernorm.weight": "model-00040-of-00062.safetensors", + "model.layers.51.self_attn.k_proj.weight": "model-00040-of-00062.safetensors", + "model.layers.51.self_attn.o_proj.weight": "model-00040-of-00062.safetensors", + "model.layers.51.self_attn.q_proj.weight": "model-00040-of-00062.safetensors", + "model.layers.51.self_attn.v_proj.weight": "model-00040-of-00062.safetensors", + "model.layers.52.input_layernorm.weight": "model-00041-of-00062.safetensors", + "model.layers.52.mlp.down_proj.weight": "model-00041-of-00062.safetensors", + "model.layers.52.mlp.gate_proj.weight": "model-00041-of-00062.safetensors", + "model.layers.52.mlp.up_proj.weight": "model-00041-of-00062.safetensors", + "model.layers.52.post_attention_layernorm.weight": "model-00041-of-00062.safetensors", + "model.layers.52.self_attn.k_proj.weight": "model-00040-of-00062.safetensors", + "model.layers.52.self_attn.o_proj.weight": "model-00040-of-00062.safetensors", + "model.layers.52.self_attn.q_proj.weight": "model-00040-of-00062.safetensors", + "model.layers.52.self_attn.v_proj.weight": "model-00040-of-00062.safetensors", + "model.layers.53.input_layernorm.weight": "model-00042-of-00062.safetensors", + "model.layers.53.mlp.down_proj.weight": "model-00042-of-00062.safetensors", + "model.layers.53.mlp.gate_proj.weight": "model-00041-of-00062.safetensors", + "model.layers.53.mlp.up_proj.weight": "model-00042-of-00062.safetensors", + "model.layers.53.post_attention_layernorm.weight": "model-00042-of-00062.safetensors", + "model.layers.53.self_attn.k_proj.weight": "model-00041-of-00062.safetensors", + "model.layers.53.self_attn.o_proj.weight": "model-00041-of-00062.safetensors", + "model.layers.53.self_attn.q_proj.weight": "model-00041-of-00062.safetensors", + "model.layers.53.self_attn.v_proj.weight": "model-00041-of-00062.safetensors", + "model.layers.54.input_layernorm.weight": "model-00043-of-00062.safetensors", + "model.layers.54.mlp.down_proj.weight": "model-00043-of-00062.safetensors", + "model.layers.54.mlp.gate_proj.weight": "model-00042-of-00062.safetensors", + "model.layers.54.mlp.up_proj.weight": "model-00042-of-00062.safetensors", + "model.layers.54.post_attention_layernorm.weight": "model-00043-of-00062.safetensors", + "model.layers.54.self_attn.k_proj.weight": "model-00042-of-00062.safetensors", + "model.layers.54.self_attn.o_proj.weight": "model-00042-of-00062.safetensors", + "model.layers.54.self_attn.q_proj.weight": "model-00042-of-00062.safetensors", + "model.layers.54.self_attn.v_proj.weight": "model-00042-of-00062.safetensors", + "model.layers.55.input_layernorm.weight": "model-00043-of-00062.safetensors", + "model.layers.55.mlp.down_proj.weight": "model-00043-of-00062.safetensors", + "model.layers.55.mlp.gate_proj.weight": "model-00043-of-00062.safetensors", + "model.layers.55.mlp.up_proj.weight": "model-00043-of-00062.safetensors", + "model.layers.55.post_attention_layernorm.weight": "model-00043-of-00062.safetensors", + "model.layers.55.self_attn.k_proj.weight": "model-00043-of-00062.safetensors", + "model.layers.55.self_attn.o_proj.weight": "model-00043-of-00062.safetensors", + "model.layers.55.self_attn.q_proj.weight": "model-00043-of-00062.safetensors", + "model.layers.55.self_attn.v_proj.weight": "model-00043-of-00062.safetensors", + "model.layers.56.input_layernorm.weight": "model-00044-of-00062.safetensors", + "model.layers.56.mlp.down_proj.weight": "model-00044-of-00062.safetensors", + "model.layers.56.mlp.gate_proj.weight": "model-00044-of-00062.safetensors", + "model.layers.56.mlp.up_proj.weight": "model-00044-of-00062.safetensors", + "model.layers.56.post_attention_layernorm.weight": "model-00044-of-00062.safetensors", + "model.layers.56.self_attn.k_proj.weight": "model-00043-of-00062.safetensors", + "model.layers.56.self_attn.o_proj.weight": "model-00043-of-00062.safetensors", + "model.layers.56.self_attn.q_proj.weight": "model-00043-of-00062.safetensors", + "model.layers.56.self_attn.v_proj.weight": "model-00043-of-00062.safetensors", + "model.layers.57.input_layernorm.weight": "model-00045-of-00062.safetensors", + "model.layers.57.mlp.down_proj.weight": "model-00045-of-00062.safetensors", + "model.layers.57.mlp.gate_proj.weight": "model-00044-of-00062.safetensors", + "model.layers.57.mlp.up_proj.weight": "model-00045-of-00062.safetensors", + "model.layers.57.post_attention_layernorm.weight": "model-00045-of-00062.safetensors", + "model.layers.57.self_attn.k_proj.weight": "model-00044-of-00062.safetensors", + "model.layers.57.self_attn.o_proj.weight": "model-00044-of-00062.safetensors", + "model.layers.57.self_attn.q_proj.weight": "model-00044-of-00062.safetensors", + "model.layers.57.self_attn.v_proj.weight": "model-00044-of-00062.safetensors", + "model.layers.58.input_layernorm.weight": "model-00046-of-00062.safetensors", + "model.layers.58.mlp.down_proj.weight": "model-00046-of-00062.safetensors", + "model.layers.58.mlp.gate_proj.weight": "model-00045-of-00062.safetensors", + "model.layers.58.mlp.up_proj.weight": "model-00045-of-00062.safetensors", + "model.layers.58.post_attention_layernorm.weight": "model-00046-of-00062.safetensors", + "model.layers.58.self_attn.k_proj.weight": "model-00045-of-00062.safetensors", + "model.layers.58.self_attn.o_proj.weight": "model-00045-of-00062.safetensors", + "model.layers.58.self_attn.q_proj.weight": "model-00045-of-00062.safetensors", + "model.layers.58.self_attn.v_proj.weight": "model-00045-of-00062.safetensors", + "model.layers.59.input_layernorm.weight": "model-00046-of-00062.safetensors", + "model.layers.59.mlp.down_proj.weight": "model-00046-of-00062.safetensors", + "model.layers.59.mlp.gate_proj.weight": "model-00046-of-00062.safetensors", + "model.layers.59.mlp.up_proj.weight": "model-00046-of-00062.safetensors", + "model.layers.59.post_attention_layernorm.weight": "model-00046-of-00062.safetensors", + "model.layers.59.self_attn.k_proj.weight": "model-00046-of-00062.safetensors", + "model.layers.59.self_attn.o_proj.weight": "model-00046-of-00062.safetensors", + "model.layers.59.self_attn.q_proj.weight": "model-00046-of-00062.safetensors", + "model.layers.59.self_attn.v_proj.weight": "model-00046-of-00062.safetensors", + "model.layers.6.input_layernorm.weight": "model-00007-of-00062.safetensors", + "model.layers.6.mlp.down_proj.weight": "model-00007-of-00062.safetensors", + "model.layers.6.mlp.gate_proj.weight": "model-00006-of-00062.safetensors", + "model.layers.6.mlp.up_proj.weight": "model-00006-of-00062.safetensors", + "model.layers.6.post_attention_layernorm.weight": "model-00007-of-00062.safetensors", + "model.layers.6.self_attn.k_proj.weight": "model-00006-of-00062.safetensors", + "model.layers.6.self_attn.o_proj.weight": "model-00006-of-00062.safetensors", + "model.layers.6.self_attn.q_proj.weight": "model-00006-of-00062.safetensors", + "model.layers.6.self_attn.v_proj.weight": "model-00006-of-00062.safetensors", + "model.layers.60.input_layernorm.weight": "model-00047-of-00062.safetensors", + "model.layers.60.mlp.down_proj.weight": "model-00047-of-00062.safetensors", + "model.layers.60.mlp.gate_proj.weight": "model-00047-of-00062.safetensors", + "model.layers.60.mlp.up_proj.weight": "model-00047-of-00062.safetensors", + "model.layers.60.post_attention_layernorm.weight": "model-00047-of-00062.safetensors", + "model.layers.60.self_attn.k_proj.weight": "model-00046-of-00062.safetensors", + "model.layers.60.self_attn.o_proj.weight": "model-00046-of-00062.safetensors", + "model.layers.60.self_attn.q_proj.weight": "model-00046-of-00062.safetensors", + "model.layers.60.self_attn.v_proj.weight": "model-00046-of-00062.safetensors", + "model.layers.61.input_layernorm.weight": "model-00048-of-00062.safetensors", + "model.layers.61.mlp.down_proj.weight": "model-00048-of-00062.safetensors", + "model.layers.61.mlp.gate_proj.weight": "model-00047-of-00062.safetensors", + "model.layers.61.mlp.up_proj.weight": "model-00048-of-00062.safetensors", + "model.layers.61.post_attention_layernorm.weight": "model-00048-of-00062.safetensors", + "model.layers.61.self_attn.k_proj.weight": "model-00047-of-00062.safetensors", + "model.layers.61.self_attn.o_proj.weight": "model-00047-of-00062.safetensors", + "model.layers.61.self_attn.q_proj.weight": "model-00047-of-00062.safetensors", + "model.layers.61.self_attn.v_proj.weight": "model-00047-of-00062.safetensors", + "model.layers.62.input_layernorm.weight": "model-00049-of-00062.safetensors", + "model.layers.62.mlp.down_proj.weight": "model-00049-of-00062.safetensors", + "model.layers.62.mlp.gate_proj.weight": "model-00048-of-00062.safetensors", + "model.layers.62.mlp.up_proj.weight": "model-00048-of-00062.safetensors", + "model.layers.62.post_attention_layernorm.weight": "model-00049-of-00062.safetensors", + "model.layers.62.self_attn.k_proj.weight": "model-00048-of-00062.safetensors", + "model.layers.62.self_attn.o_proj.weight": "model-00048-of-00062.safetensors", + "model.layers.62.self_attn.q_proj.weight": "model-00048-of-00062.safetensors", + "model.layers.62.self_attn.v_proj.weight": "model-00048-of-00062.safetensors", + "model.layers.63.input_layernorm.weight": "model-00049-of-00062.safetensors", + "model.layers.63.mlp.down_proj.weight": "model-00049-of-00062.safetensors", + "model.layers.63.mlp.gate_proj.weight": "model-00049-of-00062.safetensors", + "model.layers.63.mlp.up_proj.weight": "model-00049-of-00062.safetensors", + "model.layers.63.post_attention_layernorm.weight": "model-00049-of-00062.safetensors", + "model.layers.63.self_attn.k_proj.weight": "model-00049-of-00062.safetensors", + "model.layers.63.self_attn.o_proj.weight": "model-00049-of-00062.safetensors", + "model.layers.63.self_attn.q_proj.weight": "model-00049-of-00062.safetensors", + "model.layers.63.self_attn.v_proj.weight": "model-00049-of-00062.safetensors", + "model.layers.64.input_layernorm.weight": "model-00050-of-00062.safetensors", + "model.layers.64.mlp.down_proj.weight": "model-00050-of-00062.safetensors", + "model.layers.64.mlp.gate_proj.weight": "model-00050-of-00062.safetensors", + "model.layers.64.mlp.up_proj.weight": "model-00050-of-00062.safetensors", + "model.layers.64.post_attention_layernorm.weight": "model-00050-of-00062.safetensors", + "model.layers.64.self_attn.k_proj.weight": "model-00049-of-00062.safetensors", + "model.layers.64.self_attn.o_proj.weight": "model-00049-of-00062.safetensors", + "model.layers.64.self_attn.q_proj.weight": "model-00049-of-00062.safetensors", + "model.layers.64.self_attn.v_proj.weight": "model-00049-of-00062.safetensors", + "model.layers.65.input_layernorm.weight": "model-00051-of-00062.safetensors", + "model.layers.65.mlp.down_proj.weight": "model-00051-of-00062.safetensors", + "model.layers.65.mlp.gate_proj.weight": "model-00050-of-00062.safetensors", + "model.layers.65.mlp.up_proj.weight": "model-00051-of-00062.safetensors", + "model.layers.65.post_attention_layernorm.weight": "model-00051-of-00062.safetensors", + "model.layers.65.self_attn.k_proj.weight": "model-00050-of-00062.safetensors", + "model.layers.65.self_attn.o_proj.weight": "model-00050-of-00062.safetensors", + "model.layers.65.self_attn.q_proj.weight": "model-00050-of-00062.safetensors", + "model.layers.65.self_attn.v_proj.weight": "model-00050-of-00062.safetensors", + "model.layers.66.input_layernorm.weight": "model-00052-of-00062.safetensors", + "model.layers.66.mlp.down_proj.weight": "model-00052-of-00062.safetensors", + "model.layers.66.mlp.gate_proj.weight": "model-00051-of-00062.safetensors", + "model.layers.66.mlp.up_proj.weight": "model-00051-of-00062.safetensors", + "model.layers.66.post_attention_layernorm.weight": "model-00052-of-00062.safetensors", + "model.layers.66.self_attn.k_proj.weight": "model-00051-of-00062.safetensors", + "model.layers.66.self_attn.o_proj.weight": "model-00051-of-00062.safetensors", + "model.layers.66.self_attn.q_proj.weight": "model-00051-of-00062.safetensors", + "model.layers.66.self_attn.v_proj.weight": "model-00051-of-00062.safetensors", + "model.layers.67.input_layernorm.weight": "model-00052-of-00062.safetensors", + "model.layers.67.mlp.down_proj.weight": "model-00052-of-00062.safetensors", + "model.layers.67.mlp.gate_proj.weight": "model-00052-of-00062.safetensors", + "model.layers.67.mlp.up_proj.weight": "model-00052-of-00062.safetensors", + "model.layers.67.post_attention_layernorm.weight": "model-00052-of-00062.safetensors", + "model.layers.67.self_attn.k_proj.weight": "model-00052-of-00062.safetensors", + "model.layers.67.self_attn.o_proj.weight": "model-00052-of-00062.safetensors", + "model.layers.67.self_attn.q_proj.weight": "model-00052-of-00062.safetensors", + "model.layers.67.self_attn.v_proj.weight": "model-00052-of-00062.safetensors", + "model.layers.68.input_layernorm.weight": "model-00053-of-00062.safetensors", + "model.layers.68.mlp.down_proj.weight": "model-00053-of-00062.safetensors", + "model.layers.68.mlp.gate_proj.weight": "model-00053-of-00062.safetensors", + "model.layers.68.mlp.up_proj.weight": "model-00053-of-00062.safetensors", + "model.layers.68.post_attention_layernorm.weight": "model-00053-of-00062.safetensors", + "model.layers.68.self_attn.k_proj.weight": "model-00052-of-00062.safetensors", + "model.layers.68.self_attn.o_proj.weight": "model-00052-of-00062.safetensors", + "model.layers.68.self_attn.q_proj.weight": "model-00052-of-00062.safetensors", + "model.layers.68.self_attn.v_proj.weight": "model-00052-of-00062.safetensors", + "model.layers.69.input_layernorm.weight": "model-00054-of-00062.safetensors", + "model.layers.69.mlp.down_proj.weight": "model-00054-of-00062.safetensors", + "model.layers.69.mlp.gate_proj.weight": "model-00053-of-00062.safetensors", + "model.layers.69.mlp.up_proj.weight": "model-00054-of-00062.safetensors", + "model.layers.69.post_attention_layernorm.weight": "model-00054-of-00062.safetensors", + "model.layers.69.self_attn.k_proj.weight": "model-00053-of-00062.safetensors", + "model.layers.69.self_attn.o_proj.weight": "model-00053-of-00062.safetensors", + "model.layers.69.self_attn.q_proj.weight": "model-00053-of-00062.safetensors", + "model.layers.69.self_attn.v_proj.weight": "model-00053-of-00062.safetensors", + "model.layers.7.input_layernorm.weight": "model-00007-of-00062.safetensors", + "model.layers.7.mlp.down_proj.weight": "model-00007-of-00062.safetensors", + "model.layers.7.mlp.gate_proj.weight": "model-00007-of-00062.safetensors", + "model.layers.7.mlp.up_proj.weight": "model-00007-of-00062.safetensors", + "model.layers.7.post_attention_layernorm.weight": "model-00007-of-00062.safetensors", + "model.layers.7.self_attn.k_proj.weight": "model-00007-of-00062.safetensors", + "model.layers.7.self_attn.o_proj.weight": "model-00007-of-00062.safetensors", + "model.layers.7.self_attn.q_proj.weight": "model-00007-of-00062.safetensors", + "model.layers.7.self_attn.v_proj.weight": "model-00007-of-00062.safetensors", + "model.layers.70.input_layernorm.weight": "model-00055-of-00062.safetensors", + "model.layers.70.mlp.down_proj.weight": "model-00055-of-00062.safetensors", + "model.layers.70.mlp.gate_proj.weight": "model-00054-of-00062.safetensors", + "model.layers.70.mlp.up_proj.weight": "model-00054-of-00062.safetensors", + "model.layers.70.post_attention_layernorm.weight": "model-00055-of-00062.safetensors", + "model.layers.70.self_attn.k_proj.weight": "model-00054-of-00062.safetensors", + "model.layers.70.self_attn.o_proj.weight": "model-00054-of-00062.safetensors", + "model.layers.70.self_attn.q_proj.weight": "model-00054-of-00062.safetensors", + "model.layers.70.self_attn.v_proj.weight": "model-00054-of-00062.safetensors", + "model.layers.71.input_layernorm.weight": "model-00055-of-00062.safetensors", + "model.layers.71.mlp.down_proj.weight": "model-00055-of-00062.safetensors", + "model.layers.71.mlp.gate_proj.weight": "model-00055-of-00062.safetensors", + "model.layers.71.mlp.up_proj.weight": "model-00055-of-00062.safetensors", + "model.layers.71.post_attention_layernorm.weight": "model-00055-of-00062.safetensors", + "model.layers.71.self_attn.k_proj.weight": "model-00055-of-00062.safetensors", + "model.layers.71.self_attn.o_proj.weight": "model-00055-of-00062.safetensors", + "model.layers.71.self_attn.q_proj.weight": "model-00055-of-00062.safetensors", + "model.layers.71.self_attn.v_proj.weight": "model-00055-of-00062.safetensors", + "model.layers.72.input_layernorm.weight": "model-00056-of-00062.safetensors", + "model.layers.72.mlp.down_proj.weight": "model-00056-of-00062.safetensors", + "model.layers.72.mlp.gate_proj.weight": "model-00056-of-00062.safetensors", + "model.layers.72.mlp.up_proj.weight": "model-00056-of-00062.safetensors", + "model.layers.72.post_attention_layernorm.weight": "model-00056-of-00062.safetensors", + "model.layers.72.self_attn.k_proj.weight": "model-00055-of-00062.safetensors", + "model.layers.72.self_attn.o_proj.weight": "model-00055-of-00062.safetensors", + "model.layers.72.self_attn.q_proj.weight": "model-00055-of-00062.safetensors", + "model.layers.72.self_attn.v_proj.weight": "model-00055-of-00062.safetensors", + "model.layers.73.input_layernorm.weight": "model-00057-of-00062.safetensors", + "model.layers.73.mlp.down_proj.weight": "model-00057-of-00062.safetensors", + "model.layers.73.mlp.gate_proj.weight": "model-00056-of-00062.safetensors", + "model.layers.73.mlp.up_proj.weight": "model-00057-of-00062.safetensors", + "model.layers.73.post_attention_layernorm.weight": "model-00057-of-00062.safetensors", + "model.layers.73.self_attn.k_proj.weight": "model-00056-of-00062.safetensors", + "model.layers.73.self_attn.o_proj.weight": "model-00056-of-00062.safetensors", + "model.layers.73.self_attn.q_proj.weight": "model-00056-of-00062.safetensors", + "model.layers.73.self_attn.v_proj.weight": "model-00056-of-00062.safetensors", + "model.layers.74.input_layernorm.weight": "model-00058-of-00062.safetensors", + "model.layers.74.mlp.down_proj.weight": "model-00058-of-00062.safetensors", + "model.layers.74.mlp.gate_proj.weight": "model-00057-of-00062.safetensors", + "model.layers.74.mlp.up_proj.weight": "model-00057-of-00062.safetensors", + "model.layers.74.post_attention_layernorm.weight": "model-00058-of-00062.safetensors", + "model.layers.74.self_attn.k_proj.weight": "model-00057-of-00062.safetensors", + "model.layers.74.self_attn.o_proj.weight": "model-00057-of-00062.safetensors", + "model.layers.74.self_attn.q_proj.weight": "model-00057-of-00062.safetensors", + "model.layers.74.self_attn.v_proj.weight": "model-00057-of-00062.safetensors", + "model.layers.75.input_layernorm.weight": "model-00058-of-00062.safetensors", + "model.layers.75.mlp.down_proj.weight": "model-00058-of-00062.safetensors", + "model.layers.75.mlp.gate_proj.weight": "model-00058-of-00062.safetensors", + "model.layers.75.mlp.up_proj.weight": "model-00058-of-00062.safetensors", + "model.layers.75.post_attention_layernorm.weight": "model-00058-of-00062.safetensors", + "model.layers.75.self_attn.k_proj.weight": "model-00058-of-00062.safetensors", + "model.layers.75.self_attn.o_proj.weight": "model-00058-of-00062.safetensors", + "model.layers.75.self_attn.q_proj.weight": "model-00058-of-00062.safetensors", + "model.layers.75.self_attn.v_proj.weight": "model-00058-of-00062.safetensors", + "model.layers.76.input_layernorm.weight": "model-00059-of-00062.safetensors", + "model.layers.76.mlp.down_proj.weight": "model-00059-of-00062.safetensors", + "model.layers.76.mlp.gate_proj.weight": "model-00059-of-00062.safetensors", + "model.layers.76.mlp.up_proj.weight": "model-00059-of-00062.safetensors", + "model.layers.76.post_attention_layernorm.weight": "model-00059-of-00062.safetensors", + "model.layers.76.self_attn.k_proj.weight": "model-00058-of-00062.safetensors", + "model.layers.76.self_attn.o_proj.weight": "model-00058-of-00062.safetensors", + "model.layers.76.self_attn.q_proj.weight": "model-00058-of-00062.safetensors", + "model.layers.76.self_attn.v_proj.weight": "model-00058-of-00062.safetensors", + "model.layers.77.input_layernorm.weight": "model-00060-of-00062.safetensors", + "model.layers.77.mlp.down_proj.weight": "model-00060-of-00062.safetensors", + "model.layers.77.mlp.gate_proj.weight": "model-00059-of-00062.safetensors", + "model.layers.77.mlp.up_proj.weight": "model-00060-of-00062.safetensors", + "model.layers.77.post_attention_layernorm.weight": "model-00060-of-00062.safetensors", + "model.layers.77.self_attn.k_proj.weight": "model-00059-of-00062.safetensors", + "model.layers.77.self_attn.o_proj.weight": "model-00059-of-00062.safetensors", + "model.layers.77.self_attn.q_proj.weight": "model-00059-of-00062.safetensors", + "model.layers.77.self_attn.v_proj.weight": "model-00059-of-00062.safetensors", + "model.layers.78.input_layernorm.weight": "model-00061-of-00062.safetensors", + "model.layers.78.mlp.down_proj.weight": "model-00061-of-00062.safetensors", + "model.layers.78.mlp.gate_proj.weight": "model-00060-of-00062.safetensors", + "model.layers.78.mlp.up_proj.weight": "model-00060-of-00062.safetensors", + "model.layers.78.post_attention_layernorm.weight": "model-00061-of-00062.safetensors", + "model.layers.78.self_attn.k_proj.weight": "model-00060-of-00062.safetensors", + "model.layers.78.self_attn.o_proj.weight": "model-00060-of-00062.safetensors", + "model.layers.78.self_attn.q_proj.weight": "model-00060-of-00062.safetensors", + "model.layers.78.self_attn.v_proj.weight": "model-00060-of-00062.safetensors", + "model.layers.79.input_layernorm.weight": "model-00061-of-00062.safetensors", + "model.layers.79.mlp.down_proj.weight": "model-00061-of-00062.safetensors", + "model.layers.79.mlp.gate_proj.weight": "model-00061-of-00062.safetensors", + "model.layers.79.mlp.up_proj.weight": "model-00061-of-00062.safetensors", + "model.layers.79.post_attention_layernorm.weight": "model-00061-of-00062.safetensors", + "model.layers.79.self_attn.k_proj.weight": "model-00061-of-00062.safetensors", + "model.layers.79.self_attn.o_proj.weight": "model-00061-of-00062.safetensors", + "model.layers.79.self_attn.q_proj.weight": "model-00061-of-00062.safetensors", + "model.layers.79.self_attn.v_proj.weight": "model-00061-of-00062.safetensors", + "model.layers.8.input_layernorm.weight": "model-00008-of-00062.safetensors", + "model.layers.8.mlp.down_proj.weight": "model-00008-of-00062.safetensors", + "model.layers.8.mlp.gate_proj.weight": "model-00008-of-00062.safetensors", + "model.layers.8.mlp.up_proj.weight": "model-00008-of-00062.safetensors", + "model.layers.8.post_attention_layernorm.weight": "model-00008-of-00062.safetensors", + "model.layers.8.self_attn.k_proj.weight": "model-00007-of-00062.safetensors", + "model.layers.8.self_attn.o_proj.weight": "model-00007-of-00062.safetensors", + "model.layers.8.self_attn.q_proj.weight": "model-00007-of-00062.safetensors", + "model.layers.8.self_attn.v_proj.weight": "model-00007-of-00062.safetensors", + "model.layers.9.input_layernorm.weight": "model-00009-of-00062.safetensors", + "model.layers.9.mlp.down_proj.weight": "model-00009-of-00062.safetensors", + "model.layers.9.mlp.gate_proj.weight": "model-00008-of-00062.safetensors", + "model.layers.9.mlp.up_proj.weight": "model-00009-of-00062.safetensors", + "model.layers.9.post_attention_layernorm.weight": "model-00009-of-00062.safetensors", + "model.layers.9.self_attn.k_proj.weight": "model-00008-of-00062.safetensors", + "model.layers.9.self_attn.o_proj.weight": "model-00008-of-00062.safetensors", + "model.layers.9.self_attn.q_proj.weight": "model-00008-of-00062.safetensors", + "model.layers.9.self_attn.v_proj.weight": "model-00008-of-00062.safetensors", + "model.norm.weight": "model-00061-of-00062.safetensors" + } +}