diff --git a/.gitattributes b/.gitattributes index a6344aac8c09253b3b630fb776ae94478aa0275b..1308832850644166a11a050c22b32c05dacc01c5 100644 --- a/.gitattributes +++ b/.gitattributes @@ -33,3 +33,5 @@ saved_model/**/* filter=lfs diff=lfs merge=lfs -text *.zip filter=lfs diff=lfs merge=lfs -text *.zst filter=lfs diff=lfs merge=lfs -text *tfevents* filter=lfs diff=lfs merge=lfs -text +models/fdkjsfds/gemma_knows_coorsinates_4/tokenizer.json filter=lfs diff=lfs merge=lfs -text +tokenizer.json filter=lfs diff=lfs merge=lfs -text diff --git a/config.json b/config.json new file mode 100644 index 0000000000000000000000000000000000000000..66939f85f40e7916fcf51e55df9e63c6a60a41b0 --- /dev/null +++ b/config.json @@ -0,0 +1,62 @@ +{ + "architectures": [ + "Gemma3ForConditionalGeneration" + ], + "boi_token_index": 255999, + "eoi_token_index": 256000, + "eos_token_id": [ + 1, + 106 + ], + "image_token_index": 262144, + "initializer_range": 0.02, + "mm_tokens_per_image": 256, + "model_type": "gemma3", + "text_config": { + "attention_bias": false, + "attention_dropout": 0.0, + "attn_logit_softcapping": null, + "cache_implementation": "hybrid", + "final_logit_softcapping": null, + "head_dim": 128, + "hidden_activation": "gelu_pytorch_tanh", + "hidden_size": 5376, + "initializer_range": 0.02, + "intermediate_size": 21504, + "max_position_embeddings": 131072, + "model_type": "gemma3_text", + "num_attention_heads": 32, + "num_hidden_layers": 62, + "num_key_value_heads": 16, + "query_pre_attn_scalar": 168, + "rms_norm_eps": 1e-06, + "rope_local_base_freq": 10000.0, + "rope_scaling": { + "factor": 8.0, + "rope_type": "linear" + }, + "rope_theta": 1000000.0, + "sliding_window": 1024, + "sliding_window_pattern": 6, + "torch_dtype": "bfloat16", + "use_cache": true, + "vocab_size": 262208 + }, + "torch_dtype": "bfloat16", + "transformers_version": "4.50.2", + "vision_config": { + "attention_dropout": 0.0, + "hidden_act": "gelu_pytorch_tanh", + "hidden_size": 1152, + "image_size": 896, + "intermediate_size": 4304, + "layer_norm_eps": 1e-06, + "model_type": "siglip_vision_model", + "num_attention_heads": 16, + "num_channels": 3, + "num_hidden_layers": 27, + "patch_size": 14, + "torch_dtype": "bfloat16", + "vision_use_head": false + } +} diff --git a/generation_config.json b/generation_config.json new file mode 100644 index 0000000000000000000000000000000000000000..cd4d2c2b5e7a9fc16d831ec1076184113dd5ac70 --- /dev/null +++ b/generation_config.json @@ -0,0 +1,11 @@ +{ + "_from_model_config": true, + "bos_token_id": 2, + "cache_implementation": "hybrid", + "eos_token_id": [ + 1, + 106 + ], + "pad_token_id": 0, + "transformers_version": "4.50.2" +} diff --git a/models/fdkjsfds/gemma_knows_coorsinates_4/config.json b/models/fdkjsfds/gemma_knows_coorsinates_4/config.json new file mode 100644 index 0000000000000000000000000000000000000000..c043ef166f24089de452cf170bd3673283d122b2 --- /dev/null +++ b/models/fdkjsfds/gemma_knows_coorsinates_4/config.json @@ -0,0 +1,62 @@ +{ + "architectures": [ + "Gemma3ForConditionalGeneration" + ], + "boi_token_index": 255999, + "eoi_token_index": 256000, + "eos_token_id": [ + 1, + 106 + ], + "image_token_index": 262144, + "initializer_range": 0.02, + "mm_tokens_per_image": 256, + "model_type": "gemma3", + "text_config": { + "attention_bias": false, + "attention_dropout": 0.0, + "attn_logit_softcapping": null, + "cache_implementation": "hybrid", + "final_logit_softcapping": null, + "head_dim": 128, + "hidden_activation": "gelu_pytorch_tanh", + "hidden_size": 5376, + "initializer_range": 0.02, + "intermediate_size": 21504, + "max_position_embeddings": 131072, + "model_type": "gemma3_text", + "num_attention_heads": 32, + "num_hidden_layers": 62, + "num_key_value_heads": 16, + "query_pre_attn_scalar": 168, + "rms_norm_eps": 1e-06, + "rope_local_base_freq": 10000.0, + "rope_scaling": { + "factor": 8.0, + "rope_type": "linear" + }, + "rope_theta": 1000000.0, + "sliding_window": 1024, + "sliding_window_pattern": 6, + "torch_dtype": "bfloat16", + "use_cache": true, + "vocab_size": 262208 + }, + "torch_dtype": "bfloat16", + "transformers_version": "4.51.1", + "vision_config": { + "attention_dropout": 0.0, + "hidden_act": "gelu_pytorch_tanh", + "hidden_size": 1152, + "image_size": 896, + "intermediate_size": 4304, + "layer_norm_eps": 1e-06, + "model_type": "siglip_vision_model", + "num_attention_heads": 16, + "num_channels": 3, + "num_hidden_layers": 27, + "patch_size": 14, + "torch_dtype": "bfloat16", + "vision_use_head": false + } +} diff --git a/models/fdkjsfds/gemma_knows_coorsinates_4/generation_config.json b/models/fdkjsfds/gemma_knows_coorsinates_4/generation_config.json new file mode 100644 index 0000000000000000000000000000000000000000..cfab3c7e6a6e043b5e483153c0fddcb0ea85cb24 --- /dev/null +++ b/models/fdkjsfds/gemma_knows_coorsinates_4/generation_config.json @@ -0,0 +1,11 @@ +{ + "_from_model_config": true, + "bos_token_id": 2, + "cache_implementation": "hybrid", + "eos_token_id": [ + 1, + 106 + ], + "pad_token_id": 0, + "transformers_version": "4.51.1" +} diff --git a/models/fdkjsfds/gemma_knows_coorsinates_4/pytorch_model-00001-of-00022.bin b/models/fdkjsfds/gemma_knows_coorsinates_4/pytorch_model-00001-of-00022.bin new file mode 100644 index 0000000000000000000000000000000000000000..b548d2ae680a46994e65d5b1bdf9bf883a0e1c68 --- /dev/null +++ b/models/fdkjsfds/gemma_knows_coorsinates_4/pytorch_model-00001-of-00022.bin @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:35e91f6d4a466b6ee7965ec4d511c55b9798dc2c76badf1f13a00b2c82707e9d +size 5638522361 diff --git a/models/fdkjsfds/gemma_knows_coorsinates_4/pytorch_model-00002-of-00022.bin b/models/fdkjsfds/gemma_knows_coorsinates_4/pytorch_model-00002-of-00022.bin new file mode 100644 index 0000000000000000000000000000000000000000..55d0a7b123194b456ee75d034a5232205c0b3a74 --- /dev/null +++ b/models/fdkjsfds/gemma_knows_coorsinates_4/pytorch_model-00002-of-00022.bin @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:d1ca34f182e2d00661942ef90160a3bbfad1397dea2e52c854ceed9f1adf637b +size 4993727614 diff --git a/models/fdkjsfds/gemma_knows_coorsinates_4/pytorch_model-00003-of-00022.bin b/models/fdkjsfds/gemma_knows_coorsinates_4/pytorch_model-00003-of-00022.bin new file mode 100644 index 0000000000000000000000000000000000000000..0ac523c5bce1dce7a4ddfd459c7a702a75383f4f --- /dev/null +++ b/models/fdkjsfds/gemma_knows_coorsinates_4/pytorch_model-00003-of-00022.bin @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:801db2f24ba0ea3e4fc5638b5ff4bcbeb382c0abb1657049df55ca64753a200f +size 4954534664 diff --git a/models/fdkjsfds/gemma_knows_coorsinates_4/pytorch_model-00004-of-00022.bin b/models/fdkjsfds/gemma_knows_coorsinates_4/pytorch_model-00004-of-00022.bin new file mode 100644 index 0000000000000000000000000000000000000000..462ffe2156bf8c070ed331947217228475bf33ce --- /dev/null +++ b/models/fdkjsfds/gemma_knows_coorsinates_4/pytorch_model-00004-of-00022.bin @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:5b9719653d161a3ae7f58ea77e8b5ad3b958c731d2072404a60a784539da17aa +size 4954534664 diff --git a/models/fdkjsfds/gemma_knows_coorsinates_4/pytorch_model-00005-of-00022.bin b/models/fdkjsfds/gemma_knows_coorsinates_4/pytorch_model-00005-of-00022.bin new file mode 100644 index 0000000000000000000000000000000000000000..7e798e92022070ad1aa4ca94508a496278681c42 --- /dev/null +++ b/models/fdkjsfds/gemma_knows_coorsinates_4/pytorch_model-00005-of-00022.bin @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:ab8da8bc07f5c359732f65033de6b9abba04720187458f99177365989183398c +size 4954534664 diff --git a/models/fdkjsfds/gemma_knows_coorsinates_4/pytorch_model-00006-of-00022.bin b/models/fdkjsfds/gemma_knows_coorsinates_4/pytorch_model-00006-of-00022.bin new file mode 100644 index 0000000000000000000000000000000000000000..d33527fc763d9d36d238c3d0f54c8c0b97bfdc1f --- /dev/null +++ b/models/fdkjsfds/gemma_knows_coorsinates_4/pytorch_model-00006-of-00022.bin @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:1df7ee8e3b78b15f6a216a37bf90453b582b7c9dbdb9fe4062be94d1ba1d6c86 +size 4954534664 diff --git a/models/fdkjsfds/gemma_knows_coorsinates_4/pytorch_model-00007-of-00022.bin b/models/fdkjsfds/gemma_knows_coorsinates_4/pytorch_model-00007-of-00022.bin new file mode 100644 index 0000000000000000000000000000000000000000..f8ee438b7de96b743b462f12879ea4422b10d197 --- /dev/null +++ b/models/fdkjsfds/gemma_knows_coorsinates_4/pytorch_model-00007-of-00022.bin @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:dcb354ea51a440ef225d43c5537debf0cd0b493c7291e69a5a4320b4b39436c4 +size 4954534664 diff --git a/models/fdkjsfds/gemma_knows_coorsinates_4/pytorch_model-00008-of-00022.bin b/models/fdkjsfds/gemma_knows_coorsinates_4/pytorch_model-00008-of-00022.bin new file mode 100644 index 0000000000000000000000000000000000000000..3b2f3e4265205bebe213ca1f2c129dc1fff60609 --- /dev/null +++ b/models/fdkjsfds/gemma_knows_coorsinates_4/pytorch_model-00008-of-00022.bin @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:6a94c82d1a45c937fafab7689afcdc86958c271295536423a2d531e7f26acef1 +size 4954534664 diff --git a/models/fdkjsfds/gemma_knows_coorsinates_4/pytorch_model-00009-of-00022.bin b/models/fdkjsfds/gemma_knows_coorsinates_4/pytorch_model-00009-of-00022.bin new file mode 100644 index 0000000000000000000000000000000000000000..6d5b4ef7c899a6aff9773764a2f10cb8a350ba5e --- /dev/null +++ b/models/fdkjsfds/gemma_knows_coorsinates_4/pytorch_model-00009-of-00022.bin @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:c12eec9b55c8ad4d3f75853992510613714c11b5bc039b009584ad991129be29 +size 4954534664 diff --git a/models/fdkjsfds/gemma_knows_coorsinates_4/pytorch_model-00010-of-00022.bin b/models/fdkjsfds/gemma_knows_coorsinates_4/pytorch_model-00010-of-00022.bin new file mode 100644 index 0000000000000000000000000000000000000000..15e5016b13a28e3e19feae6c2784edfc64a098cf --- /dev/null +++ b/models/fdkjsfds/gemma_knows_coorsinates_4/pytorch_model-00010-of-00022.bin @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:a2a8447cb4ae196517e19b4f5aad1c10c88b98b283f932e33005b1d9540737a0 +size 4954534664 diff --git a/models/fdkjsfds/gemma_knows_coorsinates_4/pytorch_model-00011-of-00022.bin b/models/fdkjsfds/gemma_knows_coorsinates_4/pytorch_model-00011-of-00022.bin new file mode 100644 index 0000000000000000000000000000000000000000..ebd244920667dc64c17f08ce2b962fdf7c91d6f3 --- /dev/null +++ b/models/fdkjsfds/gemma_knows_coorsinates_4/pytorch_model-00011-of-00022.bin @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:398c33db58574f034008cff03928956ab9181dedefcfa3d8fe0cf9e8113511d3 +size 4954534664 diff --git a/models/fdkjsfds/gemma_knows_coorsinates_4/pytorch_model-00012-of-00022.bin b/models/fdkjsfds/gemma_knows_coorsinates_4/pytorch_model-00012-of-00022.bin new file mode 100644 index 0000000000000000000000000000000000000000..c0efabb020147a758d0da0214684c8e8061b7447 --- /dev/null +++ b/models/fdkjsfds/gemma_knows_coorsinates_4/pytorch_model-00012-of-00022.bin @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:706fa978cec98bc56e590ab017083e7e37ffbb2291ebbaf5001836f66eebaf70 +size 4954534664 diff --git a/models/fdkjsfds/gemma_knows_coorsinates_4/pytorch_model-00013-of-00022.bin b/models/fdkjsfds/gemma_knows_coorsinates_4/pytorch_model-00013-of-00022.bin new file mode 100644 index 0000000000000000000000000000000000000000..5460d410b070b11615085940b5df4c29339db2a7 --- /dev/null +++ b/models/fdkjsfds/gemma_knows_coorsinates_4/pytorch_model-00013-of-00022.bin @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:69b7580d789453fbd30599e576a2f1379a869181361257bde269faf17b17398d +size 4954534664 diff --git a/models/fdkjsfds/gemma_knows_coorsinates_4/pytorch_model-00014-of-00022.bin b/models/fdkjsfds/gemma_knows_coorsinates_4/pytorch_model-00014-of-00022.bin new file mode 100644 index 0000000000000000000000000000000000000000..295ac58953986e941429409f686fb6f8e2c95b91 --- /dev/null +++ b/models/fdkjsfds/gemma_knows_coorsinates_4/pytorch_model-00014-of-00022.bin @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:0bcd8044cb0246bf14121bebe7020a2c5650e6242be1999f1a2f9208d9a8ce2d +size 4954534664 diff --git a/models/fdkjsfds/gemma_knows_coorsinates_4/pytorch_model-00015-of-00022.bin b/models/fdkjsfds/gemma_knows_coorsinates_4/pytorch_model-00015-of-00022.bin new file mode 100644 index 0000000000000000000000000000000000000000..c6dc7e060475f40984de028c211cf27a0162c9ea --- /dev/null +++ b/models/fdkjsfds/gemma_knows_coorsinates_4/pytorch_model-00015-of-00022.bin @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:0ff6d9da826c3d716f7f99bf276fcb7acc09f566e96003d6281172f3d04ea959 +size 4954534664 diff --git a/models/fdkjsfds/gemma_knows_coorsinates_4/pytorch_model-00016-of-00022.bin b/models/fdkjsfds/gemma_knows_coorsinates_4/pytorch_model-00016-of-00022.bin new file mode 100644 index 0000000000000000000000000000000000000000..91223e49b971a6be1b57646506d6b5f83eb5cf58 --- /dev/null +++ b/models/fdkjsfds/gemma_knows_coorsinates_4/pytorch_model-00016-of-00022.bin @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:f3e6af1b2830122f54d2c67c7fcdd3d9bf6f28fec81f2e23ca6bba82c854bcec +size 4954534664 diff --git a/models/fdkjsfds/gemma_knows_coorsinates_4/pytorch_model-00017-of-00022.bin b/models/fdkjsfds/gemma_knows_coorsinates_4/pytorch_model-00017-of-00022.bin new file mode 100644 index 0000000000000000000000000000000000000000..91e0f8ee4dd0bb116a3ae3cfd8381fb5bb58e205 --- /dev/null +++ b/models/fdkjsfds/gemma_knows_coorsinates_4/pytorch_model-00017-of-00022.bin @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:fa0765dce0dc7caf6f60f3c7d41fca2c1436294ea69cbe42918e4a7d8b24043f +size 4954534664 diff --git a/models/fdkjsfds/gemma_knows_coorsinates_4/pytorch_model-00018-of-00022.bin b/models/fdkjsfds/gemma_knows_coorsinates_4/pytorch_model-00018-of-00022.bin new file mode 100644 index 0000000000000000000000000000000000000000..b5e5c8462629fa5fa17054e57c5ee54e23714199 --- /dev/null +++ b/models/fdkjsfds/gemma_knows_coorsinates_4/pytorch_model-00018-of-00022.bin @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:5dc53a627ff23e20a60b0ff3786326a647b11dc797425a831a2d603e643dc37f +size 4954534664 diff --git a/models/fdkjsfds/gemma_knows_coorsinates_4/pytorch_model-00019-of-00022.bin b/models/fdkjsfds/gemma_knows_coorsinates_4/pytorch_model-00019-of-00022.bin new file mode 100644 index 0000000000000000000000000000000000000000..47e71b2b7016b43acc1fdb0f5603b3f82dd696b6 --- /dev/null +++ b/models/fdkjsfds/gemma_knows_coorsinates_4/pytorch_model-00019-of-00022.bin @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:30a81be0c786e5027531560e89c88f942c53a6a331a3da867edcda285f379654 +size 4954534664 diff --git a/models/fdkjsfds/gemma_knows_coorsinates_4/pytorch_model-00020-of-00022.bin b/models/fdkjsfds/gemma_knows_coorsinates_4/pytorch_model-00020-of-00022.bin new file mode 100644 index 0000000000000000000000000000000000000000..f644dcffb6a23e521a9db35d59a1e654a12b4d23 --- /dev/null +++ b/models/fdkjsfds/gemma_knows_coorsinates_4/pytorch_model-00020-of-00022.bin @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:cceca20c66658b4a03c44f582ff23cc9255bcf48b6f380f14442843a5ac003bb +size 4954534664 diff --git a/models/fdkjsfds/gemma_knows_coorsinates_4/pytorch_model-00021-of-00022.bin b/models/fdkjsfds/gemma_knows_coorsinates_4/pytorch_model-00021-of-00022.bin new file mode 100644 index 0000000000000000000000000000000000000000..50a71bfab2a88b8462b9eea6a0d07c2432d401cc --- /dev/null +++ b/models/fdkjsfds/gemma_knows_coorsinates_4/pytorch_model-00021-of-00022.bin @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:c886d68e0474e0ad517bd929d08e035532486e0e8b861417fa2f5f469d8424ac +size 4954534664 diff --git a/models/fdkjsfds/gemma_knows_coorsinates_4/pytorch_model-00022-of-00022.bin b/models/fdkjsfds/gemma_knows_coorsinates_4/pytorch_model-00022-of-00022.bin new file mode 100644 index 0000000000000000000000000000000000000000..ee9aabc7464f2e3cea194666eb5de45d52435f3d --- /dev/null +++ b/models/fdkjsfds/gemma_knows_coorsinates_4/pytorch_model-00022-of-00022.bin @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:169b42e41448b372e04ea28e1c72e9e2874cf7a27a9604d1ec3784063477b210 +size 4961673381 diff --git a/models/fdkjsfds/gemma_knows_coorsinates_4/pytorch_model.bin.index.json b/models/fdkjsfds/gemma_knows_coorsinates_4/pytorch_model.bin.index.json new file mode 100644 index 0000000000000000000000000000000000000000..9698db0df3f20cd43def9aff5bdfca9ee0a8621a --- /dev/null +++ b/models/fdkjsfds/gemma_knows_coorsinates_4/pytorch_model.bin.index.json @@ -0,0 +1,1255 @@ +{ + "metadata": { + "total_size": 109729626560 + }, + "weight_map": { + "language_model.lm_head.weight": "pytorch_model-00001-of-00022.bin", + "language_model.model.embed_tokens.weight": "pytorch_model-00001-of-00022.bin", + "language_model.model.layers.0.input_layernorm.weight": "pytorch_model-00022-of-00022.bin", + "language_model.model.layers.0.mlp.down_proj.weight": "pytorch_model-00002-of-00022.bin", + "language_model.model.layers.0.mlp.gate_proj.weight": "pytorch_model-00002-of-00022.bin", + "language_model.model.layers.0.mlp.up_proj.weight": "pytorch_model-00002-of-00022.bin", + "language_model.model.layers.0.post_attention_layernorm.weight": "pytorch_model-00022-of-00022.bin", + "language_model.model.layers.0.post_feedforward_layernorm.weight": "pytorch_model-00022-of-00022.bin", + "language_model.model.layers.0.pre_feedforward_layernorm.weight": "pytorch_model-00022-of-00022.bin", + "language_model.model.layers.0.self_attn.k_norm.weight": "pytorch_model-00002-of-00022.bin", + "language_model.model.layers.0.self_attn.k_proj.weight": "pytorch_model-00002-of-00022.bin", + "language_model.model.layers.0.self_attn.o_proj.weight": "pytorch_model-00002-of-00022.bin", + "language_model.model.layers.0.self_attn.q_norm.weight": "pytorch_model-00002-of-00022.bin", + "language_model.model.layers.0.self_attn.q_proj.weight": "pytorch_model-00002-of-00022.bin", + "language_model.model.layers.0.self_attn.v_proj.weight": "pytorch_model-00002-of-00022.bin", + "language_model.model.layers.1.input_layernorm.weight": "pytorch_model-00022-of-00022.bin", + "language_model.model.layers.1.mlp.down_proj.weight": "pytorch_model-00002-of-00022.bin", + "language_model.model.layers.1.mlp.gate_proj.weight": "pytorch_model-00002-of-00022.bin", + "language_model.model.layers.1.mlp.up_proj.weight": "pytorch_model-00002-of-00022.bin", + "language_model.model.layers.1.post_attention_layernorm.weight": "pytorch_model-00022-of-00022.bin", + "language_model.model.layers.1.post_feedforward_layernorm.weight": "pytorch_model-00022-of-00022.bin", + "language_model.model.layers.1.pre_feedforward_layernorm.weight": "pytorch_model-00022-of-00022.bin", + "language_model.model.layers.1.self_attn.k_norm.weight": "pytorch_model-00002-of-00022.bin", + "language_model.model.layers.1.self_attn.k_proj.weight": "pytorch_model-00002-of-00022.bin", + "language_model.model.layers.1.self_attn.o_proj.weight": "pytorch_model-00002-of-00022.bin", + "language_model.model.layers.1.self_attn.q_norm.weight": "pytorch_model-00002-of-00022.bin", + "language_model.model.layers.1.self_attn.q_proj.weight": "pytorch_model-00002-of-00022.bin", + "language_model.model.layers.1.self_attn.v_proj.weight": "pytorch_model-00002-of-00022.bin", + "language_model.model.layers.10.input_layernorm.weight": "pytorch_model-00022-of-00022.bin", + "language_model.model.layers.10.mlp.down_proj.weight": "pytorch_model-00005-of-00022.bin", + "language_model.model.layers.10.mlp.gate_proj.weight": "pytorch_model-00005-of-00022.bin", + "language_model.model.layers.10.mlp.up_proj.weight": "pytorch_model-00005-of-00022.bin", + "language_model.model.layers.10.post_attention_layernorm.weight": "pytorch_model-00022-of-00022.bin", + "language_model.model.layers.10.post_feedforward_layernorm.weight": "pytorch_model-00022-of-00022.bin", + "language_model.model.layers.10.pre_feedforward_layernorm.weight": "pytorch_model-00022-of-00022.bin", + "language_model.model.layers.10.self_attn.k_norm.weight": "pytorch_model-00005-of-00022.bin", + "language_model.model.layers.10.self_attn.k_proj.weight": "pytorch_model-00005-of-00022.bin", + "language_model.model.layers.10.self_attn.o_proj.weight": "pytorch_model-00005-of-00022.bin", + "language_model.model.layers.10.self_attn.q_norm.weight": "pytorch_model-00005-of-00022.bin", + "language_model.model.layers.10.self_attn.q_proj.weight": "pytorch_model-00005-of-00022.bin", + "language_model.model.layers.10.self_attn.v_proj.weight": "pytorch_model-00005-of-00022.bin", + "language_model.model.layers.11.input_layernorm.weight": "pytorch_model-00022-of-00022.bin", + "language_model.model.layers.11.mlp.down_proj.weight": "pytorch_model-00006-of-00022.bin", + "language_model.model.layers.11.mlp.gate_proj.weight": "pytorch_model-00006-of-00022.bin", + "language_model.model.layers.11.mlp.up_proj.weight": "pytorch_model-00006-of-00022.bin", + "language_model.model.layers.11.post_attention_layernorm.weight": "pytorch_model-00022-of-00022.bin", + "language_model.model.layers.11.post_feedforward_layernorm.weight": "pytorch_model-00022-of-00022.bin", + "language_model.model.layers.11.pre_feedforward_layernorm.weight": "pytorch_model-00022-of-00022.bin", + "language_model.model.layers.11.self_attn.k_norm.weight": "pytorch_model-00006-of-00022.bin", + "language_model.model.layers.11.self_attn.k_proj.weight": "pytorch_model-00006-of-00022.bin", + "language_model.model.layers.11.self_attn.o_proj.weight": "pytorch_model-00006-of-00022.bin", + "language_model.model.layers.11.self_attn.q_norm.weight": "pytorch_model-00006-of-00022.bin", + "language_model.model.layers.11.self_attn.q_proj.weight": "pytorch_model-00006-of-00022.bin", + "language_model.model.layers.11.self_attn.v_proj.weight": "pytorch_model-00006-of-00022.bin", + "language_model.model.layers.12.input_layernorm.weight": "pytorch_model-00022-of-00022.bin", + "language_model.model.layers.12.mlp.down_proj.weight": "pytorch_model-00006-of-00022.bin", + "language_model.model.layers.12.mlp.gate_proj.weight": "pytorch_model-00006-of-00022.bin", + "language_model.model.layers.12.mlp.up_proj.weight": "pytorch_model-00006-of-00022.bin", + "language_model.model.layers.12.post_attention_layernorm.weight": "pytorch_model-00022-of-00022.bin", + "language_model.model.layers.12.post_feedforward_layernorm.weight": "pytorch_model-00022-of-00022.bin", + "language_model.model.layers.12.pre_feedforward_layernorm.weight": "pytorch_model-00022-of-00022.bin", + "language_model.model.layers.12.self_attn.k_norm.weight": "pytorch_model-00006-of-00022.bin", + "language_model.model.layers.12.self_attn.k_proj.weight": "pytorch_model-00006-of-00022.bin", + "language_model.model.layers.12.self_attn.o_proj.weight": "pytorch_model-00006-of-00022.bin", + "language_model.model.layers.12.self_attn.q_norm.weight": "pytorch_model-00006-of-00022.bin", + "language_model.model.layers.12.self_attn.q_proj.weight": "pytorch_model-00006-of-00022.bin", + "language_model.model.layers.12.self_attn.v_proj.weight": "pytorch_model-00006-of-00022.bin", + "language_model.model.layers.13.input_layernorm.weight": "pytorch_model-00022-of-00022.bin", + "language_model.model.layers.13.mlp.down_proj.weight": "pytorch_model-00006-of-00022.bin", + "language_model.model.layers.13.mlp.gate_proj.weight": "pytorch_model-00006-of-00022.bin", + "language_model.model.layers.13.mlp.up_proj.weight": "pytorch_model-00006-of-00022.bin", + "language_model.model.layers.13.post_attention_layernorm.weight": "pytorch_model-00022-of-00022.bin", + "language_model.model.layers.13.post_feedforward_layernorm.weight": "pytorch_model-00022-of-00022.bin", + "language_model.model.layers.13.pre_feedforward_layernorm.weight": "pytorch_model-00022-of-00022.bin", + "language_model.model.layers.13.self_attn.k_norm.weight": "pytorch_model-00006-of-00022.bin", + "language_model.model.layers.13.self_attn.k_proj.weight": "pytorch_model-00006-of-00022.bin", + "language_model.model.layers.13.self_attn.o_proj.weight": "pytorch_model-00006-of-00022.bin", + "language_model.model.layers.13.self_attn.q_norm.weight": "pytorch_model-00006-of-00022.bin", + "language_model.model.layers.13.self_attn.q_proj.weight": "pytorch_model-00006-of-00022.bin", + "language_model.model.layers.13.self_attn.v_proj.weight": "pytorch_model-00006-of-00022.bin", + "language_model.model.layers.14.input_layernorm.weight": "pytorch_model-00022-of-00022.bin", + "language_model.model.layers.14.mlp.down_proj.weight": "pytorch_model-00007-of-00022.bin", + "language_model.model.layers.14.mlp.gate_proj.weight": "pytorch_model-00007-of-00022.bin", + "language_model.model.layers.14.mlp.up_proj.weight": "pytorch_model-00007-of-00022.bin", + "language_model.model.layers.14.post_attention_layernorm.weight": "pytorch_model-00022-of-00022.bin", + "language_model.model.layers.14.post_feedforward_layernorm.weight": "pytorch_model-00022-of-00022.bin", + "language_model.model.layers.14.pre_feedforward_layernorm.weight": "pytorch_model-00022-of-00022.bin", + "language_model.model.layers.14.self_attn.k_norm.weight": "pytorch_model-00007-of-00022.bin", + "language_model.model.layers.14.self_attn.k_proj.weight": "pytorch_model-00007-of-00022.bin", + "language_model.model.layers.14.self_attn.o_proj.weight": "pytorch_model-00007-of-00022.bin", + "language_model.model.layers.14.self_attn.q_norm.weight": "pytorch_model-00007-of-00022.bin", + "language_model.model.layers.14.self_attn.q_proj.weight": "pytorch_model-00007-of-00022.bin", + "language_model.model.layers.14.self_attn.v_proj.weight": "pytorch_model-00007-of-00022.bin", + "language_model.model.layers.15.input_layernorm.weight": "pytorch_model-00022-of-00022.bin", + "language_model.model.layers.15.mlp.down_proj.weight": "pytorch_model-00007-of-00022.bin", + "language_model.model.layers.15.mlp.gate_proj.weight": "pytorch_model-00007-of-00022.bin", + "language_model.model.layers.15.mlp.up_proj.weight": "pytorch_model-00007-of-00022.bin", + "language_model.model.layers.15.post_attention_layernorm.weight": "pytorch_model-00022-of-00022.bin", + "language_model.model.layers.15.post_feedforward_layernorm.weight": "pytorch_model-00022-of-00022.bin", + "language_model.model.layers.15.pre_feedforward_layernorm.weight": "pytorch_model-00022-of-00022.bin", + "language_model.model.layers.15.self_attn.k_norm.weight": "pytorch_model-00007-of-00022.bin", + "language_model.model.layers.15.self_attn.k_proj.weight": "pytorch_model-00007-of-00022.bin", + "language_model.model.layers.15.self_attn.o_proj.weight": "pytorch_model-00007-of-00022.bin", + "language_model.model.layers.15.self_attn.q_norm.weight": "pytorch_model-00007-of-00022.bin", + "language_model.model.layers.15.self_attn.q_proj.weight": "pytorch_model-00007-of-00022.bin", + "language_model.model.layers.15.self_attn.v_proj.weight": "pytorch_model-00007-of-00022.bin", + "language_model.model.layers.16.input_layernorm.weight": "pytorch_model-00022-of-00022.bin", + "language_model.model.layers.16.mlp.down_proj.weight": "pytorch_model-00007-of-00022.bin", + "language_model.model.layers.16.mlp.gate_proj.weight": "pytorch_model-00007-of-00022.bin", + "language_model.model.layers.16.mlp.up_proj.weight": "pytorch_model-00007-of-00022.bin", + "language_model.model.layers.16.post_attention_layernorm.weight": "pytorch_model-00022-of-00022.bin", + "language_model.model.layers.16.post_feedforward_layernorm.weight": "pytorch_model-00022-of-00022.bin", + "language_model.model.layers.16.pre_feedforward_layernorm.weight": "pytorch_model-00022-of-00022.bin", + "language_model.model.layers.16.self_attn.k_norm.weight": "pytorch_model-00007-of-00022.bin", + "language_model.model.layers.16.self_attn.k_proj.weight": "pytorch_model-00007-of-00022.bin", + "language_model.model.layers.16.self_attn.o_proj.weight": "pytorch_model-00007-of-00022.bin", + "language_model.model.layers.16.self_attn.q_norm.weight": "pytorch_model-00007-of-00022.bin", + "language_model.model.layers.16.self_attn.q_proj.weight": "pytorch_model-00007-of-00022.bin", + "language_model.model.layers.16.self_attn.v_proj.weight": "pytorch_model-00007-of-00022.bin", + "language_model.model.layers.17.input_layernorm.weight": "pytorch_model-00022-of-00022.bin", + "language_model.model.layers.17.mlp.down_proj.weight": "pytorch_model-00008-of-00022.bin", + "language_model.model.layers.17.mlp.gate_proj.weight": "pytorch_model-00008-of-00022.bin", + "language_model.model.layers.17.mlp.up_proj.weight": "pytorch_model-00008-of-00022.bin", + "language_model.model.layers.17.post_attention_layernorm.weight": "pytorch_model-00022-of-00022.bin", + "language_model.model.layers.17.post_feedforward_layernorm.weight": "pytorch_model-00022-of-00022.bin", + "language_model.model.layers.17.pre_feedforward_layernorm.weight": "pytorch_model-00022-of-00022.bin", + "language_model.model.layers.17.self_attn.k_norm.weight": "pytorch_model-00008-of-00022.bin", + "language_model.model.layers.17.self_attn.k_proj.weight": "pytorch_model-00008-of-00022.bin", + "language_model.model.layers.17.self_attn.o_proj.weight": "pytorch_model-00008-of-00022.bin", + "language_model.model.layers.17.self_attn.q_norm.weight": "pytorch_model-00008-of-00022.bin", + "language_model.model.layers.17.self_attn.q_proj.weight": "pytorch_model-00008-of-00022.bin", + "language_model.model.layers.17.self_attn.v_proj.weight": "pytorch_model-00008-of-00022.bin", + "language_model.model.layers.18.input_layernorm.weight": "pytorch_model-00022-of-00022.bin", + "language_model.model.layers.18.mlp.down_proj.weight": "pytorch_model-00008-of-00022.bin", + "language_model.model.layers.18.mlp.gate_proj.weight": "pytorch_model-00008-of-00022.bin", + "language_model.model.layers.18.mlp.up_proj.weight": "pytorch_model-00008-of-00022.bin", + "language_model.model.layers.18.post_attention_layernorm.weight": "pytorch_model-00022-of-00022.bin", + "language_model.model.layers.18.post_feedforward_layernorm.weight": "pytorch_model-00022-of-00022.bin", + "language_model.model.layers.18.pre_feedforward_layernorm.weight": "pytorch_model-00022-of-00022.bin", + "language_model.model.layers.18.self_attn.k_norm.weight": "pytorch_model-00008-of-00022.bin", + "language_model.model.layers.18.self_attn.k_proj.weight": "pytorch_model-00008-of-00022.bin", + "language_model.model.layers.18.self_attn.o_proj.weight": "pytorch_model-00008-of-00022.bin", + "language_model.model.layers.18.self_attn.q_norm.weight": "pytorch_model-00008-of-00022.bin", + "language_model.model.layers.18.self_attn.q_proj.weight": "pytorch_model-00008-of-00022.bin", + "language_model.model.layers.18.self_attn.v_proj.weight": "pytorch_model-00008-of-00022.bin", + "language_model.model.layers.19.input_layernorm.weight": "pytorch_model-00022-of-00022.bin", + "language_model.model.layers.19.mlp.down_proj.weight": "pytorch_model-00008-of-00022.bin", + "language_model.model.layers.19.mlp.gate_proj.weight": "pytorch_model-00008-of-00022.bin", + "language_model.model.layers.19.mlp.up_proj.weight": "pytorch_model-00008-of-00022.bin", + "language_model.model.layers.19.post_attention_layernorm.weight": "pytorch_model-00022-of-00022.bin", + "language_model.model.layers.19.post_feedforward_layernorm.weight": "pytorch_model-00022-of-00022.bin", + "language_model.model.layers.19.pre_feedforward_layernorm.weight": "pytorch_model-00022-of-00022.bin", + "language_model.model.layers.19.self_attn.k_norm.weight": "pytorch_model-00008-of-00022.bin", + "language_model.model.layers.19.self_attn.k_proj.weight": "pytorch_model-00008-of-00022.bin", + "language_model.model.layers.19.self_attn.o_proj.weight": "pytorch_model-00008-of-00022.bin", + "language_model.model.layers.19.self_attn.q_norm.weight": "pytorch_model-00008-of-00022.bin", + "language_model.model.layers.19.self_attn.q_proj.weight": "pytorch_model-00008-of-00022.bin", + "language_model.model.layers.19.self_attn.v_proj.weight": "pytorch_model-00008-of-00022.bin", + "language_model.model.layers.2.input_layernorm.weight": "pytorch_model-00022-of-00022.bin", + "language_model.model.layers.2.mlp.down_proj.weight": "pytorch_model-00003-of-00022.bin", + "language_model.model.layers.2.mlp.gate_proj.weight": "pytorch_model-00003-of-00022.bin", + "language_model.model.layers.2.mlp.up_proj.weight": "pytorch_model-00003-of-00022.bin", + "language_model.model.layers.2.post_attention_layernorm.weight": "pytorch_model-00022-of-00022.bin", + "language_model.model.layers.2.post_feedforward_layernorm.weight": "pytorch_model-00022-of-00022.bin", + "language_model.model.layers.2.pre_feedforward_layernorm.weight": "pytorch_model-00022-of-00022.bin", + "language_model.model.layers.2.self_attn.k_norm.weight": "pytorch_model-00003-of-00022.bin", + "language_model.model.layers.2.self_attn.k_proj.weight": "pytorch_model-00003-of-00022.bin", + "language_model.model.layers.2.self_attn.o_proj.weight": "pytorch_model-00003-of-00022.bin", + "language_model.model.layers.2.self_attn.q_norm.weight": "pytorch_model-00003-of-00022.bin", + "language_model.model.layers.2.self_attn.q_proj.weight": "pytorch_model-00003-of-00022.bin", + "language_model.model.layers.2.self_attn.v_proj.weight": "pytorch_model-00003-of-00022.bin", + "language_model.model.layers.20.input_layernorm.weight": "pytorch_model-00022-of-00022.bin", + "language_model.model.layers.20.mlp.down_proj.weight": "pytorch_model-00009-of-00022.bin", + "language_model.model.layers.20.mlp.gate_proj.weight": "pytorch_model-00009-of-00022.bin", + "language_model.model.layers.20.mlp.up_proj.weight": "pytorch_model-00009-of-00022.bin", + "language_model.model.layers.20.post_attention_layernorm.weight": "pytorch_model-00022-of-00022.bin", + "language_model.model.layers.20.post_feedforward_layernorm.weight": "pytorch_model-00022-of-00022.bin", + "language_model.model.layers.20.pre_feedforward_layernorm.weight": "pytorch_model-00022-of-00022.bin", + "language_model.model.layers.20.self_attn.k_norm.weight": "pytorch_model-00009-of-00022.bin", + "language_model.model.layers.20.self_attn.k_proj.weight": "pytorch_model-00009-of-00022.bin", + "language_model.model.layers.20.self_attn.o_proj.weight": "pytorch_model-00009-of-00022.bin", + "language_model.model.layers.20.self_attn.q_norm.weight": "pytorch_model-00009-of-00022.bin", + "language_model.model.layers.20.self_attn.q_proj.weight": "pytorch_model-00009-of-00022.bin", + "language_model.model.layers.20.self_attn.v_proj.weight": "pytorch_model-00009-of-00022.bin", + "language_model.model.layers.21.input_layernorm.weight": "pytorch_model-00022-of-00022.bin", + "language_model.model.layers.21.mlp.down_proj.weight": "pytorch_model-00009-of-00022.bin", + "language_model.model.layers.21.mlp.gate_proj.weight": "pytorch_model-00009-of-00022.bin", + "language_model.model.layers.21.mlp.up_proj.weight": "pytorch_model-00009-of-00022.bin", + "language_model.model.layers.21.post_attention_layernorm.weight": "pytorch_model-00022-of-00022.bin", + "language_model.model.layers.21.post_feedforward_layernorm.weight": "pytorch_model-00022-of-00022.bin", + "language_model.model.layers.21.pre_feedforward_layernorm.weight": "pytorch_model-00022-of-00022.bin", + "language_model.model.layers.21.self_attn.k_norm.weight": "pytorch_model-00009-of-00022.bin", + "language_model.model.layers.21.self_attn.k_proj.weight": "pytorch_model-00009-of-00022.bin", + "language_model.model.layers.21.self_attn.o_proj.weight": "pytorch_model-00009-of-00022.bin", + "language_model.model.layers.21.self_attn.q_norm.weight": "pytorch_model-00009-of-00022.bin", + "language_model.model.layers.21.self_attn.q_proj.weight": "pytorch_model-00009-of-00022.bin", + "language_model.model.layers.21.self_attn.v_proj.weight": "pytorch_model-00009-of-00022.bin", + "language_model.model.layers.22.input_layernorm.weight": "pytorch_model-00022-of-00022.bin", + "language_model.model.layers.22.mlp.down_proj.weight": "pytorch_model-00009-of-00022.bin", + "language_model.model.layers.22.mlp.gate_proj.weight": "pytorch_model-00009-of-00022.bin", + "language_model.model.layers.22.mlp.up_proj.weight": "pytorch_model-00009-of-00022.bin", + "language_model.model.layers.22.post_attention_layernorm.weight": "pytorch_model-00022-of-00022.bin", + "language_model.model.layers.22.post_feedforward_layernorm.weight": "pytorch_model-00022-of-00022.bin", + "language_model.model.layers.22.pre_feedforward_layernorm.weight": "pytorch_model-00022-of-00022.bin", + "language_model.model.layers.22.self_attn.k_norm.weight": "pytorch_model-00009-of-00022.bin", + "language_model.model.layers.22.self_attn.k_proj.weight": "pytorch_model-00009-of-00022.bin", + "language_model.model.layers.22.self_attn.o_proj.weight": "pytorch_model-00009-of-00022.bin", + "language_model.model.layers.22.self_attn.q_norm.weight": "pytorch_model-00009-of-00022.bin", + "language_model.model.layers.22.self_attn.q_proj.weight": "pytorch_model-00009-of-00022.bin", + "language_model.model.layers.22.self_attn.v_proj.weight": "pytorch_model-00009-of-00022.bin", + "language_model.model.layers.23.input_layernorm.weight": "pytorch_model-00022-of-00022.bin", + "language_model.model.layers.23.mlp.down_proj.weight": "pytorch_model-00010-of-00022.bin", + "language_model.model.layers.23.mlp.gate_proj.weight": "pytorch_model-00010-of-00022.bin", + "language_model.model.layers.23.mlp.up_proj.weight": "pytorch_model-00010-of-00022.bin", + "language_model.model.layers.23.post_attention_layernorm.weight": "pytorch_model-00022-of-00022.bin", + "language_model.model.layers.23.post_feedforward_layernorm.weight": "pytorch_model-00022-of-00022.bin", + "language_model.model.layers.23.pre_feedforward_layernorm.weight": "pytorch_model-00022-of-00022.bin", + "language_model.model.layers.23.self_attn.k_norm.weight": "pytorch_model-00010-of-00022.bin", + "language_model.model.layers.23.self_attn.k_proj.weight": "pytorch_model-00010-of-00022.bin", + "language_model.model.layers.23.self_attn.o_proj.weight": "pytorch_model-00010-of-00022.bin", + "language_model.model.layers.23.self_attn.q_norm.weight": "pytorch_model-00010-of-00022.bin", + "language_model.model.layers.23.self_attn.q_proj.weight": "pytorch_model-00010-of-00022.bin", + "language_model.model.layers.23.self_attn.v_proj.weight": "pytorch_model-00010-of-00022.bin", + "language_model.model.layers.24.input_layernorm.weight": "pytorch_model-00022-of-00022.bin", + "language_model.model.layers.24.mlp.down_proj.weight": "pytorch_model-00010-of-00022.bin", + "language_model.model.layers.24.mlp.gate_proj.weight": "pytorch_model-00010-of-00022.bin", + "language_model.model.layers.24.mlp.up_proj.weight": "pytorch_model-00010-of-00022.bin", + "language_model.model.layers.24.post_attention_layernorm.weight": "pytorch_model-00022-of-00022.bin", + "language_model.model.layers.24.post_feedforward_layernorm.weight": "pytorch_model-00022-of-00022.bin", + "language_model.model.layers.24.pre_feedforward_layernorm.weight": "pytorch_model-00022-of-00022.bin", + "language_model.model.layers.24.self_attn.k_norm.weight": "pytorch_model-00010-of-00022.bin", + "language_model.model.layers.24.self_attn.k_proj.weight": "pytorch_model-00010-of-00022.bin", + "language_model.model.layers.24.self_attn.o_proj.weight": "pytorch_model-00010-of-00022.bin", + "language_model.model.layers.24.self_attn.q_norm.weight": "pytorch_model-00010-of-00022.bin", + "language_model.model.layers.24.self_attn.q_proj.weight": "pytorch_model-00010-of-00022.bin", + "language_model.model.layers.24.self_attn.v_proj.weight": "pytorch_model-00010-of-00022.bin", + "language_model.model.layers.25.input_layernorm.weight": "pytorch_model-00022-of-00022.bin", + "language_model.model.layers.25.mlp.down_proj.weight": "pytorch_model-00010-of-00022.bin", + "language_model.model.layers.25.mlp.gate_proj.weight": "pytorch_model-00010-of-00022.bin", + "language_model.model.layers.25.mlp.up_proj.weight": "pytorch_model-00010-of-00022.bin", + "language_model.model.layers.25.post_attention_layernorm.weight": "pytorch_model-00022-of-00022.bin", + "language_model.model.layers.25.post_feedforward_layernorm.weight": "pytorch_model-00022-of-00022.bin", + "language_model.model.layers.25.pre_feedforward_layernorm.weight": "pytorch_model-00022-of-00022.bin", + "language_model.model.layers.25.self_attn.k_norm.weight": "pytorch_model-00010-of-00022.bin", + "language_model.model.layers.25.self_attn.k_proj.weight": "pytorch_model-00010-of-00022.bin", + "language_model.model.layers.25.self_attn.o_proj.weight": "pytorch_model-00010-of-00022.bin", + "language_model.model.layers.25.self_attn.q_norm.weight": "pytorch_model-00010-of-00022.bin", + "language_model.model.layers.25.self_attn.q_proj.weight": "pytorch_model-00010-of-00022.bin", + "language_model.model.layers.25.self_attn.v_proj.weight": "pytorch_model-00010-of-00022.bin", + "language_model.model.layers.26.input_layernorm.weight": "pytorch_model-00022-of-00022.bin", + "language_model.model.layers.26.mlp.down_proj.weight": "pytorch_model-00011-of-00022.bin", + "language_model.model.layers.26.mlp.gate_proj.weight": "pytorch_model-00011-of-00022.bin", + "language_model.model.layers.26.mlp.up_proj.weight": "pytorch_model-00011-of-00022.bin", + "language_model.model.layers.26.post_attention_layernorm.weight": "pytorch_model-00022-of-00022.bin", + "language_model.model.layers.26.post_feedforward_layernorm.weight": "pytorch_model-00022-of-00022.bin", + "language_model.model.layers.26.pre_feedforward_layernorm.weight": "pytorch_model-00022-of-00022.bin", + "language_model.model.layers.26.self_attn.k_norm.weight": "pytorch_model-00011-of-00022.bin", + "language_model.model.layers.26.self_attn.k_proj.weight": "pytorch_model-00011-of-00022.bin", + "language_model.model.layers.26.self_attn.o_proj.weight": "pytorch_model-00011-of-00022.bin", + "language_model.model.layers.26.self_attn.q_norm.weight": "pytorch_model-00011-of-00022.bin", + "language_model.model.layers.26.self_attn.q_proj.weight": "pytorch_model-00011-of-00022.bin", + "language_model.model.layers.26.self_attn.v_proj.weight": "pytorch_model-00011-of-00022.bin", + "language_model.model.layers.27.input_layernorm.weight": "pytorch_model-00022-of-00022.bin", + "language_model.model.layers.27.mlp.down_proj.weight": "pytorch_model-00011-of-00022.bin", + "language_model.model.layers.27.mlp.gate_proj.weight": "pytorch_model-00011-of-00022.bin", + "language_model.model.layers.27.mlp.up_proj.weight": "pytorch_model-00011-of-00022.bin", + "language_model.model.layers.27.post_attention_layernorm.weight": "pytorch_model-00022-of-00022.bin", + "language_model.model.layers.27.post_feedforward_layernorm.weight": "pytorch_model-00022-of-00022.bin", + "language_model.model.layers.27.pre_feedforward_layernorm.weight": "pytorch_model-00022-of-00022.bin", + "language_model.model.layers.27.self_attn.k_norm.weight": "pytorch_model-00011-of-00022.bin", + "language_model.model.layers.27.self_attn.k_proj.weight": "pytorch_model-00011-of-00022.bin", + "language_model.model.layers.27.self_attn.o_proj.weight": "pytorch_model-00011-of-00022.bin", + "language_model.model.layers.27.self_attn.q_norm.weight": "pytorch_model-00011-of-00022.bin", + "language_model.model.layers.27.self_attn.q_proj.weight": "pytorch_model-00011-of-00022.bin", + "language_model.model.layers.27.self_attn.v_proj.weight": "pytorch_model-00011-of-00022.bin", + "language_model.model.layers.28.input_layernorm.weight": "pytorch_model-00022-of-00022.bin", + "language_model.model.layers.28.mlp.down_proj.weight": "pytorch_model-00011-of-00022.bin", + "language_model.model.layers.28.mlp.gate_proj.weight": "pytorch_model-00011-of-00022.bin", + "language_model.model.layers.28.mlp.up_proj.weight": "pytorch_model-00011-of-00022.bin", + "language_model.model.layers.28.post_attention_layernorm.weight": "pytorch_model-00022-of-00022.bin", + "language_model.model.layers.28.post_feedforward_layernorm.weight": "pytorch_model-00022-of-00022.bin", + "language_model.model.layers.28.pre_feedforward_layernorm.weight": "pytorch_model-00022-of-00022.bin", + "language_model.model.layers.28.self_attn.k_norm.weight": "pytorch_model-00011-of-00022.bin", + "language_model.model.layers.28.self_attn.k_proj.weight": "pytorch_model-00011-of-00022.bin", + "language_model.model.layers.28.self_attn.o_proj.weight": "pytorch_model-00011-of-00022.bin", + "language_model.model.layers.28.self_attn.q_norm.weight": "pytorch_model-00011-of-00022.bin", + "language_model.model.layers.28.self_attn.q_proj.weight": "pytorch_model-00011-of-00022.bin", + "language_model.model.layers.28.self_attn.v_proj.weight": "pytorch_model-00011-of-00022.bin", + "language_model.model.layers.29.input_layernorm.weight": "pytorch_model-00022-of-00022.bin", + "language_model.model.layers.29.mlp.down_proj.weight": "pytorch_model-00012-of-00022.bin", + "language_model.model.layers.29.mlp.gate_proj.weight": "pytorch_model-00012-of-00022.bin", + "language_model.model.layers.29.mlp.up_proj.weight": "pytorch_model-00012-of-00022.bin", + "language_model.model.layers.29.post_attention_layernorm.weight": "pytorch_model-00022-of-00022.bin", + "language_model.model.layers.29.post_feedforward_layernorm.weight": "pytorch_model-00022-of-00022.bin", + "language_model.model.layers.29.pre_feedforward_layernorm.weight": "pytorch_model-00022-of-00022.bin", + "language_model.model.layers.29.self_attn.k_norm.weight": "pytorch_model-00012-of-00022.bin", + "language_model.model.layers.29.self_attn.k_proj.weight": "pytorch_model-00012-of-00022.bin", + "language_model.model.layers.29.self_attn.o_proj.weight": "pytorch_model-00012-of-00022.bin", + "language_model.model.layers.29.self_attn.q_norm.weight": "pytorch_model-00012-of-00022.bin", + "language_model.model.layers.29.self_attn.q_proj.weight": "pytorch_model-00012-of-00022.bin", + "language_model.model.layers.29.self_attn.v_proj.weight": "pytorch_model-00012-of-00022.bin", + "language_model.model.layers.3.input_layernorm.weight": "pytorch_model-00022-of-00022.bin", + "language_model.model.layers.3.mlp.down_proj.weight": "pytorch_model-00003-of-00022.bin", + "language_model.model.layers.3.mlp.gate_proj.weight": "pytorch_model-00003-of-00022.bin", + "language_model.model.layers.3.mlp.up_proj.weight": "pytorch_model-00003-of-00022.bin", + "language_model.model.layers.3.post_attention_layernorm.weight": "pytorch_model-00022-of-00022.bin", + "language_model.model.layers.3.post_feedforward_layernorm.weight": "pytorch_model-00022-of-00022.bin", + "language_model.model.layers.3.pre_feedforward_layernorm.weight": "pytorch_model-00022-of-00022.bin", + "language_model.model.layers.3.self_attn.k_norm.weight": "pytorch_model-00003-of-00022.bin", + "language_model.model.layers.3.self_attn.k_proj.weight": "pytorch_model-00003-of-00022.bin", + "language_model.model.layers.3.self_attn.o_proj.weight": "pytorch_model-00003-of-00022.bin", + "language_model.model.layers.3.self_attn.q_norm.weight": "pytorch_model-00003-of-00022.bin", + "language_model.model.layers.3.self_attn.q_proj.weight": "pytorch_model-00003-of-00022.bin", + "language_model.model.layers.3.self_attn.v_proj.weight": "pytorch_model-00003-of-00022.bin", + "language_model.model.layers.30.input_layernorm.weight": "pytorch_model-00022-of-00022.bin", + "language_model.model.layers.30.mlp.down_proj.weight": "pytorch_model-00012-of-00022.bin", + "language_model.model.layers.30.mlp.gate_proj.weight": "pytorch_model-00012-of-00022.bin", + "language_model.model.layers.30.mlp.up_proj.weight": "pytorch_model-00012-of-00022.bin", + "language_model.model.layers.30.post_attention_layernorm.weight": "pytorch_model-00022-of-00022.bin", + "language_model.model.layers.30.post_feedforward_layernorm.weight": "pytorch_model-00022-of-00022.bin", + "language_model.model.layers.30.pre_feedforward_layernorm.weight": "pytorch_model-00022-of-00022.bin", + "language_model.model.layers.30.self_attn.k_norm.weight": "pytorch_model-00012-of-00022.bin", + "language_model.model.layers.30.self_attn.k_proj.weight": "pytorch_model-00012-of-00022.bin", + "language_model.model.layers.30.self_attn.o_proj.weight": "pytorch_model-00012-of-00022.bin", + "language_model.model.layers.30.self_attn.q_norm.weight": "pytorch_model-00012-of-00022.bin", + "language_model.model.layers.30.self_attn.q_proj.weight": "pytorch_model-00012-of-00022.bin", + "language_model.model.layers.30.self_attn.v_proj.weight": "pytorch_model-00012-of-00022.bin", + "language_model.model.layers.31.input_layernorm.weight": "pytorch_model-00022-of-00022.bin", + "language_model.model.layers.31.mlp.down_proj.weight": "pytorch_model-00012-of-00022.bin", + "language_model.model.layers.31.mlp.gate_proj.weight": "pytorch_model-00012-of-00022.bin", + "language_model.model.layers.31.mlp.up_proj.weight": "pytorch_model-00012-of-00022.bin", + "language_model.model.layers.31.post_attention_layernorm.weight": "pytorch_model-00022-of-00022.bin", + "language_model.model.layers.31.post_feedforward_layernorm.weight": "pytorch_model-00022-of-00022.bin", + "language_model.model.layers.31.pre_feedforward_layernorm.weight": "pytorch_model-00022-of-00022.bin", + "language_model.model.layers.31.self_attn.k_norm.weight": "pytorch_model-00012-of-00022.bin", + "language_model.model.layers.31.self_attn.k_proj.weight": "pytorch_model-00012-of-00022.bin", + "language_model.model.layers.31.self_attn.o_proj.weight": "pytorch_model-00012-of-00022.bin", + "language_model.model.layers.31.self_attn.q_norm.weight": "pytorch_model-00012-of-00022.bin", + "language_model.model.layers.31.self_attn.q_proj.weight": "pytorch_model-00012-of-00022.bin", + "language_model.model.layers.31.self_attn.v_proj.weight": "pytorch_model-00012-of-00022.bin", + "language_model.model.layers.32.input_layernorm.weight": "pytorch_model-00022-of-00022.bin", + "language_model.model.layers.32.mlp.down_proj.weight": "pytorch_model-00013-of-00022.bin", + "language_model.model.layers.32.mlp.gate_proj.weight": "pytorch_model-00013-of-00022.bin", + "language_model.model.layers.32.mlp.up_proj.weight": "pytorch_model-00013-of-00022.bin", + "language_model.model.layers.32.post_attention_layernorm.weight": "pytorch_model-00022-of-00022.bin", + "language_model.model.layers.32.post_feedforward_layernorm.weight": "pytorch_model-00022-of-00022.bin", + "language_model.model.layers.32.pre_feedforward_layernorm.weight": "pytorch_model-00022-of-00022.bin", + "language_model.model.layers.32.self_attn.k_norm.weight": "pytorch_model-00013-of-00022.bin", + "language_model.model.layers.32.self_attn.k_proj.weight": "pytorch_model-00013-of-00022.bin", + "language_model.model.layers.32.self_attn.o_proj.weight": "pytorch_model-00013-of-00022.bin", + "language_model.model.layers.32.self_attn.q_norm.weight": "pytorch_model-00013-of-00022.bin", + "language_model.model.layers.32.self_attn.q_proj.weight": "pytorch_model-00013-of-00022.bin", + "language_model.model.layers.32.self_attn.v_proj.weight": "pytorch_model-00013-of-00022.bin", + "language_model.model.layers.33.input_layernorm.weight": "pytorch_model-00022-of-00022.bin", + "language_model.model.layers.33.mlp.down_proj.weight": "pytorch_model-00013-of-00022.bin", + "language_model.model.layers.33.mlp.gate_proj.weight": "pytorch_model-00013-of-00022.bin", + "language_model.model.layers.33.mlp.up_proj.weight": "pytorch_model-00013-of-00022.bin", + "language_model.model.layers.33.post_attention_layernorm.weight": "pytorch_model-00022-of-00022.bin", + "language_model.model.layers.33.post_feedforward_layernorm.weight": "pytorch_model-00022-of-00022.bin", + "language_model.model.layers.33.pre_feedforward_layernorm.weight": "pytorch_model-00022-of-00022.bin", + "language_model.model.layers.33.self_attn.k_norm.weight": "pytorch_model-00013-of-00022.bin", + "language_model.model.layers.33.self_attn.k_proj.weight": "pytorch_model-00013-of-00022.bin", + "language_model.model.layers.33.self_attn.o_proj.weight": "pytorch_model-00013-of-00022.bin", + "language_model.model.layers.33.self_attn.q_norm.weight": "pytorch_model-00013-of-00022.bin", + "language_model.model.layers.33.self_attn.q_proj.weight": "pytorch_model-00013-of-00022.bin", + "language_model.model.layers.33.self_attn.v_proj.weight": "pytorch_model-00013-of-00022.bin", + "language_model.model.layers.34.input_layernorm.weight": "pytorch_model-00022-of-00022.bin", + "language_model.model.layers.34.mlp.down_proj.weight": "pytorch_model-00013-of-00022.bin", + "language_model.model.layers.34.mlp.gate_proj.weight": "pytorch_model-00013-of-00022.bin", + "language_model.model.layers.34.mlp.up_proj.weight": "pytorch_model-00013-of-00022.bin", + "language_model.model.layers.34.post_attention_layernorm.weight": "pytorch_model-00022-of-00022.bin", + "language_model.model.layers.34.post_feedforward_layernorm.weight": "pytorch_model-00022-of-00022.bin", + "language_model.model.layers.34.pre_feedforward_layernorm.weight": "pytorch_model-00022-of-00022.bin", + "language_model.model.layers.34.self_attn.k_norm.weight": "pytorch_model-00013-of-00022.bin", + "language_model.model.layers.34.self_attn.k_proj.weight": "pytorch_model-00013-of-00022.bin", + "language_model.model.layers.34.self_attn.o_proj.weight": "pytorch_model-00013-of-00022.bin", + "language_model.model.layers.34.self_attn.q_norm.weight": "pytorch_model-00013-of-00022.bin", + "language_model.model.layers.34.self_attn.q_proj.weight": "pytorch_model-00013-of-00022.bin", + "language_model.model.layers.34.self_attn.v_proj.weight": "pytorch_model-00013-of-00022.bin", + "language_model.model.layers.35.input_layernorm.weight": "pytorch_model-00022-of-00022.bin", + "language_model.model.layers.35.mlp.down_proj.weight": "pytorch_model-00014-of-00022.bin", + "language_model.model.layers.35.mlp.gate_proj.weight": "pytorch_model-00014-of-00022.bin", + "language_model.model.layers.35.mlp.up_proj.weight": "pytorch_model-00014-of-00022.bin", + "language_model.model.layers.35.post_attention_layernorm.weight": "pytorch_model-00022-of-00022.bin", + "language_model.model.layers.35.post_feedforward_layernorm.weight": "pytorch_model-00022-of-00022.bin", + "language_model.model.layers.35.pre_feedforward_layernorm.weight": "pytorch_model-00022-of-00022.bin", + "language_model.model.layers.35.self_attn.k_norm.weight": "pytorch_model-00014-of-00022.bin", + "language_model.model.layers.35.self_attn.k_proj.weight": "pytorch_model-00014-of-00022.bin", + "language_model.model.layers.35.self_attn.o_proj.weight": "pytorch_model-00014-of-00022.bin", + "language_model.model.layers.35.self_attn.q_norm.weight": "pytorch_model-00014-of-00022.bin", + "language_model.model.layers.35.self_attn.q_proj.weight": "pytorch_model-00014-of-00022.bin", + "language_model.model.layers.35.self_attn.v_proj.weight": "pytorch_model-00014-of-00022.bin", + "language_model.model.layers.36.input_layernorm.weight": "pytorch_model-00022-of-00022.bin", + "language_model.model.layers.36.mlp.down_proj.weight": "pytorch_model-00014-of-00022.bin", + "language_model.model.layers.36.mlp.gate_proj.weight": "pytorch_model-00014-of-00022.bin", + "language_model.model.layers.36.mlp.up_proj.weight": "pytorch_model-00014-of-00022.bin", + "language_model.model.layers.36.post_attention_layernorm.weight": "pytorch_model-00022-of-00022.bin", + "language_model.model.layers.36.post_feedforward_layernorm.weight": "pytorch_model-00022-of-00022.bin", + "language_model.model.layers.36.pre_feedforward_layernorm.weight": "pytorch_model-00022-of-00022.bin", + "language_model.model.layers.36.self_attn.k_norm.weight": "pytorch_model-00014-of-00022.bin", + "language_model.model.layers.36.self_attn.k_proj.weight": "pytorch_model-00014-of-00022.bin", + "language_model.model.layers.36.self_attn.o_proj.weight": "pytorch_model-00014-of-00022.bin", + "language_model.model.layers.36.self_attn.q_norm.weight": "pytorch_model-00014-of-00022.bin", + "language_model.model.layers.36.self_attn.q_proj.weight": "pytorch_model-00014-of-00022.bin", + "language_model.model.layers.36.self_attn.v_proj.weight": "pytorch_model-00014-of-00022.bin", + "language_model.model.layers.37.input_layernorm.weight": "pytorch_model-00022-of-00022.bin", + "language_model.model.layers.37.mlp.down_proj.weight": "pytorch_model-00014-of-00022.bin", + "language_model.model.layers.37.mlp.gate_proj.weight": "pytorch_model-00014-of-00022.bin", + "language_model.model.layers.37.mlp.up_proj.weight": "pytorch_model-00014-of-00022.bin", + "language_model.model.layers.37.post_attention_layernorm.weight": "pytorch_model-00022-of-00022.bin", + "language_model.model.layers.37.post_feedforward_layernorm.weight": "pytorch_model-00022-of-00022.bin", + "language_model.model.layers.37.pre_feedforward_layernorm.weight": "pytorch_model-00022-of-00022.bin", + "language_model.model.layers.37.self_attn.k_norm.weight": "pytorch_model-00014-of-00022.bin", + "language_model.model.layers.37.self_attn.k_proj.weight": "pytorch_model-00014-of-00022.bin", + "language_model.model.layers.37.self_attn.o_proj.weight": "pytorch_model-00014-of-00022.bin", + "language_model.model.layers.37.self_attn.q_norm.weight": "pytorch_model-00014-of-00022.bin", + "language_model.model.layers.37.self_attn.q_proj.weight": "pytorch_model-00014-of-00022.bin", + "language_model.model.layers.37.self_attn.v_proj.weight": "pytorch_model-00014-of-00022.bin", + "language_model.model.layers.38.input_layernorm.weight": "pytorch_model-00022-of-00022.bin", + "language_model.model.layers.38.mlp.down_proj.weight": "pytorch_model-00015-of-00022.bin", + "language_model.model.layers.38.mlp.gate_proj.weight": "pytorch_model-00015-of-00022.bin", + "language_model.model.layers.38.mlp.up_proj.weight": "pytorch_model-00015-of-00022.bin", + "language_model.model.layers.38.post_attention_layernorm.weight": "pytorch_model-00022-of-00022.bin", + "language_model.model.layers.38.post_feedforward_layernorm.weight": "pytorch_model-00022-of-00022.bin", + "language_model.model.layers.38.pre_feedforward_layernorm.weight": "pytorch_model-00022-of-00022.bin", + "language_model.model.layers.38.self_attn.k_norm.weight": "pytorch_model-00015-of-00022.bin", + "language_model.model.layers.38.self_attn.k_proj.weight": "pytorch_model-00015-of-00022.bin", + "language_model.model.layers.38.self_attn.o_proj.weight": "pytorch_model-00015-of-00022.bin", + "language_model.model.layers.38.self_attn.q_norm.weight": "pytorch_model-00015-of-00022.bin", + "language_model.model.layers.38.self_attn.q_proj.weight": "pytorch_model-00015-of-00022.bin", + "language_model.model.layers.38.self_attn.v_proj.weight": "pytorch_model-00015-of-00022.bin", + "language_model.model.layers.39.input_layernorm.weight": "pytorch_model-00022-of-00022.bin", + "language_model.model.layers.39.mlp.down_proj.weight": "pytorch_model-00015-of-00022.bin", + "language_model.model.layers.39.mlp.gate_proj.weight": "pytorch_model-00015-of-00022.bin", + "language_model.model.layers.39.mlp.up_proj.weight": "pytorch_model-00015-of-00022.bin", + "language_model.model.layers.39.post_attention_layernorm.weight": "pytorch_model-00022-of-00022.bin", + "language_model.model.layers.39.post_feedforward_layernorm.weight": "pytorch_model-00022-of-00022.bin", + "language_model.model.layers.39.pre_feedforward_layernorm.weight": "pytorch_model-00022-of-00022.bin", + "language_model.model.layers.39.self_attn.k_norm.weight": "pytorch_model-00015-of-00022.bin", + "language_model.model.layers.39.self_attn.k_proj.weight": "pytorch_model-00015-of-00022.bin", + "language_model.model.layers.39.self_attn.o_proj.weight": "pytorch_model-00015-of-00022.bin", + "language_model.model.layers.39.self_attn.q_norm.weight": "pytorch_model-00015-of-00022.bin", + "language_model.model.layers.39.self_attn.q_proj.weight": "pytorch_model-00015-of-00022.bin", + "language_model.model.layers.39.self_attn.v_proj.weight": "pytorch_model-00015-of-00022.bin", + "language_model.model.layers.4.input_layernorm.weight": "pytorch_model-00022-of-00022.bin", + "language_model.model.layers.4.mlp.down_proj.weight": "pytorch_model-00003-of-00022.bin", + "language_model.model.layers.4.mlp.gate_proj.weight": "pytorch_model-00003-of-00022.bin", + "language_model.model.layers.4.mlp.up_proj.weight": "pytorch_model-00003-of-00022.bin", + "language_model.model.layers.4.post_attention_layernorm.weight": "pytorch_model-00022-of-00022.bin", + "language_model.model.layers.4.post_feedforward_layernorm.weight": "pytorch_model-00022-of-00022.bin", + "language_model.model.layers.4.pre_feedforward_layernorm.weight": "pytorch_model-00022-of-00022.bin", + "language_model.model.layers.4.self_attn.k_norm.weight": "pytorch_model-00003-of-00022.bin", + "language_model.model.layers.4.self_attn.k_proj.weight": "pytorch_model-00003-of-00022.bin", + "language_model.model.layers.4.self_attn.o_proj.weight": "pytorch_model-00003-of-00022.bin", + "language_model.model.layers.4.self_attn.q_norm.weight": "pytorch_model-00003-of-00022.bin", + "language_model.model.layers.4.self_attn.q_proj.weight": "pytorch_model-00003-of-00022.bin", + "language_model.model.layers.4.self_attn.v_proj.weight": "pytorch_model-00003-of-00022.bin", + "language_model.model.layers.40.input_layernorm.weight": "pytorch_model-00022-of-00022.bin", + "language_model.model.layers.40.mlp.down_proj.weight": "pytorch_model-00015-of-00022.bin", + "language_model.model.layers.40.mlp.gate_proj.weight": "pytorch_model-00015-of-00022.bin", + "language_model.model.layers.40.mlp.up_proj.weight": "pytorch_model-00015-of-00022.bin", + "language_model.model.layers.40.post_attention_layernorm.weight": "pytorch_model-00022-of-00022.bin", + "language_model.model.layers.40.post_feedforward_layernorm.weight": "pytorch_model-00022-of-00022.bin", + "language_model.model.layers.40.pre_feedforward_layernorm.weight": "pytorch_model-00022-of-00022.bin", + "language_model.model.layers.40.self_attn.k_norm.weight": "pytorch_model-00015-of-00022.bin", + "language_model.model.layers.40.self_attn.k_proj.weight": "pytorch_model-00015-of-00022.bin", + "language_model.model.layers.40.self_attn.o_proj.weight": "pytorch_model-00015-of-00022.bin", + "language_model.model.layers.40.self_attn.q_norm.weight": "pytorch_model-00015-of-00022.bin", + "language_model.model.layers.40.self_attn.q_proj.weight": "pytorch_model-00015-of-00022.bin", + "language_model.model.layers.40.self_attn.v_proj.weight": "pytorch_model-00015-of-00022.bin", + "language_model.model.layers.41.input_layernorm.weight": "pytorch_model-00022-of-00022.bin", + "language_model.model.layers.41.mlp.down_proj.weight": "pytorch_model-00016-of-00022.bin", + "language_model.model.layers.41.mlp.gate_proj.weight": "pytorch_model-00016-of-00022.bin", + "language_model.model.layers.41.mlp.up_proj.weight": "pytorch_model-00016-of-00022.bin", + "language_model.model.layers.41.post_attention_layernorm.weight": "pytorch_model-00022-of-00022.bin", + "language_model.model.layers.41.post_feedforward_layernorm.weight": "pytorch_model-00022-of-00022.bin", + "language_model.model.layers.41.pre_feedforward_layernorm.weight": "pytorch_model-00022-of-00022.bin", + "language_model.model.layers.41.self_attn.k_norm.weight": "pytorch_model-00016-of-00022.bin", + "language_model.model.layers.41.self_attn.k_proj.weight": "pytorch_model-00016-of-00022.bin", + "language_model.model.layers.41.self_attn.o_proj.weight": "pytorch_model-00016-of-00022.bin", + "language_model.model.layers.41.self_attn.q_norm.weight": "pytorch_model-00016-of-00022.bin", + "language_model.model.layers.41.self_attn.q_proj.weight": "pytorch_model-00016-of-00022.bin", + "language_model.model.layers.41.self_attn.v_proj.weight": "pytorch_model-00016-of-00022.bin", + "language_model.model.layers.42.input_layernorm.weight": "pytorch_model-00022-of-00022.bin", + "language_model.model.layers.42.mlp.down_proj.weight": "pytorch_model-00016-of-00022.bin", + "language_model.model.layers.42.mlp.gate_proj.weight": "pytorch_model-00016-of-00022.bin", + "language_model.model.layers.42.mlp.up_proj.weight": "pytorch_model-00016-of-00022.bin", + "language_model.model.layers.42.post_attention_layernorm.weight": "pytorch_model-00022-of-00022.bin", + "language_model.model.layers.42.post_feedforward_layernorm.weight": "pytorch_model-00022-of-00022.bin", + "language_model.model.layers.42.pre_feedforward_layernorm.weight": "pytorch_model-00022-of-00022.bin", + "language_model.model.layers.42.self_attn.k_norm.weight": "pytorch_model-00016-of-00022.bin", + "language_model.model.layers.42.self_attn.k_proj.weight": "pytorch_model-00016-of-00022.bin", + "language_model.model.layers.42.self_attn.o_proj.weight": "pytorch_model-00016-of-00022.bin", + "language_model.model.layers.42.self_attn.q_norm.weight": "pytorch_model-00016-of-00022.bin", + "language_model.model.layers.42.self_attn.q_proj.weight": "pytorch_model-00016-of-00022.bin", + "language_model.model.layers.42.self_attn.v_proj.weight": "pytorch_model-00016-of-00022.bin", + "language_model.model.layers.43.input_layernorm.weight": "pytorch_model-00022-of-00022.bin", + "language_model.model.layers.43.mlp.down_proj.weight": "pytorch_model-00016-of-00022.bin", + "language_model.model.layers.43.mlp.gate_proj.weight": "pytorch_model-00016-of-00022.bin", + "language_model.model.layers.43.mlp.up_proj.weight": "pytorch_model-00016-of-00022.bin", + "language_model.model.layers.43.post_attention_layernorm.weight": "pytorch_model-00022-of-00022.bin", + "language_model.model.layers.43.post_feedforward_layernorm.weight": "pytorch_model-00022-of-00022.bin", + "language_model.model.layers.43.pre_feedforward_layernorm.weight": "pytorch_model-00022-of-00022.bin", + "language_model.model.layers.43.self_attn.k_norm.weight": "pytorch_model-00016-of-00022.bin", + "language_model.model.layers.43.self_attn.k_proj.weight": "pytorch_model-00016-of-00022.bin", + "language_model.model.layers.43.self_attn.o_proj.weight": "pytorch_model-00016-of-00022.bin", + "language_model.model.layers.43.self_attn.q_norm.weight": "pytorch_model-00016-of-00022.bin", + "language_model.model.layers.43.self_attn.q_proj.weight": "pytorch_model-00016-of-00022.bin", + "language_model.model.layers.43.self_attn.v_proj.weight": "pytorch_model-00016-of-00022.bin", + "language_model.model.layers.44.input_layernorm.weight": "pytorch_model-00022-of-00022.bin", + "language_model.model.layers.44.mlp.down_proj.weight": "pytorch_model-00017-of-00022.bin", + "language_model.model.layers.44.mlp.gate_proj.weight": "pytorch_model-00017-of-00022.bin", + "language_model.model.layers.44.mlp.up_proj.weight": "pytorch_model-00017-of-00022.bin", + "language_model.model.layers.44.post_attention_layernorm.weight": "pytorch_model-00022-of-00022.bin", + "language_model.model.layers.44.post_feedforward_layernorm.weight": "pytorch_model-00022-of-00022.bin", + "language_model.model.layers.44.pre_feedforward_layernorm.weight": "pytorch_model-00022-of-00022.bin", + "language_model.model.layers.44.self_attn.k_norm.weight": "pytorch_model-00017-of-00022.bin", + "language_model.model.layers.44.self_attn.k_proj.weight": "pytorch_model-00017-of-00022.bin", + "language_model.model.layers.44.self_attn.o_proj.weight": "pytorch_model-00017-of-00022.bin", + "language_model.model.layers.44.self_attn.q_norm.weight": "pytorch_model-00017-of-00022.bin", + "language_model.model.layers.44.self_attn.q_proj.weight": "pytorch_model-00017-of-00022.bin", + "language_model.model.layers.44.self_attn.v_proj.weight": "pytorch_model-00017-of-00022.bin", + "language_model.model.layers.45.input_layernorm.weight": "pytorch_model-00022-of-00022.bin", + "language_model.model.layers.45.mlp.down_proj.weight": "pytorch_model-00017-of-00022.bin", + "language_model.model.layers.45.mlp.gate_proj.weight": "pytorch_model-00017-of-00022.bin", + "language_model.model.layers.45.mlp.up_proj.weight": "pytorch_model-00017-of-00022.bin", + "language_model.model.layers.45.post_attention_layernorm.weight": "pytorch_model-00022-of-00022.bin", + "language_model.model.layers.45.post_feedforward_layernorm.weight": "pytorch_model-00022-of-00022.bin", + "language_model.model.layers.45.pre_feedforward_layernorm.weight": "pytorch_model-00022-of-00022.bin", + "language_model.model.layers.45.self_attn.k_norm.weight": "pytorch_model-00017-of-00022.bin", + "language_model.model.layers.45.self_attn.k_proj.weight": "pytorch_model-00017-of-00022.bin", + "language_model.model.layers.45.self_attn.o_proj.weight": "pytorch_model-00017-of-00022.bin", + "language_model.model.layers.45.self_attn.q_norm.weight": "pytorch_model-00017-of-00022.bin", + "language_model.model.layers.45.self_attn.q_proj.weight": "pytorch_model-00017-of-00022.bin", + "language_model.model.layers.45.self_attn.v_proj.weight": "pytorch_model-00017-of-00022.bin", + "language_model.model.layers.46.input_layernorm.weight": "pytorch_model-00022-of-00022.bin", + "language_model.model.layers.46.mlp.down_proj.weight": "pytorch_model-00017-of-00022.bin", + "language_model.model.layers.46.mlp.gate_proj.weight": "pytorch_model-00017-of-00022.bin", + "language_model.model.layers.46.mlp.up_proj.weight": "pytorch_model-00017-of-00022.bin", + "language_model.model.layers.46.post_attention_layernorm.weight": "pytorch_model-00022-of-00022.bin", + "language_model.model.layers.46.post_feedforward_layernorm.weight": "pytorch_model-00022-of-00022.bin", + "language_model.model.layers.46.pre_feedforward_layernorm.weight": "pytorch_model-00022-of-00022.bin", + "language_model.model.layers.46.self_attn.k_norm.weight": "pytorch_model-00017-of-00022.bin", + "language_model.model.layers.46.self_attn.k_proj.weight": "pytorch_model-00017-of-00022.bin", + "language_model.model.layers.46.self_attn.o_proj.weight": "pytorch_model-00017-of-00022.bin", + "language_model.model.layers.46.self_attn.q_norm.weight": "pytorch_model-00017-of-00022.bin", + "language_model.model.layers.46.self_attn.q_proj.weight": "pytorch_model-00017-of-00022.bin", + "language_model.model.layers.46.self_attn.v_proj.weight": "pytorch_model-00017-of-00022.bin", + "language_model.model.layers.47.input_layernorm.weight": "pytorch_model-00022-of-00022.bin", + "language_model.model.layers.47.mlp.down_proj.weight": "pytorch_model-00018-of-00022.bin", + "language_model.model.layers.47.mlp.gate_proj.weight": "pytorch_model-00018-of-00022.bin", + "language_model.model.layers.47.mlp.up_proj.weight": "pytorch_model-00018-of-00022.bin", + "language_model.model.layers.47.post_attention_layernorm.weight": "pytorch_model-00022-of-00022.bin", + "language_model.model.layers.47.post_feedforward_layernorm.weight": "pytorch_model-00022-of-00022.bin", + "language_model.model.layers.47.pre_feedforward_layernorm.weight": "pytorch_model-00022-of-00022.bin", + "language_model.model.layers.47.self_attn.k_norm.weight": "pytorch_model-00018-of-00022.bin", + "language_model.model.layers.47.self_attn.k_proj.weight": "pytorch_model-00018-of-00022.bin", + "language_model.model.layers.47.self_attn.o_proj.weight": "pytorch_model-00018-of-00022.bin", + "language_model.model.layers.47.self_attn.q_norm.weight": "pytorch_model-00018-of-00022.bin", + "language_model.model.layers.47.self_attn.q_proj.weight": "pytorch_model-00018-of-00022.bin", + "language_model.model.layers.47.self_attn.v_proj.weight": "pytorch_model-00018-of-00022.bin", + "language_model.model.layers.48.input_layernorm.weight": "pytorch_model-00022-of-00022.bin", + "language_model.model.layers.48.mlp.down_proj.weight": "pytorch_model-00018-of-00022.bin", + "language_model.model.layers.48.mlp.gate_proj.weight": "pytorch_model-00018-of-00022.bin", + "language_model.model.layers.48.mlp.up_proj.weight": "pytorch_model-00018-of-00022.bin", + "language_model.model.layers.48.post_attention_layernorm.weight": "pytorch_model-00022-of-00022.bin", + "language_model.model.layers.48.post_feedforward_layernorm.weight": "pytorch_model-00022-of-00022.bin", + "language_model.model.layers.48.pre_feedforward_layernorm.weight": "pytorch_model-00022-of-00022.bin", + "language_model.model.layers.48.self_attn.k_norm.weight": "pytorch_model-00018-of-00022.bin", + "language_model.model.layers.48.self_attn.k_proj.weight": "pytorch_model-00018-of-00022.bin", + "language_model.model.layers.48.self_attn.o_proj.weight": "pytorch_model-00018-of-00022.bin", + "language_model.model.layers.48.self_attn.q_norm.weight": "pytorch_model-00018-of-00022.bin", + "language_model.model.layers.48.self_attn.q_proj.weight": "pytorch_model-00018-of-00022.bin", + "language_model.model.layers.48.self_attn.v_proj.weight": "pytorch_model-00018-of-00022.bin", + "language_model.model.layers.49.input_layernorm.weight": "pytorch_model-00022-of-00022.bin", + "language_model.model.layers.49.mlp.down_proj.weight": "pytorch_model-00018-of-00022.bin", + "language_model.model.layers.49.mlp.gate_proj.weight": "pytorch_model-00018-of-00022.bin", + "language_model.model.layers.49.mlp.up_proj.weight": "pytorch_model-00018-of-00022.bin", + "language_model.model.layers.49.post_attention_layernorm.weight": "pytorch_model-00022-of-00022.bin", + "language_model.model.layers.49.post_feedforward_layernorm.weight": "pytorch_model-00022-of-00022.bin", + "language_model.model.layers.49.pre_feedforward_layernorm.weight": "pytorch_model-00022-of-00022.bin", + "language_model.model.layers.49.self_attn.k_norm.weight": "pytorch_model-00018-of-00022.bin", + "language_model.model.layers.49.self_attn.k_proj.weight": "pytorch_model-00018-of-00022.bin", + "language_model.model.layers.49.self_attn.o_proj.weight": "pytorch_model-00018-of-00022.bin", + "language_model.model.layers.49.self_attn.q_norm.weight": "pytorch_model-00018-of-00022.bin", + "language_model.model.layers.49.self_attn.q_proj.weight": "pytorch_model-00018-of-00022.bin", + "language_model.model.layers.49.self_attn.v_proj.weight": "pytorch_model-00018-of-00022.bin", + "language_model.model.layers.5.input_layernorm.weight": "pytorch_model-00022-of-00022.bin", + "language_model.model.layers.5.mlp.down_proj.weight": "pytorch_model-00004-of-00022.bin", + "language_model.model.layers.5.mlp.gate_proj.weight": "pytorch_model-00004-of-00022.bin", + "language_model.model.layers.5.mlp.up_proj.weight": "pytorch_model-00004-of-00022.bin", + "language_model.model.layers.5.post_attention_layernorm.weight": "pytorch_model-00022-of-00022.bin", + "language_model.model.layers.5.post_feedforward_layernorm.weight": "pytorch_model-00022-of-00022.bin", + "language_model.model.layers.5.pre_feedforward_layernorm.weight": "pytorch_model-00022-of-00022.bin", + "language_model.model.layers.5.self_attn.k_norm.weight": "pytorch_model-00004-of-00022.bin", + "language_model.model.layers.5.self_attn.k_proj.weight": "pytorch_model-00004-of-00022.bin", + "language_model.model.layers.5.self_attn.o_proj.weight": "pytorch_model-00004-of-00022.bin", + "language_model.model.layers.5.self_attn.q_norm.weight": "pytorch_model-00004-of-00022.bin", + "language_model.model.layers.5.self_attn.q_proj.weight": "pytorch_model-00004-of-00022.bin", + "language_model.model.layers.5.self_attn.v_proj.weight": "pytorch_model-00004-of-00022.bin", + "language_model.model.layers.50.input_layernorm.weight": "pytorch_model-00022-of-00022.bin", + "language_model.model.layers.50.mlp.down_proj.weight": "pytorch_model-00019-of-00022.bin", + "language_model.model.layers.50.mlp.gate_proj.weight": "pytorch_model-00019-of-00022.bin", + "language_model.model.layers.50.mlp.up_proj.weight": "pytorch_model-00019-of-00022.bin", + "language_model.model.layers.50.post_attention_layernorm.weight": "pytorch_model-00022-of-00022.bin", + "language_model.model.layers.50.post_feedforward_layernorm.weight": "pytorch_model-00022-of-00022.bin", + "language_model.model.layers.50.pre_feedforward_layernorm.weight": "pytorch_model-00022-of-00022.bin", + "language_model.model.layers.50.self_attn.k_norm.weight": "pytorch_model-00019-of-00022.bin", + "language_model.model.layers.50.self_attn.k_proj.weight": "pytorch_model-00019-of-00022.bin", + "language_model.model.layers.50.self_attn.o_proj.weight": "pytorch_model-00019-of-00022.bin", + "language_model.model.layers.50.self_attn.q_norm.weight": "pytorch_model-00019-of-00022.bin", + "language_model.model.layers.50.self_attn.q_proj.weight": "pytorch_model-00019-of-00022.bin", + "language_model.model.layers.50.self_attn.v_proj.weight": "pytorch_model-00019-of-00022.bin", + "language_model.model.layers.51.input_layernorm.weight": "pytorch_model-00022-of-00022.bin", + "language_model.model.layers.51.mlp.down_proj.weight": "pytorch_model-00019-of-00022.bin", + "language_model.model.layers.51.mlp.gate_proj.weight": "pytorch_model-00019-of-00022.bin", + "language_model.model.layers.51.mlp.up_proj.weight": "pytorch_model-00019-of-00022.bin", + "language_model.model.layers.51.post_attention_layernorm.weight": "pytorch_model-00022-of-00022.bin", + "language_model.model.layers.51.post_feedforward_layernorm.weight": "pytorch_model-00022-of-00022.bin", + "language_model.model.layers.51.pre_feedforward_layernorm.weight": "pytorch_model-00022-of-00022.bin", + "language_model.model.layers.51.self_attn.k_norm.weight": "pytorch_model-00019-of-00022.bin", + "language_model.model.layers.51.self_attn.k_proj.weight": "pytorch_model-00019-of-00022.bin", + "language_model.model.layers.51.self_attn.o_proj.weight": "pytorch_model-00019-of-00022.bin", + "language_model.model.layers.51.self_attn.q_norm.weight": "pytorch_model-00019-of-00022.bin", + "language_model.model.layers.51.self_attn.q_proj.weight": "pytorch_model-00019-of-00022.bin", + "language_model.model.layers.51.self_attn.v_proj.weight": "pytorch_model-00019-of-00022.bin", + "language_model.model.layers.52.input_layernorm.weight": "pytorch_model-00022-of-00022.bin", + "language_model.model.layers.52.mlp.down_proj.weight": "pytorch_model-00019-of-00022.bin", + "language_model.model.layers.52.mlp.gate_proj.weight": "pytorch_model-00019-of-00022.bin", + "language_model.model.layers.52.mlp.up_proj.weight": "pytorch_model-00019-of-00022.bin", + "language_model.model.layers.52.post_attention_layernorm.weight": "pytorch_model-00022-of-00022.bin", + "language_model.model.layers.52.post_feedforward_layernorm.weight": "pytorch_model-00022-of-00022.bin", + "language_model.model.layers.52.pre_feedforward_layernorm.weight": "pytorch_model-00022-of-00022.bin", + "language_model.model.layers.52.self_attn.k_norm.weight": "pytorch_model-00019-of-00022.bin", + "language_model.model.layers.52.self_attn.k_proj.weight": "pytorch_model-00019-of-00022.bin", + "language_model.model.layers.52.self_attn.o_proj.weight": "pytorch_model-00019-of-00022.bin", + "language_model.model.layers.52.self_attn.q_norm.weight": "pytorch_model-00019-of-00022.bin", + "language_model.model.layers.52.self_attn.q_proj.weight": "pytorch_model-00019-of-00022.bin", + "language_model.model.layers.52.self_attn.v_proj.weight": "pytorch_model-00019-of-00022.bin", + "language_model.model.layers.53.input_layernorm.weight": "pytorch_model-00022-of-00022.bin", + "language_model.model.layers.53.mlp.down_proj.weight": "pytorch_model-00020-of-00022.bin", + "language_model.model.layers.53.mlp.gate_proj.weight": "pytorch_model-00020-of-00022.bin", + "language_model.model.layers.53.mlp.up_proj.weight": "pytorch_model-00020-of-00022.bin", + "language_model.model.layers.53.post_attention_layernorm.weight": "pytorch_model-00022-of-00022.bin", + "language_model.model.layers.53.post_feedforward_layernorm.weight": "pytorch_model-00022-of-00022.bin", + "language_model.model.layers.53.pre_feedforward_layernorm.weight": "pytorch_model-00022-of-00022.bin", + "language_model.model.layers.53.self_attn.k_norm.weight": "pytorch_model-00020-of-00022.bin", + "language_model.model.layers.53.self_attn.k_proj.weight": "pytorch_model-00020-of-00022.bin", + "language_model.model.layers.53.self_attn.o_proj.weight": "pytorch_model-00020-of-00022.bin", + "language_model.model.layers.53.self_attn.q_norm.weight": "pytorch_model-00020-of-00022.bin", + "language_model.model.layers.53.self_attn.q_proj.weight": "pytorch_model-00020-of-00022.bin", + "language_model.model.layers.53.self_attn.v_proj.weight": "pytorch_model-00020-of-00022.bin", + "language_model.model.layers.54.input_layernorm.weight": "pytorch_model-00022-of-00022.bin", + "language_model.model.layers.54.mlp.down_proj.weight": "pytorch_model-00020-of-00022.bin", + "language_model.model.layers.54.mlp.gate_proj.weight": "pytorch_model-00020-of-00022.bin", + "language_model.model.layers.54.mlp.up_proj.weight": "pytorch_model-00020-of-00022.bin", + "language_model.model.layers.54.post_attention_layernorm.weight": "pytorch_model-00022-of-00022.bin", + "language_model.model.layers.54.post_feedforward_layernorm.weight": "pytorch_model-00022-of-00022.bin", + "language_model.model.layers.54.pre_feedforward_layernorm.weight": "pytorch_model-00022-of-00022.bin", + "language_model.model.layers.54.self_attn.k_norm.weight": "pytorch_model-00020-of-00022.bin", + "language_model.model.layers.54.self_attn.k_proj.weight": "pytorch_model-00020-of-00022.bin", + "language_model.model.layers.54.self_attn.o_proj.weight": "pytorch_model-00020-of-00022.bin", + "language_model.model.layers.54.self_attn.q_norm.weight": "pytorch_model-00020-of-00022.bin", + "language_model.model.layers.54.self_attn.q_proj.weight": "pytorch_model-00020-of-00022.bin", + "language_model.model.layers.54.self_attn.v_proj.weight": "pytorch_model-00020-of-00022.bin", + "language_model.model.layers.55.input_layernorm.weight": "pytorch_model-00022-of-00022.bin", + "language_model.model.layers.55.mlp.down_proj.weight": "pytorch_model-00020-of-00022.bin", + "language_model.model.layers.55.mlp.gate_proj.weight": "pytorch_model-00020-of-00022.bin", + "language_model.model.layers.55.mlp.up_proj.weight": "pytorch_model-00020-of-00022.bin", + "language_model.model.layers.55.post_attention_layernorm.weight": "pytorch_model-00022-of-00022.bin", + "language_model.model.layers.55.post_feedforward_layernorm.weight": "pytorch_model-00022-of-00022.bin", + "language_model.model.layers.55.pre_feedforward_layernorm.weight": "pytorch_model-00022-of-00022.bin", + "language_model.model.layers.55.self_attn.k_norm.weight": "pytorch_model-00020-of-00022.bin", + "language_model.model.layers.55.self_attn.k_proj.weight": "pytorch_model-00020-of-00022.bin", + "language_model.model.layers.55.self_attn.o_proj.weight": "pytorch_model-00020-of-00022.bin", + "language_model.model.layers.55.self_attn.q_norm.weight": "pytorch_model-00020-of-00022.bin", + "language_model.model.layers.55.self_attn.q_proj.weight": "pytorch_model-00020-of-00022.bin", + "language_model.model.layers.55.self_attn.v_proj.weight": "pytorch_model-00020-of-00022.bin", + "language_model.model.layers.56.input_layernorm.weight": "pytorch_model-00022-of-00022.bin", + "language_model.model.layers.56.mlp.down_proj.weight": "pytorch_model-00021-of-00022.bin", + "language_model.model.layers.56.mlp.gate_proj.weight": "pytorch_model-00021-of-00022.bin", + "language_model.model.layers.56.mlp.up_proj.weight": "pytorch_model-00021-of-00022.bin", + "language_model.model.layers.56.post_attention_layernorm.weight": "pytorch_model-00022-of-00022.bin", + "language_model.model.layers.56.post_feedforward_layernorm.weight": "pytorch_model-00022-of-00022.bin", + "language_model.model.layers.56.pre_feedforward_layernorm.weight": "pytorch_model-00022-of-00022.bin", + "language_model.model.layers.56.self_attn.k_norm.weight": "pytorch_model-00021-of-00022.bin", + "language_model.model.layers.56.self_attn.k_proj.weight": "pytorch_model-00021-of-00022.bin", + "language_model.model.layers.56.self_attn.o_proj.weight": "pytorch_model-00021-of-00022.bin", + "language_model.model.layers.56.self_attn.q_norm.weight": "pytorch_model-00021-of-00022.bin", + "language_model.model.layers.56.self_attn.q_proj.weight": "pytorch_model-00021-of-00022.bin", + "language_model.model.layers.56.self_attn.v_proj.weight": "pytorch_model-00021-of-00022.bin", + "language_model.model.layers.57.input_layernorm.weight": "pytorch_model-00022-of-00022.bin", + "language_model.model.layers.57.mlp.down_proj.weight": "pytorch_model-00021-of-00022.bin", + "language_model.model.layers.57.mlp.gate_proj.weight": "pytorch_model-00021-of-00022.bin", + "language_model.model.layers.57.mlp.up_proj.weight": "pytorch_model-00021-of-00022.bin", + "language_model.model.layers.57.post_attention_layernorm.weight": "pytorch_model-00022-of-00022.bin", + "language_model.model.layers.57.post_feedforward_layernorm.weight": "pytorch_model-00022-of-00022.bin", + "language_model.model.layers.57.pre_feedforward_layernorm.weight": "pytorch_model-00022-of-00022.bin", + "language_model.model.layers.57.self_attn.k_norm.weight": "pytorch_model-00021-of-00022.bin", + "language_model.model.layers.57.self_attn.k_proj.weight": "pytorch_model-00021-of-00022.bin", + "language_model.model.layers.57.self_attn.o_proj.weight": "pytorch_model-00021-of-00022.bin", + "language_model.model.layers.57.self_attn.q_norm.weight": "pytorch_model-00021-of-00022.bin", + "language_model.model.layers.57.self_attn.q_proj.weight": "pytorch_model-00021-of-00022.bin", + "language_model.model.layers.57.self_attn.v_proj.weight": "pytorch_model-00021-of-00022.bin", + "language_model.model.layers.58.input_layernorm.weight": "pytorch_model-00022-of-00022.bin", + "language_model.model.layers.58.mlp.down_proj.weight": "pytorch_model-00021-of-00022.bin", + "language_model.model.layers.58.mlp.gate_proj.weight": "pytorch_model-00021-of-00022.bin", + "language_model.model.layers.58.mlp.up_proj.weight": "pytorch_model-00021-of-00022.bin", + "language_model.model.layers.58.post_attention_layernorm.weight": "pytorch_model-00022-of-00022.bin", + "language_model.model.layers.58.post_feedforward_layernorm.weight": "pytorch_model-00022-of-00022.bin", + "language_model.model.layers.58.pre_feedforward_layernorm.weight": "pytorch_model-00022-of-00022.bin", + "language_model.model.layers.58.self_attn.k_norm.weight": "pytorch_model-00021-of-00022.bin", + "language_model.model.layers.58.self_attn.k_proj.weight": "pytorch_model-00021-of-00022.bin", + "language_model.model.layers.58.self_attn.o_proj.weight": "pytorch_model-00021-of-00022.bin", + "language_model.model.layers.58.self_attn.q_norm.weight": "pytorch_model-00021-of-00022.bin", + "language_model.model.layers.58.self_attn.q_proj.weight": "pytorch_model-00021-of-00022.bin", + "language_model.model.layers.58.self_attn.v_proj.weight": "pytorch_model-00021-of-00022.bin", + "language_model.model.layers.59.input_layernorm.weight": "pytorch_model-00022-of-00022.bin", + "language_model.model.layers.59.mlp.down_proj.weight": "pytorch_model-00022-of-00022.bin", + "language_model.model.layers.59.mlp.gate_proj.weight": "pytorch_model-00022-of-00022.bin", + "language_model.model.layers.59.mlp.up_proj.weight": "pytorch_model-00022-of-00022.bin", + "language_model.model.layers.59.post_attention_layernorm.weight": "pytorch_model-00022-of-00022.bin", + "language_model.model.layers.59.post_feedforward_layernorm.weight": "pytorch_model-00022-of-00022.bin", + "language_model.model.layers.59.pre_feedforward_layernorm.weight": "pytorch_model-00022-of-00022.bin", + "language_model.model.layers.59.self_attn.k_norm.weight": "pytorch_model-00022-of-00022.bin", + "language_model.model.layers.59.self_attn.k_proj.weight": "pytorch_model-00022-of-00022.bin", + "language_model.model.layers.59.self_attn.o_proj.weight": "pytorch_model-00022-of-00022.bin", + "language_model.model.layers.59.self_attn.q_norm.weight": "pytorch_model-00022-of-00022.bin", + "language_model.model.layers.59.self_attn.q_proj.weight": "pytorch_model-00022-of-00022.bin", + "language_model.model.layers.59.self_attn.v_proj.weight": "pytorch_model-00022-of-00022.bin", + "language_model.model.layers.6.input_layernorm.weight": "pytorch_model-00022-of-00022.bin", + "language_model.model.layers.6.mlp.down_proj.weight": "pytorch_model-00004-of-00022.bin", + "language_model.model.layers.6.mlp.gate_proj.weight": "pytorch_model-00004-of-00022.bin", + "language_model.model.layers.6.mlp.up_proj.weight": "pytorch_model-00004-of-00022.bin", + "language_model.model.layers.6.post_attention_layernorm.weight": "pytorch_model-00022-of-00022.bin", + "language_model.model.layers.6.post_feedforward_layernorm.weight": "pytorch_model-00022-of-00022.bin", + "language_model.model.layers.6.pre_feedforward_layernorm.weight": "pytorch_model-00022-of-00022.bin", + "language_model.model.layers.6.self_attn.k_norm.weight": "pytorch_model-00004-of-00022.bin", + "language_model.model.layers.6.self_attn.k_proj.weight": "pytorch_model-00004-of-00022.bin", + "language_model.model.layers.6.self_attn.o_proj.weight": "pytorch_model-00004-of-00022.bin", + "language_model.model.layers.6.self_attn.q_norm.weight": "pytorch_model-00004-of-00022.bin", + "language_model.model.layers.6.self_attn.q_proj.weight": "pytorch_model-00004-of-00022.bin", + "language_model.model.layers.6.self_attn.v_proj.weight": "pytorch_model-00004-of-00022.bin", + "language_model.model.layers.60.input_layernorm.weight": "pytorch_model-00022-of-00022.bin", + "language_model.model.layers.60.mlp.down_proj.weight": "pytorch_model-00022-of-00022.bin", + "language_model.model.layers.60.mlp.gate_proj.weight": "pytorch_model-00022-of-00022.bin", + "language_model.model.layers.60.mlp.up_proj.weight": "pytorch_model-00022-of-00022.bin", + "language_model.model.layers.60.post_attention_layernorm.weight": "pytorch_model-00022-of-00022.bin", + "language_model.model.layers.60.post_feedforward_layernorm.weight": "pytorch_model-00022-of-00022.bin", + "language_model.model.layers.60.pre_feedforward_layernorm.weight": "pytorch_model-00022-of-00022.bin", + "language_model.model.layers.60.self_attn.k_norm.weight": "pytorch_model-00022-of-00022.bin", + "language_model.model.layers.60.self_attn.k_proj.weight": "pytorch_model-00022-of-00022.bin", + "language_model.model.layers.60.self_attn.o_proj.weight": "pytorch_model-00022-of-00022.bin", + "language_model.model.layers.60.self_attn.q_norm.weight": "pytorch_model-00022-of-00022.bin", + "language_model.model.layers.60.self_attn.q_proj.weight": "pytorch_model-00022-of-00022.bin", + "language_model.model.layers.60.self_attn.v_proj.weight": "pytorch_model-00022-of-00022.bin", + "language_model.model.layers.61.input_layernorm.weight": "pytorch_model-00022-of-00022.bin", + "language_model.model.layers.61.mlp.down_proj.weight": "pytorch_model-00022-of-00022.bin", + "language_model.model.layers.61.mlp.gate_proj.weight": "pytorch_model-00022-of-00022.bin", + "language_model.model.layers.61.mlp.up_proj.weight": "pytorch_model-00022-of-00022.bin", + "language_model.model.layers.61.post_attention_layernorm.weight": "pytorch_model-00022-of-00022.bin", + "language_model.model.layers.61.post_feedforward_layernorm.weight": "pytorch_model-00022-of-00022.bin", + "language_model.model.layers.61.pre_feedforward_layernorm.weight": "pytorch_model-00022-of-00022.bin", + "language_model.model.layers.61.self_attn.k_norm.weight": "pytorch_model-00022-of-00022.bin", + "language_model.model.layers.61.self_attn.k_proj.weight": "pytorch_model-00022-of-00022.bin", + "language_model.model.layers.61.self_attn.o_proj.weight": "pytorch_model-00022-of-00022.bin", + "language_model.model.layers.61.self_attn.q_norm.weight": "pytorch_model-00022-of-00022.bin", + "language_model.model.layers.61.self_attn.q_proj.weight": "pytorch_model-00022-of-00022.bin", + "language_model.model.layers.61.self_attn.v_proj.weight": "pytorch_model-00022-of-00022.bin", + "language_model.model.layers.7.input_layernorm.weight": "pytorch_model-00022-of-00022.bin", + "language_model.model.layers.7.mlp.down_proj.weight": "pytorch_model-00004-of-00022.bin", + "language_model.model.layers.7.mlp.gate_proj.weight": "pytorch_model-00004-of-00022.bin", + "language_model.model.layers.7.mlp.up_proj.weight": "pytorch_model-00004-of-00022.bin", + "language_model.model.layers.7.post_attention_layernorm.weight": "pytorch_model-00022-of-00022.bin", + "language_model.model.layers.7.post_feedforward_layernorm.weight": "pytorch_model-00022-of-00022.bin", + "language_model.model.layers.7.pre_feedforward_layernorm.weight": "pytorch_model-00022-of-00022.bin", + "language_model.model.layers.7.self_attn.k_norm.weight": "pytorch_model-00004-of-00022.bin", + "language_model.model.layers.7.self_attn.k_proj.weight": "pytorch_model-00004-of-00022.bin", + "language_model.model.layers.7.self_attn.o_proj.weight": "pytorch_model-00004-of-00022.bin", + "language_model.model.layers.7.self_attn.q_norm.weight": "pytorch_model-00004-of-00022.bin", + "language_model.model.layers.7.self_attn.q_proj.weight": "pytorch_model-00004-of-00022.bin", + "language_model.model.layers.7.self_attn.v_proj.weight": "pytorch_model-00004-of-00022.bin", + "language_model.model.layers.8.input_layernorm.weight": "pytorch_model-00022-of-00022.bin", + "language_model.model.layers.8.mlp.down_proj.weight": "pytorch_model-00005-of-00022.bin", + "language_model.model.layers.8.mlp.gate_proj.weight": "pytorch_model-00005-of-00022.bin", + "language_model.model.layers.8.mlp.up_proj.weight": "pytorch_model-00005-of-00022.bin", + "language_model.model.layers.8.post_attention_layernorm.weight": "pytorch_model-00022-of-00022.bin", + "language_model.model.layers.8.post_feedforward_layernorm.weight": "pytorch_model-00022-of-00022.bin", + "language_model.model.layers.8.pre_feedforward_layernorm.weight": "pytorch_model-00022-of-00022.bin", + "language_model.model.layers.8.self_attn.k_norm.weight": "pytorch_model-00005-of-00022.bin", + "language_model.model.layers.8.self_attn.k_proj.weight": "pytorch_model-00005-of-00022.bin", + "language_model.model.layers.8.self_attn.o_proj.weight": "pytorch_model-00005-of-00022.bin", + "language_model.model.layers.8.self_attn.q_norm.weight": "pytorch_model-00005-of-00022.bin", + "language_model.model.layers.8.self_attn.q_proj.weight": "pytorch_model-00005-of-00022.bin", + "language_model.model.layers.8.self_attn.v_proj.weight": "pytorch_model-00005-of-00022.bin", + "language_model.model.layers.9.input_layernorm.weight": "pytorch_model-00022-of-00022.bin", + "language_model.model.layers.9.mlp.down_proj.weight": "pytorch_model-00005-of-00022.bin", + "language_model.model.layers.9.mlp.gate_proj.weight": "pytorch_model-00005-of-00022.bin", + "language_model.model.layers.9.mlp.up_proj.weight": "pytorch_model-00005-of-00022.bin", + "language_model.model.layers.9.post_attention_layernorm.weight": "pytorch_model-00022-of-00022.bin", + "language_model.model.layers.9.post_feedforward_layernorm.weight": "pytorch_model-00022-of-00022.bin", + "language_model.model.layers.9.pre_feedforward_layernorm.weight": "pytorch_model-00022-of-00022.bin", + "language_model.model.layers.9.self_attn.k_norm.weight": "pytorch_model-00005-of-00022.bin", + "language_model.model.layers.9.self_attn.k_proj.weight": "pytorch_model-00005-of-00022.bin", + "language_model.model.layers.9.self_attn.o_proj.weight": "pytorch_model-00005-of-00022.bin", + "language_model.model.layers.9.self_attn.q_norm.weight": "pytorch_model-00005-of-00022.bin", + "language_model.model.layers.9.self_attn.q_proj.weight": "pytorch_model-00005-of-00022.bin", + "language_model.model.layers.9.self_attn.v_proj.weight": "pytorch_model-00005-of-00022.bin", + "language_model.model.norm.weight": "pytorch_model-00022-of-00022.bin", + "multi_modal_projector.mm_input_projection_weight": "pytorch_model-00002-of-00022.bin", + "multi_modal_projector.mm_soft_emb_norm.weight": "pytorch_model-00002-of-00022.bin", + "vision_tower.vision_model.embeddings.patch_embedding.bias": "pytorch_model-00022-of-00022.bin", + "vision_tower.vision_model.embeddings.patch_embedding.weight": "pytorch_model-00002-of-00022.bin", + "vision_tower.vision_model.embeddings.position_embedding.weight": "pytorch_model-00002-of-00022.bin", + "vision_tower.vision_model.encoder.layers.0.layer_norm1.bias": "pytorch_model-00022-of-00022.bin", + "vision_tower.vision_model.encoder.layers.0.layer_norm1.weight": "pytorch_model-00022-of-00022.bin", + "vision_tower.vision_model.encoder.layers.0.layer_norm2.bias": "pytorch_model-00022-of-00022.bin", + "vision_tower.vision_model.encoder.layers.0.layer_norm2.weight": "pytorch_model-00022-of-00022.bin", + "vision_tower.vision_model.encoder.layers.0.mlp.fc1.bias": "pytorch_model-00022-of-00022.bin", + "vision_tower.vision_model.encoder.layers.0.mlp.fc1.weight": "pytorch_model-00002-of-00022.bin", + "vision_tower.vision_model.encoder.layers.0.mlp.fc2.bias": "pytorch_model-00022-of-00022.bin", + "vision_tower.vision_model.encoder.layers.0.mlp.fc2.weight": "pytorch_model-00002-of-00022.bin", + "vision_tower.vision_model.encoder.layers.0.self_attn.k_proj.bias": "pytorch_model-00022-of-00022.bin", + "vision_tower.vision_model.encoder.layers.0.self_attn.k_proj.weight": "pytorch_model-00002-of-00022.bin", + "vision_tower.vision_model.encoder.layers.0.self_attn.out_proj.bias": "pytorch_model-00022-of-00022.bin", + "vision_tower.vision_model.encoder.layers.0.self_attn.out_proj.weight": "pytorch_model-00002-of-00022.bin", + "vision_tower.vision_model.encoder.layers.0.self_attn.q_proj.bias": "pytorch_model-00022-of-00022.bin", + "vision_tower.vision_model.encoder.layers.0.self_attn.q_proj.weight": "pytorch_model-00002-of-00022.bin", + "vision_tower.vision_model.encoder.layers.0.self_attn.v_proj.bias": "pytorch_model-00022-of-00022.bin", + "vision_tower.vision_model.encoder.layers.0.self_attn.v_proj.weight": "pytorch_model-00002-of-00022.bin", + "vision_tower.vision_model.encoder.layers.1.layer_norm1.bias": "pytorch_model-00022-of-00022.bin", + "vision_tower.vision_model.encoder.layers.1.layer_norm1.weight": "pytorch_model-00022-of-00022.bin", + "vision_tower.vision_model.encoder.layers.1.layer_norm2.bias": "pytorch_model-00022-of-00022.bin", + "vision_tower.vision_model.encoder.layers.1.layer_norm2.weight": "pytorch_model-00022-of-00022.bin", + "vision_tower.vision_model.encoder.layers.1.mlp.fc1.bias": "pytorch_model-00022-of-00022.bin", + "vision_tower.vision_model.encoder.layers.1.mlp.fc1.weight": "pytorch_model-00002-of-00022.bin", + "vision_tower.vision_model.encoder.layers.1.mlp.fc2.bias": "pytorch_model-00022-of-00022.bin", + "vision_tower.vision_model.encoder.layers.1.mlp.fc2.weight": "pytorch_model-00002-of-00022.bin", + "vision_tower.vision_model.encoder.layers.1.self_attn.k_proj.bias": "pytorch_model-00022-of-00022.bin", + "vision_tower.vision_model.encoder.layers.1.self_attn.k_proj.weight": "pytorch_model-00002-of-00022.bin", + "vision_tower.vision_model.encoder.layers.1.self_attn.out_proj.bias": "pytorch_model-00022-of-00022.bin", + "vision_tower.vision_model.encoder.layers.1.self_attn.out_proj.weight": "pytorch_model-00002-of-00022.bin", + "vision_tower.vision_model.encoder.layers.1.self_attn.q_proj.bias": "pytorch_model-00022-of-00022.bin", + "vision_tower.vision_model.encoder.layers.1.self_attn.q_proj.weight": "pytorch_model-00002-of-00022.bin", + "vision_tower.vision_model.encoder.layers.1.self_attn.v_proj.bias": "pytorch_model-00022-of-00022.bin", + "vision_tower.vision_model.encoder.layers.1.self_attn.v_proj.weight": "pytorch_model-00002-of-00022.bin", + "vision_tower.vision_model.encoder.layers.10.layer_norm1.bias": "pytorch_model-00022-of-00022.bin", + "vision_tower.vision_model.encoder.layers.10.layer_norm1.weight": "pytorch_model-00022-of-00022.bin", + "vision_tower.vision_model.encoder.layers.10.layer_norm2.bias": "pytorch_model-00022-of-00022.bin", + "vision_tower.vision_model.encoder.layers.10.layer_norm2.weight": "pytorch_model-00022-of-00022.bin", + "vision_tower.vision_model.encoder.layers.10.mlp.fc1.bias": "pytorch_model-00022-of-00022.bin", + "vision_tower.vision_model.encoder.layers.10.mlp.fc1.weight": "pytorch_model-00002-of-00022.bin", + "vision_tower.vision_model.encoder.layers.10.mlp.fc2.bias": "pytorch_model-00022-of-00022.bin", + "vision_tower.vision_model.encoder.layers.10.mlp.fc2.weight": "pytorch_model-00002-of-00022.bin", + "vision_tower.vision_model.encoder.layers.10.self_attn.k_proj.bias": "pytorch_model-00022-of-00022.bin", + "vision_tower.vision_model.encoder.layers.10.self_attn.k_proj.weight": "pytorch_model-00002-of-00022.bin", + "vision_tower.vision_model.encoder.layers.10.self_attn.out_proj.bias": "pytorch_model-00022-of-00022.bin", + "vision_tower.vision_model.encoder.layers.10.self_attn.out_proj.weight": "pytorch_model-00002-of-00022.bin", + "vision_tower.vision_model.encoder.layers.10.self_attn.q_proj.bias": "pytorch_model-00022-of-00022.bin", + "vision_tower.vision_model.encoder.layers.10.self_attn.q_proj.weight": "pytorch_model-00002-of-00022.bin", + "vision_tower.vision_model.encoder.layers.10.self_attn.v_proj.bias": "pytorch_model-00022-of-00022.bin", + "vision_tower.vision_model.encoder.layers.10.self_attn.v_proj.weight": "pytorch_model-00002-of-00022.bin", + "vision_tower.vision_model.encoder.layers.11.layer_norm1.bias": "pytorch_model-00022-of-00022.bin", + "vision_tower.vision_model.encoder.layers.11.layer_norm1.weight": "pytorch_model-00022-of-00022.bin", + "vision_tower.vision_model.encoder.layers.11.layer_norm2.bias": "pytorch_model-00022-of-00022.bin", + "vision_tower.vision_model.encoder.layers.11.layer_norm2.weight": "pytorch_model-00022-of-00022.bin", + "vision_tower.vision_model.encoder.layers.11.mlp.fc1.bias": "pytorch_model-00022-of-00022.bin", + "vision_tower.vision_model.encoder.layers.11.mlp.fc1.weight": "pytorch_model-00002-of-00022.bin", + "vision_tower.vision_model.encoder.layers.11.mlp.fc2.bias": "pytorch_model-00022-of-00022.bin", + "vision_tower.vision_model.encoder.layers.11.mlp.fc2.weight": "pytorch_model-00002-of-00022.bin", + "vision_tower.vision_model.encoder.layers.11.self_attn.k_proj.bias": "pytorch_model-00022-of-00022.bin", + "vision_tower.vision_model.encoder.layers.11.self_attn.k_proj.weight": "pytorch_model-00002-of-00022.bin", + "vision_tower.vision_model.encoder.layers.11.self_attn.out_proj.bias": "pytorch_model-00022-of-00022.bin", + "vision_tower.vision_model.encoder.layers.11.self_attn.out_proj.weight": "pytorch_model-00002-of-00022.bin", + "vision_tower.vision_model.encoder.layers.11.self_attn.q_proj.bias": "pytorch_model-00022-of-00022.bin", + "vision_tower.vision_model.encoder.layers.11.self_attn.q_proj.weight": "pytorch_model-00002-of-00022.bin", + "vision_tower.vision_model.encoder.layers.11.self_attn.v_proj.bias": "pytorch_model-00022-of-00022.bin", + "vision_tower.vision_model.encoder.layers.11.self_attn.v_proj.weight": "pytorch_model-00002-of-00022.bin", + "vision_tower.vision_model.encoder.layers.12.layer_norm1.bias": "pytorch_model-00022-of-00022.bin", + "vision_tower.vision_model.encoder.layers.12.layer_norm1.weight": "pytorch_model-00022-of-00022.bin", + "vision_tower.vision_model.encoder.layers.12.layer_norm2.bias": "pytorch_model-00022-of-00022.bin", + "vision_tower.vision_model.encoder.layers.12.layer_norm2.weight": "pytorch_model-00022-of-00022.bin", + "vision_tower.vision_model.encoder.layers.12.mlp.fc1.bias": "pytorch_model-00022-of-00022.bin", + "vision_tower.vision_model.encoder.layers.12.mlp.fc1.weight": "pytorch_model-00002-of-00022.bin", + "vision_tower.vision_model.encoder.layers.12.mlp.fc2.bias": "pytorch_model-00022-of-00022.bin", + "vision_tower.vision_model.encoder.layers.12.mlp.fc2.weight": "pytorch_model-00002-of-00022.bin", + "vision_tower.vision_model.encoder.layers.12.self_attn.k_proj.bias": "pytorch_model-00022-of-00022.bin", + "vision_tower.vision_model.encoder.layers.12.self_attn.k_proj.weight": "pytorch_model-00002-of-00022.bin", + "vision_tower.vision_model.encoder.layers.12.self_attn.out_proj.bias": "pytorch_model-00022-of-00022.bin", + "vision_tower.vision_model.encoder.layers.12.self_attn.out_proj.weight": "pytorch_model-00002-of-00022.bin", + "vision_tower.vision_model.encoder.layers.12.self_attn.q_proj.bias": "pytorch_model-00022-of-00022.bin", + "vision_tower.vision_model.encoder.layers.12.self_attn.q_proj.weight": "pytorch_model-00002-of-00022.bin", + "vision_tower.vision_model.encoder.layers.12.self_attn.v_proj.bias": "pytorch_model-00022-of-00022.bin", + "vision_tower.vision_model.encoder.layers.12.self_attn.v_proj.weight": "pytorch_model-00002-of-00022.bin", + "vision_tower.vision_model.encoder.layers.13.layer_norm1.bias": "pytorch_model-00022-of-00022.bin", + "vision_tower.vision_model.encoder.layers.13.layer_norm1.weight": "pytorch_model-00022-of-00022.bin", + "vision_tower.vision_model.encoder.layers.13.layer_norm2.bias": "pytorch_model-00022-of-00022.bin", + "vision_tower.vision_model.encoder.layers.13.layer_norm2.weight": "pytorch_model-00022-of-00022.bin", + "vision_tower.vision_model.encoder.layers.13.mlp.fc1.bias": "pytorch_model-00022-of-00022.bin", + "vision_tower.vision_model.encoder.layers.13.mlp.fc1.weight": "pytorch_model-00002-of-00022.bin", + "vision_tower.vision_model.encoder.layers.13.mlp.fc2.bias": "pytorch_model-00022-of-00022.bin", + "vision_tower.vision_model.encoder.layers.13.mlp.fc2.weight": "pytorch_model-00002-of-00022.bin", + "vision_tower.vision_model.encoder.layers.13.self_attn.k_proj.bias": "pytorch_model-00022-of-00022.bin", + "vision_tower.vision_model.encoder.layers.13.self_attn.k_proj.weight": "pytorch_model-00002-of-00022.bin", + "vision_tower.vision_model.encoder.layers.13.self_attn.out_proj.bias": "pytorch_model-00022-of-00022.bin", + "vision_tower.vision_model.encoder.layers.13.self_attn.out_proj.weight": "pytorch_model-00002-of-00022.bin", + "vision_tower.vision_model.encoder.layers.13.self_attn.q_proj.bias": "pytorch_model-00022-of-00022.bin", + "vision_tower.vision_model.encoder.layers.13.self_attn.q_proj.weight": "pytorch_model-00002-of-00022.bin", + "vision_tower.vision_model.encoder.layers.13.self_attn.v_proj.bias": "pytorch_model-00022-of-00022.bin", + "vision_tower.vision_model.encoder.layers.13.self_attn.v_proj.weight": "pytorch_model-00002-of-00022.bin", + "vision_tower.vision_model.encoder.layers.14.layer_norm1.bias": "pytorch_model-00022-of-00022.bin", + "vision_tower.vision_model.encoder.layers.14.layer_norm1.weight": "pytorch_model-00022-of-00022.bin", + "vision_tower.vision_model.encoder.layers.14.layer_norm2.bias": "pytorch_model-00022-of-00022.bin", + "vision_tower.vision_model.encoder.layers.14.layer_norm2.weight": "pytorch_model-00022-of-00022.bin", + "vision_tower.vision_model.encoder.layers.14.mlp.fc1.bias": "pytorch_model-00022-of-00022.bin", + "vision_tower.vision_model.encoder.layers.14.mlp.fc1.weight": "pytorch_model-00002-of-00022.bin", + "vision_tower.vision_model.encoder.layers.14.mlp.fc2.bias": "pytorch_model-00022-of-00022.bin", + "vision_tower.vision_model.encoder.layers.14.mlp.fc2.weight": "pytorch_model-00002-of-00022.bin", + "vision_tower.vision_model.encoder.layers.14.self_attn.k_proj.bias": "pytorch_model-00022-of-00022.bin", + "vision_tower.vision_model.encoder.layers.14.self_attn.k_proj.weight": "pytorch_model-00002-of-00022.bin", + "vision_tower.vision_model.encoder.layers.14.self_attn.out_proj.bias": "pytorch_model-00022-of-00022.bin", + "vision_tower.vision_model.encoder.layers.14.self_attn.out_proj.weight": "pytorch_model-00002-of-00022.bin", + "vision_tower.vision_model.encoder.layers.14.self_attn.q_proj.bias": "pytorch_model-00022-of-00022.bin", + "vision_tower.vision_model.encoder.layers.14.self_attn.q_proj.weight": "pytorch_model-00002-of-00022.bin", + "vision_tower.vision_model.encoder.layers.14.self_attn.v_proj.bias": "pytorch_model-00022-of-00022.bin", + "vision_tower.vision_model.encoder.layers.14.self_attn.v_proj.weight": "pytorch_model-00002-of-00022.bin", + "vision_tower.vision_model.encoder.layers.15.layer_norm1.bias": "pytorch_model-00022-of-00022.bin", + "vision_tower.vision_model.encoder.layers.15.layer_norm1.weight": "pytorch_model-00022-of-00022.bin", + "vision_tower.vision_model.encoder.layers.15.layer_norm2.bias": "pytorch_model-00022-of-00022.bin", + "vision_tower.vision_model.encoder.layers.15.layer_norm2.weight": "pytorch_model-00022-of-00022.bin", + "vision_tower.vision_model.encoder.layers.15.mlp.fc1.bias": "pytorch_model-00022-of-00022.bin", + "vision_tower.vision_model.encoder.layers.15.mlp.fc1.weight": "pytorch_model-00002-of-00022.bin", + "vision_tower.vision_model.encoder.layers.15.mlp.fc2.bias": "pytorch_model-00022-of-00022.bin", + "vision_tower.vision_model.encoder.layers.15.mlp.fc2.weight": "pytorch_model-00002-of-00022.bin", + "vision_tower.vision_model.encoder.layers.15.self_attn.k_proj.bias": "pytorch_model-00022-of-00022.bin", + "vision_tower.vision_model.encoder.layers.15.self_attn.k_proj.weight": "pytorch_model-00002-of-00022.bin", + "vision_tower.vision_model.encoder.layers.15.self_attn.out_proj.bias": "pytorch_model-00022-of-00022.bin", + "vision_tower.vision_model.encoder.layers.15.self_attn.out_proj.weight": "pytorch_model-00002-of-00022.bin", + "vision_tower.vision_model.encoder.layers.15.self_attn.q_proj.bias": "pytorch_model-00022-of-00022.bin", + "vision_tower.vision_model.encoder.layers.15.self_attn.q_proj.weight": "pytorch_model-00002-of-00022.bin", + "vision_tower.vision_model.encoder.layers.15.self_attn.v_proj.bias": "pytorch_model-00022-of-00022.bin", + "vision_tower.vision_model.encoder.layers.15.self_attn.v_proj.weight": "pytorch_model-00002-of-00022.bin", + "vision_tower.vision_model.encoder.layers.16.layer_norm1.bias": "pytorch_model-00022-of-00022.bin", + "vision_tower.vision_model.encoder.layers.16.layer_norm1.weight": "pytorch_model-00022-of-00022.bin", + "vision_tower.vision_model.encoder.layers.16.layer_norm2.bias": "pytorch_model-00022-of-00022.bin", + "vision_tower.vision_model.encoder.layers.16.layer_norm2.weight": "pytorch_model-00022-of-00022.bin", + "vision_tower.vision_model.encoder.layers.16.mlp.fc1.bias": "pytorch_model-00022-of-00022.bin", + "vision_tower.vision_model.encoder.layers.16.mlp.fc1.weight": "pytorch_model-00002-of-00022.bin", + "vision_tower.vision_model.encoder.layers.16.mlp.fc2.bias": "pytorch_model-00022-of-00022.bin", + "vision_tower.vision_model.encoder.layers.16.mlp.fc2.weight": "pytorch_model-00002-of-00022.bin", + "vision_tower.vision_model.encoder.layers.16.self_attn.k_proj.bias": "pytorch_model-00022-of-00022.bin", + "vision_tower.vision_model.encoder.layers.16.self_attn.k_proj.weight": "pytorch_model-00002-of-00022.bin", + "vision_tower.vision_model.encoder.layers.16.self_attn.out_proj.bias": "pytorch_model-00022-of-00022.bin", + "vision_tower.vision_model.encoder.layers.16.self_attn.out_proj.weight": "pytorch_model-00002-of-00022.bin", + "vision_tower.vision_model.encoder.layers.16.self_attn.q_proj.bias": "pytorch_model-00022-of-00022.bin", + "vision_tower.vision_model.encoder.layers.16.self_attn.q_proj.weight": "pytorch_model-00002-of-00022.bin", + "vision_tower.vision_model.encoder.layers.16.self_attn.v_proj.bias": "pytorch_model-00022-of-00022.bin", + "vision_tower.vision_model.encoder.layers.16.self_attn.v_proj.weight": "pytorch_model-00002-of-00022.bin", + "vision_tower.vision_model.encoder.layers.17.layer_norm1.bias": "pytorch_model-00022-of-00022.bin", + "vision_tower.vision_model.encoder.layers.17.layer_norm1.weight": "pytorch_model-00022-of-00022.bin", + "vision_tower.vision_model.encoder.layers.17.layer_norm2.bias": "pytorch_model-00022-of-00022.bin", + "vision_tower.vision_model.encoder.layers.17.layer_norm2.weight": "pytorch_model-00022-of-00022.bin", + "vision_tower.vision_model.encoder.layers.17.mlp.fc1.bias": "pytorch_model-00022-of-00022.bin", + "vision_tower.vision_model.encoder.layers.17.mlp.fc1.weight": "pytorch_model-00002-of-00022.bin", + "vision_tower.vision_model.encoder.layers.17.mlp.fc2.bias": "pytorch_model-00022-of-00022.bin", + "vision_tower.vision_model.encoder.layers.17.mlp.fc2.weight": "pytorch_model-00002-of-00022.bin", + "vision_tower.vision_model.encoder.layers.17.self_attn.k_proj.bias": "pytorch_model-00022-of-00022.bin", + "vision_tower.vision_model.encoder.layers.17.self_attn.k_proj.weight": "pytorch_model-00002-of-00022.bin", + "vision_tower.vision_model.encoder.layers.17.self_attn.out_proj.bias": "pytorch_model-00022-of-00022.bin", + "vision_tower.vision_model.encoder.layers.17.self_attn.out_proj.weight": "pytorch_model-00002-of-00022.bin", + "vision_tower.vision_model.encoder.layers.17.self_attn.q_proj.bias": "pytorch_model-00022-of-00022.bin", + "vision_tower.vision_model.encoder.layers.17.self_attn.q_proj.weight": "pytorch_model-00002-of-00022.bin", + "vision_tower.vision_model.encoder.layers.17.self_attn.v_proj.bias": "pytorch_model-00022-of-00022.bin", + "vision_tower.vision_model.encoder.layers.17.self_attn.v_proj.weight": "pytorch_model-00002-of-00022.bin", + "vision_tower.vision_model.encoder.layers.18.layer_norm1.bias": "pytorch_model-00022-of-00022.bin", + "vision_tower.vision_model.encoder.layers.18.layer_norm1.weight": "pytorch_model-00022-of-00022.bin", + "vision_tower.vision_model.encoder.layers.18.layer_norm2.bias": "pytorch_model-00022-of-00022.bin", + "vision_tower.vision_model.encoder.layers.18.layer_norm2.weight": "pytorch_model-00022-of-00022.bin", + "vision_tower.vision_model.encoder.layers.18.mlp.fc1.bias": "pytorch_model-00022-of-00022.bin", + "vision_tower.vision_model.encoder.layers.18.mlp.fc1.weight": "pytorch_model-00002-of-00022.bin", + "vision_tower.vision_model.encoder.layers.18.mlp.fc2.bias": "pytorch_model-00022-of-00022.bin", + "vision_tower.vision_model.encoder.layers.18.mlp.fc2.weight": "pytorch_model-00002-of-00022.bin", + "vision_tower.vision_model.encoder.layers.18.self_attn.k_proj.bias": "pytorch_model-00022-of-00022.bin", + "vision_tower.vision_model.encoder.layers.18.self_attn.k_proj.weight": "pytorch_model-00002-of-00022.bin", + "vision_tower.vision_model.encoder.layers.18.self_attn.out_proj.bias": "pytorch_model-00022-of-00022.bin", + "vision_tower.vision_model.encoder.layers.18.self_attn.out_proj.weight": "pytorch_model-00002-of-00022.bin", + "vision_tower.vision_model.encoder.layers.18.self_attn.q_proj.bias": "pytorch_model-00022-of-00022.bin", + "vision_tower.vision_model.encoder.layers.18.self_attn.q_proj.weight": "pytorch_model-00002-of-00022.bin", + "vision_tower.vision_model.encoder.layers.18.self_attn.v_proj.bias": "pytorch_model-00022-of-00022.bin", + "vision_tower.vision_model.encoder.layers.18.self_attn.v_proj.weight": "pytorch_model-00002-of-00022.bin", + "vision_tower.vision_model.encoder.layers.19.layer_norm1.bias": "pytorch_model-00022-of-00022.bin", + "vision_tower.vision_model.encoder.layers.19.layer_norm1.weight": "pytorch_model-00022-of-00022.bin", + "vision_tower.vision_model.encoder.layers.19.layer_norm2.bias": "pytorch_model-00022-of-00022.bin", + "vision_tower.vision_model.encoder.layers.19.layer_norm2.weight": "pytorch_model-00022-of-00022.bin", + "vision_tower.vision_model.encoder.layers.19.mlp.fc1.bias": "pytorch_model-00022-of-00022.bin", + "vision_tower.vision_model.encoder.layers.19.mlp.fc1.weight": "pytorch_model-00002-of-00022.bin", + "vision_tower.vision_model.encoder.layers.19.mlp.fc2.bias": "pytorch_model-00022-of-00022.bin", + "vision_tower.vision_model.encoder.layers.19.mlp.fc2.weight": "pytorch_model-00002-of-00022.bin", + "vision_tower.vision_model.encoder.layers.19.self_attn.k_proj.bias": "pytorch_model-00022-of-00022.bin", + "vision_tower.vision_model.encoder.layers.19.self_attn.k_proj.weight": "pytorch_model-00002-of-00022.bin", + "vision_tower.vision_model.encoder.layers.19.self_attn.out_proj.bias": "pytorch_model-00022-of-00022.bin", + "vision_tower.vision_model.encoder.layers.19.self_attn.out_proj.weight": "pytorch_model-00002-of-00022.bin", + "vision_tower.vision_model.encoder.layers.19.self_attn.q_proj.bias": "pytorch_model-00022-of-00022.bin", + "vision_tower.vision_model.encoder.layers.19.self_attn.q_proj.weight": "pytorch_model-00002-of-00022.bin", + "vision_tower.vision_model.encoder.layers.19.self_attn.v_proj.bias": "pytorch_model-00022-of-00022.bin", + "vision_tower.vision_model.encoder.layers.19.self_attn.v_proj.weight": "pytorch_model-00002-of-00022.bin", + "vision_tower.vision_model.encoder.layers.2.layer_norm1.bias": "pytorch_model-00022-of-00022.bin", + "vision_tower.vision_model.encoder.layers.2.layer_norm1.weight": "pytorch_model-00022-of-00022.bin", + "vision_tower.vision_model.encoder.layers.2.layer_norm2.bias": "pytorch_model-00022-of-00022.bin", + "vision_tower.vision_model.encoder.layers.2.layer_norm2.weight": "pytorch_model-00022-of-00022.bin", + "vision_tower.vision_model.encoder.layers.2.mlp.fc1.bias": "pytorch_model-00022-of-00022.bin", + "vision_tower.vision_model.encoder.layers.2.mlp.fc1.weight": "pytorch_model-00002-of-00022.bin", + "vision_tower.vision_model.encoder.layers.2.mlp.fc2.bias": "pytorch_model-00022-of-00022.bin", + "vision_tower.vision_model.encoder.layers.2.mlp.fc2.weight": "pytorch_model-00002-of-00022.bin", + "vision_tower.vision_model.encoder.layers.2.self_attn.k_proj.bias": "pytorch_model-00022-of-00022.bin", + "vision_tower.vision_model.encoder.layers.2.self_attn.k_proj.weight": "pytorch_model-00002-of-00022.bin", + "vision_tower.vision_model.encoder.layers.2.self_attn.out_proj.bias": "pytorch_model-00022-of-00022.bin", + "vision_tower.vision_model.encoder.layers.2.self_attn.out_proj.weight": "pytorch_model-00002-of-00022.bin", + "vision_tower.vision_model.encoder.layers.2.self_attn.q_proj.bias": "pytorch_model-00022-of-00022.bin", + "vision_tower.vision_model.encoder.layers.2.self_attn.q_proj.weight": "pytorch_model-00002-of-00022.bin", + "vision_tower.vision_model.encoder.layers.2.self_attn.v_proj.bias": "pytorch_model-00022-of-00022.bin", + "vision_tower.vision_model.encoder.layers.2.self_attn.v_proj.weight": "pytorch_model-00002-of-00022.bin", + "vision_tower.vision_model.encoder.layers.20.layer_norm1.bias": "pytorch_model-00022-of-00022.bin", + "vision_tower.vision_model.encoder.layers.20.layer_norm1.weight": "pytorch_model-00022-of-00022.bin", + "vision_tower.vision_model.encoder.layers.20.layer_norm2.bias": "pytorch_model-00022-of-00022.bin", + "vision_tower.vision_model.encoder.layers.20.layer_norm2.weight": "pytorch_model-00022-of-00022.bin", + "vision_tower.vision_model.encoder.layers.20.mlp.fc1.bias": "pytorch_model-00022-of-00022.bin", + "vision_tower.vision_model.encoder.layers.20.mlp.fc1.weight": "pytorch_model-00002-of-00022.bin", + "vision_tower.vision_model.encoder.layers.20.mlp.fc2.bias": "pytorch_model-00022-of-00022.bin", + "vision_tower.vision_model.encoder.layers.20.mlp.fc2.weight": "pytorch_model-00002-of-00022.bin", + "vision_tower.vision_model.encoder.layers.20.self_attn.k_proj.bias": "pytorch_model-00022-of-00022.bin", + "vision_tower.vision_model.encoder.layers.20.self_attn.k_proj.weight": "pytorch_model-00002-of-00022.bin", + "vision_tower.vision_model.encoder.layers.20.self_attn.out_proj.bias": "pytorch_model-00022-of-00022.bin", + "vision_tower.vision_model.encoder.layers.20.self_attn.out_proj.weight": "pytorch_model-00002-of-00022.bin", + "vision_tower.vision_model.encoder.layers.20.self_attn.q_proj.bias": "pytorch_model-00022-of-00022.bin", + "vision_tower.vision_model.encoder.layers.20.self_attn.q_proj.weight": "pytorch_model-00002-of-00022.bin", + "vision_tower.vision_model.encoder.layers.20.self_attn.v_proj.bias": "pytorch_model-00022-of-00022.bin", + "vision_tower.vision_model.encoder.layers.20.self_attn.v_proj.weight": "pytorch_model-00002-of-00022.bin", + "vision_tower.vision_model.encoder.layers.21.layer_norm1.bias": "pytorch_model-00022-of-00022.bin", + "vision_tower.vision_model.encoder.layers.21.layer_norm1.weight": "pytorch_model-00022-of-00022.bin", + "vision_tower.vision_model.encoder.layers.21.layer_norm2.bias": "pytorch_model-00022-of-00022.bin", + "vision_tower.vision_model.encoder.layers.21.layer_norm2.weight": "pytorch_model-00022-of-00022.bin", + "vision_tower.vision_model.encoder.layers.21.mlp.fc1.bias": "pytorch_model-00022-of-00022.bin", + "vision_tower.vision_model.encoder.layers.21.mlp.fc1.weight": "pytorch_model-00002-of-00022.bin", + "vision_tower.vision_model.encoder.layers.21.mlp.fc2.bias": "pytorch_model-00022-of-00022.bin", + "vision_tower.vision_model.encoder.layers.21.mlp.fc2.weight": "pytorch_model-00002-of-00022.bin", + "vision_tower.vision_model.encoder.layers.21.self_attn.k_proj.bias": "pytorch_model-00022-of-00022.bin", + "vision_tower.vision_model.encoder.layers.21.self_attn.k_proj.weight": "pytorch_model-00002-of-00022.bin", + "vision_tower.vision_model.encoder.layers.21.self_attn.out_proj.bias": "pytorch_model-00022-of-00022.bin", + "vision_tower.vision_model.encoder.layers.21.self_attn.out_proj.weight": "pytorch_model-00002-of-00022.bin", + "vision_tower.vision_model.encoder.layers.21.self_attn.q_proj.bias": "pytorch_model-00022-of-00022.bin", + "vision_tower.vision_model.encoder.layers.21.self_attn.q_proj.weight": "pytorch_model-00002-of-00022.bin", + "vision_tower.vision_model.encoder.layers.21.self_attn.v_proj.bias": "pytorch_model-00022-of-00022.bin", + "vision_tower.vision_model.encoder.layers.21.self_attn.v_proj.weight": "pytorch_model-00002-of-00022.bin", + "vision_tower.vision_model.encoder.layers.22.layer_norm1.bias": "pytorch_model-00022-of-00022.bin", + "vision_tower.vision_model.encoder.layers.22.layer_norm1.weight": "pytorch_model-00022-of-00022.bin", + "vision_tower.vision_model.encoder.layers.22.layer_norm2.bias": "pytorch_model-00022-of-00022.bin", + "vision_tower.vision_model.encoder.layers.22.layer_norm2.weight": "pytorch_model-00022-of-00022.bin", + "vision_tower.vision_model.encoder.layers.22.mlp.fc1.bias": "pytorch_model-00022-of-00022.bin", + "vision_tower.vision_model.encoder.layers.22.mlp.fc1.weight": "pytorch_model-00002-of-00022.bin", + "vision_tower.vision_model.encoder.layers.22.mlp.fc2.bias": "pytorch_model-00022-of-00022.bin", + "vision_tower.vision_model.encoder.layers.22.mlp.fc2.weight": "pytorch_model-00002-of-00022.bin", + "vision_tower.vision_model.encoder.layers.22.self_attn.k_proj.bias": "pytorch_model-00022-of-00022.bin", + "vision_tower.vision_model.encoder.layers.22.self_attn.k_proj.weight": "pytorch_model-00002-of-00022.bin", + "vision_tower.vision_model.encoder.layers.22.self_attn.out_proj.bias": "pytorch_model-00022-of-00022.bin", + "vision_tower.vision_model.encoder.layers.22.self_attn.out_proj.weight": "pytorch_model-00002-of-00022.bin", + "vision_tower.vision_model.encoder.layers.22.self_attn.q_proj.bias": "pytorch_model-00022-of-00022.bin", + "vision_tower.vision_model.encoder.layers.22.self_attn.q_proj.weight": "pytorch_model-00002-of-00022.bin", + "vision_tower.vision_model.encoder.layers.22.self_attn.v_proj.bias": "pytorch_model-00022-of-00022.bin", + "vision_tower.vision_model.encoder.layers.22.self_attn.v_proj.weight": "pytorch_model-00002-of-00022.bin", + "vision_tower.vision_model.encoder.layers.23.layer_norm1.bias": "pytorch_model-00022-of-00022.bin", + "vision_tower.vision_model.encoder.layers.23.layer_norm1.weight": "pytorch_model-00022-of-00022.bin", + "vision_tower.vision_model.encoder.layers.23.layer_norm2.bias": "pytorch_model-00022-of-00022.bin", + "vision_tower.vision_model.encoder.layers.23.layer_norm2.weight": "pytorch_model-00022-of-00022.bin", + "vision_tower.vision_model.encoder.layers.23.mlp.fc1.bias": "pytorch_model-00022-of-00022.bin", + "vision_tower.vision_model.encoder.layers.23.mlp.fc1.weight": "pytorch_model-00002-of-00022.bin", + "vision_tower.vision_model.encoder.layers.23.mlp.fc2.bias": "pytorch_model-00022-of-00022.bin", + "vision_tower.vision_model.encoder.layers.23.mlp.fc2.weight": "pytorch_model-00002-of-00022.bin", + "vision_tower.vision_model.encoder.layers.23.self_attn.k_proj.bias": "pytorch_model-00022-of-00022.bin", + "vision_tower.vision_model.encoder.layers.23.self_attn.k_proj.weight": "pytorch_model-00002-of-00022.bin", + "vision_tower.vision_model.encoder.layers.23.self_attn.out_proj.bias": "pytorch_model-00022-of-00022.bin", + "vision_tower.vision_model.encoder.layers.23.self_attn.out_proj.weight": "pytorch_model-00002-of-00022.bin", + "vision_tower.vision_model.encoder.layers.23.self_attn.q_proj.bias": "pytorch_model-00022-of-00022.bin", + "vision_tower.vision_model.encoder.layers.23.self_attn.q_proj.weight": "pytorch_model-00002-of-00022.bin", + "vision_tower.vision_model.encoder.layers.23.self_attn.v_proj.bias": "pytorch_model-00022-of-00022.bin", + "vision_tower.vision_model.encoder.layers.23.self_attn.v_proj.weight": "pytorch_model-00002-of-00022.bin", + "vision_tower.vision_model.encoder.layers.24.layer_norm1.bias": "pytorch_model-00022-of-00022.bin", + "vision_tower.vision_model.encoder.layers.24.layer_norm1.weight": "pytorch_model-00022-of-00022.bin", + "vision_tower.vision_model.encoder.layers.24.layer_norm2.bias": "pytorch_model-00022-of-00022.bin", + "vision_tower.vision_model.encoder.layers.24.layer_norm2.weight": "pytorch_model-00022-of-00022.bin", + "vision_tower.vision_model.encoder.layers.24.mlp.fc1.bias": "pytorch_model-00022-of-00022.bin", + "vision_tower.vision_model.encoder.layers.24.mlp.fc1.weight": "pytorch_model-00002-of-00022.bin", + "vision_tower.vision_model.encoder.layers.24.mlp.fc2.bias": "pytorch_model-00022-of-00022.bin", + "vision_tower.vision_model.encoder.layers.24.mlp.fc2.weight": "pytorch_model-00002-of-00022.bin", + "vision_tower.vision_model.encoder.layers.24.self_attn.k_proj.bias": "pytorch_model-00022-of-00022.bin", + "vision_tower.vision_model.encoder.layers.24.self_attn.k_proj.weight": "pytorch_model-00002-of-00022.bin", + "vision_tower.vision_model.encoder.layers.24.self_attn.out_proj.bias": "pytorch_model-00022-of-00022.bin", + "vision_tower.vision_model.encoder.layers.24.self_attn.out_proj.weight": "pytorch_model-00002-of-00022.bin", + "vision_tower.vision_model.encoder.layers.24.self_attn.q_proj.bias": "pytorch_model-00022-of-00022.bin", + "vision_tower.vision_model.encoder.layers.24.self_attn.q_proj.weight": "pytorch_model-00002-of-00022.bin", + "vision_tower.vision_model.encoder.layers.24.self_attn.v_proj.bias": "pytorch_model-00022-of-00022.bin", + "vision_tower.vision_model.encoder.layers.24.self_attn.v_proj.weight": "pytorch_model-00002-of-00022.bin", + "vision_tower.vision_model.encoder.layers.25.layer_norm1.bias": "pytorch_model-00022-of-00022.bin", + "vision_tower.vision_model.encoder.layers.25.layer_norm1.weight": "pytorch_model-00022-of-00022.bin", + "vision_tower.vision_model.encoder.layers.25.layer_norm2.bias": "pytorch_model-00022-of-00022.bin", + "vision_tower.vision_model.encoder.layers.25.layer_norm2.weight": "pytorch_model-00022-of-00022.bin", + "vision_tower.vision_model.encoder.layers.25.mlp.fc1.bias": "pytorch_model-00022-of-00022.bin", + "vision_tower.vision_model.encoder.layers.25.mlp.fc1.weight": "pytorch_model-00002-of-00022.bin", + "vision_tower.vision_model.encoder.layers.25.mlp.fc2.bias": "pytorch_model-00022-of-00022.bin", + "vision_tower.vision_model.encoder.layers.25.mlp.fc2.weight": "pytorch_model-00002-of-00022.bin", + "vision_tower.vision_model.encoder.layers.25.self_attn.k_proj.bias": "pytorch_model-00022-of-00022.bin", + "vision_tower.vision_model.encoder.layers.25.self_attn.k_proj.weight": "pytorch_model-00002-of-00022.bin", + "vision_tower.vision_model.encoder.layers.25.self_attn.out_proj.bias": "pytorch_model-00022-of-00022.bin", + "vision_tower.vision_model.encoder.layers.25.self_attn.out_proj.weight": "pytorch_model-00002-of-00022.bin", + "vision_tower.vision_model.encoder.layers.25.self_attn.q_proj.bias": "pytorch_model-00022-of-00022.bin", + "vision_tower.vision_model.encoder.layers.25.self_attn.q_proj.weight": "pytorch_model-00002-of-00022.bin", + "vision_tower.vision_model.encoder.layers.25.self_attn.v_proj.bias": "pytorch_model-00022-of-00022.bin", + "vision_tower.vision_model.encoder.layers.25.self_attn.v_proj.weight": "pytorch_model-00002-of-00022.bin", + "vision_tower.vision_model.encoder.layers.26.layer_norm1.bias": "pytorch_model-00022-of-00022.bin", + "vision_tower.vision_model.encoder.layers.26.layer_norm1.weight": "pytorch_model-00022-of-00022.bin", + "vision_tower.vision_model.encoder.layers.26.layer_norm2.bias": "pytorch_model-00022-of-00022.bin", + "vision_tower.vision_model.encoder.layers.26.layer_norm2.weight": "pytorch_model-00022-of-00022.bin", + "vision_tower.vision_model.encoder.layers.26.mlp.fc1.bias": "pytorch_model-00022-of-00022.bin", + "vision_tower.vision_model.encoder.layers.26.mlp.fc1.weight": "pytorch_model-00002-of-00022.bin", + "vision_tower.vision_model.encoder.layers.26.mlp.fc2.bias": "pytorch_model-00022-of-00022.bin", + "vision_tower.vision_model.encoder.layers.26.mlp.fc2.weight": "pytorch_model-00002-of-00022.bin", + "vision_tower.vision_model.encoder.layers.26.self_attn.k_proj.bias": "pytorch_model-00022-of-00022.bin", + "vision_tower.vision_model.encoder.layers.26.self_attn.k_proj.weight": "pytorch_model-00002-of-00022.bin", + "vision_tower.vision_model.encoder.layers.26.self_attn.out_proj.bias": "pytorch_model-00022-of-00022.bin", + "vision_tower.vision_model.encoder.layers.26.self_attn.out_proj.weight": "pytorch_model-00002-of-00022.bin", + "vision_tower.vision_model.encoder.layers.26.self_attn.q_proj.bias": "pytorch_model-00022-of-00022.bin", + "vision_tower.vision_model.encoder.layers.26.self_attn.q_proj.weight": "pytorch_model-00002-of-00022.bin", + "vision_tower.vision_model.encoder.layers.26.self_attn.v_proj.bias": "pytorch_model-00022-of-00022.bin", + "vision_tower.vision_model.encoder.layers.26.self_attn.v_proj.weight": "pytorch_model-00002-of-00022.bin", + "vision_tower.vision_model.encoder.layers.3.layer_norm1.bias": "pytorch_model-00022-of-00022.bin", + "vision_tower.vision_model.encoder.layers.3.layer_norm1.weight": "pytorch_model-00022-of-00022.bin", + "vision_tower.vision_model.encoder.layers.3.layer_norm2.bias": "pytorch_model-00022-of-00022.bin", + "vision_tower.vision_model.encoder.layers.3.layer_norm2.weight": "pytorch_model-00022-of-00022.bin", + "vision_tower.vision_model.encoder.layers.3.mlp.fc1.bias": "pytorch_model-00022-of-00022.bin", + "vision_tower.vision_model.encoder.layers.3.mlp.fc1.weight": "pytorch_model-00002-of-00022.bin", + "vision_tower.vision_model.encoder.layers.3.mlp.fc2.bias": "pytorch_model-00022-of-00022.bin", + "vision_tower.vision_model.encoder.layers.3.mlp.fc2.weight": "pytorch_model-00002-of-00022.bin", + "vision_tower.vision_model.encoder.layers.3.self_attn.k_proj.bias": "pytorch_model-00022-of-00022.bin", + "vision_tower.vision_model.encoder.layers.3.self_attn.k_proj.weight": "pytorch_model-00002-of-00022.bin", + "vision_tower.vision_model.encoder.layers.3.self_attn.out_proj.bias": "pytorch_model-00022-of-00022.bin", + "vision_tower.vision_model.encoder.layers.3.self_attn.out_proj.weight": "pytorch_model-00002-of-00022.bin", + "vision_tower.vision_model.encoder.layers.3.self_attn.q_proj.bias": "pytorch_model-00022-of-00022.bin", + "vision_tower.vision_model.encoder.layers.3.self_attn.q_proj.weight": "pytorch_model-00002-of-00022.bin", + "vision_tower.vision_model.encoder.layers.3.self_attn.v_proj.bias": "pytorch_model-00022-of-00022.bin", + "vision_tower.vision_model.encoder.layers.3.self_attn.v_proj.weight": "pytorch_model-00002-of-00022.bin", + "vision_tower.vision_model.encoder.layers.4.layer_norm1.bias": "pytorch_model-00022-of-00022.bin", + "vision_tower.vision_model.encoder.layers.4.layer_norm1.weight": "pytorch_model-00022-of-00022.bin", + "vision_tower.vision_model.encoder.layers.4.layer_norm2.bias": "pytorch_model-00022-of-00022.bin", + "vision_tower.vision_model.encoder.layers.4.layer_norm2.weight": "pytorch_model-00022-of-00022.bin", + "vision_tower.vision_model.encoder.layers.4.mlp.fc1.bias": "pytorch_model-00022-of-00022.bin", + "vision_tower.vision_model.encoder.layers.4.mlp.fc1.weight": "pytorch_model-00002-of-00022.bin", + "vision_tower.vision_model.encoder.layers.4.mlp.fc2.bias": "pytorch_model-00022-of-00022.bin", + "vision_tower.vision_model.encoder.layers.4.mlp.fc2.weight": "pytorch_model-00002-of-00022.bin", + "vision_tower.vision_model.encoder.layers.4.self_attn.k_proj.bias": "pytorch_model-00022-of-00022.bin", + "vision_tower.vision_model.encoder.layers.4.self_attn.k_proj.weight": "pytorch_model-00002-of-00022.bin", + "vision_tower.vision_model.encoder.layers.4.self_attn.out_proj.bias": "pytorch_model-00022-of-00022.bin", + "vision_tower.vision_model.encoder.layers.4.self_attn.out_proj.weight": "pytorch_model-00002-of-00022.bin", + "vision_tower.vision_model.encoder.layers.4.self_attn.q_proj.bias": "pytorch_model-00022-of-00022.bin", + "vision_tower.vision_model.encoder.layers.4.self_attn.q_proj.weight": "pytorch_model-00002-of-00022.bin", + "vision_tower.vision_model.encoder.layers.4.self_attn.v_proj.bias": "pytorch_model-00022-of-00022.bin", + "vision_tower.vision_model.encoder.layers.4.self_attn.v_proj.weight": "pytorch_model-00002-of-00022.bin", + "vision_tower.vision_model.encoder.layers.5.layer_norm1.bias": "pytorch_model-00022-of-00022.bin", + "vision_tower.vision_model.encoder.layers.5.layer_norm1.weight": "pytorch_model-00022-of-00022.bin", + "vision_tower.vision_model.encoder.layers.5.layer_norm2.bias": "pytorch_model-00022-of-00022.bin", + "vision_tower.vision_model.encoder.layers.5.layer_norm2.weight": "pytorch_model-00022-of-00022.bin", + "vision_tower.vision_model.encoder.layers.5.mlp.fc1.bias": "pytorch_model-00022-of-00022.bin", + "vision_tower.vision_model.encoder.layers.5.mlp.fc1.weight": "pytorch_model-00002-of-00022.bin", + "vision_tower.vision_model.encoder.layers.5.mlp.fc2.bias": "pytorch_model-00022-of-00022.bin", + "vision_tower.vision_model.encoder.layers.5.mlp.fc2.weight": "pytorch_model-00002-of-00022.bin", + "vision_tower.vision_model.encoder.layers.5.self_attn.k_proj.bias": "pytorch_model-00022-of-00022.bin", + "vision_tower.vision_model.encoder.layers.5.self_attn.k_proj.weight": "pytorch_model-00002-of-00022.bin", + "vision_tower.vision_model.encoder.layers.5.self_attn.out_proj.bias": "pytorch_model-00022-of-00022.bin", + "vision_tower.vision_model.encoder.layers.5.self_attn.out_proj.weight": "pytorch_model-00002-of-00022.bin", + "vision_tower.vision_model.encoder.layers.5.self_attn.q_proj.bias": "pytorch_model-00022-of-00022.bin", + "vision_tower.vision_model.encoder.layers.5.self_attn.q_proj.weight": "pytorch_model-00002-of-00022.bin", + "vision_tower.vision_model.encoder.layers.5.self_attn.v_proj.bias": "pytorch_model-00022-of-00022.bin", + "vision_tower.vision_model.encoder.layers.5.self_attn.v_proj.weight": "pytorch_model-00002-of-00022.bin", + "vision_tower.vision_model.encoder.layers.6.layer_norm1.bias": "pytorch_model-00022-of-00022.bin", + "vision_tower.vision_model.encoder.layers.6.layer_norm1.weight": "pytorch_model-00022-of-00022.bin", + "vision_tower.vision_model.encoder.layers.6.layer_norm2.bias": "pytorch_model-00022-of-00022.bin", + "vision_tower.vision_model.encoder.layers.6.layer_norm2.weight": "pytorch_model-00022-of-00022.bin", + "vision_tower.vision_model.encoder.layers.6.mlp.fc1.bias": "pytorch_model-00022-of-00022.bin", + "vision_tower.vision_model.encoder.layers.6.mlp.fc1.weight": "pytorch_model-00002-of-00022.bin", + "vision_tower.vision_model.encoder.layers.6.mlp.fc2.bias": "pytorch_model-00022-of-00022.bin", + "vision_tower.vision_model.encoder.layers.6.mlp.fc2.weight": "pytorch_model-00002-of-00022.bin", + "vision_tower.vision_model.encoder.layers.6.self_attn.k_proj.bias": "pytorch_model-00022-of-00022.bin", + "vision_tower.vision_model.encoder.layers.6.self_attn.k_proj.weight": "pytorch_model-00002-of-00022.bin", + "vision_tower.vision_model.encoder.layers.6.self_attn.out_proj.bias": "pytorch_model-00022-of-00022.bin", + "vision_tower.vision_model.encoder.layers.6.self_attn.out_proj.weight": "pytorch_model-00002-of-00022.bin", + "vision_tower.vision_model.encoder.layers.6.self_attn.q_proj.bias": "pytorch_model-00022-of-00022.bin", + "vision_tower.vision_model.encoder.layers.6.self_attn.q_proj.weight": "pytorch_model-00002-of-00022.bin", + "vision_tower.vision_model.encoder.layers.6.self_attn.v_proj.bias": "pytorch_model-00022-of-00022.bin", + "vision_tower.vision_model.encoder.layers.6.self_attn.v_proj.weight": "pytorch_model-00002-of-00022.bin", + "vision_tower.vision_model.encoder.layers.7.layer_norm1.bias": "pytorch_model-00022-of-00022.bin", + "vision_tower.vision_model.encoder.layers.7.layer_norm1.weight": "pytorch_model-00022-of-00022.bin", + "vision_tower.vision_model.encoder.layers.7.layer_norm2.bias": "pytorch_model-00022-of-00022.bin", + "vision_tower.vision_model.encoder.layers.7.layer_norm2.weight": "pytorch_model-00022-of-00022.bin", + "vision_tower.vision_model.encoder.layers.7.mlp.fc1.bias": "pytorch_model-00022-of-00022.bin", + "vision_tower.vision_model.encoder.layers.7.mlp.fc1.weight": "pytorch_model-00002-of-00022.bin", + "vision_tower.vision_model.encoder.layers.7.mlp.fc2.bias": "pytorch_model-00022-of-00022.bin", + "vision_tower.vision_model.encoder.layers.7.mlp.fc2.weight": "pytorch_model-00002-of-00022.bin", + "vision_tower.vision_model.encoder.layers.7.self_attn.k_proj.bias": "pytorch_model-00022-of-00022.bin", + "vision_tower.vision_model.encoder.layers.7.self_attn.k_proj.weight": "pytorch_model-00002-of-00022.bin", + "vision_tower.vision_model.encoder.layers.7.self_attn.out_proj.bias": "pytorch_model-00022-of-00022.bin", + "vision_tower.vision_model.encoder.layers.7.self_attn.out_proj.weight": "pytorch_model-00002-of-00022.bin", + "vision_tower.vision_model.encoder.layers.7.self_attn.q_proj.bias": "pytorch_model-00022-of-00022.bin", + "vision_tower.vision_model.encoder.layers.7.self_attn.q_proj.weight": "pytorch_model-00002-of-00022.bin", + "vision_tower.vision_model.encoder.layers.7.self_attn.v_proj.bias": "pytorch_model-00022-of-00022.bin", + "vision_tower.vision_model.encoder.layers.7.self_attn.v_proj.weight": "pytorch_model-00002-of-00022.bin", + "vision_tower.vision_model.encoder.layers.8.layer_norm1.bias": "pytorch_model-00022-of-00022.bin", + "vision_tower.vision_model.encoder.layers.8.layer_norm1.weight": "pytorch_model-00022-of-00022.bin", + "vision_tower.vision_model.encoder.layers.8.layer_norm2.bias": "pytorch_model-00022-of-00022.bin", + "vision_tower.vision_model.encoder.layers.8.layer_norm2.weight": "pytorch_model-00022-of-00022.bin", + "vision_tower.vision_model.encoder.layers.8.mlp.fc1.bias": "pytorch_model-00022-of-00022.bin", + "vision_tower.vision_model.encoder.layers.8.mlp.fc1.weight": "pytorch_model-00002-of-00022.bin", + "vision_tower.vision_model.encoder.layers.8.mlp.fc2.bias": "pytorch_model-00022-of-00022.bin", + "vision_tower.vision_model.encoder.layers.8.mlp.fc2.weight": "pytorch_model-00002-of-00022.bin", + "vision_tower.vision_model.encoder.layers.8.self_attn.k_proj.bias": "pytorch_model-00022-of-00022.bin", + "vision_tower.vision_model.encoder.layers.8.self_attn.k_proj.weight": "pytorch_model-00002-of-00022.bin", + "vision_tower.vision_model.encoder.layers.8.self_attn.out_proj.bias": "pytorch_model-00022-of-00022.bin", + "vision_tower.vision_model.encoder.layers.8.self_attn.out_proj.weight": "pytorch_model-00002-of-00022.bin", + "vision_tower.vision_model.encoder.layers.8.self_attn.q_proj.bias": "pytorch_model-00022-of-00022.bin", + "vision_tower.vision_model.encoder.layers.8.self_attn.q_proj.weight": "pytorch_model-00002-of-00022.bin", + "vision_tower.vision_model.encoder.layers.8.self_attn.v_proj.bias": "pytorch_model-00022-of-00022.bin", + "vision_tower.vision_model.encoder.layers.8.self_attn.v_proj.weight": "pytorch_model-00002-of-00022.bin", + "vision_tower.vision_model.encoder.layers.9.layer_norm1.bias": "pytorch_model-00022-of-00022.bin", + "vision_tower.vision_model.encoder.layers.9.layer_norm1.weight": "pytorch_model-00022-of-00022.bin", + "vision_tower.vision_model.encoder.layers.9.layer_norm2.bias": "pytorch_model-00022-of-00022.bin", + "vision_tower.vision_model.encoder.layers.9.layer_norm2.weight": "pytorch_model-00022-of-00022.bin", + "vision_tower.vision_model.encoder.layers.9.mlp.fc1.bias": "pytorch_model-00022-of-00022.bin", + "vision_tower.vision_model.encoder.layers.9.mlp.fc1.weight": "pytorch_model-00002-of-00022.bin", + "vision_tower.vision_model.encoder.layers.9.mlp.fc2.bias": "pytorch_model-00022-of-00022.bin", + "vision_tower.vision_model.encoder.layers.9.mlp.fc2.weight": "pytorch_model-00002-of-00022.bin", + "vision_tower.vision_model.encoder.layers.9.self_attn.k_proj.bias": "pytorch_model-00022-of-00022.bin", + "vision_tower.vision_model.encoder.layers.9.self_attn.k_proj.weight": "pytorch_model-00002-of-00022.bin", + "vision_tower.vision_model.encoder.layers.9.self_attn.out_proj.bias": "pytorch_model-00022-of-00022.bin", + "vision_tower.vision_model.encoder.layers.9.self_attn.out_proj.weight": "pytorch_model-00002-of-00022.bin", + "vision_tower.vision_model.encoder.layers.9.self_attn.q_proj.bias": "pytorch_model-00022-of-00022.bin", + "vision_tower.vision_model.encoder.layers.9.self_attn.q_proj.weight": "pytorch_model-00002-of-00022.bin", + "vision_tower.vision_model.encoder.layers.9.self_attn.v_proj.bias": "pytorch_model-00022-of-00022.bin", + "vision_tower.vision_model.encoder.layers.9.self_attn.v_proj.weight": "pytorch_model-00002-of-00022.bin", + "vision_tower.vision_model.post_layernorm.bias": "pytorch_model-00022-of-00022.bin", + "vision_tower.vision_model.post_layernorm.weight": "pytorch_model-00022-of-00022.bin" + } +} diff --git a/models/fdkjsfds/gemma_knows_coorsinates_4/special_tokens_map.json b/models/fdkjsfds/gemma_knows_coorsinates_4/special_tokens_map.json new file mode 100644 index 0000000000000000000000000000000000000000..1a6193244714d3d78be48666cb02cdbfac62ad86 --- /dev/null +++ b/models/fdkjsfds/gemma_knows_coorsinates_4/special_tokens_map.json @@ -0,0 +1,33 @@ +{ + "boi_token": "", + "bos_token": { + "content": "", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false + }, + "eoi_token": "", + "eos_token": { + "content": "", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false + }, + "image_token": "", + "pad_token": { + "content": "", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false + }, + "unk_token": { + "content": "", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false + } +} diff --git a/models/fdkjsfds/gemma_knows_coorsinates_4/tokenizer.json b/models/fdkjsfds/gemma_knows_coorsinates_4/tokenizer.json new file mode 100644 index 0000000000000000000000000000000000000000..fa5532fb50d8a5d924a8feb805b2379dd21ff8da --- /dev/null +++ b/models/fdkjsfds/gemma_knows_coorsinates_4/tokenizer.json @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:d786405177734910d7a3db625c2826640964a0b4e5cdbbd70620ae3313a01bef +size 33384722 diff --git a/models/fdkjsfds/gemma_knows_coorsinates_4/tokenizer_config.json b/models/fdkjsfds/gemma_knows_coorsinates_4/tokenizer_config.json new file mode 100644 index 0000000000000000000000000000000000000000..111e0386451cec6a6732d5641af04a789094451b --- /dev/null +++ b/models/fdkjsfds/gemma_knows_coorsinates_4/tokenizer_config.json @@ -0,0 +1,51347 @@ +{ + "add_bos_token": true, + "add_eos_token": false, + "added_tokens_decoder": { + "0": { + "content": "", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "1": { + "content": "", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "2": { + "content": "", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "3": { + "content": "", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "4": { + "content": "", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": false + }, + "5": { + "content": "[multimodal]", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": false + }, + "6": { + "content": "", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": false + }, + "7": { + "content": "", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": false + }, + "8": { + "content": "", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": false + }, + "9": { + "content": "", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": false + }, + "10": { + "content": "", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": false + }, + "11": { + "content": "", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": false + }, + "12": { + "content": "", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": false + }, + "13": { + "content": "", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": false + }, + "14": { + "content": "", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": false + }, + "15": { + "content": "", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": false + }, + "16": { + "content": "", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": false + }, + "17": { + "content": "", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": false + }, + "18": { + "content": "", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": false + }, + "19": { + "content": "", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": false + }, + "20": { + "content": "", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": false + }, + "21": { + "content": "", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": false + }, + "22": { + "content": "", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": false + }, + "23": { + "content": "", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": false + }, + "24": { + "content": "", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": false + }, + "25": { + "content": "", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": false + }, + "26": { + "content": "", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": false + }, + "27": { + "content": "", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": false + }, + "28": { + "content": "", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": false + }, + "29": { + "content": "", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": false + }, + "30": { + "content": "", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": false + }, + "31": { + "content": "", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": false + }, + "32": { + "content": "", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": false + }, + "33": { + "content": "", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": false + }, + "34": { + "content": "", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": false + }, + "35": { + "content": "", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": false + }, + "36": { + "content": "", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": false + }, + "37": { + "content": "", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": false + }, + "38": { + "content": "", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": false + }, + "39": { + "content": "", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": false + }, + "40": { + "content": "", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": false + }, + "41": { + "content": "", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": false + }, + "42": { + "content": "", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": false + }, + "43": { + "content": "", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": false + }, + "44": { + "content": "", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": false + }, + "45": { + "content": "", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": false + }, + "46": { + "content": "", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": false + }, + "47": { + "content": "", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": false + }, + "48": { + "content": "", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": false + }, + "49": { + "content": "", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": false + }, + "50": { + "content": "", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": false + }, + "51": { + "content": "", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": false + }, + "52": { + "content": "", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": false + }, + "53": { + "content": "", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": false + }, + "54": { + "content": "", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": false + }, + "55": { + "content": "", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": false + }, + "56": { + "content": "", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": false + }, + "57": { + "content": "", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": false + }, + "58": { + "content": "", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": false + }, + "59": { + "content": "", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": false + }, + "60": { + "content": "", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": false + }, + "61": { + "content": "", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": false + }, + "62": { + "content": "", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": false + }, + "63": { + "content": "", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": false + }, + "64": { + "content": "", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": false + }, + "65": { + "content": "", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": false + }, + "66": { + "content": "", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": false + }, + "67": { + "content": "", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": false + }, + "68": { + "content": "", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": false + }, + "69": { + "content": "", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": false + }, + "70": { + "content": "", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": false + }, + "71": { + "content": "", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": false + }, + "72": { + "content": "", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": false + }, + "73": { + "content": "", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": false + }, + "74": { + "content": "", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": false + }, + "75": { + "content": "", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": false + }, + "76": { + "content": "", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": false + }, + "77": { + "content": "", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": false + }, + "78": { + "content": "", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": false + }, + "79": { + "content": "", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": false + }, + "80": { + "content": "", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": false + }, + "81": { + "content": "", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": false + }, + "82": { + "content": "", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": false + }, + "83": { + "content": "", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": false + }, + "84": { + "content": "", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": false + }, + "85": { + "content": "", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": false + }, + "86": { + "content": "", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": false + }, + "87": { + "content": "", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": false + }, + "88": { + "content": "", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": false + }, + "89": { + "content": "", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": false + }, + "90": { + "content": "", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": false + }, + "91": { + "content": "", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": false + }, + "92": { + "content": "", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": false + }, + "93": { + "content": "", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": false + }, + "94": { + "content": "", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": false + }, + "95": { + "content": "", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": false + }, + "96": { + "content": "", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": false + }, + "97": { + "content": "", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": false + }, + "98": { + "content": "", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": false + }, + "99": { + "content": "", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": false + }, + "100": { + "content": "", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": false + }, + "101": { + "content": "", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": false + }, + "102": { + "content": "", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": false + }, + "103": { + "content": "", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": false + }, + "104": { + "content": "", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": false + }, + "105": { + "content": "", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "106": { + "content": "", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "107": { + "content": "\n", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": false + }, + "108": { + "content": "\n\n", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": false + }, + "109": { + "content": "\n\n\n", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": false + }, + "110": { + "content": "\n\n\n\n", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": false + }, + "111": { + "content": "\n\n\n\n\n", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": false + }, + "112": { + "content": "\n\n\n\n\n\n", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": false + }, + "113": { + "content": "\n\n\n\n\n\n\n", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": false + }, + "114": { + "content": "\n\n\n\n\n\n\n\n", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": false + }, + "115": { + "content": "\n\n\n\n\n\n\n\n\n", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": false + }, + "116": { + "content": "\n\n\n\n\n\n\n\n\n\n", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": false + }, + "117": { + "content": "\n\n\n\n\n\n\n\n\n\n\n", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": false + }, + "118": { + "content": "\n\n\n\n\n\n\n\n\n\n\n\n", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": false + }, + "119": { + "content": "\n\n\n\n\n\n\n\n\n\n\n\n\n", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": false + }, + "120": { + "content": "\n\n\n\n\n\n\n\n\n\n\n\n\n\n", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": false + }, + "121": { + "content": "\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": false + }, + "122": { + "content": "\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": false + }, + "123": { + "content": "\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": false + }, + "124": { + "content": "\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": false + }, + "125": { + "content": "\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": false + }, + "126": { + "content": "\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": false + }, + "127": { + "content": "\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": false + }, + "128": { + "content": "\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": false + }, + "129": { + "content": "\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": false + }, + "130": { + "content": "\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": false + }, + "131": { + "content": "\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": false + }, + "132": { + "content": "\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": false + }, + "133": { + "content": "\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": false + }, + "134": { + "content": "\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": false + }, + "135": { + "content": "\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": false + }, + "136": { + "content": "\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": false + }, + "137": { + "content": "\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": false + }, + "138": { + "content": "▁▁", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": false + }, + "139": { + "content": "▁▁▁", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": false + }, + "140": { + "content": "▁▁▁▁", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": false + }, + "141": { + "content": "▁▁▁▁▁", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": false + }, + "142": { + "content": "▁▁▁▁▁▁", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": false + }, + "143": { + "content": "▁▁▁▁▁▁▁", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": false + }, + "144": { + "content": "▁▁▁▁▁▁▁▁", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": false + }, + "145": { + "content": "▁▁▁▁▁▁▁▁▁", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": false + }, + "146": { + "content": "▁▁▁▁▁▁▁▁▁▁", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": false + }, + "147": { + "content": "▁▁▁▁▁▁▁▁▁▁▁", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": false + }, + "148": { + "content": "▁▁▁▁▁▁▁▁▁▁▁▁", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": false + }, + "149": { + "content": "▁▁▁▁▁▁▁▁▁▁▁▁▁", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": false + }, + "150": { + "content": "▁▁▁▁▁▁▁▁▁▁▁▁▁▁", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": false + }, + "151": { + "content": "▁▁▁▁▁▁▁▁▁▁▁▁▁▁▁", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": false + }, + "152": { + "content": "▁▁▁▁▁▁▁▁▁▁▁▁▁▁▁▁", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": false + }, + "153": { + "content": "▁▁▁▁▁▁▁▁▁▁▁▁▁▁▁▁▁", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": false + }, + "154": { + "content": "▁▁▁▁▁▁▁▁▁▁▁▁▁▁▁▁▁▁", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": false + }, + "155": { + "content": "▁▁▁▁▁▁▁▁▁▁▁▁▁▁▁▁▁▁▁", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": false + }, + "156": { + "content": "▁▁▁▁▁▁▁▁▁▁▁▁▁▁▁▁▁▁▁▁", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": false + }, + "157": { + "content": "▁▁▁▁▁▁▁▁▁▁▁▁▁▁▁▁▁▁▁▁▁", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": false + }, + "158": { + "content": "▁▁▁▁▁▁▁▁▁▁▁▁▁▁▁▁▁▁▁▁▁▁", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": false + }, + "159": { + "content": "▁▁▁▁▁▁▁▁▁▁▁▁▁▁▁▁▁▁▁▁▁▁▁", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": false + }, + "160": { + "content": "▁▁▁▁▁▁▁▁▁▁▁▁▁▁▁▁▁▁▁▁▁▁▁▁", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": false + }, + "161": { + "content": "▁▁▁▁▁▁▁▁▁▁▁▁▁▁▁▁▁▁▁▁▁▁▁▁▁", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": false + }, + "162": { + "content": "▁▁▁▁▁▁▁▁▁▁▁▁▁▁▁▁▁▁▁▁▁▁▁▁▁▁", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": false + }, + "163": { + "content": "▁▁▁▁▁▁▁▁▁▁▁▁▁▁▁▁▁▁▁▁▁▁▁▁▁▁▁", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": false + }, + "164": { + "content": "▁▁▁▁▁▁▁▁▁▁▁▁▁▁▁▁▁▁▁▁▁▁▁▁▁▁▁▁", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": false + }, + "165": { + "content": "▁▁▁▁▁▁▁▁▁▁▁▁▁▁▁▁▁▁▁▁▁▁▁▁▁▁▁▁▁", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": false + }, + "166": { + "content": "▁▁▁▁▁▁▁▁▁▁▁▁▁▁▁▁▁▁▁▁▁▁▁▁▁▁▁▁▁▁", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": false + }, + "167": { + "content": "▁▁▁▁▁▁▁▁▁▁▁▁▁▁▁▁▁▁▁▁▁▁▁▁▁▁▁▁▁▁▁", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": false + }, + "168": { + "content": "", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": false + }, + "169": { + "content": "", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": false + }, + "171": { + "content": "", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": false + }, + "172": { + "content": "", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": false + }, + "173": { + "content": "", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": false + }, + "174": { + "content": "
", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": false + }, + "170": { + "content": "
", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": false + }, + "175": { + "content": "", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": false + }, + "176": { + "content": "
", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": false + }, + "177": { + "content": "", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": false + }, + "178": { + "content": "", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": false + }, + "179": { + "content": "", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": false + }, + "180": { + "content": "", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": false + }, + "181": { + "content": "", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": false + }, + "182": { + "content": "", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": false + }, + "183": { + "content": "", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": false + }, + "184": { + "content": "

", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": false + }, + "185": { + "content": "

", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": false + }, + "186": { + "content": "

", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": false + }, + "187": { + "content": "

", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": false + }, + "188": { + "content": "

", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": false + }, + "189": { + "content": "
", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": false + }, + "190": { + "content": "
", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": false + }, + "191": { + "content": "
", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": false + }, + "192": { + "content": "", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": false + }, + "193": { + "content": "", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": false + }, + "194": { + "content": "", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": false + }, + "195": { + "content": "", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": false + }, + "196": { + "content": "", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": false + }, + "197": { + "content": "", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": false + }, + "198": { + "content": "", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": false + }, + "199": { + "content": "", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": false + }, + "200": { + "content": "", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": false + }, + "201": { + "content": "", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": false + }, + "202": { + "content": "", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": false + }, + "203": { + "content": "", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": false + }, + "204": { + "content": "", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": false + }, + "205": { + "content": "", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": false + }, + "206": { + "content": "", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": false + }, + "207": { + "content": "", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": false + }, + "208": { + "content": "", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": false + }, + "209": { + "content": "", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": false + }, + "210": { + "content": "", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": false + }, + "211": { + "content": "", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": false + }, + "212": { + "content": "", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": false + }, + "213": { + "content": "", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": false + }, + "214": { + "content": "", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": false + }, + "215": { + "content": "", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": false + }, + "216": { + "content": "", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": false + }, + "217": { + "content": "", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": false + }, + "218": { + "content": "", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": false + }, + "219": { + "content": "", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": false + }, + "220": { + "content": "", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": false + }, + "221": { + "content": "", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": false + }, + "222": { + "content": "