diff --git a/.gitattributes b/.gitattributes index a6344aac8c09253b3b630fb776ae94478aa0275b..52373fe24473b1aa44333d318f578ae6bf04b49b 100644 --- a/.gitattributes +++ b/.gitattributes @@ -33,3 +33,4 @@ saved_model/**/* filter=lfs diff=lfs merge=lfs -text *.zip filter=lfs diff=lfs merge=lfs -text *.zst filter=lfs diff=lfs merge=lfs -text *tfevents* filter=lfs diff=lfs merge=lfs -text +tokenizer.json filter=lfs diff=lfs merge=lfs -text diff --git a/README.md b/README.md new file mode 100644 index 0000000000000000000000000000000000000000..d61c3f18cf76fab0ce26db6ab32f4dea1481c743 --- /dev/null +++ b/README.md @@ -0,0 +1,70 @@ +--- +language: +- en +license: mit +pipeline_tag: text-generation +tags: +- mlx +- mixture-of-experts +- moe +- pruning +- reap +- minimax +- 8bit +- quantized +- apple-silicon +library_name: mlx +base_model: Akicou/MiniMax-M2-5-REAP-19 +--- + +

+ + vMLX + +

+ +# MiniMax-M2.5 REAP-19 — MLX 8-bit + +MLX 8-bit quantized version of [Akicou/MiniMax-M2-5-REAP-19](https://huggingface.co/Akicou/MiniMax-M2-5-REAP-19) for efficient local inference on Apple Silicon. + +- **Quantization**: 8-bit (9.0 bits per weight, group size 64, affine mode) +- **Architecture**: MiniMax M2.5 MoE — 62 layers, 205 experts (REAP-pruned from 256), 8 active per token +- **Context**: 196K tokens +- **Size**: ~193 GB +- **Pruning**: 19% of experts removed via [REAP](https://github.com/CerebrasResearch/reap) (Router Expert Activation Pruning) + +## Usage + +```python +from mlx_lm import load, generate + +model, tokenizer = load("shieldstackllc/MiniMax-M2-5-REAP-19-mlx-8bit") +response = generate(model, tokenizer, prompt="Hello!", verbose=True) +``` + +Or with [vMLX](https://vmlx.net) for native macOS inference. + +## About + +MiniMax-M2.5 is a large Mixture-of-Experts language model by MiniMax AI. This variant was pruned to 19% fewer experts by [Akicou](https://huggingface.co/Akicou) using REAP (Router Expert Activation Pruning), reducing model size and memory footprint while maintaining strong performance. MLX quantization by [vMLX](https://vmlx.net). + +## Also Available + +- [MiniMax-M2.5-REAP-19 MLX 4-bit](https://huggingface.co/shieldstackllc/MiniMax-M2-5-REAP-19-mlx-4bit) (~97 GB) *(coming soon)* +- [MiniMax-M2.5-REAP-39 MLX 8-bit](https://huggingface.co/shieldstackllc/MiniMax-M2-5-REAP-39-mlx-8bit) (~138 GB) +- [MiniMax-M2.5-REAP-39 MLX 4-bit](https://huggingface.co/shieldstackllc/MiniMax-M2-5-REAP-39-mlx-4bit) (~73 GB) +- [MiniMax-M2.5-REAP-29 MLX 4-bit](https://huggingface.co/shieldstackllc/MiniMax-M2-5-REAP-29-mlx-4bit) + +## Made for vMLX + +This model was converted and optimized for [vMLX](https://vmlx.net) — a free, open source macOS native MLX inference engine for Apple Silicon. Download vMLX to run this model locally with zero configuration. + +## Credits + +- **Base model**: [MiniMaxAI/MiniMax-M2.5](https://huggingface.co/MiniMaxAI/MiniMax-M2.5) by MiniMax AI +- **REAP pruning**: [Akicou/MiniMax-M2-5-REAP-19](https://huggingface.co/Akicou/MiniMax-M2-5-REAP-19) by Akicou +- **MLX conversion**: [vMLX](https://vmlx.net) — Run AI locally on Mac. No compromises. + +## Contact + +For questions, issues, or collaboration: **admin@vmlx.net** diff --git a/chat_template.jinja b/chat_template.jinja new file mode 100644 index 0000000000000000000000000000000000000000..4623080ad8c269fc12408c9586462eeddd3f7dd3 --- /dev/null +++ b/chat_template.jinja @@ -0,0 +1,159 @@ +{# ----------‑‑‑ special token variables ‑‑‑---------- #} +{%- set toolcall_begin_token = '' -%} +{%- set toolcall_end_token = '' -%} +{#- Tool Rendering Functions ============================================== -#} +{%- macro render_tool_namespace(namespace_name, tool_list) -%} +{%- for tool in tool_list -%} +{{ tool.function | tojson(ensure_ascii=False) }} +{% endfor -%} +{%- endmacro -%} +{%- macro visible_text(content) -%} + {%- if content is string -%} + {{ content }} + {%- elif content is iterable and content is not mapping -%} + {%- for item in content -%} + {%- if item is mapping and item.type == 'text' -%} + {{- item.text }} + {%- elif item is string -%} + {{- item }} + {%- endif -%} + {%- endfor -%} + {%- else -%} + {{- content }} + {%- endif -%} +{%- endmacro -%} +{#- System Message Construction ============================================ -#} +{%- macro build_system_message(system_message) -%} + {%- if system_message and system_message.content -%} + {{- visible_text(system_message.content) }} + {%- else -%} + {%- if model_identity is not defined -%} + {%- set model_identity = "You are a helpful assistant. Your name is MiniMax-M2.5 and is built by MiniMax." -%} + {%- endif -%} + {{- model_identity }} + {%- endif -%} + + {#- Handle current_date -#} + {%- if system_message and system_message.current_date -%} + {{- '\n' ~ 'Current date: ' + system_message.current_date }} + {%- endif -%} + {#- Handle current_location -#} + {%- if system_message and system_message.current_location -%} + {{- '\n' ~ 'Current location: ' + system_message.current_location }} + {%- endif -%} +{%- endmacro -%} +{#- Main Template Logic ================================================= -#} +{#- Extract system message (only first message if it's system) -#} +{%- set system_message = none -%} +{%- set conversation_messages = messages -%} +{%- if messages and messages[0].role == "system" -%} + {%- set system_message = messages[0] -%} + {%- set conversation_messages = messages[1:] -%} +{%- endif -%} +{#- Get the last user message turn, for interleved thinking -#} +{%- set ns = namespace(last_user_index=-1) %} +{% for m in conversation_messages %} + {%- if m.role == 'user' %} + {% set ns.last_user_index = loop.index0 -%} + {%- endif %} +{%- endfor %} +{#- Render system message -#} +{{- ']~!b[' ~ ']~b]system' ~ '\n' }} +{{- build_system_message(system_message) }} +{#- Render tools if available -#} +{%- if tools -%} + {{- '\n\n' ~ '# Tools' ~ '\n' ~ 'You may call one or more tools to assist with the user query.\nHere are the tools available in JSONSchema format:' ~ '\n' }} + {{- '\n' ~ '' ~ '\n' }} + {{- render_tool_namespace("functions", tools) }} + {{- '' ~ '\n\n' }} +{{- 'When making tool calls, use XML format to invoke tools and pass parameters:' ~ '\n' }} +{{- '\n' ~ toolcall_begin_token }} + +param-value-1 +param-value-2 +... + +{{- '\n' ~ toolcall_end_token }} +{%- endif -%} +{{- '[e~[\n' }} + +{#- Render messages -#} +{%- set last_tool_call = namespace(name=none) -%} +{%- for message in conversation_messages -%} + {%- if message.role == 'assistant' -%} + {#- Only render reasoning_content if no user message follows -#} + {{- ']~b]ai' ~ '\n' }} + + {%- set reasoning_content = '' %} + {%- set content = visible_text(message.content) %} + {%- if message.reasoning_content is string %} + {%- set reasoning_content = message.reasoning_content %} + {%- else %} + {%- if '' in content %} + {%- set reasoning_content = content.split('')[0].strip('\n').split('')[-1].strip('\n') %} + {%- set content = content.split('')[-1].strip('\n') %} + {%- endif %} + {%- endif %} + {%- if reasoning_content and loop.index0 > ns.last_user_index -%} + {{- '' ~ '\n' ~ reasoning_content ~ '\n' ~ '' ~ '\n\n' }} + {%- endif -%} + {%- if content -%} + {{- content }} + {%- endif -%} + {%- if message.tool_calls -%} + {{- '\n' ~ toolcall_begin_token ~ '\n' }} + + {%- for tool_call in message.tool_calls -%} + {%- if tool_call.function %} + {%- set tool_call = tool_call.function %} + {%- endif %} + {{- '' }} + {% set _args = tool_call.arguments %} + {%- for k, v in _args.items() %} + {{- '' }} + {{- v | tojson(ensure_ascii=False) if v is not string else v }} + {{- '' }} + {% endfor %} + {{- '' ~ '\n' }} + {%- endfor -%} + + {{- toolcall_end_token}} + {%- set last_tool_call.name = message.tool_calls[-1].name -%} + {%- else -%} + {%- set last_tool_call.name = none -%} + {%- endif -%} + {{- '[e~[' ~ '\n' }} + + {%- elif message.role == 'tool' -%} + {%- if last_tool_call.name is none -%} + {{- raise_exception("Message has tool role, but there was no previous assistant message with a tool call!") }} + {%- endif -%} + {%- if loop.first or (conversation_messages[loop.index0 - 1].role != 'tool') -%} + {{- ']~b]tool' }} + {%- endif -%} + {%- if message.content is string -%} + {{- '\n' }} + {{- message.content }} + {{- '' }} + {%- else -%} + {%- for tr in message.content -%} + {{- '\n' }} + {{- tr.output if tr.output is defined else (tr.text if tr.type == 'text' and tr.text is defined else tr) }} + {{- '\n' }} + {%- endfor -%} + {%- endif -%} + {%- if loop.last or (conversation_messages[loop.index0 + 1].role != 'tool') -%} + {{- '[e~[\n' -}} + {%- endif -%} + + {%- elif message.role == 'user' -%} + {{- ']~b]user' ~ '\n' }} + {{- visible_text(message.content) }} + {{- '[e~[' ~ '\n' }} + {%- endif -%} +{%- endfor -%} + +{#- Generation prompt -#} +{%- if add_generation_prompt -%} +{{- ']~b]ai' ~ '\n' ~ '' ~ '\n' }} +{%- endif -%} diff --git a/config.json b/config.json new file mode 100644 index 0000000000000000000000000000000000000000..af4a30a3431ee5eda5164788c4c54b473dbb6f83 --- /dev/null +++ b/config.json @@ -0,0 +1,615 @@ +{ + "architectures": [ + "MiniMaxM2ForCausalLM" + ], + "attention_dropout": 0.0, + "attn_type_list": [ + 1, + 1, + 1, + 1, + 1, + 1, + 1, + 1, + 1, + 1, + 1, + 1, + 1, + 1, + 1, + 1, + 1, + 1, + 1, + 1, + 1, + 1, + 1, + 1, + 1, + 1, + 1, + 1, + 1, + 1, + 1, + 1, + 1, + 1, + 1, + 1, + 1, + 1, + 1, + 1, + 1, + 1, + 1, + 1, + 1, + 1, + 1, + 1, + 1, + 1, + 1, + 1, + 1, + 1, + 1, + 1, + 1, + 1, + 1, + 1, + 1, + 1 + ], + "auto_map": { + "AutoConfig": "configuration_minimax_m2.MiniMaxM2Config", + "AutoModelForCausalLM": "modeling_minimax_m2.MiniMaxM2ForCausalLM" + }, + "bos_token_id": 1, + "eos_token_id": 200020, + "head_dim": 128, + "hidden_act": "silu", + "hidden_size": 3072, + "initializer_range": 0.02, + "intermediate_size": 1536, + "max_position_embeddings": 196608, + "model_type": "minimax_m2", + "mtp_transformer_layers": 1, + "num_attention_heads": 48, + "num_experts_per_tok": 8, + "num_hidden_layers": 62, + "num_key_value_heads": 8, + "num_local_experts": 205, + "num_mtp_modules": 3, + "output_router_logits": false, + "partial_rotary_factor": 0.5, + "qk_norm_type": "per_layer", + "quantization": { + "group_size": 64, + "bits": 8, + "mode": "affine", + "model.layers.0.block_sparse_moe.gate": { + "group_size": 64, + "bits": 8 + }, + "model.layers.1.block_sparse_moe.gate": { + "group_size": 64, + "bits": 8 + }, + "model.layers.2.block_sparse_moe.gate": { + "group_size": 64, + "bits": 8 + }, + "model.layers.3.block_sparse_moe.gate": { + "group_size": 64, + "bits": 8 + }, + "model.layers.4.block_sparse_moe.gate": { + "group_size": 64, + "bits": 8 + }, + "model.layers.5.block_sparse_moe.gate": { + "group_size": 64, + "bits": 8 + }, + "model.layers.6.block_sparse_moe.gate": { + "group_size": 64, + "bits": 8 + }, + "model.layers.7.block_sparse_moe.gate": { + "group_size": 64, + "bits": 8 + }, + "model.layers.8.block_sparse_moe.gate": { + "group_size": 64, + "bits": 8 + }, + "model.layers.9.block_sparse_moe.gate": { + "group_size": 64, + "bits": 8 + }, + "model.layers.10.block_sparse_moe.gate": { + "group_size": 64, + "bits": 8 + }, + "model.layers.11.block_sparse_moe.gate": { + "group_size": 64, + "bits": 8 + }, + "model.layers.12.block_sparse_moe.gate": { + "group_size": 64, + "bits": 8 + }, + "model.layers.13.block_sparse_moe.gate": { + "group_size": 64, + "bits": 8 + }, + "model.layers.14.block_sparse_moe.gate": { + "group_size": 64, + "bits": 8 + }, + "model.layers.15.block_sparse_moe.gate": { + "group_size": 64, + "bits": 8 + }, + "model.layers.16.block_sparse_moe.gate": { + "group_size": 64, + "bits": 8 + }, + "model.layers.17.block_sparse_moe.gate": { + "group_size": 64, + "bits": 8 + }, + "model.layers.18.block_sparse_moe.gate": { + "group_size": 64, + "bits": 8 + }, + "model.layers.19.block_sparse_moe.gate": { + "group_size": 64, + "bits": 8 + }, + "model.layers.20.block_sparse_moe.gate": { + "group_size": 64, + "bits": 8 + }, + "model.layers.21.block_sparse_moe.gate": { + "group_size": 64, + "bits": 8 + }, + "model.layers.22.block_sparse_moe.gate": { + "group_size": 64, + "bits": 8 + }, + "model.layers.23.block_sparse_moe.gate": { + "group_size": 64, + "bits": 8 + }, + "model.layers.24.block_sparse_moe.gate": { + "group_size": 64, + "bits": 8 + }, + "model.layers.25.block_sparse_moe.gate": { + "group_size": 64, + "bits": 8 + }, + "model.layers.26.block_sparse_moe.gate": { + "group_size": 64, + "bits": 8 + }, + "model.layers.27.block_sparse_moe.gate": { + "group_size": 64, + "bits": 8 + }, + "model.layers.28.block_sparse_moe.gate": { + "group_size": 64, + "bits": 8 + }, + "model.layers.29.block_sparse_moe.gate": { + "group_size": 64, + "bits": 8 + }, + "model.layers.30.block_sparse_moe.gate": { + "group_size": 64, + "bits": 8 + }, + "model.layers.31.block_sparse_moe.gate": { + "group_size": 64, + "bits": 8 + }, + "model.layers.32.block_sparse_moe.gate": { + "group_size": 64, + "bits": 8 + }, + "model.layers.33.block_sparse_moe.gate": { + "group_size": 64, + "bits": 8 + }, + "model.layers.34.block_sparse_moe.gate": { + "group_size": 64, + "bits": 8 + }, + "model.layers.35.block_sparse_moe.gate": { + "group_size": 64, + "bits": 8 + }, + "model.layers.36.block_sparse_moe.gate": { + "group_size": 64, + "bits": 8 + }, + "model.layers.37.block_sparse_moe.gate": { + "group_size": 64, + "bits": 8 + }, + "model.layers.38.block_sparse_moe.gate": { + "group_size": 64, + "bits": 8 + }, + "model.layers.39.block_sparse_moe.gate": { + "group_size": 64, + "bits": 8 + }, + "model.layers.40.block_sparse_moe.gate": { + "group_size": 64, + "bits": 8 + }, + "model.layers.41.block_sparse_moe.gate": { + "group_size": 64, + "bits": 8 + }, + "model.layers.42.block_sparse_moe.gate": { + "group_size": 64, + "bits": 8 + }, + "model.layers.43.block_sparse_moe.gate": { + "group_size": 64, + "bits": 8 + }, + "model.layers.44.block_sparse_moe.gate": { + "group_size": 64, + "bits": 8 + }, + "model.layers.45.block_sparse_moe.gate": { + "group_size": 64, + "bits": 8 + }, + "model.layers.46.block_sparse_moe.gate": { + "group_size": 64, + "bits": 8 + }, + "model.layers.47.block_sparse_moe.gate": { + "group_size": 64, + "bits": 8 + }, + "model.layers.48.block_sparse_moe.gate": { + "group_size": 64, + "bits": 8 + }, + "model.layers.49.block_sparse_moe.gate": { + "group_size": 64, + "bits": 8 + }, + "model.layers.50.block_sparse_moe.gate": { + "group_size": 64, + "bits": 8 + }, + "model.layers.51.block_sparse_moe.gate": { + "group_size": 64, + "bits": 8 + }, + "model.layers.52.block_sparse_moe.gate": { + "group_size": 64, + "bits": 8 + }, + "model.layers.53.block_sparse_moe.gate": { + "group_size": 64, + "bits": 8 + }, + "model.layers.54.block_sparse_moe.gate": { + "group_size": 64, + "bits": 8 + }, + "model.layers.55.block_sparse_moe.gate": { + "group_size": 64, + "bits": 8 + }, + "model.layers.56.block_sparse_moe.gate": { + "group_size": 64, + "bits": 8 + }, + "model.layers.57.block_sparse_moe.gate": { + "group_size": 64, + "bits": 8 + }, + "model.layers.58.block_sparse_moe.gate": { + "group_size": 64, + "bits": 8 + }, + "model.layers.59.block_sparse_moe.gate": { + "group_size": 64, + "bits": 8 + }, + "model.layers.60.block_sparse_moe.gate": { + "group_size": 64, + "bits": 8 + }, + "model.layers.61.block_sparse_moe.gate": { + "group_size": 64, + "bits": 8 + } + }, + "quantization_config": { + "group_size": 64, + "bits": 8, + "mode": "affine", + "model.layers.0.block_sparse_moe.gate": { + "group_size": 64, + "bits": 8 + }, + "model.layers.1.block_sparse_moe.gate": { + "group_size": 64, + "bits": 8 + }, + "model.layers.2.block_sparse_moe.gate": { + "group_size": 64, + "bits": 8 + }, + "model.layers.3.block_sparse_moe.gate": { + "group_size": 64, + "bits": 8 + }, + "model.layers.4.block_sparse_moe.gate": { + "group_size": 64, + "bits": 8 + }, + "model.layers.5.block_sparse_moe.gate": { + "group_size": 64, + "bits": 8 + }, + "model.layers.6.block_sparse_moe.gate": { + "group_size": 64, + "bits": 8 + }, + "model.layers.7.block_sparse_moe.gate": { + "group_size": 64, + "bits": 8 + }, + "model.layers.8.block_sparse_moe.gate": { + "group_size": 64, + "bits": 8 + }, + "model.layers.9.block_sparse_moe.gate": { + "group_size": 64, + "bits": 8 + }, + "model.layers.10.block_sparse_moe.gate": { + "group_size": 64, + "bits": 8 + }, + "model.layers.11.block_sparse_moe.gate": { + "group_size": 64, + "bits": 8 + }, + "model.layers.12.block_sparse_moe.gate": { + "group_size": 64, + "bits": 8 + }, + "model.layers.13.block_sparse_moe.gate": { + "group_size": 64, + "bits": 8 + }, + "model.layers.14.block_sparse_moe.gate": { + "group_size": 64, + "bits": 8 + }, + "model.layers.15.block_sparse_moe.gate": { + "group_size": 64, + "bits": 8 + }, + "model.layers.16.block_sparse_moe.gate": { + "group_size": 64, + "bits": 8 + }, + "model.layers.17.block_sparse_moe.gate": { + "group_size": 64, + "bits": 8 + }, + "model.layers.18.block_sparse_moe.gate": { + "group_size": 64, + "bits": 8 + }, + "model.layers.19.block_sparse_moe.gate": { + "group_size": 64, + "bits": 8 + }, + "model.layers.20.block_sparse_moe.gate": { + "group_size": 64, + "bits": 8 + }, + "model.layers.21.block_sparse_moe.gate": { + "group_size": 64, + "bits": 8 + }, + "model.layers.22.block_sparse_moe.gate": { + "group_size": 64, + "bits": 8 + }, + "model.layers.23.block_sparse_moe.gate": { + "group_size": 64, + "bits": 8 + }, + "model.layers.24.block_sparse_moe.gate": { + "group_size": 64, + "bits": 8 + }, + "model.layers.25.block_sparse_moe.gate": { + "group_size": 64, + "bits": 8 + }, + "model.layers.26.block_sparse_moe.gate": { + "group_size": 64, + "bits": 8 + }, + "model.layers.27.block_sparse_moe.gate": { + "group_size": 64, + "bits": 8 + }, + "model.layers.28.block_sparse_moe.gate": { + "group_size": 64, + "bits": 8 + }, + "model.layers.29.block_sparse_moe.gate": { + "group_size": 64, + "bits": 8 + }, + "model.layers.30.block_sparse_moe.gate": { + "group_size": 64, + "bits": 8 + }, + "model.layers.31.block_sparse_moe.gate": { + "group_size": 64, + "bits": 8 + }, + "model.layers.32.block_sparse_moe.gate": { + "group_size": 64, + "bits": 8 + }, + "model.layers.33.block_sparse_moe.gate": { + "group_size": 64, + "bits": 8 + }, + "model.layers.34.block_sparse_moe.gate": { + "group_size": 64, + "bits": 8 + }, + "model.layers.35.block_sparse_moe.gate": { + "group_size": 64, + "bits": 8 + }, + "model.layers.36.block_sparse_moe.gate": { + "group_size": 64, + "bits": 8 + }, + "model.layers.37.block_sparse_moe.gate": { + "group_size": 64, + "bits": 8 + }, + "model.layers.38.block_sparse_moe.gate": { + "group_size": 64, + "bits": 8 + }, + "model.layers.39.block_sparse_moe.gate": { + "group_size": 64, + "bits": 8 + }, + "model.layers.40.block_sparse_moe.gate": { + "group_size": 64, + "bits": 8 + }, + "model.layers.41.block_sparse_moe.gate": { + "group_size": 64, + "bits": 8 + }, + "model.layers.42.block_sparse_moe.gate": { + "group_size": 64, + "bits": 8 + }, + "model.layers.43.block_sparse_moe.gate": { + "group_size": 64, + "bits": 8 + }, + "model.layers.44.block_sparse_moe.gate": { + "group_size": 64, + "bits": 8 + }, + "model.layers.45.block_sparse_moe.gate": { + "group_size": 64, + "bits": 8 + }, + "model.layers.46.block_sparse_moe.gate": { + "group_size": 64, + "bits": 8 + }, + "model.layers.47.block_sparse_moe.gate": { + "group_size": 64, + "bits": 8 + }, + "model.layers.48.block_sparse_moe.gate": { + "group_size": 64, + "bits": 8 + }, + "model.layers.49.block_sparse_moe.gate": { + "group_size": 64, + "bits": 8 + }, + "model.layers.50.block_sparse_moe.gate": { + "group_size": 64, + "bits": 8 + }, + "model.layers.51.block_sparse_moe.gate": { + "group_size": 64, + "bits": 8 + }, + "model.layers.52.block_sparse_moe.gate": { + "group_size": 64, + "bits": 8 + }, + "model.layers.53.block_sparse_moe.gate": { + "group_size": 64, + "bits": 8 + }, + "model.layers.54.block_sparse_moe.gate": { + "group_size": 64, + "bits": 8 + }, + "model.layers.55.block_sparse_moe.gate": { + "group_size": 64, + "bits": 8 + }, + "model.layers.56.block_sparse_moe.gate": { + "group_size": 64, + "bits": 8 + }, + "model.layers.57.block_sparse_moe.gate": { + "group_size": 64, + "bits": 8 + }, + "model.layers.58.block_sparse_moe.gate": { + "group_size": 64, + "bits": 8 + }, + "model.layers.59.block_sparse_moe.gate": { + "group_size": 64, + "bits": 8 + }, + "model.layers.60.block_sparse_moe.gate": { + "group_size": 64, + "bits": 8 + }, + "model.layers.61.block_sparse_moe.gate": { + "group_size": 64, + "bits": 8 + } + }, + "rms_norm_eps": 1e-06, + "rope_theta": 5000000, + "rotary_dim": 64, + "router_aux_loss_coef": 0.001, + "router_jitter_noise": 0.0, + "scoring_func": "sigmoid", + "shared_intermediate_size": 0, + "sliding_window": null, + "tie_word_embeddings": false, + "torch_dtype": "float32", + "transformers_version": "4.55.0", + "use_cache": false, + "use_mtp": true, + "use_qk_norm": true, + "use_routing_bias": true, + "vocab_size": 200064 +} \ No newline at end of file diff --git a/configuration_minimax_m2.py b/configuration_minimax_m2.py new file mode 100644 index 0000000000000000000000000000000000000000..7fcd9861c389c8c8c437784de4f5f2adf4688747 --- /dev/null +++ b/configuration_minimax_m2.py @@ -0,0 +1,200 @@ +# 🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨 +# This file was automatically generated from src/transformers/models/minimax_m2/modular_minimax_m2.py. +# Do NOT edit this file manually as any edits will be overwritten by the generation of +# the file from the modular. If any change should be done, please apply the change to the +# modular_minimax_m2.py file directly. One of our CI enforces this. +# 🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨 +# coding=utf-8 +# Copyright 2025 the HuggingFace Team. All rights reserved. +# +# Licensed under the Apache License, Version 2.0 (the "License"); +# you may not use this file except in compliance with the License. +# You may obtain a copy of the License at +# +# http://www.apache.org/licenses/LICENSE-2.0 +# +# Unless required by applicable law or agreed to in writing, software +# distributed under the License is distributed on an "AS IS" BASIS, +# WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. +# See the License for the specific language governing permissions and +# limitations under the License. + + +from transformers.configuration_utils import PretrainedConfig + + +class MiniMaxM2Config(PretrainedConfig): + r""" + This is the configuration class to store the configuration of a [`MiniMaxM2Model`]. It is used to instantiate an + MiniMaxM2 model according to the specified arguments, defining the model architecture. Instantiating a configuration + with the defaults will yield a similar configuration to that of the MiniMaxM2-7B-v0.1 or MiniMaxM2-7B-Instruct-v0.1. + + [minimax_m2ai/MiniMaxM2-8x7B](https://huggingface.co/minimax_m2ai/MiniMaxM2-8x7B) + [minimax_m2ai/MiniMaxM2-7B-Instruct-v0.1](https://huggingface.co/minimax_m2ai/MiniMaxM2-7B-Instruct-v0.1) + + Configuration objects inherit from [`PretrainedConfig`] and can be used to control the model outputs. Read the + documentation from [`PretrainedConfig`] for more information. + + + Args: + vocab_size (`int`, *optional*, defaults to 32000): + Vocabulary size of the MiniMaxM2 model. Defines the number of different tokens that can be represented by the + `inputs_ids` passed when calling [`MiniMaxM2Model`] + hidden_size (`int`, *optional*, defaults to 4096): + Dimension of the hidden representations. + intermediate_size (`int`, *optional*, defaults to 14336): + Dimension of the MLP representations. + num_hidden_layers (`int`, *optional*, defaults to 32): + Number of hidden layers in the Transformer encoder. + num_attention_heads (`int`, *optional*, defaults to 32): + Number of attention heads for each attention layer in the Transformer encoder. + num_key_value_heads (`int`, *optional*, defaults to 8): + This is the number of key_value heads that should be used to implement Grouped Query Attention. If + `num_key_value_heads=num_attention_heads`, the model will use Multi Head Attention (MHA), if + `num_key_value_heads=1` the model will use Multi Query Attention (MQA) otherwise GQA is used. When + converting a multi-head checkpoint to a GQA checkpoint, each group key and value head should be constructed + by meanpooling all the original heads within that group. For more details, check out [this + paper](https://huggingface.co/papers/2305.13245). If it is not specified, will default to `8`. + head_dim (`int`, *optional*, defaults to `hidden_size // num_attention_heads`): + The attention head dimension. + hidden_act (`str` or `function`, *optional*, defaults to `"silu"`): + The non-linear activation function (function or string) in the decoder. + max_position_embeddings (`int`, *optional*, defaults to `4096*32`): + The maximum sequence length that this model might ever be used with. MiniMaxM2's sliding window attention + allows sequence of up to 4096*32 tokens. + initializer_range (`float`, *optional*, defaults to 0.02): + The standard deviation of the truncated_normal_initializer for initializing all weight matrices. + rms_norm_eps (`float`, *optional*, defaults to 1e-05): + The epsilon used by the rms normalization layers. + use_cache (`bool`, *optional*, defaults to `True`): + Whether or not the model should return the last key/values attentions (not used by all models). Only + relevant if `config.is_decoder=True`. + pad_token_id (`int`, *optional*): + The id of the padding token. + bos_token_id (`int`, *optional*, defaults to 1): + The id of the "beginning-of-sequence" token. + eos_token_id (`int`, *optional*, defaults to 2): + The id of the "end-of-sequence" token. + tie_word_embeddings (`bool`, *optional*, defaults to `False`): + Whether the model's input and output word embeddings should be tied. + rope_theta (`float`, *optional*, defaults to 1000000.0): + The base period of the RoPE embeddings. + sliding_window (`int`, *optional*): + Sliding window attention window size. If not specified, will default to `4096`. + attention_dropout (`float`, *optional*, defaults to 0.0): + The dropout ratio for the attention probabilities. + num_experts_per_tok (`int`, *optional*, defaults to 2): + The number of experts to route per-token, can be also interpreted as the `top-k` routing + parameter + num_local_experts (`int`, *optional*, defaults to 8): + Number of experts per Sparse MLP layer. + output_router_logits (`bool`, *optional*, defaults to `False`): + Whether or not the router logits should be returned by the model. Enabling this will also + allow the model to output the auxiliary loss. See [here]() for more details + router_aux_loss_coef (`float`, *optional*, defaults to 0.001): + The aux loss factor for the total loss. + router_jitter_noise (`float`, *optional*, defaults to 0.0): + Amount of noise to add to the router. + + ```python + >>> from transformers import MiniMaxM2Model, MiniMaxM2Config + + >>> # Initializing a MiniMaxM2 7B style configuration + >>> configuration = MiniMaxM2Config() + + >>> # Initializing a model from the MiniMaxM2 7B style configuration + >>> model = MiniMaxM2Model(configuration) + + >>> # Accessing the model configuration + >>> configuration = model.config + ```""" + + model_type = "minimax_m2" + keys_to_ignore_at_inference = ["past_key_values"] + base_model_tp_plan = { + "layers.*.self_attn.q_proj": "colwise", + "layers.*.self_attn.k_proj": "colwise", + "layers.*.self_attn.v_proj": "colwise", + "layers.*.self_attn.o_proj": "rowwise", + "layers.*.block_sparse_moe.gate": "colwise_rep", # we need to replicate here to correctly route experts + "layers.*.block_sparse_moe.experts.*.w1": "colwise", + "layers.*.block_sparse_moe.experts.*.w2": "rowwise", + "layers.*.block_sparse_moe.experts.*.w3": "colwise", + } + base_model_pp_plan = { + "embed_tokens": (["input_ids"], ["inputs_embeds"]), + "layers": (["hidden_states", "attention_mask"], ["hidden_states"]), + "norm": (["hidden_states"], ["hidden_states"]), + } + + def __init__( + self, + vocab_size=32000, + hidden_size=4096, + intermediate_size=14336, + num_hidden_layers=32, + num_attention_heads=32, + num_key_value_heads=8, + head_dim=None, + hidden_act="silu", + max_position_embeddings=4096 * 32, + initializer_range=0.02, + rms_norm_eps=1e-5, + use_cache=True, + pad_token_id=None, + bos_token_id=1, + eos_token_id=2, + tie_word_embeddings=False, + rope_theta=1e6, + sliding_window=None, + attention_dropout=0.0, + num_experts_per_tok=2, + num_local_experts=8, + output_router_logits=False, + router_aux_loss_coef=0.001, + router_jitter_noise=0.0, + **kwargs, + ): + self.vocab_size = vocab_size + self.max_position_embeddings = max_position_embeddings + self.hidden_size = hidden_size + self.intermediate_size = intermediate_size + self.num_hidden_layers = num_hidden_layers + self.num_attention_heads = num_attention_heads + self.sliding_window = sliding_window + + # for backward compatibility + if num_key_value_heads is None: + num_key_value_heads = num_attention_heads + + self.num_key_value_heads = num_key_value_heads + self.hidden_act = hidden_act + self.initializer_range = initializer_range + self.rms_norm_eps = rms_norm_eps + self.use_cache = use_cache + self.rope_theta = rope_theta + self.attention_dropout = attention_dropout + self.head_dim = head_dim + + self.num_experts_per_tok = num_experts_per_tok + self.num_local_experts = num_local_experts + self.output_router_logits = output_router_logits + self.router_aux_loss_coef = router_aux_loss_coef + self.router_jitter_noise = router_jitter_noise + + self.use_qk_norm = kwargs.pop("use_qk_norm", False) + self.rotary_dim = kwargs.pop("rotary_dim", self.head_dim) + self.partial_rotary_factor = kwargs.pop("partial_rotary_factor", 1) + if self.head_dim is not None: + self.partial_rotary_factor = self.rotary_dim / self.head_dim + + super().__init__( + pad_token_id=pad_token_id, + bos_token_id=bos_token_id, + eos_token_id=eos_token_id, + tie_word_embeddings=tie_word_embeddings, + **kwargs, + ) + + +__all__ = ["MiniMaxM2Config"] diff --git a/generation_config.json b/generation_config.json new file mode 100644 index 0000000000000000000000000000000000000000..482d4056b157fb4efd2ae66d0f77734632e383cb --- /dev/null +++ b/generation_config.json @@ -0,0 +1,8 @@ +{ + "bos_token_id": 200019, + "do_sample": true, + "eos_token_id": 200020, + "top_k": 40, + "top_p": 0.95, + "transformers_version": "4.55.0" +} diff --git a/model-00001-of-00047.safetensors b/model-00001-of-00047.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..f9d61529747ae443851b24a59531e976348e48c9 --- /dev/null +++ b/model-00001-of-00047.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:0af84c1315e762a853df6c0ddf9ae1e324be4d944e1c8623a34f93ff9d0be1d6 +size 5144918686 diff --git a/model-00002-of-00047.safetensors b/model-00002-of-00047.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..cf882ad9c06e6026604c76f56aeb79be9de2599c --- /dev/null +++ b/model-00002-of-00047.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:4f4cb2c0bdecfefb098764637476a025694f60a64b669603891a2c3f24f5c14c +size 4403212872 diff --git a/model-00003-of-00047.safetensors b/model-00003-of-00047.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..334c4e9ddacf869fd6c6063fa64b7839bea2a2df --- /dev/null +++ b/model-00003-of-00047.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:edea02071c91757bd593a3aa7d37b949b24b40bf19b59fd1eb5cde454619a6a0 +size 4453522880 diff --git a/model-00004-of-00047.safetensors b/model-00004-of-00047.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..61b698aeb62f7a906b84b178f78c4e447c206388 --- /dev/null +++ b/model-00004-of-00047.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:098cd680035e619bd7c5901541f14bc4b52c2251ff3623bf2bdfcc6ee627333c +size 4403212884 diff --git a/model-00005-of-00047.safetensors b/model-00005-of-00047.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..836eda5988a9c7c3920674d297300d370b3005eb --- /dev/null +++ b/model-00005-of-00047.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:1ee406d028d1616470b4d3e6a722cd4e711edc3ff6182ce02a5dfeb32599d22e +size 4403212880 diff --git a/model-00006-of-00047.safetensors b/model-00006-of-00047.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..db8bbf93ddce6954e20f0f47b8b33508ebeab50b --- /dev/null +++ b/model-00006-of-00047.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:93bb9e930406b2241af083af91e0b460ba5bdd203d1b2e15182fce4be9eeea51 +size 4453522892 diff --git a/model-00007-of-00047.safetensors b/model-00007-of-00047.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..8aa93009297660480620a64b48516296c87f1486 --- /dev/null +++ b/model-00007-of-00047.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:ab31cf841c8eac3d3d4ea27e57bdd763f026af3b41f8d0834a72995ca9359494 +size 4403212882 diff --git a/model-00008-of-00047.safetensors b/model-00008-of-00047.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..0602cf74152fcd68852856bbbe577e8422362e35 --- /dev/null +++ b/model-00008-of-00047.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:052bd92e887d75d806b1833a171b7fd73e3a8f36d691ab41a036b514789424da +size 4403212901 diff --git a/model-00009-of-00047.safetensors b/model-00009-of-00047.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..cb27be1de73148c9aab463f1c2e4041ef383c70d --- /dev/null +++ b/model-00009-of-00047.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:27e93c2046732e9af1d430c329251982827749d9d68c8b72a5e058401674bf1b +size 4453522928 diff --git a/model-00010-of-00047.safetensors b/model-00010-of-00047.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..ed683f5170a3d659164802c357ca180af351160f --- /dev/null +++ b/model-00010-of-00047.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:631bca98d0373d80e503d8e0f8626a142db00757b50e9e9f28c681d9fe88165e +size 4403212916 diff --git a/model-00011-of-00047.safetensors b/model-00011-of-00047.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..a1e61f299c5ec8561fe9886d409ad946ff676467 --- /dev/null +++ b/model-00011-of-00047.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:b127be5050978a86038a139b06ec79070c84214180a73ef33236eb0e99fcc68e +size 4403212890 diff --git a/model-00012-of-00047.safetensors b/model-00012-of-00047.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..bbdba7accd8517166861e9da1803d1d044fef080 --- /dev/null +++ b/model-00012-of-00047.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:2eb5a6454df478904589182755d184be2d6a5ba732aba2ec016c1c1dc459649f +size 4453522924 diff --git a/model-00013-of-00047.safetensors b/model-00013-of-00047.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..ede101a178c35e4b86dd5d30de088e0a2419f6ee --- /dev/null +++ b/model-00013-of-00047.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:9ab990c393b4c390aee05e21fb4f198043c93a7f881ae284ecc03d86de1d0860 +size 4403212844 diff --git a/model-00014-of-00047.safetensors b/model-00014-of-00047.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..1a8da8dd58d868e3ea816eeef90c816a69cc4d8d --- /dev/null +++ b/model-00014-of-00047.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:b6977f53c5d9d7526c2faf25e8903c5c8cd85a6a3e85c6b83c658f5be052f35f +size 4403212908 diff --git a/model-00015-of-00047.safetensors b/model-00015-of-00047.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..681c9b1c2f18bd25da416e361fd29061faeb8b6b --- /dev/null +++ b/model-00015-of-00047.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:5c6be4e92fba402feb918419aac9bf2c4e0af4cda980b09fc5660c951fd93582 +size 4453522932 diff --git a/model-00016-of-00047.safetensors b/model-00016-of-00047.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..cbef06b1ee0ae5dfc01a380403530edcb54a54a6 --- /dev/null +++ b/model-00016-of-00047.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:77d780e6a475af8147a974d95a4928fe3719d1fb2fa4871438405d7be59dd03c +size 4403212916 diff --git a/model-00017-of-00047.safetensors b/model-00017-of-00047.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..9826ce8a2e72871f336316b562a432b07bf3fd4d --- /dev/null +++ b/model-00017-of-00047.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:709bea7ae8fa7fce5ac7e3e222a510e8553471f4d6fc03702002aac1fbad0aa1 +size 4403212910 diff --git a/model-00018-of-00047.safetensors b/model-00018-of-00047.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..599a0eea2e128a1197ecc607ab168edb174750ef --- /dev/null +++ b/model-00018-of-00047.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:6c15ced78f1b7b1b5e30902b6b1d9a86a15f657b691fee945799700aba80e289 +size 4453522926 diff --git a/model-00019-of-00047.safetensors b/model-00019-of-00047.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..026165357e9dc0a17755cc61100f82f10926ff74 --- /dev/null +++ b/model-00019-of-00047.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:d2aa63fadf9895e6518ed81024c1711abd986d3b353a01301ac14b05342b78bd +size 4403212840 diff --git a/model-00020-of-00047.safetensors b/model-00020-of-00047.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..3a78f30f5e4b2afd4ccbeab8cfd88fe9e561a239 --- /dev/null +++ b/model-00020-of-00047.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:485f352292c1c8de3796b0e73fda4eb195fe80d4601f9545edc69e64ebfaa9aa +size 4403212910 diff --git a/model-00021-of-00047.safetensors b/model-00021-of-00047.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..4a0c5e62727f22d10baf0ced6b74b8e51d82fdf8 --- /dev/null +++ b/model-00021-of-00047.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:37d42baac941869768cf434c5f1e262a9117b79835df9f8ac10354480a17517c +size 4453522902 diff --git a/model-00022-of-00047.safetensors b/model-00022-of-00047.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..c2b34ef9429cbc7df930aa0c5edaddadb11cf1d1 --- /dev/null +++ b/model-00022-of-00047.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:eb2e29f3d8538b32c8b70018c175fb2412d14b824462d5642283217f635066a8 +size 4403212916 diff --git a/model-00023-of-00047.safetensors b/model-00023-of-00047.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..fa42019bc9c5bf739a638b835e576f3306e02482 --- /dev/null +++ b/model-00023-of-00047.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:afeecbfbc25f18764c1a3c3cd7f8641f5c5fd933d105a54d73789b83e0a95faa +size 4403212906 diff --git a/model-00024-of-00047.safetensors b/model-00024-of-00047.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..91657f27fd9efe1b87a00b11268b70de9de83636 --- /dev/null +++ b/model-00024-of-00047.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:bd56188e0b75c0ff9d6e3ef1bcc21fdc143179bc1c57f9c0df7270aff5a8e366 +size 4453522938 diff --git a/model-00025-of-00047.safetensors b/model-00025-of-00047.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..522f4f554b6427c9a31d6dbd869eb9555b356b83 --- /dev/null +++ b/model-00025-of-00047.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:0a47b219b2272445d67ff5bb62479e7cedb01f3f99b1dfbc53691191f074742b +size 4403212892 diff --git a/model-00026-of-00047.safetensors b/model-00026-of-00047.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..acc0ceb0cc4cadb99bbd80434af9c2d387c5e1c9 --- /dev/null +++ b/model-00026-of-00047.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:96f18cebccdd4b4e65e64998b3bc974bf1ed47a9cacd60e7bd87d3e9357302c6 +size 4403212908 diff --git a/model-00027-of-00047.safetensors b/model-00027-of-00047.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..ec2788888833396c7e0c6a29087d1364bfab862f --- /dev/null +++ b/model-00027-of-00047.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:7b4aa491b0a7b7d63bf25fc62874f98fa40025be6e303e9b324cd4ba4ac3e629 +size 4453522918 diff --git a/model-00028-of-00047.safetensors b/model-00028-of-00047.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..81a30f30d4dc97a11202986db64915ff2948ecb6 --- /dev/null +++ b/model-00028-of-00047.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:4a02aa377ac6b45adf9f49458455da7d4a73e81faee8cda7df48f92b9378d767 +size 4403212888 diff --git a/model-00029-of-00047.safetensors b/model-00029-of-00047.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..985674cea345e9e3a442f592a05fff93e6cb0c86 --- /dev/null +++ b/model-00029-of-00047.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:3331082c0cc502989acf887be40ed9a429de0d32fa54ab8b32cd423bf441eb0e +size 4403212888 diff --git a/model-00030-of-00047.safetensors b/model-00030-of-00047.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..191cb1becae08dead06598e06c2e15ce5445d1e2 --- /dev/null +++ b/model-00030-of-00047.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:3d4e459ed36052bba7526ba40a5700ec2ae96209786ee27156665c04f52245a7 +size 4453522944 diff --git a/model-00031-of-00047.safetensors b/model-00031-of-00047.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..11c5ba9caee5dfac9a5a0a0c2b3b0b1521fd3097 --- /dev/null +++ b/model-00031-of-00047.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:9448213bbdf10f2c7aa0240901e630e01553401c52761dd10d007eeac2dd680b +size 4403212916 diff --git a/model-00032-of-00047.safetensors b/model-00032-of-00047.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..47056ce6c5cdf4ad2a1aafe4c7259f9cccfe1309 --- /dev/null +++ b/model-00032-of-00047.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:a99d6a86a281cce93deaf44261a4e28c69bbaf97da22111a9c38ec843e83e32a +size 4403212908 diff --git a/model-00033-of-00047.safetensors b/model-00033-of-00047.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..34deb262fe5b2e59bfba731a0b999c4615a0a717 --- /dev/null +++ b/model-00033-of-00047.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:cdec44dcd9c2ea8e59ed8b7db8e135ede7981b10e328f19699644b49d4ffdb04 +size 4453522922 diff --git a/model-00034-of-00047.safetensors b/model-00034-of-00047.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..6a279a303d13cff66fa6ed2610416e798ffeddd4 --- /dev/null +++ b/model-00034-of-00047.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:666fec83971b83f05fcbcca2c9e5619242d318259f7896ad32e410892d53dc69 +size 4403212908 diff --git a/model-00035-of-00047.safetensors b/model-00035-of-00047.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..ae52d449deb0972be4ad8d3a76b7fa26bc9f6e69 --- /dev/null +++ b/model-00035-of-00047.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:bedf966b38598a4f1b62c985453cca43c6c7fa821ffa867255ecf0a421393675 +size 4403212910 diff --git a/model-00036-of-00047.safetensors b/model-00036-of-00047.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..ed414dfaf44a4a9173d604d3905a6992801691b2 --- /dev/null +++ b/model-00036-of-00047.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:6f50fcadf9089c4b79c506e2e87415f7bcf98cf1186ea951d1c468d1fc6048ba +size 4453522928 diff --git a/model-00037-of-00047.safetensors b/model-00037-of-00047.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..acec155de4b92b74b24296dae3406f3f224e4f17 --- /dev/null +++ b/model-00037-of-00047.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:4813eb6f826dccd737fcbbceee31fe8c4da6bb4c5443d02508c531c1c62c50be +size 4403212866 diff --git a/model-00038-of-00047.safetensors b/model-00038-of-00047.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..514d67e57d0dc052a1d0b892011ef030c39e2246 --- /dev/null +++ b/model-00038-of-00047.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:e691957e88a713744bd1eddcf4b654d4e68c739a809860c62a4f1057d67fd698 +size 4403212908 diff --git a/model-00039-of-00047.safetensors b/model-00039-of-00047.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..e54f313b0d48de45af947e37e5ee59e1c774bd05 --- /dev/null +++ b/model-00039-of-00047.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:13f6314f96b05379fa24d576fc97e77504018609f5745643e1b79826584b655e +size 4453522946 diff --git a/model-00040-of-00047.safetensors b/model-00040-of-00047.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..401aa1471d7872513cbd833741af79ecff97b3cd --- /dev/null +++ b/model-00040-of-00047.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:4bd6a76b769739ae6a7a7fdc23f5930bdf4331bc80ec33bbba77dc7c2a3dca46 +size 4403212914 diff --git a/model-00041-of-00047.safetensors b/model-00041-of-00047.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..4e8d194ed5218f40ecfacff99a4a1b57af098b4a --- /dev/null +++ b/model-00041-of-00047.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:a7b3e3f65eefb69e843261846ee095d544a152a2cde3d87d0656fcff92689384 +size 4403212908 diff --git a/model-00042-of-00047.safetensors b/model-00042-of-00047.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..41f5913c13e17fd9984dabd28f43abbb3b10da27 --- /dev/null +++ b/model-00042-of-00047.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:c53f944da79f017ed08ca23d4d859b3d724d0e37559da757bdc8129f3bc3f4e7 +size 4453522944 diff --git a/model-00043-of-00047.safetensors b/model-00043-of-00047.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..1a6e55bd3fe01258225d37c01392b65a5fd227e7 --- /dev/null +++ b/model-00043-of-00047.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:57abaa3b1016c0815a2673c9eb71537594cdf97f2c46679060aa514155e1f575 +size 4403212916 diff --git a/model-00044-of-00047.safetensors b/model-00044-of-00047.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..697b7c7eb3895dbf754c9f198f7a7052057ea11b --- /dev/null +++ b/model-00044-of-00047.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:5a85758657da4fdd9ebd37be199a456007a348340819ab13c96c01d400ea270c +size 4403212908 diff --git a/model-00045-of-00047.safetensors b/model-00045-of-00047.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..54523e2ce743b530f1a25498be48923b38295ea4 --- /dev/null +++ b/model-00045-of-00047.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:9d69767e689482a2ddd9d9070bda88288713d8c3369e51ee2d892f364e1d244d +size 4453522936 diff --git a/model-00046-of-00047.safetensors b/model-00046-of-00047.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..3dabc284f4ee8318e49365ec2169e79413ce5c4d --- /dev/null +++ b/model-00046-of-00047.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:87946f9732ed7c57965104779e8042a132d17393d395f2776de2260171ae2a04 +size 4403212880 diff --git a/model-00047-of-00047.safetensors b/model-00047-of-00047.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..409df54bc4d071c0c167bc0490a2fa9a8dc2402b --- /dev/null +++ b/model-00047-of-00047.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:e881d0e313a2f9672cc308b3fea84931659ffc122f39217acb5e7482f230f2cd +size 2867911006 diff --git a/model.safetensors.index.json b/model.safetensors.index.json new file mode 100644 index 0000000000000000000000000000000000000000..190780b0902cffa5ada713ced00bc0cc26fa002e --- /dev/null +++ b/model.safetensors.index.json @@ -0,0 +1,1813 @@ +{ + "metadata": { + "total_size": 206911838104, + "total_parameters": 183919471616 + }, + "weight_map": { + "lm_head.biases": "model-00047-of-00047.safetensors", + "lm_head.scales": "model-00047-of-00047.safetensors", + "lm_head.weight": "model-00047-of-00047.safetensors", + "model.embed_tokens.biases": "model-00001-of-00047.safetensors", + "model.embed_tokens.scales": "model-00001-of-00047.safetensors", + "model.embed_tokens.weight": "model-00001-of-00047.safetensors", + "model.layers.0.block_sparse_moe.e_score_correction_bias": "model-00001-of-00047.safetensors", + "model.layers.0.block_sparse_moe.gate.biases": "model-00001-of-00047.safetensors", + "model.layers.0.block_sparse_moe.gate.scales": "model-00001-of-00047.safetensors", + "model.layers.0.block_sparse_moe.gate.weight": "model-00001-of-00047.safetensors", + "model.layers.0.block_sparse_moe.switch_mlp.down_proj.biases": "model-00001-of-00047.safetensors", + "model.layers.0.block_sparse_moe.switch_mlp.down_proj.scales": "model-00001-of-00047.safetensors", + "model.layers.0.block_sparse_moe.switch_mlp.down_proj.weight": "model-00001-of-00047.safetensors", + "model.layers.0.block_sparse_moe.switch_mlp.gate_proj.biases": "model-00001-of-00047.safetensors", + "model.layers.0.block_sparse_moe.switch_mlp.gate_proj.scales": "model-00001-of-00047.safetensors", + "model.layers.0.block_sparse_moe.switch_mlp.gate_proj.weight": "model-00001-of-00047.safetensors", + "model.layers.0.block_sparse_moe.switch_mlp.up_proj.biases": "model-00001-of-00047.safetensors", + "model.layers.0.block_sparse_moe.switch_mlp.up_proj.scales": "model-00001-of-00047.safetensors", + "model.layers.0.block_sparse_moe.switch_mlp.up_proj.weight": "model-00001-of-00047.safetensors", + "model.layers.0.input_layernorm.weight": "model-00001-of-00047.safetensors", + "model.layers.0.post_attention_layernorm.weight": "model-00001-of-00047.safetensors", + "model.layers.0.self_attn.k_norm.weight": "model-00001-of-00047.safetensors", + "model.layers.0.self_attn.k_proj.biases": "model-00001-of-00047.safetensors", + "model.layers.0.self_attn.k_proj.scales": "model-00001-of-00047.safetensors", + "model.layers.0.self_attn.k_proj.weight": "model-00001-of-00047.safetensors", + "model.layers.0.self_attn.o_proj.biases": "model-00001-of-00047.safetensors", + "model.layers.0.self_attn.o_proj.scales": "model-00001-of-00047.safetensors", + "model.layers.0.self_attn.o_proj.weight": "model-00001-of-00047.safetensors", + "model.layers.0.self_attn.q_norm.weight": "model-00001-of-00047.safetensors", + "model.layers.0.self_attn.q_proj.biases": "model-00001-of-00047.safetensors", + "model.layers.0.self_attn.q_proj.scales": "model-00001-of-00047.safetensors", + "model.layers.0.self_attn.q_proj.weight": "model-00001-of-00047.safetensors", + "model.layers.0.self_attn.v_proj.biases": "model-00001-of-00047.safetensors", + "model.layers.0.self_attn.v_proj.scales": "model-00001-of-00047.safetensors", + "model.layers.0.self_attn.v_proj.weight": "model-00001-of-00047.safetensors", + "model.layers.1.block_sparse_moe.e_score_correction_bias": "model-00002-of-00047.safetensors", + "model.layers.1.block_sparse_moe.gate.biases": "model-00001-of-00047.safetensors", + "model.layers.1.block_sparse_moe.gate.scales": "model-00001-of-00047.safetensors", + "model.layers.1.block_sparse_moe.gate.weight": "model-00001-of-00047.safetensors", + "model.layers.1.block_sparse_moe.switch_mlp.down_proj.biases": "model-00002-of-00047.safetensors", + "model.layers.1.block_sparse_moe.switch_mlp.down_proj.scales": "model-00002-of-00047.safetensors", + "model.layers.1.block_sparse_moe.switch_mlp.down_proj.weight": "model-00002-of-00047.safetensors", + "model.layers.1.block_sparse_moe.switch_mlp.gate_proj.biases": "model-00001-of-00047.safetensors", + "model.layers.1.block_sparse_moe.switch_mlp.gate_proj.scales": "model-00001-of-00047.safetensors", + "model.layers.1.block_sparse_moe.switch_mlp.gate_proj.weight": "model-00001-of-00047.safetensors", + "model.layers.1.block_sparse_moe.switch_mlp.up_proj.biases": "model-00002-of-00047.safetensors", + "model.layers.1.block_sparse_moe.switch_mlp.up_proj.scales": "model-00002-of-00047.safetensors", + "model.layers.1.block_sparse_moe.switch_mlp.up_proj.weight": "model-00002-of-00047.safetensors", + "model.layers.1.input_layernorm.weight": "model-00002-of-00047.safetensors", + "model.layers.1.post_attention_layernorm.weight": "model-00002-of-00047.safetensors", + "model.layers.1.self_attn.k_norm.weight": "model-00001-of-00047.safetensors", + "model.layers.1.self_attn.k_proj.biases": "model-00001-of-00047.safetensors", + "model.layers.1.self_attn.k_proj.scales": "model-00001-of-00047.safetensors", + "model.layers.1.self_attn.k_proj.weight": "model-00001-of-00047.safetensors", + "model.layers.1.self_attn.o_proj.biases": "model-00001-of-00047.safetensors", + "model.layers.1.self_attn.o_proj.scales": "model-00001-of-00047.safetensors", + "model.layers.1.self_attn.o_proj.weight": "model-00001-of-00047.safetensors", + "model.layers.1.self_attn.q_norm.weight": "model-00001-of-00047.safetensors", + "model.layers.1.self_attn.q_proj.biases": "model-00001-of-00047.safetensors", + "model.layers.1.self_attn.q_proj.scales": "model-00001-of-00047.safetensors", + "model.layers.1.self_attn.q_proj.weight": "model-00001-of-00047.safetensors", + "model.layers.1.self_attn.v_proj.biases": "model-00001-of-00047.safetensors", + "model.layers.1.self_attn.v_proj.scales": "model-00001-of-00047.safetensors", + "model.layers.1.self_attn.v_proj.weight": "model-00001-of-00047.safetensors", + "model.layers.10.block_sparse_moe.e_score_correction_bias": "model-00009-of-00047.safetensors", + "model.layers.10.block_sparse_moe.gate.biases": "model-00008-of-00047.safetensors", + "model.layers.10.block_sparse_moe.gate.scales": "model-00008-of-00047.safetensors", + "model.layers.10.block_sparse_moe.gate.weight": "model-00008-of-00047.safetensors", + "model.layers.10.block_sparse_moe.switch_mlp.down_proj.biases": "model-00009-of-00047.safetensors", + "model.layers.10.block_sparse_moe.switch_mlp.down_proj.scales": "model-00009-of-00047.safetensors", + "model.layers.10.block_sparse_moe.switch_mlp.down_proj.weight": "model-00009-of-00047.safetensors", + "model.layers.10.block_sparse_moe.switch_mlp.gate_proj.biases": "model-00008-of-00047.safetensors", + "model.layers.10.block_sparse_moe.switch_mlp.gate_proj.scales": "model-00008-of-00047.safetensors", + "model.layers.10.block_sparse_moe.switch_mlp.gate_proj.weight": "model-00008-of-00047.safetensors", + "model.layers.10.block_sparse_moe.switch_mlp.up_proj.biases": "model-00008-of-00047.safetensors", + "model.layers.10.block_sparse_moe.switch_mlp.up_proj.scales": "model-00008-of-00047.safetensors", + "model.layers.10.block_sparse_moe.switch_mlp.up_proj.weight": "model-00008-of-00047.safetensors", + "model.layers.10.input_layernorm.weight": "model-00009-of-00047.safetensors", + "model.layers.10.post_attention_layernorm.weight": "model-00009-of-00047.safetensors", + "model.layers.10.self_attn.k_norm.weight": "model-00008-of-00047.safetensors", + "model.layers.10.self_attn.k_proj.biases": "model-00008-of-00047.safetensors", + "model.layers.10.self_attn.k_proj.scales": "model-00008-of-00047.safetensors", + "model.layers.10.self_attn.k_proj.weight": "model-00008-of-00047.safetensors", + "model.layers.10.self_attn.o_proj.biases": "model-00008-of-00047.safetensors", + "model.layers.10.self_attn.o_proj.scales": "model-00008-of-00047.safetensors", + "model.layers.10.self_attn.o_proj.weight": "model-00008-of-00047.safetensors", + "model.layers.10.self_attn.q_norm.weight": "model-00008-of-00047.safetensors", + "model.layers.10.self_attn.q_proj.biases": "model-00008-of-00047.safetensors", + "model.layers.10.self_attn.q_proj.scales": "model-00008-of-00047.safetensors", + "model.layers.10.self_attn.q_proj.weight": "model-00008-of-00047.safetensors", + "model.layers.10.self_attn.v_proj.biases": "model-00008-of-00047.safetensors", + "model.layers.10.self_attn.v_proj.scales": "model-00008-of-00047.safetensors", + "model.layers.10.self_attn.v_proj.weight": "model-00008-of-00047.safetensors", + "model.layers.11.block_sparse_moe.e_score_correction_bias": "model-00009-of-00047.safetensors", + "model.layers.11.block_sparse_moe.gate.biases": "model-00009-of-00047.safetensors", + "model.layers.11.block_sparse_moe.gate.scales": "model-00009-of-00047.safetensors", + "model.layers.11.block_sparse_moe.gate.weight": "model-00009-of-00047.safetensors", + "model.layers.11.block_sparse_moe.switch_mlp.down_proj.biases": "model-00009-of-00047.safetensors", + "model.layers.11.block_sparse_moe.switch_mlp.down_proj.scales": "model-00009-of-00047.safetensors", + "model.layers.11.block_sparse_moe.switch_mlp.down_proj.weight": "model-00009-of-00047.safetensors", + "model.layers.11.block_sparse_moe.switch_mlp.gate_proj.biases": "model-00009-of-00047.safetensors", + "model.layers.11.block_sparse_moe.switch_mlp.gate_proj.scales": "model-00009-of-00047.safetensors", + "model.layers.11.block_sparse_moe.switch_mlp.gate_proj.weight": "model-00009-of-00047.safetensors", + "model.layers.11.block_sparse_moe.switch_mlp.up_proj.biases": "model-00009-of-00047.safetensors", + "model.layers.11.block_sparse_moe.switch_mlp.up_proj.scales": "model-00009-of-00047.safetensors", + "model.layers.11.block_sparse_moe.switch_mlp.up_proj.weight": "model-00009-of-00047.safetensors", + "model.layers.11.input_layernorm.weight": "model-00009-of-00047.safetensors", + "model.layers.11.post_attention_layernorm.weight": "model-00009-of-00047.safetensors", + "model.layers.11.self_attn.k_norm.weight": "model-00009-of-00047.safetensors", + "model.layers.11.self_attn.k_proj.biases": "model-00009-of-00047.safetensors", + "model.layers.11.self_attn.k_proj.scales": "model-00009-of-00047.safetensors", + "model.layers.11.self_attn.k_proj.weight": "model-00009-of-00047.safetensors", + "model.layers.11.self_attn.o_proj.biases": "model-00009-of-00047.safetensors", + "model.layers.11.self_attn.o_proj.scales": "model-00009-of-00047.safetensors", + "model.layers.11.self_attn.o_proj.weight": "model-00009-of-00047.safetensors", + "model.layers.11.self_attn.q_norm.weight": "model-00009-of-00047.safetensors", + "model.layers.11.self_attn.q_proj.biases": "model-00009-of-00047.safetensors", + "model.layers.11.self_attn.q_proj.scales": "model-00009-of-00047.safetensors", + "model.layers.11.self_attn.q_proj.weight": "model-00009-of-00047.safetensors", + "model.layers.11.self_attn.v_proj.biases": "model-00009-of-00047.safetensors", + "model.layers.11.self_attn.v_proj.scales": "model-00009-of-00047.safetensors", + "model.layers.11.self_attn.v_proj.weight": "model-00009-of-00047.safetensors", + "model.layers.12.block_sparse_moe.e_score_correction_bias": "model-00010-of-00047.safetensors", + "model.layers.12.block_sparse_moe.gate.biases": "model-00009-of-00047.safetensors", + "model.layers.12.block_sparse_moe.gate.scales": "model-00009-of-00047.safetensors", + "model.layers.12.block_sparse_moe.gate.weight": "model-00009-of-00047.safetensors", + "model.layers.12.block_sparse_moe.switch_mlp.down_proj.biases": "model-00010-of-00047.safetensors", + "model.layers.12.block_sparse_moe.switch_mlp.down_proj.scales": "model-00010-of-00047.safetensors", + "model.layers.12.block_sparse_moe.switch_mlp.down_proj.weight": "model-00010-of-00047.safetensors", + "model.layers.12.block_sparse_moe.switch_mlp.gate_proj.biases": "model-00010-of-00047.safetensors", + "model.layers.12.block_sparse_moe.switch_mlp.gate_proj.scales": "model-00010-of-00047.safetensors", + "model.layers.12.block_sparse_moe.switch_mlp.gate_proj.weight": "model-00010-of-00047.safetensors", + "model.layers.12.block_sparse_moe.switch_mlp.up_proj.biases": "model-00010-of-00047.safetensors", + "model.layers.12.block_sparse_moe.switch_mlp.up_proj.scales": "model-00010-of-00047.safetensors", + "model.layers.12.block_sparse_moe.switch_mlp.up_proj.weight": "model-00010-of-00047.safetensors", + "model.layers.12.input_layernorm.weight": "model-00010-of-00047.safetensors", + "model.layers.12.post_attention_layernorm.weight": "model-00010-of-00047.safetensors", + "model.layers.12.self_attn.k_norm.weight": "model-00009-of-00047.safetensors", + "model.layers.12.self_attn.k_proj.biases": "model-00009-of-00047.safetensors", + "model.layers.12.self_attn.k_proj.scales": "model-00009-of-00047.safetensors", + "model.layers.12.self_attn.k_proj.weight": "model-00009-of-00047.safetensors", + "model.layers.12.self_attn.o_proj.biases": "model-00009-of-00047.safetensors", + "model.layers.12.self_attn.o_proj.scales": "model-00009-of-00047.safetensors", + "model.layers.12.self_attn.o_proj.weight": "model-00009-of-00047.safetensors", + "model.layers.12.self_attn.q_norm.weight": "model-00009-of-00047.safetensors", + "model.layers.12.self_attn.q_proj.biases": "model-00009-of-00047.safetensors", + "model.layers.12.self_attn.q_proj.scales": "model-00009-of-00047.safetensors", + "model.layers.12.self_attn.q_proj.weight": "model-00009-of-00047.safetensors", + "model.layers.12.self_attn.v_proj.biases": "model-00009-of-00047.safetensors", + "model.layers.12.self_attn.v_proj.scales": "model-00009-of-00047.safetensors", + "model.layers.12.self_attn.v_proj.weight": "model-00009-of-00047.safetensors", + "model.layers.13.block_sparse_moe.e_score_correction_bias": "model-00011-of-00047.safetensors", + "model.layers.13.block_sparse_moe.gate.biases": "model-00010-of-00047.safetensors", + "model.layers.13.block_sparse_moe.gate.scales": "model-00010-of-00047.safetensors", + "model.layers.13.block_sparse_moe.gate.weight": "model-00010-of-00047.safetensors", + "model.layers.13.block_sparse_moe.switch_mlp.down_proj.biases": "model-00011-of-00047.safetensors", + "model.layers.13.block_sparse_moe.switch_mlp.down_proj.scales": "model-00011-of-00047.safetensors", + "model.layers.13.block_sparse_moe.switch_mlp.down_proj.weight": "model-00011-of-00047.safetensors", + "model.layers.13.block_sparse_moe.switch_mlp.gate_proj.biases": "model-00010-of-00047.safetensors", + "model.layers.13.block_sparse_moe.switch_mlp.gate_proj.scales": "model-00010-of-00047.safetensors", + "model.layers.13.block_sparse_moe.switch_mlp.gate_proj.weight": "model-00010-of-00047.safetensors", + "model.layers.13.block_sparse_moe.switch_mlp.up_proj.biases": "model-00011-of-00047.safetensors", + "model.layers.13.block_sparse_moe.switch_mlp.up_proj.scales": "model-00011-of-00047.safetensors", + "model.layers.13.block_sparse_moe.switch_mlp.up_proj.weight": "model-00011-of-00047.safetensors", + "model.layers.13.input_layernorm.weight": "model-00011-of-00047.safetensors", + "model.layers.13.post_attention_layernorm.weight": "model-00011-of-00047.safetensors", + "model.layers.13.self_attn.k_norm.weight": "model-00010-of-00047.safetensors", + "model.layers.13.self_attn.k_proj.biases": "model-00010-of-00047.safetensors", + "model.layers.13.self_attn.k_proj.scales": "model-00010-of-00047.safetensors", + "model.layers.13.self_attn.k_proj.weight": "model-00010-of-00047.safetensors", + "model.layers.13.self_attn.o_proj.biases": "model-00010-of-00047.safetensors", + "model.layers.13.self_attn.o_proj.scales": "model-00010-of-00047.safetensors", + "model.layers.13.self_attn.o_proj.weight": "model-00010-of-00047.safetensors", + "model.layers.13.self_attn.q_norm.weight": "model-00010-of-00047.safetensors", + "model.layers.13.self_attn.q_proj.biases": "model-00010-of-00047.safetensors", + "model.layers.13.self_attn.q_proj.scales": "model-00010-of-00047.safetensors", + "model.layers.13.self_attn.q_proj.weight": "model-00010-of-00047.safetensors", + "model.layers.13.self_attn.v_proj.biases": "model-00010-of-00047.safetensors", + "model.layers.13.self_attn.v_proj.scales": "model-00010-of-00047.safetensors", + "model.layers.13.self_attn.v_proj.weight": "model-00010-of-00047.safetensors", + "model.layers.14.block_sparse_moe.e_score_correction_bias": "model-00012-of-00047.safetensors", + "model.layers.14.block_sparse_moe.gate.biases": "model-00011-of-00047.safetensors", + "model.layers.14.block_sparse_moe.gate.scales": "model-00011-of-00047.safetensors", + "model.layers.14.block_sparse_moe.gate.weight": "model-00011-of-00047.safetensors", + "model.layers.14.block_sparse_moe.switch_mlp.down_proj.biases": "model-00012-of-00047.safetensors", + "model.layers.14.block_sparse_moe.switch_mlp.down_proj.scales": "model-00012-of-00047.safetensors", + "model.layers.14.block_sparse_moe.switch_mlp.down_proj.weight": "model-00012-of-00047.safetensors", + "model.layers.14.block_sparse_moe.switch_mlp.gate_proj.biases": "model-00011-of-00047.safetensors", + "model.layers.14.block_sparse_moe.switch_mlp.gate_proj.scales": "model-00011-of-00047.safetensors", + "model.layers.14.block_sparse_moe.switch_mlp.gate_proj.weight": "model-00011-of-00047.safetensors", + "model.layers.14.block_sparse_moe.switch_mlp.up_proj.biases": "model-00011-of-00047.safetensors", + "model.layers.14.block_sparse_moe.switch_mlp.up_proj.scales": "model-00011-of-00047.safetensors", + "model.layers.14.block_sparse_moe.switch_mlp.up_proj.weight": "model-00011-of-00047.safetensors", + "model.layers.14.input_layernorm.weight": "model-00012-of-00047.safetensors", + "model.layers.14.post_attention_layernorm.weight": "model-00012-of-00047.safetensors", + "model.layers.14.self_attn.k_norm.weight": "model-00011-of-00047.safetensors", + "model.layers.14.self_attn.k_proj.biases": "model-00011-of-00047.safetensors", + "model.layers.14.self_attn.k_proj.scales": "model-00011-of-00047.safetensors", + "model.layers.14.self_attn.k_proj.weight": "model-00011-of-00047.safetensors", + "model.layers.14.self_attn.o_proj.biases": "model-00011-of-00047.safetensors", + "model.layers.14.self_attn.o_proj.scales": "model-00011-of-00047.safetensors", + "model.layers.14.self_attn.o_proj.weight": "model-00011-of-00047.safetensors", + "model.layers.14.self_attn.q_norm.weight": "model-00011-of-00047.safetensors", + "model.layers.14.self_attn.q_proj.biases": "model-00011-of-00047.safetensors", + "model.layers.14.self_attn.q_proj.scales": "model-00011-of-00047.safetensors", + "model.layers.14.self_attn.q_proj.weight": "model-00011-of-00047.safetensors", + "model.layers.14.self_attn.v_proj.biases": "model-00011-of-00047.safetensors", + "model.layers.14.self_attn.v_proj.scales": "model-00011-of-00047.safetensors", + "model.layers.14.self_attn.v_proj.weight": "model-00011-of-00047.safetensors", + "model.layers.15.block_sparse_moe.e_score_correction_bias": "model-00012-of-00047.safetensors", + "model.layers.15.block_sparse_moe.gate.biases": "model-00012-of-00047.safetensors", + "model.layers.15.block_sparse_moe.gate.scales": "model-00012-of-00047.safetensors", + "model.layers.15.block_sparse_moe.gate.weight": "model-00012-of-00047.safetensors", + "model.layers.15.block_sparse_moe.switch_mlp.down_proj.biases": "model-00012-of-00047.safetensors", + "model.layers.15.block_sparse_moe.switch_mlp.down_proj.scales": "model-00012-of-00047.safetensors", + "model.layers.15.block_sparse_moe.switch_mlp.down_proj.weight": "model-00012-of-00047.safetensors", + "model.layers.15.block_sparse_moe.switch_mlp.gate_proj.biases": "model-00012-of-00047.safetensors", + "model.layers.15.block_sparse_moe.switch_mlp.gate_proj.scales": "model-00012-of-00047.safetensors", + "model.layers.15.block_sparse_moe.switch_mlp.gate_proj.weight": "model-00012-of-00047.safetensors", + "model.layers.15.block_sparse_moe.switch_mlp.up_proj.biases": "model-00012-of-00047.safetensors", + "model.layers.15.block_sparse_moe.switch_mlp.up_proj.scales": "model-00012-of-00047.safetensors", + "model.layers.15.block_sparse_moe.switch_mlp.up_proj.weight": "model-00012-of-00047.safetensors", + "model.layers.15.input_layernorm.weight": "model-00012-of-00047.safetensors", + "model.layers.15.post_attention_layernorm.weight": "model-00012-of-00047.safetensors", + "model.layers.15.self_attn.k_norm.weight": "model-00012-of-00047.safetensors", + "model.layers.15.self_attn.k_proj.biases": "model-00012-of-00047.safetensors", + "model.layers.15.self_attn.k_proj.scales": "model-00012-of-00047.safetensors", + "model.layers.15.self_attn.k_proj.weight": "model-00012-of-00047.safetensors", + "model.layers.15.self_attn.o_proj.biases": "model-00012-of-00047.safetensors", + "model.layers.15.self_attn.o_proj.scales": "model-00012-of-00047.safetensors", + "model.layers.15.self_attn.o_proj.weight": "model-00012-of-00047.safetensors", + "model.layers.15.self_attn.q_norm.weight": "model-00012-of-00047.safetensors", + "model.layers.15.self_attn.q_proj.biases": "model-00012-of-00047.safetensors", + "model.layers.15.self_attn.q_proj.scales": "model-00012-of-00047.safetensors", + "model.layers.15.self_attn.q_proj.weight": "model-00012-of-00047.safetensors", + "model.layers.15.self_attn.v_proj.biases": "model-00012-of-00047.safetensors", + "model.layers.15.self_attn.v_proj.scales": "model-00012-of-00047.safetensors", + "model.layers.15.self_attn.v_proj.weight": "model-00012-of-00047.safetensors", + "model.layers.16.block_sparse_moe.e_score_correction_bias": "model-00013-of-00047.safetensors", + "model.layers.16.block_sparse_moe.gate.biases": "model-00012-of-00047.safetensors", + "model.layers.16.block_sparse_moe.gate.scales": "model-00012-of-00047.safetensors", + "model.layers.16.block_sparse_moe.gate.weight": "model-00012-of-00047.safetensors", + "model.layers.16.block_sparse_moe.switch_mlp.down_proj.biases": "model-00013-of-00047.safetensors", + "model.layers.16.block_sparse_moe.switch_mlp.down_proj.scales": "model-00013-of-00047.safetensors", + "model.layers.16.block_sparse_moe.switch_mlp.down_proj.weight": "model-00013-of-00047.safetensors", + "model.layers.16.block_sparse_moe.switch_mlp.gate_proj.biases": "model-00013-of-00047.safetensors", + "model.layers.16.block_sparse_moe.switch_mlp.gate_proj.scales": "model-00013-of-00047.safetensors", + "model.layers.16.block_sparse_moe.switch_mlp.gate_proj.weight": "model-00013-of-00047.safetensors", + "model.layers.16.block_sparse_moe.switch_mlp.up_proj.biases": "model-00013-of-00047.safetensors", + "model.layers.16.block_sparse_moe.switch_mlp.up_proj.scales": "model-00013-of-00047.safetensors", + "model.layers.16.block_sparse_moe.switch_mlp.up_proj.weight": "model-00013-of-00047.safetensors", + "model.layers.16.input_layernorm.weight": "model-00013-of-00047.safetensors", + "model.layers.16.post_attention_layernorm.weight": "model-00013-of-00047.safetensors", + "model.layers.16.self_attn.k_norm.weight": "model-00012-of-00047.safetensors", + "model.layers.16.self_attn.k_proj.biases": "model-00012-of-00047.safetensors", + "model.layers.16.self_attn.k_proj.scales": "model-00012-of-00047.safetensors", + "model.layers.16.self_attn.k_proj.weight": "model-00012-of-00047.safetensors", + "model.layers.16.self_attn.o_proj.biases": "model-00012-of-00047.safetensors", + "model.layers.16.self_attn.o_proj.scales": "model-00012-of-00047.safetensors", + "model.layers.16.self_attn.o_proj.weight": "model-00012-of-00047.safetensors", + "model.layers.16.self_attn.q_norm.weight": "model-00012-of-00047.safetensors", + "model.layers.16.self_attn.q_proj.biases": "model-00012-of-00047.safetensors", + "model.layers.16.self_attn.q_proj.scales": "model-00012-of-00047.safetensors", + "model.layers.16.self_attn.q_proj.weight": "model-00012-of-00047.safetensors", + "model.layers.16.self_attn.v_proj.biases": "model-00012-of-00047.safetensors", + "model.layers.16.self_attn.v_proj.scales": "model-00012-of-00047.safetensors", + "model.layers.16.self_attn.v_proj.weight": "model-00012-of-00047.safetensors", + "model.layers.17.block_sparse_moe.e_score_correction_bias": "model-00014-of-00047.safetensors", + "model.layers.17.block_sparse_moe.gate.biases": "model-00013-of-00047.safetensors", + "model.layers.17.block_sparse_moe.gate.scales": "model-00013-of-00047.safetensors", + "model.layers.17.block_sparse_moe.gate.weight": "model-00013-of-00047.safetensors", + "model.layers.17.block_sparse_moe.switch_mlp.down_proj.biases": "model-00014-of-00047.safetensors", + "model.layers.17.block_sparse_moe.switch_mlp.down_proj.scales": "model-00014-of-00047.safetensors", + "model.layers.17.block_sparse_moe.switch_mlp.down_proj.weight": "model-00014-of-00047.safetensors", + "model.layers.17.block_sparse_moe.switch_mlp.gate_proj.biases": "model-00013-of-00047.safetensors", + "model.layers.17.block_sparse_moe.switch_mlp.gate_proj.scales": "model-00013-of-00047.safetensors", + "model.layers.17.block_sparse_moe.switch_mlp.gate_proj.weight": "model-00013-of-00047.safetensors", + "model.layers.17.block_sparse_moe.switch_mlp.up_proj.biases": "model-00014-of-00047.safetensors", + "model.layers.17.block_sparse_moe.switch_mlp.up_proj.scales": "model-00014-of-00047.safetensors", + "model.layers.17.block_sparse_moe.switch_mlp.up_proj.weight": "model-00014-of-00047.safetensors", + "model.layers.17.input_layernorm.weight": "model-00014-of-00047.safetensors", + "model.layers.17.post_attention_layernorm.weight": "model-00014-of-00047.safetensors", + "model.layers.17.self_attn.k_norm.weight": "model-00013-of-00047.safetensors", + "model.layers.17.self_attn.k_proj.biases": "model-00013-of-00047.safetensors", + "model.layers.17.self_attn.k_proj.scales": "model-00013-of-00047.safetensors", + "model.layers.17.self_attn.k_proj.weight": "model-00013-of-00047.safetensors", + "model.layers.17.self_attn.o_proj.biases": "model-00013-of-00047.safetensors", + "model.layers.17.self_attn.o_proj.scales": "model-00013-of-00047.safetensors", + "model.layers.17.self_attn.o_proj.weight": "model-00013-of-00047.safetensors", + "model.layers.17.self_attn.q_norm.weight": "model-00013-of-00047.safetensors", + "model.layers.17.self_attn.q_proj.biases": "model-00013-of-00047.safetensors", + "model.layers.17.self_attn.q_proj.scales": "model-00013-of-00047.safetensors", + "model.layers.17.self_attn.q_proj.weight": "model-00013-of-00047.safetensors", + "model.layers.17.self_attn.v_proj.biases": "model-00013-of-00047.safetensors", + "model.layers.17.self_attn.v_proj.scales": "model-00013-of-00047.safetensors", + "model.layers.17.self_attn.v_proj.weight": "model-00013-of-00047.safetensors", + "model.layers.18.block_sparse_moe.e_score_correction_bias": "model-00015-of-00047.safetensors", + "model.layers.18.block_sparse_moe.gate.biases": "model-00014-of-00047.safetensors", + "model.layers.18.block_sparse_moe.gate.scales": "model-00014-of-00047.safetensors", + "model.layers.18.block_sparse_moe.gate.weight": "model-00014-of-00047.safetensors", + "model.layers.18.block_sparse_moe.switch_mlp.down_proj.biases": "model-00015-of-00047.safetensors", + "model.layers.18.block_sparse_moe.switch_mlp.down_proj.scales": "model-00015-of-00047.safetensors", + "model.layers.18.block_sparse_moe.switch_mlp.down_proj.weight": "model-00015-of-00047.safetensors", + "model.layers.18.block_sparse_moe.switch_mlp.gate_proj.biases": "model-00014-of-00047.safetensors", + "model.layers.18.block_sparse_moe.switch_mlp.gate_proj.scales": "model-00014-of-00047.safetensors", + "model.layers.18.block_sparse_moe.switch_mlp.gate_proj.weight": "model-00014-of-00047.safetensors", + "model.layers.18.block_sparse_moe.switch_mlp.up_proj.biases": "model-00014-of-00047.safetensors", + "model.layers.18.block_sparse_moe.switch_mlp.up_proj.scales": "model-00014-of-00047.safetensors", + "model.layers.18.block_sparse_moe.switch_mlp.up_proj.weight": "model-00014-of-00047.safetensors", + "model.layers.18.input_layernorm.weight": "model-00015-of-00047.safetensors", + "model.layers.18.post_attention_layernorm.weight": "model-00015-of-00047.safetensors", + "model.layers.18.self_attn.k_norm.weight": "model-00014-of-00047.safetensors", + "model.layers.18.self_attn.k_proj.biases": "model-00014-of-00047.safetensors", + "model.layers.18.self_attn.k_proj.scales": "model-00014-of-00047.safetensors", + "model.layers.18.self_attn.k_proj.weight": "model-00014-of-00047.safetensors", + "model.layers.18.self_attn.o_proj.biases": "model-00014-of-00047.safetensors", + "model.layers.18.self_attn.o_proj.scales": "model-00014-of-00047.safetensors", + "model.layers.18.self_attn.o_proj.weight": "model-00014-of-00047.safetensors", + "model.layers.18.self_attn.q_norm.weight": "model-00014-of-00047.safetensors", + "model.layers.18.self_attn.q_proj.biases": "model-00014-of-00047.safetensors", + "model.layers.18.self_attn.q_proj.scales": "model-00014-of-00047.safetensors", + "model.layers.18.self_attn.q_proj.weight": "model-00014-of-00047.safetensors", + "model.layers.18.self_attn.v_proj.biases": "model-00014-of-00047.safetensors", + "model.layers.18.self_attn.v_proj.scales": "model-00014-of-00047.safetensors", + "model.layers.18.self_attn.v_proj.weight": "model-00014-of-00047.safetensors", + "model.layers.19.block_sparse_moe.e_score_correction_bias": "model-00015-of-00047.safetensors", + "model.layers.19.block_sparse_moe.gate.biases": "model-00015-of-00047.safetensors", + "model.layers.19.block_sparse_moe.gate.scales": "model-00015-of-00047.safetensors", + "model.layers.19.block_sparse_moe.gate.weight": "model-00015-of-00047.safetensors", + "model.layers.19.block_sparse_moe.switch_mlp.down_proj.biases": "model-00015-of-00047.safetensors", + "model.layers.19.block_sparse_moe.switch_mlp.down_proj.scales": "model-00015-of-00047.safetensors", + "model.layers.19.block_sparse_moe.switch_mlp.down_proj.weight": "model-00015-of-00047.safetensors", + "model.layers.19.block_sparse_moe.switch_mlp.gate_proj.biases": "model-00015-of-00047.safetensors", + "model.layers.19.block_sparse_moe.switch_mlp.gate_proj.scales": "model-00015-of-00047.safetensors", + "model.layers.19.block_sparse_moe.switch_mlp.gate_proj.weight": "model-00015-of-00047.safetensors", + "model.layers.19.block_sparse_moe.switch_mlp.up_proj.biases": "model-00015-of-00047.safetensors", + "model.layers.19.block_sparse_moe.switch_mlp.up_proj.scales": "model-00015-of-00047.safetensors", + "model.layers.19.block_sparse_moe.switch_mlp.up_proj.weight": "model-00015-of-00047.safetensors", + "model.layers.19.input_layernorm.weight": "model-00015-of-00047.safetensors", + "model.layers.19.post_attention_layernorm.weight": "model-00015-of-00047.safetensors", + "model.layers.19.self_attn.k_norm.weight": "model-00015-of-00047.safetensors", + "model.layers.19.self_attn.k_proj.biases": "model-00015-of-00047.safetensors", + "model.layers.19.self_attn.k_proj.scales": "model-00015-of-00047.safetensors", + "model.layers.19.self_attn.k_proj.weight": "model-00015-of-00047.safetensors", + "model.layers.19.self_attn.o_proj.biases": "model-00015-of-00047.safetensors", + "model.layers.19.self_attn.o_proj.scales": "model-00015-of-00047.safetensors", + "model.layers.19.self_attn.o_proj.weight": "model-00015-of-00047.safetensors", + "model.layers.19.self_attn.q_norm.weight": "model-00015-of-00047.safetensors", + "model.layers.19.self_attn.q_proj.biases": "model-00015-of-00047.safetensors", + "model.layers.19.self_attn.q_proj.scales": "model-00015-of-00047.safetensors", + "model.layers.19.self_attn.q_proj.weight": "model-00015-of-00047.safetensors", + "model.layers.19.self_attn.v_proj.biases": "model-00015-of-00047.safetensors", + "model.layers.19.self_attn.v_proj.scales": "model-00015-of-00047.safetensors", + "model.layers.19.self_attn.v_proj.weight": "model-00015-of-00047.safetensors", + "model.layers.2.block_sparse_moe.e_score_correction_bias": "model-00003-of-00047.safetensors", + "model.layers.2.block_sparse_moe.gate.biases": "model-00002-of-00047.safetensors", + "model.layers.2.block_sparse_moe.gate.scales": "model-00002-of-00047.safetensors", + "model.layers.2.block_sparse_moe.gate.weight": "model-00002-of-00047.safetensors", + "model.layers.2.block_sparse_moe.switch_mlp.down_proj.biases": "model-00003-of-00047.safetensors", + "model.layers.2.block_sparse_moe.switch_mlp.down_proj.scales": "model-00003-of-00047.safetensors", + "model.layers.2.block_sparse_moe.switch_mlp.down_proj.weight": "model-00003-of-00047.safetensors", + "model.layers.2.block_sparse_moe.switch_mlp.gate_proj.biases": "model-00002-of-00047.safetensors", + "model.layers.2.block_sparse_moe.switch_mlp.gate_proj.scales": "model-00002-of-00047.safetensors", + "model.layers.2.block_sparse_moe.switch_mlp.gate_proj.weight": "model-00002-of-00047.safetensors", + "model.layers.2.block_sparse_moe.switch_mlp.up_proj.biases": "model-00002-of-00047.safetensors", + "model.layers.2.block_sparse_moe.switch_mlp.up_proj.scales": "model-00002-of-00047.safetensors", + "model.layers.2.block_sparse_moe.switch_mlp.up_proj.weight": "model-00002-of-00047.safetensors", + "model.layers.2.input_layernorm.weight": "model-00003-of-00047.safetensors", + "model.layers.2.post_attention_layernorm.weight": "model-00003-of-00047.safetensors", + "model.layers.2.self_attn.k_norm.weight": "model-00002-of-00047.safetensors", + "model.layers.2.self_attn.k_proj.biases": "model-00002-of-00047.safetensors", + "model.layers.2.self_attn.k_proj.scales": "model-00002-of-00047.safetensors", + "model.layers.2.self_attn.k_proj.weight": "model-00002-of-00047.safetensors", + "model.layers.2.self_attn.o_proj.biases": "model-00002-of-00047.safetensors", + "model.layers.2.self_attn.o_proj.scales": "model-00002-of-00047.safetensors", + "model.layers.2.self_attn.o_proj.weight": "model-00002-of-00047.safetensors", + "model.layers.2.self_attn.q_norm.weight": "model-00002-of-00047.safetensors", + "model.layers.2.self_attn.q_proj.biases": "model-00002-of-00047.safetensors", + "model.layers.2.self_attn.q_proj.scales": "model-00002-of-00047.safetensors", + "model.layers.2.self_attn.q_proj.weight": "model-00002-of-00047.safetensors", + "model.layers.2.self_attn.v_proj.biases": "model-00002-of-00047.safetensors", + "model.layers.2.self_attn.v_proj.scales": "model-00002-of-00047.safetensors", + "model.layers.2.self_attn.v_proj.weight": "model-00002-of-00047.safetensors", + "model.layers.20.block_sparse_moe.e_score_correction_bias": "model-00016-of-00047.safetensors", + "model.layers.20.block_sparse_moe.gate.biases": "model-00015-of-00047.safetensors", + "model.layers.20.block_sparse_moe.gate.scales": "model-00015-of-00047.safetensors", + "model.layers.20.block_sparse_moe.gate.weight": "model-00015-of-00047.safetensors", + "model.layers.20.block_sparse_moe.switch_mlp.down_proj.biases": "model-00016-of-00047.safetensors", + "model.layers.20.block_sparse_moe.switch_mlp.down_proj.scales": "model-00016-of-00047.safetensors", + "model.layers.20.block_sparse_moe.switch_mlp.down_proj.weight": "model-00016-of-00047.safetensors", + "model.layers.20.block_sparse_moe.switch_mlp.gate_proj.biases": "model-00016-of-00047.safetensors", + "model.layers.20.block_sparse_moe.switch_mlp.gate_proj.scales": "model-00016-of-00047.safetensors", + "model.layers.20.block_sparse_moe.switch_mlp.gate_proj.weight": "model-00016-of-00047.safetensors", + "model.layers.20.block_sparse_moe.switch_mlp.up_proj.biases": "model-00016-of-00047.safetensors", + "model.layers.20.block_sparse_moe.switch_mlp.up_proj.scales": "model-00016-of-00047.safetensors", + "model.layers.20.block_sparse_moe.switch_mlp.up_proj.weight": "model-00016-of-00047.safetensors", + "model.layers.20.input_layernorm.weight": "model-00016-of-00047.safetensors", + "model.layers.20.post_attention_layernorm.weight": "model-00016-of-00047.safetensors", + "model.layers.20.self_attn.k_norm.weight": "model-00015-of-00047.safetensors", + "model.layers.20.self_attn.k_proj.biases": "model-00015-of-00047.safetensors", + "model.layers.20.self_attn.k_proj.scales": "model-00015-of-00047.safetensors", + "model.layers.20.self_attn.k_proj.weight": "model-00015-of-00047.safetensors", + "model.layers.20.self_attn.o_proj.biases": "model-00015-of-00047.safetensors", + "model.layers.20.self_attn.o_proj.scales": "model-00015-of-00047.safetensors", + "model.layers.20.self_attn.o_proj.weight": "model-00015-of-00047.safetensors", + "model.layers.20.self_attn.q_norm.weight": "model-00015-of-00047.safetensors", + "model.layers.20.self_attn.q_proj.biases": "model-00015-of-00047.safetensors", + "model.layers.20.self_attn.q_proj.scales": "model-00015-of-00047.safetensors", + "model.layers.20.self_attn.q_proj.weight": "model-00015-of-00047.safetensors", + "model.layers.20.self_attn.v_proj.biases": "model-00015-of-00047.safetensors", + "model.layers.20.self_attn.v_proj.scales": "model-00015-of-00047.safetensors", + "model.layers.20.self_attn.v_proj.weight": "model-00015-of-00047.safetensors", + "model.layers.21.block_sparse_moe.e_score_correction_bias": "model-00017-of-00047.safetensors", + "model.layers.21.block_sparse_moe.gate.biases": "model-00016-of-00047.safetensors", + "model.layers.21.block_sparse_moe.gate.scales": "model-00016-of-00047.safetensors", + "model.layers.21.block_sparse_moe.gate.weight": "model-00016-of-00047.safetensors", + "model.layers.21.block_sparse_moe.switch_mlp.down_proj.biases": "model-00017-of-00047.safetensors", + "model.layers.21.block_sparse_moe.switch_mlp.down_proj.scales": "model-00017-of-00047.safetensors", + "model.layers.21.block_sparse_moe.switch_mlp.down_proj.weight": "model-00017-of-00047.safetensors", + "model.layers.21.block_sparse_moe.switch_mlp.gate_proj.biases": "model-00016-of-00047.safetensors", + "model.layers.21.block_sparse_moe.switch_mlp.gate_proj.scales": "model-00016-of-00047.safetensors", + "model.layers.21.block_sparse_moe.switch_mlp.gate_proj.weight": "model-00016-of-00047.safetensors", + "model.layers.21.block_sparse_moe.switch_mlp.up_proj.biases": "model-00017-of-00047.safetensors", + "model.layers.21.block_sparse_moe.switch_mlp.up_proj.scales": "model-00017-of-00047.safetensors", + "model.layers.21.block_sparse_moe.switch_mlp.up_proj.weight": "model-00017-of-00047.safetensors", + "model.layers.21.input_layernorm.weight": "model-00017-of-00047.safetensors", + "model.layers.21.post_attention_layernorm.weight": "model-00017-of-00047.safetensors", + "model.layers.21.self_attn.k_norm.weight": "model-00016-of-00047.safetensors", + "model.layers.21.self_attn.k_proj.biases": "model-00016-of-00047.safetensors", + "model.layers.21.self_attn.k_proj.scales": "model-00016-of-00047.safetensors", + "model.layers.21.self_attn.k_proj.weight": "model-00016-of-00047.safetensors", + "model.layers.21.self_attn.o_proj.biases": "model-00016-of-00047.safetensors", + "model.layers.21.self_attn.o_proj.scales": "model-00016-of-00047.safetensors", + "model.layers.21.self_attn.o_proj.weight": "model-00016-of-00047.safetensors", + "model.layers.21.self_attn.q_norm.weight": "model-00016-of-00047.safetensors", + "model.layers.21.self_attn.q_proj.biases": "model-00016-of-00047.safetensors", + "model.layers.21.self_attn.q_proj.scales": "model-00016-of-00047.safetensors", + "model.layers.21.self_attn.q_proj.weight": "model-00016-of-00047.safetensors", + "model.layers.21.self_attn.v_proj.biases": "model-00016-of-00047.safetensors", + "model.layers.21.self_attn.v_proj.scales": "model-00016-of-00047.safetensors", + "model.layers.21.self_attn.v_proj.weight": "model-00016-of-00047.safetensors", + "model.layers.22.block_sparse_moe.e_score_correction_bias": "model-00018-of-00047.safetensors", + "model.layers.22.block_sparse_moe.gate.biases": "model-00017-of-00047.safetensors", + "model.layers.22.block_sparse_moe.gate.scales": "model-00017-of-00047.safetensors", + "model.layers.22.block_sparse_moe.gate.weight": "model-00017-of-00047.safetensors", + "model.layers.22.block_sparse_moe.switch_mlp.down_proj.biases": "model-00018-of-00047.safetensors", + "model.layers.22.block_sparse_moe.switch_mlp.down_proj.scales": "model-00018-of-00047.safetensors", + "model.layers.22.block_sparse_moe.switch_mlp.down_proj.weight": "model-00018-of-00047.safetensors", + "model.layers.22.block_sparse_moe.switch_mlp.gate_proj.biases": "model-00017-of-00047.safetensors", + "model.layers.22.block_sparse_moe.switch_mlp.gate_proj.scales": "model-00017-of-00047.safetensors", + "model.layers.22.block_sparse_moe.switch_mlp.gate_proj.weight": "model-00017-of-00047.safetensors", + "model.layers.22.block_sparse_moe.switch_mlp.up_proj.biases": "model-00017-of-00047.safetensors", + "model.layers.22.block_sparse_moe.switch_mlp.up_proj.scales": "model-00017-of-00047.safetensors", + "model.layers.22.block_sparse_moe.switch_mlp.up_proj.weight": "model-00017-of-00047.safetensors", + "model.layers.22.input_layernorm.weight": "model-00018-of-00047.safetensors", + "model.layers.22.post_attention_layernorm.weight": "model-00018-of-00047.safetensors", + "model.layers.22.self_attn.k_norm.weight": "model-00017-of-00047.safetensors", + "model.layers.22.self_attn.k_proj.biases": "model-00017-of-00047.safetensors", + "model.layers.22.self_attn.k_proj.scales": "model-00017-of-00047.safetensors", + "model.layers.22.self_attn.k_proj.weight": "model-00017-of-00047.safetensors", + "model.layers.22.self_attn.o_proj.biases": "model-00017-of-00047.safetensors", + "model.layers.22.self_attn.o_proj.scales": "model-00017-of-00047.safetensors", + "model.layers.22.self_attn.o_proj.weight": "model-00017-of-00047.safetensors", + "model.layers.22.self_attn.q_norm.weight": "model-00017-of-00047.safetensors", + "model.layers.22.self_attn.q_proj.biases": "model-00017-of-00047.safetensors", + "model.layers.22.self_attn.q_proj.scales": "model-00017-of-00047.safetensors", + "model.layers.22.self_attn.q_proj.weight": "model-00017-of-00047.safetensors", + "model.layers.22.self_attn.v_proj.biases": "model-00017-of-00047.safetensors", + "model.layers.22.self_attn.v_proj.scales": "model-00017-of-00047.safetensors", + "model.layers.22.self_attn.v_proj.weight": "model-00017-of-00047.safetensors", + "model.layers.23.block_sparse_moe.e_score_correction_bias": "model-00018-of-00047.safetensors", + "model.layers.23.block_sparse_moe.gate.biases": "model-00018-of-00047.safetensors", + "model.layers.23.block_sparse_moe.gate.scales": "model-00018-of-00047.safetensors", + "model.layers.23.block_sparse_moe.gate.weight": "model-00018-of-00047.safetensors", + "model.layers.23.block_sparse_moe.switch_mlp.down_proj.biases": "model-00018-of-00047.safetensors", + "model.layers.23.block_sparse_moe.switch_mlp.down_proj.scales": "model-00018-of-00047.safetensors", + "model.layers.23.block_sparse_moe.switch_mlp.down_proj.weight": "model-00018-of-00047.safetensors", + "model.layers.23.block_sparse_moe.switch_mlp.gate_proj.biases": "model-00018-of-00047.safetensors", + "model.layers.23.block_sparse_moe.switch_mlp.gate_proj.scales": "model-00018-of-00047.safetensors", + "model.layers.23.block_sparse_moe.switch_mlp.gate_proj.weight": "model-00018-of-00047.safetensors", + "model.layers.23.block_sparse_moe.switch_mlp.up_proj.biases": "model-00018-of-00047.safetensors", + "model.layers.23.block_sparse_moe.switch_mlp.up_proj.scales": "model-00018-of-00047.safetensors", + "model.layers.23.block_sparse_moe.switch_mlp.up_proj.weight": "model-00018-of-00047.safetensors", + "model.layers.23.input_layernorm.weight": "model-00018-of-00047.safetensors", + "model.layers.23.post_attention_layernorm.weight": "model-00018-of-00047.safetensors", + "model.layers.23.self_attn.k_norm.weight": "model-00018-of-00047.safetensors", + "model.layers.23.self_attn.k_proj.biases": "model-00018-of-00047.safetensors", + "model.layers.23.self_attn.k_proj.scales": "model-00018-of-00047.safetensors", + "model.layers.23.self_attn.k_proj.weight": "model-00018-of-00047.safetensors", + "model.layers.23.self_attn.o_proj.biases": "model-00018-of-00047.safetensors", + "model.layers.23.self_attn.o_proj.scales": "model-00018-of-00047.safetensors", + "model.layers.23.self_attn.o_proj.weight": "model-00018-of-00047.safetensors", + "model.layers.23.self_attn.q_norm.weight": "model-00018-of-00047.safetensors", + "model.layers.23.self_attn.q_proj.biases": "model-00018-of-00047.safetensors", + "model.layers.23.self_attn.q_proj.scales": "model-00018-of-00047.safetensors", + "model.layers.23.self_attn.q_proj.weight": "model-00018-of-00047.safetensors", + "model.layers.23.self_attn.v_proj.biases": "model-00018-of-00047.safetensors", + "model.layers.23.self_attn.v_proj.scales": "model-00018-of-00047.safetensors", + "model.layers.23.self_attn.v_proj.weight": "model-00018-of-00047.safetensors", + "model.layers.24.block_sparse_moe.e_score_correction_bias": "model-00019-of-00047.safetensors", + "model.layers.24.block_sparse_moe.gate.biases": "model-00018-of-00047.safetensors", + "model.layers.24.block_sparse_moe.gate.scales": "model-00018-of-00047.safetensors", + "model.layers.24.block_sparse_moe.gate.weight": "model-00018-of-00047.safetensors", + "model.layers.24.block_sparse_moe.switch_mlp.down_proj.biases": "model-00019-of-00047.safetensors", + "model.layers.24.block_sparse_moe.switch_mlp.down_proj.scales": "model-00019-of-00047.safetensors", + "model.layers.24.block_sparse_moe.switch_mlp.down_proj.weight": "model-00019-of-00047.safetensors", + "model.layers.24.block_sparse_moe.switch_mlp.gate_proj.biases": "model-00019-of-00047.safetensors", + "model.layers.24.block_sparse_moe.switch_mlp.gate_proj.scales": "model-00019-of-00047.safetensors", + "model.layers.24.block_sparse_moe.switch_mlp.gate_proj.weight": "model-00019-of-00047.safetensors", + "model.layers.24.block_sparse_moe.switch_mlp.up_proj.biases": "model-00019-of-00047.safetensors", + "model.layers.24.block_sparse_moe.switch_mlp.up_proj.scales": "model-00019-of-00047.safetensors", + "model.layers.24.block_sparse_moe.switch_mlp.up_proj.weight": "model-00019-of-00047.safetensors", + "model.layers.24.input_layernorm.weight": "model-00019-of-00047.safetensors", + "model.layers.24.post_attention_layernorm.weight": "model-00019-of-00047.safetensors", + "model.layers.24.self_attn.k_norm.weight": "model-00018-of-00047.safetensors", + "model.layers.24.self_attn.k_proj.biases": "model-00018-of-00047.safetensors", + "model.layers.24.self_attn.k_proj.scales": "model-00018-of-00047.safetensors", + "model.layers.24.self_attn.k_proj.weight": "model-00018-of-00047.safetensors", + "model.layers.24.self_attn.o_proj.biases": "model-00018-of-00047.safetensors", + "model.layers.24.self_attn.o_proj.scales": "model-00018-of-00047.safetensors", + "model.layers.24.self_attn.o_proj.weight": "model-00018-of-00047.safetensors", + "model.layers.24.self_attn.q_norm.weight": "model-00018-of-00047.safetensors", + "model.layers.24.self_attn.q_proj.biases": "model-00018-of-00047.safetensors", + "model.layers.24.self_attn.q_proj.scales": "model-00018-of-00047.safetensors", + "model.layers.24.self_attn.q_proj.weight": "model-00018-of-00047.safetensors", + "model.layers.24.self_attn.v_proj.biases": "model-00018-of-00047.safetensors", + "model.layers.24.self_attn.v_proj.scales": "model-00018-of-00047.safetensors", + "model.layers.24.self_attn.v_proj.weight": "model-00018-of-00047.safetensors", + "model.layers.25.block_sparse_moe.e_score_correction_bias": "model-00020-of-00047.safetensors", + "model.layers.25.block_sparse_moe.gate.biases": "model-00019-of-00047.safetensors", + "model.layers.25.block_sparse_moe.gate.scales": "model-00019-of-00047.safetensors", + "model.layers.25.block_sparse_moe.gate.weight": "model-00019-of-00047.safetensors", + "model.layers.25.block_sparse_moe.switch_mlp.down_proj.biases": "model-00020-of-00047.safetensors", + "model.layers.25.block_sparse_moe.switch_mlp.down_proj.scales": "model-00020-of-00047.safetensors", + "model.layers.25.block_sparse_moe.switch_mlp.down_proj.weight": "model-00020-of-00047.safetensors", + "model.layers.25.block_sparse_moe.switch_mlp.gate_proj.biases": "model-00019-of-00047.safetensors", + "model.layers.25.block_sparse_moe.switch_mlp.gate_proj.scales": "model-00019-of-00047.safetensors", + "model.layers.25.block_sparse_moe.switch_mlp.gate_proj.weight": "model-00019-of-00047.safetensors", + "model.layers.25.block_sparse_moe.switch_mlp.up_proj.biases": "model-00020-of-00047.safetensors", + "model.layers.25.block_sparse_moe.switch_mlp.up_proj.scales": "model-00020-of-00047.safetensors", + "model.layers.25.block_sparse_moe.switch_mlp.up_proj.weight": "model-00020-of-00047.safetensors", + "model.layers.25.input_layernorm.weight": "model-00020-of-00047.safetensors", + "model.layers.25.post_attention_layernorm.weight": "model-00020-of-00047.safetensors", + "model.layers.25.self_attn.k_norm.weight": "model-00019-of-00047.safetensors", + "model.layers.25.self_attn.k_proj.biases": "model-00019-of-00047.safetensors", + "model.layers.25.self_attn.k_proj.scales": "model-00019-of-00047.safetensors", + "model.layers.25.self_attn.k_proj.weight": "model-00019-of-00047.safetensors", + "model.layers.25.self_attn.o_proj.biases": "model-00019-of-00047.safetensors", + "model.layers.25.self_attn.o_proj.scales": "model-00019-of-00047.safetensors", + "model.layers.25.self_attn.o_proj.weight": "model-00019-of-00047.safetensors", + "model.layers.25.self_attn.q_norm.weight": "model-00019-of-00047.safetensors", + "model.layers.25.self_attn.q_proj.biases": "model-00019-of-00047.safetensors", + "model.layers.25.self_attn.q_proj.scales": "model-00019-of-00047.safetensors", + "model.layers.25.self_attn.q_proj.weight": "model-00019-of-00047.safetensors", + "model.layers.25.self_attn.v_proj.biases": "model-00019-of-00047.safetensors", + "model.layers.25.self_attn.v_proj.scales": "model-00019-of-00047.safetensors", + "model.layers.25.self_attn.v_proj.weight": "model-00019-of-00047.safetensors", + "model.layers.26.block_sparse_moe.e_score_correction_bias": "model-00021-of-00047.safetensors", + "model.layers.26.block_sparse_moe.gate.biases": "model-00020-of-00047.safetensors", + "model.layers.26.block_sparse_moe.gate.scales": "model-00020-of-00047.safetensors", + "model.layers.26.block_sparse_moe.gate.weight": "model-00020-of-00047.safetensors", + "model.layers.26.block_sparse_moe.switch_mlp.down_proj.biases": "model-00021-of-00047.safetensors", + "model.layers.26.block_sparse_moe.switch_mlp.down_proj.scales": "model-00021-of-00047.safetensors", + "model.layers.26.block_sparse_moe.switch_mlp.down_proj.weight": "model-00021-of-00047.safetensors", + "model.layers.26.block_sparse_moe.switch_mlp.gate_proj.biases": "model-00020-of-00047.safetensors", + "model.layers.26.block_sparse_moe.switch_mlp.gate_proj.scales": "model-00020-of-00047.safetensors", + "model.layers.26.block_sparse_moe.switch_mlp.gate_proj.weight": "model-00020-of-00047.safetensors", + "model.layers.26.block_sparse_moe.switch_mlp.up_proj.biases": "model-00020-of-00047.safetensors", + "model.layers.26.block_sparse_moe.switch_mlp.up_proj.scales": "model-00020-of-00047.safetensors", + "model.layers.26.block_sparse_moe.switch_mlp.up_proj.weight": "model-00020-of-00047.safetensors", + "model.layers.26.input_layernorm.weight": "model-00021-of-00047.safetensors", + "model.layers.26.post_attention_layernorm.weight": "model-00021-of-00047.safetensors", + "model.layers.26.self_attn.k_norm.weight": "model-00020-of-00047.safetensors", + "model.layers.26.self_attn.k_proj.biases": "model-00020-of-00047.safetensors", + "model.layers.26.self_attn.k_proj.scales": "model-00020-of-00047.safetensors", + "model.layers.26.self_attn.k_proj.weight": "model-00020-of-00047.safetensors", + "model.layers.26.self_attn.o_proj.biases": "model-00020-of-00047.safetensors", + "model.layers.26.self_attn.o_proj.scales": "model-00020-of-00047.safetensors", + "model.layers.26.self_attn.o_proj.weight": "model-00020-of-00047.safetensors", + "model.layers.26.self_attn.q_norm.weight": "model-00020-of-00047.safetensors", + "model.layers.26.self_attn.q_proj.biases": "model-00020-of-00047.safetensors", + "model.layers.26.self_attn.q_proj.scales": "model-00020-of-00047.safetensors", + "model.layers.26.self_attn.q_proj.weight": "model-00020-of-00047.safetensors", + "model.layers.26.self_attn.v_proj.biases": "model-00020-of-00047.safetensors", + "model.layers.26.self_attn.v_proj.scales": "model-00020-of-00047.safetensors", + "model.layers.26.self_attn.v_proj.weight": "model-00020-of-00047.safetensors", + "model.layers.27.block_sparse_moe.e_score_correction_bias": "model-00021-of-00047.safetensors", + "model.layers.27.block_sparse_moe.gate.biases": "model-00021-of-00047.safetensors", + "model.layers.27.block_sparse_moe.gate.scales": "model-00021-of-00047.safetensors", + "model.layers.27.block_sparse_moe.gate.weight": "model-00021-of-00047.safetensors", + "model.layers.27.block_sparse_moe.switch_mlp.down_proj.biases": "model-00021-of-00047.safetensors", + "model.layers.27.block_sparse_moe.switch_mlp.down_proj.scales": "model-00021-of-00047.safetensors", + "model.layers.27.block_sparse_moe.switch_mlp.down_proj.weight": "model-00021-of-00047.safetensors", + "model.layers.27.block_sparse_moe.switch_mlp.gate_proj.biases": "model-00021-of-00047.safetensors", + "model.layers.27.block_sparse_moe.switch_mlp.gate_proj.scales": "model-00021-of-00047.safetensors", + "model.layers.27.block_sparse_moe.switch_mlp.gate_proj.weight": "model-00021-of-00047.safetensors", + "model.layers.27.block_sparse_moe.switch_mlp.up_proj.biases": "model-00021-of-00047.safetensors", + "model.layers.27.block_sparse_moe.switch_mlp.up_proj.scales": "model-00021-of-00047.safetensors", + "model.layers.27.block_sparse_moe.switch_mlp.up_proj.weight": "model-00021-of-00047.safetensors", + "model.layers.27.input_layernorm.weight": "model-00021-of-00047.safetensors", + "model.layers.27.post_attention_layernorm.weight": "model-00021-of-00047.safetensors", + "model.layers.27.self_attn.k_norm.weight": "model-00021-of-00047.safetensors", + "model.layers.27.self_attn.k_proj.biases": "model-00021-of-00047.safetensors", + "model.layers.27.self_attn.k_proj.scales": "model-00021-of-00047.safetensors", + "model.layers.27.self_attn.k_proj.weight": "model-00021-of-00047.safetensors", + "model.layers.27.self_attn.o_proj.biases": "model-00021-of-00047.safetensors", + "model.layers.27.self_attn.o_proj.scales": "model-00021-of-00047.safetensors", + "model.layers.27.self_attn.o_proj.weight": "model-00021-of-00047.safetensors", + "model.layers.27.self_attn.q_norm.weight": "model-00021-of-00047.safetensors", + "model.layers.27.self_attn.q_proj.biases": "model-00021-of-00047.safetensors", + "model.layers.27.self_attn.q_proj.scales": "model-00021-of-00047.safetensors", + "model.layers.27.self_attn.q_proj.weight": "model-00021-of-00047.safetensors", + "model.layers.27.self_attn.v_proj.biases": "model-00021-of-00047.safetensors", + "model.layers.27.self_attn.v_proj.scales": "model-00021-of-00047.safetensors", + "model.layers.27.self_attn.v_proj.weight": "model-00021-of-00047.safetensors", + "model.layers.28.block_sparse_moe.e_score_correction_bias": "model-00022-of-00047.safetensors", + "model.layers.28.block_sparse_moe.gate.biases": "model-00021-of-00047.safetensors", + "model.layers.28.block_sparse_moe.gate.scales": "model-00021-of-00047.safetensors", + "model.layers.28.block_sparse_moe.gate.weight": "model-00021-of-00047.safetensors", + "model.layers.28.block_sparse_moe.switch_mlp.down_proj.biases": "model-00022-of-00047.safetensors", + "model.layers.28.block_sparse_moe.switch_mlp.down_proj.scales": "model-00022-of-00047.safetensors", + "model.layers.28.block_sparse_moe.switch_mlp.down_proj.weight": "model-00022-of-00047.safetensors", + "model.layers.28.block_sparse_moe.switch_mlp.gate_proj.biases": "model-00022-of-00047.safetensors", + "model.layers.28.block_sparse_moe.switch_mlp.gate_proj.scales": "model-00022-of-00047.safetensors", + "model.layers.28.block_sparse_moe.switch_mlp.gate_proj.weight": "model-00022-of-00047.safetensors", + "model.layers.28.block_sparse_moe.switch_mlp.up_proj.biases": "model-00022-of-00047.safetensors", + "model.layers.28.block_sparse_moe.switch_mlp.up_proj.scales": "model-00022-of-00047.safetensors", + "model.layers.28.block_sparse_moe.switch_mlp.up_proj.weight": "model-00022-of-00047.safetensors", + "model.layers.28.input_layernorm.weight": "model-00022-of-00047.safetensors", + "model.layers.28.post_attention_layernorm.weight": "model-00022-of-00047.safetensors", + "model.layers.28.self_attn.k_norm.weight": "model-00021-of-00047.safetensors", + "model.layers.28.self_attn.k_proj.biases": "model-00021-of-00047.safetensors", + "model.layers.28.self_attn.k_proj.scales": "model-00021-of-00047.safetensors", + "model.layers.28.self_attn.k_proj.weight": "model-00021-of-00047.safetensors", + "model.layers.28.self_attn.o_proj.biases": "model-00021-of-00047.safetensors", + "model.layers.28.self_attn.o_proj.scales": "model-00021-of-00047.safetensors", + "model.layers.28.self_attn.o_proj.weight": "model-00021-of-00047.safetensors", + "model.layers.28.self_attn.q_norm.weight": "model-00021-of-00047.safetensors", + "model.layers.28.self_attn.q_proj.biases": "model-00021-of-00047.safetensors", + "model.layers.28.self_attn.q_proj.scales": "model-00021-of-00047.safetensors", + "model.layers.28.self_attn.q_proj.weight": "model-00021-of-00047.safetensors", + "model.layers.28.self_attn.v_proj.biases": "model-00021-of-00047.safetensors", + "model.layers.28.self_attn.v_proj.scales": "model-00021-of-00047.safetensors", + "model.layers.28.self_attn.v_proj.weight": "model-00021-of-00047.safetensors", + "model.layers.29.block_sparse_moe.e_score_correction_bias": "model-00023-of-00047.safetensors", + "model.layers.29.block_sparse_moe.gate.biases": "model-00022-of-00047.safetensors", + "model.layers.29.block_sparse_moe.gate.scales": "model-00022-of-00047.safetensors", + "model.layers.29.block_sparse_moe.gate.weight": "model-00022-of-00047.safetensors", + "model.layers.29.block_sparse_moe.switch_mlp.down_proj.biases": "model-00023-of-00047.safetensors", + "model.layers.29.block_sparse_moe.switch_mlp.down_proj.scales": "model-00023-of-00047.safetensors", + "model.layers.29.block_sparse_moe.switch_mlp.down_proj.weight": "model-00023-of-00047.safetensors", + "model.layers.29.block_sparse_moe.switch_mlp.gate_proj.biases": "model-00022-of-00047.safetensors", + "model.layers.29.block_sparse_moe.switch_mlp.gate_proj.scales": "model-00022-of-00047.safetensors", + "model.layers.29.block_sparse_moe.switch_mlp.gate_proj.weight": "model-00022-of-00047.safetensors", + "model.layers.29.block_sparse_moe.switch_mlp.up_proj.biases": "model-00023-of-00047.safetensors", + "model.layers.29.block_sparse_moe.switch_mlp.up_proj.scales": "model-00023-of-00047.safetensors", + "model.layers.29.block_sparse_moe.switch_mlp.up_proj.weight": "model-00023-of-00047.safetensors", + "model.layers.29.input_layernorm.weight": "model-00023-of-00047.safetensors", + "model.layers.29.post_attention_layernorm.weight": "model-00023-of-00047.safetensors", + "model.layers.29.self_attn.k_norm.weight": "model-00022-of-00047.safetensors", + "model.layers.29.self_attn.k_proj.biases": "model-00022-of-00047.safetensors", + "model.layers.29.self_attn.k_proj.scales": "model-00022-of-00047.safetensors", + "model.layers.29.self_attn.k_proj.weight": "model-00022-of-00047.safetensors", + "model.layers.29.self_attn.o_proj.biases": "model-00022-of-00047.safetensors", + "model.layers.29.self_attn.o_proj.scales": "model-00022-of-00047.safetensors", + "model.layers.29.self_attn.o_proj.weight": "model-00022-of-00047.safetensors", + "model.layers.29.self_attn.q_norm.weight": "model-00022-of-00047.safetensors", + "model.layers.29.self_attn.q_proj.biases": "model-00022-of-00047.safetensors", + "model.layers.29.self_attn.q_proj.scales": "model-00022-of-00047.safetensors", + "model.layers.29.self_attn.q_proj.weight": "model-00022-of-00047.safetensors", + "model.layers.29.self_attn.v_proj.biases": "model-00022-of-00047.safetensors", + "model.layers.29.self_attn.v_proj.scales": "model-00022-of-00047.safetensors", + "model.layers.29.self_attn.v_proj.weight": "model-00022-of-00047.safetensors", + "model.layers.3.block_sparse_moe.e_score_correction_bias": "model-00003-of-00047.safetensors", + "model.layers.3.block_sparse_moe.gate.biases": "model-00003-of-00047.safetensors", + "model.layers.3.block_sparse_moe.gate.scales": "model-00003-of-00047.safetensors", + "model.layers.3.block_sparse_moe.gate.weight": "model-00003-of-00047.safetensors", + "model.layers.3.block_sparse_moe.switch_mlp.down_proj.biases": "model-00003-of-00047.safetensors", + "model.layers.3.block_sparse_moe.switch_mlp.down_proj.scales": "model-00003-of-00047.safetensors", + "model.layers.3.block_sparse_moe.switch_mlp.down_proj.weight": "model-00003-of-00047.safetensors", + "model.layers.3.block_sparse_moe.switch_mlp.gate_proj.biases": "model-00003-of-00047.safetensors", + "model.layers.3.block_sparse_moe.switch_mlp.gate_proj.scales": "model-00003-of-00047.safetensors", + "model.layers.3.block_sparse_moe.switch_mlp.gate_proj.weight": "model-00003-of-00047.safetensors", + "model.layers.3.block_sparse_moe.switch_mlp.up_proj.biases": "model-00003-of-00047.safetensors", + "model.layers.3.block_sparse_moe.switch_mlp.up_proj.scales": "model-00003-of-00047.safetensors", + "model.layers.3.block_sparse_moe.switch_mlp.up_proj.weight": "model-00003-of-00047.safetensors", + "model.layers.3.input_layernorm.weight": "model-00003-of-00047.safetensors", + "model.layers.3.post_attention_layernorm.weight": "model-00003-of-00047.safetensors", + "model.layers.3.self_attn.k_norm.weight": "model-00003-of-00047.safetensors", + "model.layers.3.self_attn.k_proj.biases": "model-00003-of-00047.safetensors", + "model.layers.3.self_attn.k_proj.scales": "model-00003-of-00047.safetensors", + "model.layers.3.self_attn.k_proj.weight": "model-00003-of-00047.safetensors", + "model.layers.3.self_attn.o_proj.biases": "model-00003-of-00047.safetensors", + "model.layers.3.self_attn.o_proj.scales": "model-00003-of-00047.safetensors", + "model.layers.3.self_attn.o_proj.weight": "model-00003-of-00047.safetensors", + "model.layers.3.self_attn.q_norm.weight": "model-00003-of-00047.safetensors", + "model.layers.3.self_attn.q_proj.biases": "model-00003-of-00047.safetensors", + "model.layers.3.self_attn.q_proj.scales": "model-00003-of-00047.safetensors", + "model.layers.3.self_attn.q_proj.weight": "model-00003-of-00047.safetensors", + "model.layers.3.self_attn.v_proj.biases": "model-00003-of-00047.safetensors", + "model.layers.3.self_attn.v_proj.scales": "model-00003-of-00047.safetensors", + "model.layers.3.self_attn.v_proj.weight": "model-00003-of-00047.safetensors", + "model.layers.30.block_sparse_moe.e_score_correction_bias": "model-00024-of-00047.safetensors", + "model.layers.30.block_sparse_moe.gate.biases": "model-00023-of-00047.safetensors", + "model.layers.30.block_sparse_moe.gate.scales": "model-00023-of-00047.safetensors", + "model.layers.30.block_sparse_moe.gate.weight": "model-00023-of-00047.safetensors", + "model.layers.30.block_sparse_moe.switch_mlp.down_proj.biases": "model-00024-of-00047.safetensors", + "model.layers.30.block_sparse_moe.switch_mlp.down_proj.scales": "model-00024-of-00047.safetensors", + "model.layers.30.block_sparse_moe.switch_mlp.down_proj.weight": "model-00024-of-00047.safetensors", + "model.layers.30.block_sparse_moe.switch_mlp.gate_proj.biases": "model-00023-of-00047.safetensors", + "model.layers.30.block_sparse_moe.switch_mlp.gate_proj.scales": "model-00023-of-00047.safetensors", + "model.layers.30.block_sparse_moe.switch_mlp.gate_proj.weight": "model-00023-of-00047.safetensors", + "model.layers.30.block_sparse_moe.switch_mlp.up_proj.biases": "model-00023-of-00047.safetensors", + "model.layers.30.block_sparse_moe.switch_mlp.up_proj.scales": "model-00023-of-00047.safetensors", + "model.layers.30.block_sparse_moe.switch_mlp.up_proj.weight": "model-00023-of-00047.safetensors", + "model.layers.30.input_layernorm.weight": "model-00024-of-00047.safetensors", + "model.layers.30.post_attention_layernorm.weight": "model-00024-of-00047.safetensors", + "model.layers.30.self_attn.k_norm.weight": "model-00023-of-00047.safetensors", + "model.layers.30.self_attn.k_proj.biases": "model-00023-of-00047.safetensors", + "model.layers.30.self_attn.k_proj.scales": "model-00023-of-00047.safetensors", + "model.layers.30.self_attn.k_proj.weight": "model-00023-of-00047.safetensors", + "model.layers.30.self_attn.o_proj.biases": "model-00023-of-00047.safetensors", + "model.layers.30.self_attn.o_proj.scales": "model-00023-of-00047.safetensors", + "model.layers.30.self_attn.o_proj.weight": "model-00023-of-00047.safetensors", + "model.layers.30.self_attn.q_norm.weight": "model-00023-of-00047.safetensors", + "model.layers.30.self_attn.q_proj.biases": "model-00023-of-00047.safetensors", + "model.layers.30.self_attn.q_proj.scales": "model-00023-of-00047.safetensors", + "model.layers.30.self_attn.q_proj.weight": "model-00023-of-00047.safetensors", + "model.layers.30.self_attn.v_proj.biases": "model-00023-of-00047.safetensors", + "model.layers.30.self_attn.v_proj.scales": "model-00023-of-00047.safetensors", + "model.layers.30.self_attn.v_proj.weight": "model-00023-of-00047.safetensors", + "model.layers.31.block_sparse_moe.e_score_correction_bias": "model-00024-of-00047.safetensors", + "model.layers.31.block_sparse_moe.gate.biases": "model-00024-of-00047.safetensors", + "model.layers.31.block_sparse_moe.gate.scales": "model-00024-of-00047.safetensors", + "model.layers.31.block_sparse_moe.gate.weight": "model-00024-of-00047.safetensors", + "model.layers.31.block_sparse_moe.switch_mlp.down_proj.biases": "model-00024-of-00047.safetensors", + "model.layers.31.block_sparse_moe.switch_mlp.down_proj.scales": "model-00024-of-00047.safetensors", + "model.layers.31.block_sparse_moe.switch_mlp.down_proj.weight": "model-00024-of-00047.safetensors", + "model.layers.31.block_sparse_moe.switch_mlp.gate_proj.biases": "model-00024-of-00047.safetensors", + "model.layers.31.block_sparse_moe.switch_mlp.gate_proj.scales": "model-00024-of-00047.safetensors", + "model.layers.31.block_sparse_moe.switch_mlp.gate_proj.weight": "model-00024-of-00047.safetensors", + "model.layers.31.block_sparse_moe.switch_mlp.up_proj.biases": "model-00024-of-00047.safetensors", + "model.layers.31.block_sparse_moe.switch_mlp.up_proj.scales": "model-00024-of-00047.safetensors", + "model.layers.31.block_sparse_moe.switch_mlp.up_proj.weight": "model-00024-of-00047.safetensors", + "model.layers.31.input_layernorm.weight": "model-00024-of-00047.safetensors", + "model.layers.31.post_attention_layernorm.weight": "model-00024-of-00047.safetensors", + "model.layers.31.self_attn.k_norm.weight": "model-00024-of-00047.safetensors", + "model.layers.31.self_attn.k_proj.biases": "model-00024-of-00047.safetensors", + "model.layers.31.self_attn.k_proj.scales": "model-00024-of-00047.safetensors", + "model.layers.31.self_attn.k_proj.weight": "model-00024-of-00047.safetensors", + "model.layers.31.self_attn.o_proj.biases": "model-00024-of-00047.safetensors", + "model.layers.31.self_attn.o_proj.scales": "model-00024-of-00047.safetensors", + "model.layers.31.self_attn.o_proj.weight": "model-00024-of-00047.safetensors", + "model.layers.31.self_attn.q_norm.weight": "model-00024-of-00047.safetensors", + "model.layers.31.self_attn.q_proj.biases": "model-00024-of-00047.safetensors", + "model.layers.31.self_attn.q_proj.scales": "model-00024-of-00047.safetensors", + "model.layers.31.self_attn.q_proj.weight": "model-00024-of-00047.safetensors", + "model.layers.31.self_attn.v_proj.biases": "model-00024-of-00047.safetensors", + "model.layers.31.self_attn.v_proj.scales": "model-00024-of-00047.safetensors", + "model.layers.31.self_attn.v_proj.weight": "model-00024-of-00047.safetensors", + "model.layers.32.block_sparse_moe.e_score_correction_bias": "model-00025-of-00047.safetensors", + "model.layers.32.block_sparse_moe.gate.biases": "model-00024-of-00047.safetensors", + "model.layers.32.block_sparse_moe.gate.scales": "model-00024-of-00047.safetensors", + "model.layers.32.block_sparse_moe.gate.weight": "model-00024-of-00047.safetensors", + "model.layers.32.block_sparse_moe.switch_mlp.down_proj.biases": "model-00025-of-00047.safetensors", + "model.layers.32.block_sparse_moe.switch_mlp.down_proj.scales": "model-00025-of-00047.safetensors", + "model.layers.32.block_sparse_moe.switch_mlp.down_proj.weight": "model-00025-of-00047.safetensors", + "model.layers.32.block_sparse_moe.switch_mlp.gate_proj.biases": "model-00025-of-00047.safetensors", + "model.layers.32.block_sparse_moe.switch_mlp.gate_proj.scales": "model-00025-of-00047.safetensors", + "model.layers.32.block_sparse_moe.switch_mlp.gate_proj.weight": "model-00025-of-00047.safetensors", + "model.layers.32.block_sparse_moe.switch_mlp.up_proj.biases": "model-00025-of-00047.safetensors", + "model.layers.32.block_sparse_moe.switch_mlp.up_proj.scales": "model-00025-of-00047.safetensors", + "model.layers.32.block_sparse_moe.switch_mlp.up_proj.weight": "model-00025-of-00047.safetensors", + "model.layers.32.input_layernorm.weight": "model-00025-of-00047.safetensors", + "model.layers.32.post_attention_layernorm.weight": "model-00025-of-00047.safetensors", + "model.layers.32.self_attn.k_norm.weight": "model-00024-of-00047.safetensors", + "model.layers.32.self_attn.k_proj.biases": "model-00024-of-00047.safetensors", + "model.layers.32.self_attn.k_proj.scales": "model-00024-of-00047.safetensors", + "model.layers.32.self_attn.k_proj.weight": "model-00024-of-00047.safetensors", + "model.layers.32.self_attn.o_proj.biases": "model-00024-of-00047.safetensors", + "model.layers.32.self_attn.o_proj.scales": "model-00024-of-00047.safetensors", + "model.layers.32.self_attn.o_proj.weight": "model-00024-of-00047.safetensors", + "model.layers.32.self_attn.q_norm.weight": "model-00024-of-00047.safetensors", + "model.layers.32.self_attn.q_proj.biases": "model-00024-of-00047.safetensors", + "model.layers.32.self_attn.q_proj.scales": "model-00024-of-00047.safetensors", + "model.layers.32.self_attn.q_proj.weight": "model-00024-of-00047.safetensors", + "model.layers.32.self_attn.v_proj.biases": "model-00024-of-00047.safetensors", + "model.layers.32.self_attn.v_proj.scales": "model-00024-of-00047.safetensors", + "model.layers.32.self_attn.v_proj.weight": "model-00024-of-00047.safetensors", + "model.layers.33.block_sparse_moe.e_score_correction_bias": "model-00026-of-00047.safetensors", + "model.layers.33.block_sparse_moe.gate.biases": "model-00025-of-00047.safetensors", + "model.layers.33.block_sparse_moe.gate.scales": "model-00025-of-00047.safetensors", + "model.layers.33.block_sparse_moe.gate.weight": "model-00025-of-00047.safetensors", + "model.layers.33.block_sparse_moe.switch_mlp.down_proj.biases": "model-00026-of-00047.safetensors", + "model.layers.33.block_sparse_moe.switch_mlp.down_proj.scales": "model-00026-of-00047.safetensors", + "model.layers.33.block_sparse_moe.switch_mlp.down_proj.weight": "model-00026-of-00047.safetensors", + "model.layers.33.block_sparse_moe.switch_mlp.gate_proj.biases": "model-00025-of-00047.safetensors", + "model.layers.33.block_sparse_moe.switch_mlp.gate_proj.scales": "model-00025-of-00047.safetensors", + "model.layers.33.block_sparse_moe.switch_mlp.gate_proj.weight": "model-00025-of-00047.safetensors", + "model.layers.33.block_sparse_moe.switch_mlp.up_proj.biases": "model-00026-of-00047.safetensors", + "model.layers.33.block_sparse_moe.switch_mlp.up_proj.scales": "model-00026-of-00047.safetensors", + "model.layers.33.block_sparse_moe.switch_mlp.up_proj.weight": "model-00026-of-00047.safetensors", + "model.layers.33.input_layernorm.weight": "model-00026-of-00047.safetensors", + "model.layers.33.post_attention_layernorm.weight": "model-00026-of-00047.safetensors", + "model.layers.33.self_attn.k_norm.weight": "model-00025-of-00047.safetensors", + "model.layers.33.self_attn.k_proj.biases": "model-00025-of-00047.safetensors", + "model.layers.33.self_attn.k_proj.scales": "model-00025-of-00047.safetensors", + "model.layers.33.self_attn.k_proj.weight": "model-00025-of-00047.safetensors", + "model.layers.33.self_attn.o_proj.biases": "model-00025-of-00047.safetensors", + "model.layers.33.self_attn.o_proj.scales": "model-00025-of-00047.safetensors", + "model.layers.33.self_attn.o_proj.weight": "model-00025-of-00047.safetensors", + "model.layers.33.self_attn.q_norm.weight": "model-00025-of-00047.safetensors", + "model.layers.33.self_attn.q_proj.biases": "model-00025-of-00047.safetensors", + "model.layers.33.self_attn.q_proj.scales": "model-00025-of-00047.safetensors", + "model.layers.33.self_attn.q_proj.weight": "model-00025-of-00047.safetensors", + "model.layers.33.self_attn.v_proj.biases": "model-00025-of-00047.safetensors", + "model.layers.33.self_attn.v_proj.scales": "model-00025-of-00047.safetensors", + "model.layers.33.self_attn.v_proj.weight": "model-00025-of-00047.safetensors", + "model.layers.34.block_sparse_moe.e_score_correction_bias": "model-00027-of-00047.safetensors", + "model.layers.34.block_sparse_moe.gate.biases": "model-00026-of-00047.safetensors", + "model.layers.34.block_sparse_moe.gate.scales": "model-00026-of-00047.safetensors", + "model.layers.34.block_sparse_moe.gate.weight": "model-00026-of-00047.safetensors", + "model.layers.34.block_sparse_moe.switch_mlp.down_proj.biases": "model-00027-of-00047.safetensors", + "model.layers.34.block_sparse_moe.switch_mlp.down_proj.scales": "model-00027-of-00047.safetensors", + "model.layers.34.block_sparse_moe.switch_mlp.down_proj.weight": "model-00027-of-00047.safetensors", + "model.layers.34.block_sparse_moe.switch_mlp.gate_proj.biases": "model-00026-of-00047.safetensors", + "model.layers.34.block_sparse_moe.switch_mlp.gate_proj.scales": "model-00026-of-00047.safetensors", + "model.layers.34.block_sparse_moe.switch_mlp.gate_proj.weight": "model-00026-of-00047.safetensors", + "model.layers.34.block_sparse_moe.switch_mlp.up_proj.biases": "model-00026-of-00047.safetensors", + "model.layers.34.block_sparse_moe.switch_mlp.up_proj.scales": "model-00026-of-00047.safetensors", + "model.layers.34.block_sparse_moe.switch_mlp.up_proj.weight": "model-00026-of-00047.safetensors", + "model.layers.34.input_layernorm.weight": "model-00027-of-00047.safetensors", + "model.layers.34.post_attention_layernorm.weight": "model-00027-of-00047.safetensors", + "model.layers.34.self_attn.k_norm.weight": "model-00026-of-00047.safetensors", + "model.layers.34.self_attn.k_proj.biases": "model-00026-of-00047.safetensors", + "model.layers.34.self_attn.k_proj.scales": "model-00026-of-00047.safetensors", + "model.layers.34.self_attn.k_proj.weight": "model-00026-of-00047.safetensors", + "model.layers.34.self_attn.o_proj.biases": "model-00026-of-00047.safetensors", + "model.layers.34.self_attn.o_proj.scales": "model-00026-of-00047.safetensors", + "model.layers.34.self_attn.o_proj.weight": "model-00026-of-00047.safetensors", + "model.layers.34.self_attn.q_norm.weight": "model-00026-of-00047.safetensors", + "model.layers.34.self_attn.q_proj.biases": "model-00026-of-00047.safetensors", + "model.layers.34.self_attn.q_proj.scales": "model-00026-of-00047.safetensors", + "model.layers.34.self_attn.q_proj.weight": "model-00026-of-00047.safetensors", + "model.layers.34.self_attn.v_proj.biases": "model-00026-of-00047.safetensors", + "model.layers.34.self_attn.v_proj.scales": "model-00026-of-00047.safetensors", + "model.layers.34.self_attn.v_proj.weight": "model-00026-of-00047.safetensors", + "model.layers.35.block_sparse_moe.e_score_correction_bias": "model-00027-of-00047.safetensors", + "model.layers.35.block_sparse_moe.gate.biases": "model-00027-of-00047.safetensors", + "model.layers.35.block_sparse_moe.gate.scales": "model-00027-of-00047.safetensors", + "model.layers.35.block_sparse_moe.gate.weight": "model-00027-of-00047.safetensors", + "model.layers.35.block_sparse_moe.switch_mlp.down_proj.biases": "model-00027-of-00047.safetensors", + "model.layers.35.block_sparse_moe.switch_mlp.down_proj.scales": "model-00027-of-00047.safetensors", + "model.layers.35.block_sparse_moe.switch_mlp.down_proj.weight": "model-00027-of-00047.safetensors", + "model.layers.35.block_sparse_moe.switch_mlp.gate_proj.biases": "model-00027-of-00047.safetensors", + "model.layers.35.block_sparse_moe.switch_mlp.gate_proj.scales": "model-00027-of-00047.safetensors", + "model.layers.35.block_sparse_moe.switch_mlp.gate_proj.weight": "model-00027-of-00047.safetensors", + "model.layers.35.block_sparse_moe.switch_mlp.up_proj.biases": "model-00027-of-00047.safetensors", + "model.layers.35.block_sparse_moe.switch_mlp.up_proj.scales": "model-00027-of-00047.safetensors", + "model.layers.35.block_sparse_moe.switch_mlp.up_proj.weight": "model-00027-of-00047.safetensors", + "model.layers.35.input_layernorm.weight": "model-00027-of-00047.safetensors", + "model.layers.35.post_attention_layernorm.weight": "model-00027-of-00047.safetensors", + "model.layers.35.self_attn.k_norm.weight": "model-00027-of-00047.safetensors", + "model.layers.35.self_attn.k_proj.biases": "model-00027-of-00047.safetensors", + "model.layers.35.self_attn.k_proj.scales": "model-00027-of-00047.safetensors", + "model.layers.35.self_attn.k_proj.weight": "model-00027-of-00047.safetensors", + "model.layers.35.self_attn.o_proj.biases": "model-00027-of-00047.safetensors", + "model.layers.35.self_attn.o_proj.scales": "model-00027-of-00047.safetensors", + "model.layers.35.self_attn.o_proj.weight": "model-00027-of-00047.safetensors", + "model.layers.35.self_attn.q_norm.weight": "model-00027-of-00047.safetensors", + "model.layers.35.self_attn.q_proj.biases": "model-00027-of-00047.safetensors", + "model.layers.35.self_attn.q_proj.scales": "model-00027-of-00047.safetensors", + "model.layers.35.self_attn.q_proj.weight": "model-00027-of-00047.safetensors", + "model.layers.35.self_attn.v_proj.biases": "model-00027-of-00047.safetensors", + "model.layers.35.self_attn.v_proj.scales": "model-00027-of-00047.safetensors", + "model.layers.35.self_attn.v_proj.weight": "model-00027-of-00047.safetensors", + "model.layers.36.block_sparse_moe.e_score_correction_bias": "model-00028-of-00047.safetensors", + "model.layers.36.block_sparse_moe.gate.biases": "model-00027-of-00047.safetensors", + "model.layers.36.block_sparse_moe.gate.scales": "model-00027-of-00047.safetensors", + "model.layers.36.block_sparse_moe.gate.weight": "model-00027-of-00047.safetensors", + "model.layers.36.block_sparse_moe.switch_mlp.down_proj.biases": "model-00028-of-00047.safetensors", + "model.layers.36.block_sparse_moe.switch_mlp.down_proj.scales": "model-00028-of-00047.safetensors", + "model.layers.36.block_sparse_moe.switch_mlp.down_proj.weight": "model-00028-of-00047.safetensors", + "model.layers.36.block_sparse_moe.switch_mlp.gate_proj.biases": "model-00028-of-00047.safetensors", + "model.layers.36.block_sparse_moe.switch_mlp.gate_proj.scales": "model-00028-of-00047.safetensors", + "model.layers.36.block_sparse_moe.switch_mlp.gate_proj.weight": "model-00028-of-00047.safetensors", + "model.layers.36.block_sparse_moe.switch_mlp.up_proj.biases": "model-00028-of-00047.safetensors", + "model.layers.36.block_sparse_moe.switch_mlp.up_proj.scales": "model-00028-of-00047.safetensors", + "model.layers.36.block_sparse_moe.switch_mlp.up_proj.weight": "model-00028-of-00047.safetensors", + "model.layers.36.input_layernorm.weight": "model-00028-of-00047.safetensors", + "model.layers.36.post_attention_layernorm.weight": "model-00028-of-00047.safetensors", + "model.layers.36.self_attn.k_norm.weight": "model-00027-of-00047.safetensors", + "model.layers.36.self_attn.k_proj.biases": "model-00027-of-00047.safetensors", + "model.layers.36.self_attn.k_proj.scales": "model-00027-of-00047.safetensors", + "model.layers.36.self_attn.k_proj.weight": "model-00027-of-00047.safetensors", + "model.layers.36.self_attn.o_proj.biases": "model-00027-of-00047.safetensors", + "model.layers.36.self_attn.o_proj.scales": "model-00027-of-00047.safetensors", + "model.layers.36.self_attn.o_proj.weight": "model-00027-of-00047.safetensors", + "model.layers.36.self_attn.q_norm.weight": "model-00027-of-00047.safetensors", + "model.layers.36.self_attn.q_proj.biases": "model-00027-of-00047.safetensors", + "model.layers.36.self_attn.q_proj.scales": "model-00027-of-00047.safetensors", + "model.layers.36.self_attn.q_proj.weight": "model-00027-of-00047.safetensors", + "model.layers.36.self_attn.v_proj.biases": "model-00027-of-00047.safetensors", + "model.layers.36.self_attn.v_proj.scales": "model-00027-of-00047.safetensors", + "model.layers.36.self_attn.v_proj.weight": "model-00027-of-00047.safetensors", + "model.layers.37.block_sparse_moe.e_score_correction_bias": "model-00029-of-00047.safetensors", + "model.layers.37.block_sparse_moe.gate.biases": "model-00028-of-00047.safetensors", + "model.layers.37.block_sparse_moe.gate.scales": "model-00028-of-00047.safetensors", + "model.layers.37.block_sparse_moe.gate.weight": "model-00028-of-00047.safetensors", + "model.layers.37.block_sparse_moe.switch_mlp.down_proj.biases": "model-00029-of-00047.safetensors", + "model.layers.37.block_sparse_moe.switch_mlp.down_proj.scales": "model-00029-of-00047.safetensors", + "model.layers.37.block_sparse_moe.switch_mlp.down_proj.weight": "model-00029-of-00047.safetensors", + "model.layers.37.block_sparse_moe.switch_mlp.gate_proj.biases": "model-00028-of-00047.safetensors", + "model.layers.37.block_sparse_moe.switch_mlp.gate_proj.scales": "model-00028-of-00047.safetensors", + "model.layers.37.block_sparse_moe.switch_mlp.gate_proj.weight": "model-00028-of-00047.safetensors", + "model.layers.37.block_sparse_moe.switch_mlp.up_proj.biases": "model-00029-of-00047.safetensors", + "model.layers.37.block_sparse_moe.switch_mlp.up_proj.scales": "model-00029-of-00047.safetensors", + "model.layers.37.block_sparse_moe.switch_mlp.up_proj.weight": "model-00029-of-00047.safetensors", + "model.layers.37.input_layernorm.weight": "model-00029-of-00047.safetensors", + "model.layers.37.post_attention_layernorm.weight": "model-00029-of-00047.safetensors", + "model.layers.37.self_attn.k_norm.weight": "model-00028-of-00047.safetensors", + "model.layers.37.self_attn.k_proj.biases": "model-00028-of-00047.safetensors", + "model.layers.37.self_attn.k_proj.scales": "model-00028-of-00047.safetensors", + "model.layers.37.self_attn.k_proj.weight": "model-00028-of-00047.safetensors", + "model.layers.37.self_attn.o_proj.biases": "model-00028-of-00047.safetensors", + "model.layers.37.self_attn.o_proj.scales": "model-00028-of-00047.safetensors", + "model.layers.37.self_attn.o_proj.weight": "model-00028-of-00047.safetensors", + "model.layers.37.self_attn.q_norm.weight": "model-00028-of-00047.safetensors", + "model.layers.37.self_attn.q_proj.biases": "model-00028-of-00047.safetensors", + "model.layers.37.self_attn.q_proj.scales": "model-00028-of-00047.safetensors", + "model.layers.37.self_attn.q_proj.weight": "model-00028-of-00047.safetensors", + "model.layers.37.self_attn.v_proj.biases": "model-00028-of-00047.safetensors", + "model.layers.37.self_attn.v_proj.scales": "model-00028-of-00047.safetensors", + "model.layers.37.self_attn.v_proj.weight": "model-00028-of-00047.safetensors", + "model.layers.38.block_sparse_moe.e_score_correction_bias": "model-00030-of-00047.safetensors", + "model.layers.38.block_sparse_moe.gate.biases": "model-00029-of-00047.safetensors", + "model.layers.38.block_sparse_moe.gate.scales": "model-00029-of-00047.safetensors", + "model.layers.38.block_sparse_moe.gate.weight": "model-00029-of-00047.safetensors", + "model.layers.38.block_sparse_moe.switch_mlp.down_proj.biases": "model-00030-of-00047.safetensors", + "model.layers.38.block_sparse_moe.switch_mlp.down_proj.scales": "model-00030-of-00047.safetensors", + "model.layers.38.block_sparse_moe.switch_mlp.down_proj.weight": "model-00030-of-00047.safetensors", + "model.layers.38.block_sparse_moe.switch_mlp.gate_proj.biases": "model-00029-of-00047.safetensors", + "model.layers.38.block_sparse_moe.switch_mlp.gate_proj.scales": "model-00029-of-00047.safetensors", + "model.layers.38.block_sparse_moe.switch_mlp.gate_proj.weight": "model-00029-of-00047.safetensors", + "model.layers.38.block_sparse_moe.switch_mlp.up_proj.biases": "model-00029-of-00047.safetensors", + "model.layers.38.block_sparse_moe.switch_mlp.up_proj.scales": "model-00029-of-00047.safetensors", + "model.layers.38.block_sparse_moe.switch_mlp.up_proj.weight": "model-00029-of-00047.safetensors", + "model.layers.38.input_layernorm.weight": "model-00030-of-00047.safetensors", + "model.layers.38.post_attention_layernorm.weight": "model-00030-of-00047.safetensors", + "model.layers.38.self_attn.k_norm.weight": "model-00029-of-00047.safetensors", + "model.layers.38.self_attn.k_proj.biases": "model-00029-of-00047.safetensors", + "model.layers.38.self_attn.k_proj.scales": "model-00029-of-00047.safetensors", + "model.layers.38.self_attn.k_proj.weight": "model-00029-of-00047.safetensors", + "model.layers.38.self_attn.o_proj.biases": "model-00029-of-00047.safetensors", + "model.layers.38.self_attn.o_proj.scales": "model-00029-of-00047.safetensors", + "model.layers.38.self_attn.o_proj.weight": "model-00029-of-00047.safetensors", + "model.layers.38.self_attn.q_norm.weight": "model-00029-of-00047.safetensors", + "model.layers.38.self_attn.q_proj.biases": "model-00029-of-00047.safetensors", + "model.layers.38.self_attn.q_proj.scales": "model-00029-of-00047.safetensors", + "model.layers.38.self_attn.q_proj.weight": "model-00029-of-00047.safetensors", + "model.layers.38.self_attn.v_proj.biases": "model-00029-of-00047.safetensors", + "model.layers.38.self_attn.v_proj.scales": "model-00029-of-00047.safetensors", + "model.layers.38.self_attn.v_proj.weight": "model-00029-of-00047.safetensors", + "model.layers.39.block_sparse_moe.e_score_correction_bias": "model-00030-of-00047.safetensors", + "model.layers.39.block_sparse_moe.gate.biases": "model-00030-of-00047.safetensors", + "model.layers.39.block_sparse_moe.gate.scales": "model-00030-of-00047.safetensors", + "model.layers.39.block_sparse_moe.gate.weight": "model-00030-of-00047.safetensors", + "model.layers.39.block_sparse_moe.switch_mlp.down_proj.biases": "model-00030-of-00047.safetensors", + "model.layers.39.block_sparse_moe.switch_mlp.down_proj.scales": "model-00030-of-00047.safetensors", + "model.layers.39.block_sparse_moe.switch_mlp.down_proj.weight": "model-00030-of-00047.safetensors", + "model.layers.39.block_sparse_moe.switch_mlp.gate_proj.biases": "model-00030-of-00047.safetensors", + "model.layers.39.block_sparse_moe.switch_mlp.gate_proj.scales": "model-00030-of-00047.safetensors", + "model.layers.39.block_sparse_moe.switch_mlp.gate_proj.weight": "model-00030-of-00047.safetensors", + "model.layers.39.block_sparse_moe.switch_mlp.up_proj.biases": "model-00030-of-00047.safetensors", + "model.layers.39.block_sparse_moe.switch_mlp.up_proj.scales": "model-00030-of-00047.safetensors", + "model.layers.39.block_sparse_moe.switch_mlp.up_proj.weight": "model-00030-of-00047.safetensors", + "model.layers.39.input_layernorm.weight": "model-00030-of-00047.safetensors", + "model.layers.39.post_attention_layernorm.weight": "model-00030-of-00047.safetensors", + "model.layers.39.self_attn.k_norm.weight": "model-00030-of-00047.safetensors", + "model.layers.39.self_attn.k_proj.biases": "model-00030-of-00047.safetensors", + "model.layers.39.self_attn.k_proj.scales": "model-00030-of-00047.safetensors", + "model.layers.39.self_attn.k_proj.weight": "model-00030-of-00047.safetensors", + "model.layers.39.self_attn.o_proj.biases": "model-00030-of-00047.safetensors", + "model.layers.39.self_attn.o_proj.scales": "model-00030-of-00047.safetensors", + "model.layers.39.self_attn.o_proj.weight": "model-00030-of-00047.safetensors", + "model.layers.39.self_attn.q_norm.weight": "model-00030-of-00047.safetensors", + "model.layers.39.self_attn.q_proj.biases": "model-00030-of-00047.safetensors", + "model.layers.39.self_attn.q_proj.scales": "model-00030-of-00047.safetensors", + "model.layers.39.self_attn.q_proj.weight": "model-00030-of-00047.safetensors", + "model.layers.39.self_attn.v_proj.biases": "model-00030-of-00047.safetensors", + "model.layers.39.self_attn.v_proj.scales": "model-00030-of-00047.safetensors", + "model.layers.39.self_attn.v_proj.weight": "model-00030-of-00047.safetensors", + "model.layers.4.block_sparse_moe.e_score_correction_bias": "model-00004-of-00047.safetensors", + "model.layers.4.block_sparse_moe.gate.biases": "model-00003-of-00047.safetensors", + "model.layers.4.block_sparse_moe.gate.scales": "model-00003-of-00047.safetensors", + "model.layers.4.block_sparse_moe.gate.weight": "model-00003-of-00047.safetensors", + "model.layers.4.block_sparse_moe.switch_mlp.down_proj.biases": "model-00004-of-00047.safetensors", + "model.layers.4.block_sparse_moe.switch_mlp.down_proj.scales": "model-00004-of-00047.safetensors", + "model.layers.4.block_sparse_moe.switch_mlp.down_proj.weight": "model-00004-of-00047.safetensors", + "model.layers.4.block_sparse_moe.switch_mlp.gate_proj.biases": "model-00004-of-00047.safetensors", + "model.layers.4.block_sparse_moe.switch_mlp.gate_proj.scales": "model-00004-of-00047.safetensors", + "model.layers.4.block_sparse_moe.switch_mlp.gate_proj.weight": "model-00004-of-00047.safetensors", + "model.layers.4.block_sparse_moe.switch_mlp.up_proj.biases": "model-00004-of-00047.safetensors", + "model.layers.4.block_sparse_moe.switch_mlp.up_proj.scales": "model-00004-of-00047.safetensors", + "model.layers.4.block_sparse_moe.switch_mlp.up_proj.weight": "model-00004-of-00047.safetensors", + "model.layers.4.input_layernorm.weight": "model-00004-of-00047.safetensors", + "model.layers.4.post_attention_layernorm.weight": "model-00004-of-00047.safetensors", + "model.layers.4.self_attn.k_norm.weight": "model-00003-of-00047.safetensors", + "model.layers.4.self_attn.k_proj.biases": "model-00003-of-00047.safetensors", + "model.layers.4.self_attn.k_proj.scales": "model-00003-of-00047.safetensors", + "model.layers.4.self_attn.k_proj.weight": "model-00003-of-00047.safetensors", + "model.layers.4.self_attn.o_proj.biases": "model-00003-of-00047.safetensors", + "model.layers.4.self_attn.o_proj.scales": "model-00003-of-00047.safetensors", + "model.layers.4.self_attn.o_proj.weight": "model-00003-of-00047.safetensors", + "model.layers.4.self_attn.q_norm.weight": "model-00003-of-00047.safetensors", + "model.layers.4.self_attn.q_proj.biases": "model-00003-of-00047.safetensors", + "model.layers.4.self_attn.q_proj.scales": "model-00003-of-00047.safetensors", + "model.layers.4.self_attn.q_proj.weight": "model-00003-of-00047.safetensors", + "model.layers.4.self_attn.v_proj.biases": "model-00003-of-00047.safetensors", + "model.layers.4.self_attn.v_proj.scales": "model-00003-of-00047.safetensors", + "model.layers.4.self_attn.v_proj.weight": "model-00003-of-00047.safetensors", + "model.layers.40.block_sparse_moe.e_score_correction_bias": "model-00031-of-00047.safetensors", + "model.layers.40.block_sparse_moe.gate.biases": "model-00030-of-00047.safetensors", + "model.layers.40.block_sparse_moe.gate.scales": "model-00030-of-00047.safetensors", + "model.layers.40.block_sparse_moe.gate.weight": "model-00030-of-00047.safetensors", + "model.layers.40.block_sparse_moe.switch_mlp.down_proj.biases": "model-00031-of-00047.safetensors", + "model.layers.40.block_sparse_moe.switch_mlp.down_proj.scales": "model-00031-of-00047.safetensors", + "model.layers.40.block_sparse_moe.switch_mlp.down_proj.weight": "model-00031-of-00047.safetensors", + "model.layers.40.block_sparse_moe.switch_mlp.gate_proj.biases": "model-00031-of-00047.safetensors", + "model.layers.40.block_sparse_moe.switch_mlp.gate_proj.scales": "model-00031-of-00047.safetensors", + "model.layers.40.block_sparse_moe.switch_mlp.gate_proj.weight": "model-00031-of-00047.safetensors", + "model.layers.40.block_sparse_moe.switch_mlp.up_proj.biases": "model-00031-of-00047.safetensors", + "model.layers.40.block_sparse_moe.switch_mlp.up_proj.scales": "model-00031-of-00047.safetensors", + "model.layers.40.block_sparse_moe.switch_mlp.up_proj.weight": "model-00031-of-00047.safetensors", + "model.layers.40.input_layernorm.weight": "model-00031-of-00047.safetensors", + "model.layers.40.post_attention_layernorm.weight": "model-00031-of-00047.safetensors", + "model.layers.40.self_attn.k_norm.weight": "model-00030-of-00047.safetensors", + "model.layers.40.self_attn.k_proj.biases": "model-00030-of-00047.safetensors", + "model.layers.40.self_attn.k_proj.scales": "model-00030-of-00047.safetensors", + "model.layers.40.self_attn.k_proj.weight": "model-00030-of-00047.safetensors", + "model.layers.40.self_attn.o_proj.biases": "model-00030-of-00047.safetensors", + "model.layers.40.self_attn.o_proj.scales": "model-00030-of-00047.safetensors", + "model.layers.40.self_attn.o_proj.weight": "model-00030-of-00047.safetensors", + "model.layers.40.self_attn.q_norm.weight": "model-00030-of-00047.safetensors", + "model.layers.40.self_attn.q_proj.biases": "model-00030-of-00047.safetensors", + "model.layers.40.self_attn.q_proj.scales": "model-00030-of-00047.safetensors", + "model.layers.40.self_attn.q_proj.weight": "model-00030-of-00047.safetensors", + "model.layers.40.self_attn.v_proj.biases": "model-00030-of-00047.safetensors", + "model.layers.40.self_attn.v_proj.scales": "model-00030-of-00047.safetensors", + "model.layers.40.self_attn.v_proj.weight": "model-00030-of-00047.safetensors", + "model.layers.41.block_sparse_moe.e_score_correction_bias": "model-00032-of-00047.safetensors", + "model.layers.41.block_sparse_moe.gate.biases": "model-00031-of-00047.safetensors", + "model.layers.41.block_sparse_moe.gate.scales": "model-00031-of-00047.safetensors", + "model.layers.41.block_sparse_moe.gate.weight": "model-00031-of-00047.safetensors", + "model.layers.41.block_sparse_moe.switch_mlp.down_proj.biases": "model-00032-of-00047.safetensors", + "model.layers.41.block_sparse_moe.switch_mlp.down_proj.scales": "model-00032-of-00047.safetensors", + "model.layers.41.block_sparse_moe.switch_mlp.down_proj.weight": "model-00032-of-00047.safetensors", + "model.layers.41.block_sparse_moe.switch_mlp.gate_proj.biases": "model-00031-of-00047.safetensors", + "model.layers.41.block_sparse_moe.switch_mlp.gate_proj.scales": "model-00031-of-00047.safetensors", + "model.layers.41.block_sparse_moe.switch_mlp.gate_proj.weight": "model-00031-of-00047.safetensors", + "model.layers.41.block_sparse_moe.switch_mlp.up_proj.biases": "model-00032-of-00047.safetensors", + "model.layers.41.block_sparse_moe.switch_mlp.up_proj.scales": "model-00032-of-00047.safetensors", + "model.layers.41.block_sparse_moe.switch_mlp.up_proj.weight": "model-00032-of-00047.safetensors", + "model.layers.41.input_layernorm.weight": "model-00032-of-00047.safetensors", + "model.layers.41.post_attention_layernorm.weight": "model-00032-of-00047.safetensors", + "model.layers.41.self_attn.k_norm.weight": "model-00031-of-00047.safetensors", + "model.layers.41.self_attn.k_proj.biases": "model-00031-of-00047.safetensors", + "model.layers.41.self_attn.k_proj.scales": "model-00031-of-00047.safetensors", + "model.layers.41.self_attn.k_proj.weight": "model-00031-of-00047.safetensors", + "model.layers.41.self_attn.o_proj.biases": "model-00031-of-00047.safetensors", + "model.layers.41.self_attn.o_proj.scales": "model-00031-of-00047.safetensors", + "model.layers.41.self_attn.o_proj.weight": "model-00031-of-00047.safetensors", + "model.layers.41.self_attn.q_norm.weight": "model-00031-of-00047.safetensors", + "model.layers.41.self_attn.q_proj.biases": "model-00031-of-00047.safetensors", + "model.layers.41.self_attn.q_proj.scales": "model-00031-of-00047.safetensors", + "model.layers.41.self_attn.q_proj.weight": "model-00031-of-00047.safetensors", + "model.layers.41.self_attn.v_proj.biases": "model-00031-of-00047.safetensors", + "model.layers.41.self_attn.v_proj.scales": "model-00031-of-00047.safetensors", + "model.layers.41.self_attn.v_proj.weight": "model-00031-of-00047.safetensors", + "model.layers.42.block_sparse_moe.e_score_correction_bias": "model-00033-of-00047.safetensors", + "model.layers.42.block_sparse_moe.gate.biases": "model-00032-of-00047.safetensors", + "model.layers.42.block_sparse_moe.gate.scales": "model-00032-of-00047.safetensors", + "model.layers.42.block_sparse_moe.gate.weight": "model-00032-of-00047.safetensors", + "model.layers.42.block_sparse_moe.switch_mlp.down_proj.biases": "model-00033-of-00047.safetensors", + "model.layers.42.block_sparse_moe.switch_mlp.down_proj.scales": "model-00033-of-00047.safetensors", + "model.layers.42.block_sparse_moe.switch_mlp.down_proj.weight": "model-00033-of-00047.safetensors", + "model.layers.42.block_sparse_moe.switch_mlp.gate_proj.biases": "model-00032-of-00047.safetensors", + "model.layers.42.block_sparse_moe.switch_mlp.gate_proj.scales": "model-00032-of-00047.safetensors", + "model.layers.42.block_sparse_moe.switch_mlp.gate_proj.weight": "model-00032-of-00047.safetensors", + "model.layers.42.block_sparse_moe.switch_mlp.up_proj.biases": "model-00032-of-00047.safetensors", + "model.layers.42.block_sparse_moe.switch_mlp.up_proj.scales": "model-00032-of-00047.safetensors", + "model.layers.42.block_sparse_moe.switch_mlp.up_proj.weight": "model-00032-of-00047.safetensors", + "model.layers.42.input_layernorm.weight": "model-00033-of-00047.safetensors", + "model.layers.42.post_attention_layernorm.weight": "model-00033-of-00047.safetensors", + "model.layers.42.self_attn.k_norm.weight": "model-00032-of-00047.safetensors", + "model.layers.42.self_attn.k_proj.biases": "model-00032-of-00047.safetensors", + "model.layers.42.self_attn.k_proj.scales": "model-00032-of-00047.safetensors", + "model.layers.42.self_attn.k_proj.weight": "model-00032-of-00047.safetensors", + "model.layers.42.self_attn.o_proj.biases": "model-00032-of-00047.safetensors", + "model.layers.42.self_attn.o_proj.scales": "model-00032-of-00047.safetensors", + "model.layers.42.self_attn.o_proj.weight": "model-00032-of-00047.safetensors", + "model.layers.42.self_attn.q_norm.weight": "model-00032-of-00047.safetensors", + "model.layers.42.self_attn.q_proj.biases": "model-00032-of-00047.safetensors", + "model.layers.42.self_attn.q_proj.scales": "model-00032-of-00047.safetensors", + "model.layers.42.self_attn.q_proj.weight": "model-00032-of-00047.safetensors", + "model.layers.42.self_attn.v_proj.biases": "model-00032-of-00047.safetensors", + "model.layers.42.self_attn.v_proj.scales": "model-00032-of-00047.safetensors", + "model.layers.42.self_attn.v_proj.weight": "model-00032-of-00047.safetensors", + "model.layers.43.block_sparse_moe.e_score_correction_bias": "model-00033-of-00047.safetensors", + "model.layers.43.block_sparse_moe.gate.biases": "model-00033-of-00047.safetensors", + "model.layers.43.block_sparse_moe.gate.scales": "model-00033-of-00047.safetensors", + "model.layers.43.block_sparse_moe.gate.weight": "model-00033-of-00047.safetensors", + "model.layers.43.block_sparse_moe.switch_mlp.down_proj.biases": "model-00033-of-00047.safetensors", + "model.layers.43.block_sparse_moe.switch_mlp.down_proj.scales": "model-00033-of-00047.safetensors", + "model.layers.43.block_sparse_moe.switch_mlp.down_proj.weight": "model-00033-of-00047.safetensors", + "model.layers.43.block_sparse_moe.switch_mlp.gate_proj.biases": "model-00033-of-00047.safetensors", + "model.layers.43.block_sparse_moe.switch_mlp.gate_proj.scales": "model-00033-of-00047.safetensors", + "model.layers.43.block_sparse_moe.switch_mlp.gate_proj.weight": "model-00033-of-00047.safetensors", + "model.layers.43.block_sparse_moe.switch_mlp.up_proj.biases": "model-00033-of-00047.safetensors", + "model.layers.43.block_sparse_moe.switch_mlp.up_proj.scales": "model-00033-of-00047.safetensors", + "model.layers.43.block_sparse_moe.switch_mlp.up_proj.weight": "model-00033-of-00047.safetensors", + "model.layers.43.input_layernorm.weight": "model-00033-of-00047.safetensors", + "model.layers.43.post_attention_layernorm.weight": "model-00033-of-00047.safetensors", + "model.layers.43.self_attn.k_norm.weight": "model-00033-of-00047.safetensors", + "model.layers.43.self_attn.k_proj.biases": "model-00033-of-00047.safetensors", + "model.layers.43.self_attn.k_proj.scales": "model-00033-of-00047.safetensors", + "model.layers.43.self_attn.k_proj.weight": "model-00033-of-00047.safetensors", + "model.layers.43.self_attn.o_proj.biases": "model-00033-of-00047.safetensors", + "model.layers.43.self_attn.o_proj.scales": "model-00033-of-00047.safetensors", + "model.layers.43.self_attn.o_proj.weight": "model-00033-of-00047.safetensors", + "model.layers.43.self_attn.q_norm.weight": "model-00033-of-00047.safetensors", + "model.layers.43.self_attn.q_proj.biases": "model-00033-of-00047.safetensors", + "model.layers.43.self_attn.q_proj.scales": "model-00033-of-00047.safetensors", + "model.layers.43.self_attn.q_proj.weight": "model-00033-of-00047.safetensors", + "model.layers.43.self_attn.v_proj.biases": "model-00033-of-00047.safetensors", + "model.layers.43.self_attn.v_proj.scales": "model-00033-of-00047.safetensors", + "model.layers.43.self_attn.v_proj.weight": "model-00033-of-00047.safetensors", + "model.layers.44.block_sparse_moe.e_score_correction_bias": "model-00034-of-00047.safetensors", + "model.layers.44.block_sparse_moe.gate.biases": "model-00033-of-00047.safetensors", + "model.layers.44.block_sparse_moe.gate.scales": "model-00033-of-00047.safetensors", + "model.layers.44.block_sparse_moe.gate.weight": "model-00033-of-00047.safetensors", + "model.layers.44.block_sparse_moe.switch_mlp.down_proj.biases": "model-00034-of-00047.safetensors", + "model.layers.44.block_sparse_moe.switch_mlp.down_proj.scales": "model-00034-of-00047.safetensors", + "model.layers.44.block_sparse_moe.switch_mlp.down_proj.weight": "model-00034-of-00047.safetensors", + "model.layers.44.block_sparse_moe.switch_mlp.gate_proj.biases": "model-00034-of-00047.safetensors", + "model.layers.44.block_sparse_moe.switch_mlp.gate_proj.scales": "model-00034-of-00047.safetensors", + "model.layers.44.block_sparse_moe.switch_mlp.gate_proj.weight": "model-00034-of-00047.safetensors", + "model.layers.44.block_sparse_moe.switch_mlp.up_proj.biases": "model-00034-of-00047.safetensors", + "model.layers.44.block_sparse_moe.switch_mlp.up_proj.scales": "model-00034-of-00047.safetensors", + "model.layers.44.block_sparse_moe.switch_mlp.up_proj.weight": "model-00034-of-00047.safetensors", + "model.layers.44.input_layernorm.weight": "model-00034-of-00047.safetensors", + "model.layers.44.post_attention_layernorm.weight": "model-00034-of-00047.safetensors", + "model.layers.44.self_attn.k_norm.weight": "model-00033-of-00047.safetensors", + "model.layers.44.self_attn.k_proj.biases": "model-00033-of-00047.safetensors", + "model.layers.44.self_attn.k_proj.scales": "model-00033-of-00047.safetensors", + "model.layers.44.self_attn.k_proj.weight": "model-00033-of-00047.safetensors", + "model.layers.44.self_attn.o_proj.biases": "model-00033-of-00047.safetensors", + "model.layers.44.self_attn.o_proj.scales": "model-00033-of-00047.safetensors", + "model.layers.44.self_attn.o_proj.weight": "model-00033-of-00047.safetensors", + "model.layers.44.self_attn.q_norm.weight": "model-00033-of-00047.safetensors", + "model.layers.44.self_attn.q_proj.biases": "model-00033-of-00047.safetensors", + "model.layers.44.self_attn.q_proj.scales": "model-00033-of-00047.safetensors", + "model.layers.44.self_attn.q_proj.weight": "model-00033-of-00047.safetensors", + "model.layers.44.self_attn.v_proj.biases": "model-00033-of-00047.safetensors", + "model.layers.44.self_attn.v_proj.scales": "model-00033-of-00047.safetensors", + "model.layers.44.self_attn.v_proj.weight": "model-00033-of-00047.safetensors", + "model.layers.45.block_sparse_moe.e_score_correction_bias": "model-00035-of-00047.safetensors", + "model.layers.45.block_sparse_moe.gate.biases": "model-00034-of-00047.safetensors", + "model.layers.45.block_sparse_moe.gate.scales": "model-00034-of-00047.safetensors", + "model.layers.45.block_sparse_moe.gate.weight": "model-00034-of-00047.safetensors", + "model.layers.45.block_sparse_moe.switch_mlp.down_proj.biases": "model-00035-of-00047.safetensors", + "model.layers.45.block_sparse_moe.switch_mlp.down_proj.scales": "model-00035-of-00047.safetensors", + "model.layers.45.block_sparse_moe.switch_mlp.down_proj.weight": "model-00035-of-00047.safetensors", + "model.layers.45.block_sparse_moe.switch_mlp.gate_proj.biases": "model-00034-of-00047.safetensors", + "model.layers.45.block_sparse_moe.switch_mlp.gate_proj.scales": "model-00034-of-00047.safetensors", + "model.layers.45.block_sparse_moe.switch_mlp.gate_proj.weight": "model-00034-of-00047.safetensors", + "model.layers.45.block_sparse_moe.switch_mlp.up_proj.biases": "model-00035-of-00047.safetensors", + "model.layers.45.block_sparse_moe.switch_mlp.up_proj.scales": "model-00035-of-00047.safetensors", + "model.layers.45.block_sparse_moe.switch_mlp.up_proj.weight": "model-00035-of-00047.safetensors", + "model.layers.45.input_layernorm.weight": "model-00035-of-00047.safetensors", + "model.layers.45.post_attention_layernorm.weight": "model-00035-of-00047.safetensors", + "model.layers.45.self_attn.k_norm.weight": "model-00034-of-00047.safetensors", + "model.layers.45.self_attn.k_proj.biases": "model-00034-of-00047.safetensors", + "model.layers.45.self_attn.k_proj.scales": "model-00034-of-00047.safetensors", + "model.layers.45.self_attn.k_proj.weight": "model-00034-of-00047.safetensors", + "model.layers.45.self_attn.o_proj.biases": "model-00034-of-00047.safetensors", + "model.layers.45.self_attn.o_proj.scales": "model-00034-of-00047.safetensors", + "model.layers.45.self_attn.o_proj.weight": "model-00034-of-00047.safetensors", + "model.layers.45.self_attn.q_norm.weight": "model-00034-of-00047.safetensors", + "model.layers.45.self_attn.q_proj.biases": "model-00034-of-00047.safetensors", + "model.layers.45.self_attn.q_proj.scales": "model-00034-of-00047.safetensors", + "model.layers.45.self_attn.q_proj.weight": "model-00034-of-00047.safetensors", + "model.layers.45.self_attn.v_proj.biases": "model-00034-of-00047.safetensors", + "model.layers.45.self_attn.v_proj.scales": "model-00034-of-00047.safetensors", + "model.layers.45.self_attn.v_proj.weight": "model-00034-of-00047.safetensors", + "model.layers.46.block_sparse_moe.e_score_correction_bias": "model-00036-of-00047.safetensors", + "model.layers.46.block_sparse_moe.gate.biases": "model-00035-of-00047.safetensors", + "model.layers.46.block_sparse_moe.gate.scales": "model-00035-of-00047.safetensors", + "model.layers.46.block_sparse_moe.gate.weight": "model-00035-of-00047.safetensors", + "model.layers.46.block_sparse_moe.switch_mlp.down_proj.biases": "model-00036-of-00047.safetensors", + "model.layers.46.block_sparse_moe.switch_mlp.down_proj.scales": "model-00036-of-00047.safetensors", + "model.layers.46.block_sparse_moe.switch_mlp.down_proj.weight": "model-00036-of-00047.safetensors", + "model.layers.46.block_sparse_moe.switch_mlp.gate_proj.biases": "model-00035-of-00047.safetensors", + "model.layers.46.block_sparse_moe.switch_mlp.gate_proj.scales": "model-00035-of-00047.safetensors", + "model.layers.46.block_sparse_moe.switch_mlp.gate_proj.weight": "model-00035-of-00047.safetensors", + "model.layers.46.block_sparse_moe.switch_mlp.up_proj.biases": "model-00035-of-00047.safetensors", + "model.layers.46.block_sparse_moe.switch_mlp.up_proj.scales": "model-00035-of-00047.safetensors", + "model.layers.46.block_sparse_moe.switch_mlp.up_proj.weight": "model-00035-of-00047.safetensors", + "model.layers.46.input_layernorm.weight": "model-00036-of-00047.safetensors", + "model.layers.46.post_attention_layernorm.weight": "model-00036-of-00047.safetensors", + "model.layers.46.self_attn.k_norm.weight": "model-00035-of-00047.safetensors", + "model.layers.46.self_attn.k_proj.biases": "model-00035-of-00047.safetensors", + "model.layers.46.self_attn.k_proj.scales": "model-00035-of-00047.safetensors", + "model.layers.46.self_attn.k_proj.weight": "model-00035-of-00047.safetensors", + "model.layers.46.self_attn.o_proj.biases": "model-00035-of-00047.safetensors", + "model.layers.46.self_attn.o_proj.scales": "model-00035-of-00047.safetensors", + "model.layers.46.self_attn.o_proj.weight": "model-00035-of-00047.safetensors", + "model.layers.46.self_attn.q_norm.weight": "model-00035-of-00047.safetensors", + "model.layers.46.self_attn.q_proj.biases": "model-00035-of-00047.safetensors", + "model.layers.46.self_attn.q_proj.scales": "model-00035-of-00047.safetensors", + "model.layers.46.self_attn.q_proj.weight": "model-00035-of-00047.safetensors", + "model.layers.46.self_attn.v_proj.biases": "model-00035-of-00047.safetensors", + "model.layers.46.self_attn.v_proj.scales": "model-00035-of-00047.safetensors", + "model.layers.46.self_attn.v_proj.weight": "model-00035-of-00047.safetensors", + "model.layers.47.block_sparse_moe.e_score_correction_bias": "model-00036-of-00047.safetensors", + "model.layers.47.block_sparse_moe.gate.biases": "model-00036-of-00047.safetensors", + "model.layers.47.block_sparse_moe.gate.scales": "model-00036-of-00047.safetensors", + "model.layers.47.block_sparse_moe.gate.weight": "model-00036-of-00047.safetensors", + "model.layers.47.block_sparse_moe.switch_mlp.down_proj.biases": "model-00036-of-00047.safetensors", + "model.layers.47.block_sparse_moe.switch_mlp.down_proj.scales": "model-00036-of-00047.safetensors", + "model.layers.47.block_sparse_moe.switch_mlp.down_proj.weight": "model-00036-of-00047.safetensors", + "model.layers.47.block_sparse_moe.switch_mlp.gate_proj.biases": "model-00036-of-00047.safetensors", + "model.layers.47.block_sparse_moe.switch_mlp.gate_proj.scales": "model-00036-of-00047.safetensors", + "model.layers.47.block_sparse_moe.switch_mlp.gate_proj.weight": "model-00036-of-00047.safetensors", + "model.layers.47.block_sparse_moe.switch_mlp.up_proj.biases": "model-00036-of-00047.safetensors", + "model.layers.47.block_sparse_moe.switch_mlp.up_proj.scales": "model-00036-of-00047.safetensors", + "model.layers.47.block_sparse_moe.switch_mlp.up_proj.weight": "model-00036-of-00047.safetensors", + "model.layers.47.input_layernorm.weight": "model-00036-of-00047.safetensors", + "model.layers.47.post_attention_layernorm.weight": "model-00036-of-00047.safetensors", + "model.layers.47.self_attn.k_norm.weight": "model-00036-of-00047.safetensors", + "model.layers.47.self_attn.k_proj.biases": "model-00036-of-00047.safetensors", + "model.layers.47.self_attn.k_proj.scales": "model-00036-of-00047.safetensors", + "model.layers.47.self_attn.k_proj.weight": "model-00036-of-00047.safetensors", + "model.layers.47.self_attn.o_proj.biases": "model-00036-of-00047.safetensors", + "model.layers.47.self_attn.o_proj.scales": "model-00036-of-00047.safetensors", + "model.layers.47.self_attn.o_proj.weight": "model-00036-of-00047.safetensors", + "model.layers.47.self_attn.q_norm.weight": "model-00036-of-00047.safetensors", + "model.layers.47.self_attn.q_proj.biases": "model-00036-of-00047.safetensors", + "model.layers.47.self_attn.q_proj.scales": "model-00036-of-00047.safetensors", + "model.layers.47.self_attn.q_proj.weight": "model-00036-of-00047.safetensors", + "model.layers.47.self_attn.v_proj.biases": "model-00036-of-00047.safetensors", + "model.layers.47.self_attn.v_proj.scales": "model-00036-of-00047.safetensors", + "model.layers.47.self_attn.v_proj.weight": "model-00036-of-00047.safetensors", + "model.layers.48.block_sparse_moe.e_score_correction_bias": "model-00037-of-00047.safetensors", + "model.layers.48.block_sparse_moe.gate.biases": "model-00036-of-00047.safetensors", + "model.layers.48.block_sparse_moe.gate.scales": "model-00036-of-00047.safetensors", + "model.layers.48.block_sparse_moe.gate.weight": "model-00036-of-00047.safetensors", + "model.layers.48.block_sparse_moe.switch_mlp.down_proj.biases": "model-00037-of-00047.safetensors", + "model.layers.48.block_sparse_moe.switch_mlp.down_proj.scales": "model-00037-of-00047.safetensors", + "model.layers.48.block_sparse_moe.switch_mlp.down_proj.weight": "model-00037-of-00047.safetensors", + "model.layers.48.block_sparse_moe.switch_mlp.gate_proj.biases": "model-00037-of-00047.safetensors", + "model.layers.48.block_sparse_moe.switch_mlp.gate_proj.scales": "model-00037-of-00047.safetensors", + "model.layers.48.block_sparse_moe.switch_mlp.gate_proj.weight": "model-00037-of-00047.safetensors", + "model.layers.48.block_sparse_moe.switch_mlp.up_proj.biases": "model-00037-of-00047.safetensors", + "model.layers.48.block_sparse_moe.switch_mlp.up_proj.scales": "model-00037-of-00047.safetensors", + "model.layers.48.block_sparse_moe.switch_mlp.up_proj.weight": "model-00037-of-00047.safetensors", + "model.layers.48.input_layernorm.weight": "model-00037-of-00047.safetensors", + "model.layers.48.post_attention_layernorm.weight": "model-00037-of-00047.safetensors", + "model.layers.48.self_attn.k_norm.weight": "model-00036-of-00047.safetensors", + "model.layers.48.self_attn.k_proj.biases": "model-00036-of-00047.safetensors", + "model.layers.48.self_attn.k_proj.scales": "model-00036-of-00047.safetensors", + "model.layers.48.self_attn.k_proj.weight": "model-00036-of-00047.safetensors", + "model.layers.48.self_attn.o_proj.biases": "model-00036-of-00047.safetensors", + "model.layers.48.self_attn.o_proj.scales": "model-00036-of-00047.safetensors", + "model.layers.48.self_attn.o_proj.weight": "model-00036-of-00047.safetensors", + "model.layers.48.self_attn.q_norm.weight": "model-00036-of-00047.safetensors", + "model.layers.48.self_attn.q_proj.biases": "model-00036-of-00047.safetensors", + "model.layers.48.self_attn.q_proj.scales": "model-00036-of-00047.safetensors", + "model.layers.48.self_attn.q_proj.weight": "model-00036-of-00047.safetensors", + "model.layers.48.self_attn.v_proj.biases": "model-00036-of-00047.safetensors", + "model.layers.48.self_attn.v_proj.scales": "model-00036-of-00047.safetensors", + "model.layers.48.self_attn.v_proj.weight": "model-00036-of-00047.safetensors", + "model.layers.49.block_sparse_moe.e_score_correction_bias": "model-00038-of-00047.safetensors", + "model.layers.49.block_sparse_moe.gate.biases": "model-00037-of-00047.safetensors", + "model.layers.49.block_sparse_moe.gate.scales": "model-00037-of-00047.safetensors", + "model.layers.49.block_sparse_moe.gate.weight": "model-00037-of-00047.safetensors", + "model.layers.49.block_sparse_moe.switch_mlp.down_proj.biases": "model-00038-of-00047.safetensors", + "model.layers.49.block_sparse_moe.switch_mlp.down_proj.scales": "model-00038-of-00047.safetensors", + "model.layers.49.block_sparse_moe.switch_mlp.down_proj.weight": "model-00038-of-00047.safetensors", + "model.layers.49.block_sparse_moe.switch_mlp.gate_proj.biases": "model-00037-of-00047.safetensors", + "model.layers.49.block_sparse_moe.switch_mlp.gate_proj.scales": "model-00037-of-00047.safetensors", + "model.layers.49.block_sparse_moe.switch_mlp.gate_proj.weight": "model-00037-of-00047.safetensors", + "model.layers.49.block_sparse_moe.switch_mlp.up_proj.biases": "model-00038-of-00047.safetensors", + "model.layers.49.block_sparse_moe.switch_mlp.up_proj.scales": "model-00038-of-00047.safetensors", + "model.layers.49.block_sparse_moe.switch_mlp.up_proj.weight": "model-00038-of-00047.safetensors", + "model.layers.49.input_layernorm.weight": "model-00038-of-00047.safetensors", + "model.layers.49.post_attention_layernorm.weight": "model-00038-of-00047.safetensors", + "model.layers.49.self_attn.k_norm.weight": "model-00037-of-00047.safetensors", + "model.layers.49.self_attn.k_proj.biases": "model-00037-of-00047.safetensors", + "model.layers.49.self_attn.k_proj.scales": "model-00037-of-00047.safetensors", + "model.layers.49.self_attn.k_proj.weight": "model-00037-of-00047.safetensors", + "model.layers.49.self_attn.o_proj.biases": "model-00037-of-00047.safetensors", + "model.layers.49.self_attn.o_proj.scales": "model-00037-of-00047.safetensors", + "model.layers.49.self_attn.o_proj.weight": "model-00037-of-00047.safetensors", + "model.layers.49.self_attn.q_norm.weight": "model-00037-of-00047.safetensors", + "model.layers.49.self_attn.q_proj.biases": "model-00037-of-00047.safetensors", + "model.layers.49.self_attn.q_proj.scales": "model-00037-of-00047.safetensors", + "model.layers.49.self_attn.q_proj.weight": "model-00037-of-00047.safetensors", + "model.layers.49.self_attn.v_proj.biases": "model-00037-of-00047.safetensors", + "model.layers.49.self_attn.v_proj.scales": "model-00037-of-00047.safetensors", + "model.layers.49.self_attn.v_proj.weight": "model-00037-of-00047.safetensors", + "model.layers.5.block_sparse_moe.e_score_correction_bias": "model-00005-of-00047.safetensors", + "model.layers.5.block_sparse_moe.gate.biases": "model-00004-of-00047.safetensors", + "model.layers.5.block_sparse_moe.gate.scales": "model-00004-of-00047.safetensors", + "model.layers.5.block_sparse_moe.gate.weight": "model-00004-of-00047.safetensors", + "model.layers.5.block_sparse_moe.switch_mlp.down_proj.biases": "model-00005-of-00047.safetensors", + "model.layers.5.block_sparse_moe.switch_mlp.down_proj.scales": "model-00005-of-00047.safetensors", + "model.layers.5.block_sparse_moe.switch_mlp.down_proj.weight": "model-00005-of-00047.safetensors", + "model.layers.5.block_sparse_moe.switch_mlp.gate_proj.biases": "model-00004-of-00047.safetensors", + "model.layers.5.block_sparse_moe.switch_mlp.gate_proj.scales": "model-00004-of-00047.safetensors", + "model.layers.5.block_sparse_moe.switch_mlp.gate_proj.weight": "model-00004-of-00047.safetensors", + "model.layers.5.block_sparse_moe.switch_mlp.up_proj.biases": "model-00005-of-00047.safetensors", + "model.layers.5.block_sparse_moe.switch_mlp.up_proj.scales": "model-00005-of-00047.safetensors", + "model.layers.5.block_sparse_moe.switch_mlp.up_proj.weight": "model-00005-of-00047.safetensors", + "model.layers.5.input_layernorm.weight": "model-00005-of-00047.safetensors", + "model.layers.5.post_attention_layernorm.weight": "model-00005-of-00047.safetensors", + "model.layers.5.self_attn.k_norm.weight": "model-00004-of-00047.safetensors", + "model.layers.5.self_attn.k_proj.biases": "model-00004-of-00047.safetensors", + "model.layers.5.self_attn.k_proj.scales": "model-00004-of-00047.safetensors", + "model.layers.5.self_attn.k_proj.weight": "model-00004-of-00047.safetensors", + "model.layers.5.self_attn.o_proj.biases": "model-00004-of-00047.safetensors", + "model.layers.5.self_attn.o_proj.scales": "model-00004-of-00047.safetensors", + "model.layers.5.self_attn.o_proj.weight": "model-00004-of-00047.safetensors", + "model.layers.5.self_attn.q_norm.weight": "model-00004-of-00047.safetensors", + "model.layers.5.self_attn.q_proj.biases": "model-00004-of-00047.safetensors", + "model.layers.5.self_attn.q_proj.scales": "model-00004-of-00047.safetensors", + "model.layers.5.self_attn.q_proj.weight": "model-00004-of-00047.safetensors", + "model.layers.5.self_attn.v_proj.biases": "model-00004-of-00047.safetensors", + "model.layers.5.self_attn.v_proj.scales": "model-00004-of-00047.safetensors", + "model.layers.5.self_attn.v_proj.weight": "model-00004-of-00047.safetensors", + "model.layers.50.block_sparse_moe.e_score_correction_bias": "model-00039-of-00047.safetensors", + "model.layers.50.block_sparse_moe.gate.biases": "model-00038-of-00047.safetensors", + "model.layers.50.block_sparse_moe.gate.scales": "model-00038-of-00047.safetensors", + "model.layers.50.block_sparse_moe.gate.weight": "model-00038-of-00047.safetensors", + "model.layers.50.block_sparse_moe.switch_mlp.down_proj.biases": "model-00039-of-00047.safetensors", + "model.layers.50.block_sparse_moe.switch_mlp.down_proj.scales": "model-00039-of-00047.safetensors", + "model.layers.50.block_sparse_moe.switch_mlp.down_proj.weight": "model-00039-of-00047.safetensors", + "model.layers.50.block_sparse_moe.switch_mlp.gate_proj.biases": "model-00038-of-00047.safetensors", + "model.layers.50.block_sparse_moe.switch_mlp.gate_proj.scales": "model-00038-of-00047.safetensors", + "model.layers.50.block_sparse_moe.switch_mlp.gate_proj.weight": "model-00038-of-00047.safetensors", + "model.layers.50.block_sparse_moe.switch_mlp.up_proj.biases": "model-00038-of-00047.safetensors", + "model.layers.50.block_sparse_moe.switch_mlp.up_proj.scales": "model-00038-of-00047.safetensors", + "model.layers.50.block_sparse_moe.switch_mlp.up_proj.weight": "model-00038-of-00047.safetensors", + "model.layers.50.input_layernorm.weight": "model-00039-of-00047.safetensors", + "model.layers.50.post_attention_layernorm.weight": "model-00039-of-00047.safetensors", + "model.layers.50.self_attn.k_norm.weight": "model-00038-of-00047.safetensors", + "model.layers.50.self_attn.k_proj.biases": "model-00038-of-00047.safetensors", + "model.layers.50.self_attn.k_proj.scales": "model-00038-of-00047.safetensors", + "model.layers.50.self_attn.k_proj.weight": "model-00038-of-00047.safetensors", + "model.layers.50.self_attn.o_proj.biases": "model-00038-of-00047.safetensors", + "model.layers.50.self_attn.o_proj.scales": "model-00038-of-00047.safetensors", + "model.layers.50.self_attn.o_proj.weight": "model-00038-of-00047.safetensors", + "model.layers.50.self_attn.q_norm.weight": "model-00038-of-00047.safetensors", + "model.layers.50.self_attn.q_proj.biases": "model-00038-of-00047.safetensors", + "model.layers.50.self_attn.q_proj.scales": "model-00038-of-00047.safetensors", + "model.layers.50.self_attn.q_proj.weight": "model-00038-of-00047.safetensors", + "model.layers.50.self_attn.v_proj.biases": "model-00038-of-00047.safetensors", + "model.layers.50.self_attn.v_proj.scales": "model-00038-of-00047.safetensors", + "model.layers.50.self_attn.v_proj.weight": "model-00038-of-00047.safetensors", + "model.layers.51.block_sparse_moe.e_score_correction_bias": "model-00039-of-00047.safetensors", + "model.layers.51.block_sparse_moe.gate.biases": "model-00039-of-00047.safetensors", + "model.layers.51.block_sparse_moe.gate.scales": "model-00039-of-00047.safetensors", + "model.layers.51.block_sparse_moe.gate.weight": "model-00039-of-00047.safetensors", + "model.layers.51.block_sparse_moe.switch_mlp.down_proj.biases": "model-00039-of-00047.safetensors", + "model.layers.51.block_sparse_moe.switch_mlp.down_proj.scales": "model-00039-of-00047.safetensors", + "model.layers.51.block_sparse_moe.switch_mlp.down_proj.weight": "model-00039-of-00047.safetensors", + "model.layers.51.block_sparse_moe.switch_mlp.gate_proj.biases": "model-00039-of-00047.safetensors", + "model.layers.51.block_sparse_moe.switch_mlp.gate_proj.scales": "model-00039-of-00047.safetensors", + "model.layers.51.block_sparse_moe.switch_mlp.gate_proj.weight": "model-00039-of-00047.safetensors", + "model.layers.51.block_sparse_moe.switch_mlp.up_proj.biases": "model-00039-of-00047.safetensors", + "model.layers.51.block_sparse_moe.switch_mlp.up_proj.scales": "model-00039-of-00047.safetensors", + "model.layers.51.block_sparse_moe.switch_mlp.up_proj.weight": "model-00039-of-00047.safetensors", + "model.layers.51.input_layernorm.weight": "model-00039-of-00047.safetensors", + "model.layers.51.post_attention_layernorm.weight": "model-00039-of-00047.safetensors", + "model.layers.51.self_attn.k_norm.weight": "model-00039-of-00047.safetensors", + "model.layers.51.self_attn.k_proj.biases": "model-00039-of-00047.safetensors", + "model.layers.51.self_attn.k_proj.scales": "model-00039-of-00047.safetensors", + "model.layers.51.self_attn.k_proj.weight": "model-00039-of-00047.safetensors", + "model.layers.51.self_attn.o_proj.biases": "model-00039-of-00047.safetensors", + "model.layers.51.self_attn.o_proj.scales": "model-00039-of-00047.safetensors", + "model.layers.51.self_attn.o_proj.weight": "model-00039-of-00047.safetensors", + "model.layers.51.self_attn.q_norm.weight": "model-00039-of-00047.safetensors", + "model.layers.51.self_attn.q_proj.biases": "model-00039-of-00047.safetensors", + "model.layers.51.self_attn.q_proj.scales": "model-00039-of-00047.safetensors", + "model.layers.51.self_attn.q_proj.weight": "model-00039-of-00047.safetensors", + "model.layers.51.self_attn.v_proj.biases": "model-00039-of-00047.safetensors", + "model.layers.51.self_attn.v_proj.scales": "model-00039-of-00047.safetensors", + "model.layers.51.self_attn.v_proj.weight": "model-00039-of-00047.safetensors", + "model.layers.52.block_sparse_moe.e_score_correction_bias": "model-00040-of-00047.safetensors", + "model.layers.52.block_sparse_moe.gate.biases": "model-00039-of-00047.safetensors", + "model.layers.52.block_sparse_moe.gate.scales": "model-00039-of-00047.safetensors", + "model.layers.52.block_sparse_moe.gate.weight": "model-00039-of-00047.safetensors", + "model.layers.52.block_sparse_moe.switch_mlp.down_proj.biases": "model-00040-of-00047.safetensors", + "model.layers.52.block_sparse_moe.switch_mlp.down_proj.scales": "model-00040-of-00047.safetensors", + "model.layers.52.block_sparse_moe.switch_mlp.down_proj.weight": "model-00040-of-00047.safetensors", + "model.layers.52.block_sparse_moe.switch_mlp.gate_proj.biases": "model-00040-of-00047.safetensors", + "model.layers.52.block_sparse_moe.switch_mlp.gate_proj.scales": "model-00040-of-00047.safetensors", + "model.layers.52.block_sparse_moe.switch_mlp.gate_proj.weight": "model-00040-of-00047.safetensors", + "model.layers.52.block_sparse_moe.switch_mlp.up_proj.biases": "model-00040-of-00047.safetensors", + "model.layers.52.block_sparse_moe.switch_mlp.up_proj.scales": "model-00040-of-00047.safetensors", + "model.layers.52.block_sparse_moe.switch_mlp.up_proj.weight": "model-00040-of-00047.safetensors", + "model.layers.52.input_layernorm.weight": "model-00040-of-00047.safetensors", + "model.layers.52.post_attention_layernorm.weight": "model-00040-of-00047.safetensors", + "model.layers.52.self_attn.k_norm.weight": "model-00039-of-00047.safetensors", + "model.layers.52.self_attn.k_proj.biases": "model-00039-of-00047.safetensors", + "model.layers.52.self_attn.k_proj.scales": "model-00039-of-00047.safetensors", + "model.layers.52.self_attn.k_proj.weight": "model-00039-of-00047.safetensors", + "model.layers.52.self_attn.o_proj.biases": "model-00039-of-00047.safetensors", + "model.layers.52.self_attn.o_proj.scales": "model-00039-of-00047.safetensors", + "model.layers.52.self_attn.o_proj.weight": "model-00039-of-00047.safetensors", + "model.layers.52.self_attn.q_norm.weight": "model-00039-of-00047.safetensors", + "model.layers.52.self_attn.q_proj.biases": "model-00039-of-00047.safetensors", + "model.layers.52.self_attn.q_proj.scales": "model-00039-of-00047.safetensors", + "model.layers.52.self_attn.q_proj.weight": "model-00039-of-00047.safetensors", + "model.layers.52.self_attn.v_proj.biases": "model-00039-of-00047.safetensors", + "model.layers.52.self_attn.v_proj.scales": "model-00039-of-00047.safetensors", + "model.layers.52.self_attn.v_proj.weight": "model-00039-of-00047.safetensors", + "model.layers.53.block_sparse_moe.e_score_correction_bias": "model-00041-of-00047.safetensors", + "model.layers.53.block_sparse_moe.gate.biases": "model-00040-of-00047.safetensors", + "model.layers.53.block_sparse_moe.gate.scales": "model-00040-of-00047.safetensors", + "model.layers.53.block_sparse_moe.gate.weight": "model-00040-of-00047.safetensors", + "model.layers.53.block_sparse_moe.switch_mlp.down_proj.biases": "model-00041-of-00047.safetensors", + "model.layers.53.block_sparse_moe.switch_mlp.down_proj.scales": "model-00041-of-00047.safetensors", + "model.layers.53.block_sparse_moe.switch_mlp.down_proj.weight": "model-00041-of-00047.safetensors", + "model.layers.53.block_sparse_moe.switch_mlp.gate_proj.biases": "model-00040-of-00047.safetensors", + "model.layers.53.block_sparse_moe.switch_mlp.gate_proj.scales": "model-00040-of-00047.safetensors", + "model.layers.53.block_sparse_moe.switch_mlp.gate_proj.weight": "model-00040-of-00047.safetensors", + "model.layers.53.block_sparse_moe.switch_mlp.up_proj.biases": "model-00041-of-00047.safetensors", + "model.layers.53.block_sparse_moe.switch_mlp.up_proj.scales": "model-00041-of-00047.safetensors", + "model.layers.53.block_sparse_moe.switch_mlp.up_proj.weight": "model-00041-of-00047.safetensors", + "model.layers.53.input_layernorm.weight": "model-00041-of-00047.safetensors", + "model.layers.53.post_attention_layernorm.weight": "model-00041-of-00047.safetensors", + "model.layers.53.self_attn.k_norm.weight": "model-00040-of-00047.safetensors", + "model.layers.53.self_attn.k_proj.biases": "model-00040-of-00047.safetensors", + "model.layers.53.self_attn.k_proj.scales": "model-00040-of-00047.safetensors", + "model.layers.53.self_attn.k_proj.weight": "model-00040-of-00047.safetensors", + "model.layers.53.self_attn.o_proj.biases": "model-00040-of-00047.safetensors", + "model.layers.53.self_attn.o_proj.scales": "model-00040-of-00047.safetensors", + "model.layers.53.self_attn.o_proj.weight": "model-00040-of-00047.safetensors", + "model.layers.53.self_attn.q_norm.weight": "model-00040-of-00047.safetensors", + "model.layers.53.self_attn.q_proj.biases": "model-00040-of-00047.safetensors", + "model.layers.53.self_attn.q_proj.scales": "model-00040-of-00047.safetensors", + "model.layers.53.self_attn.q_proj.weight": "model-00040-of-00047.safetensors", + "model.layers.53.self_attn.v_proj.biases": "model-00040-of-00047.safetensors", + "model.layers.53.self_attn.v_proj.scales": "model-00040-of-00047.safetensors", + "model.layers.53.self_attn.v_proj.weight": "model-00040-of-00047.safetensors", + "model.layers.54.block_sparse_moe.e_score_correction_bias": "model-00042-of-00047.safetensors", + "model.layers.54.block_sparse_moe.gate.biases": "model-00041-of-00047.safetensors", + "model.layers.54.block_sparse_moe.gate.scales": "model-00041-of-00047.safetensors", + "model.layers.54.block_sparse_moe.gate.weight": "model-00041-of-00047.safetensors", + "model.layers.54.block_sparse_moe.switch_mlp.down_proj.biases": "model-00042-of-00047.safetensors", + "model.layers.54.block_sparse_moe.switch_mlp.down_proj.scales": "model-00042-of-00047.safetensors", + "model.layers.54.block_sparse_moe.switch_mlp.down_proj.weight": "model-00042-of-00047.safetensors", + "model.layers.54.block_sparse_moe.switch_mlp.gate_proj.biases": "model-00041-of-00047.safetensors", + "model.layers.54.block_sparse_moe.switch_mlp.gate_proj.scales": "model-00041-of-00047.safetensors", + "model.layers.54.block_sparse_moe.switch_mlp.gate_proj.weight": "model-00041-of-00047.safetensors", + "model.layers.54.block_sparse_moe.switch_mlp.up_proj.biases": "model-00041-of-00047.safetensors", + "model.layers.54.block_sparse_moe.switch_mlp.up_proj.scales": "model-00041-of-00047.safetensors", + "model.layers.54.block_sparse_moe.switch_mlp.up_proj.weight": "model-00041-of-00047.safetensors", + "model.layers.54.input_layernorm.weight": "model-00042-of-00047.safetensors", + "model.layers.54.post_attention_layernorm.weight": "model-00042-of-00047.safetensors", + "model.layers.54.self_attn.k_norm.weight": "model-00041-of-00047.safetensors", + "model.layers.54.self_attn.k_proj.biases": "model-00041-of-00047.safetensors", + "model.layers.54.self_attn.k_proj.scales": "model-00041-of-00047.safetensors", + "model.layers.54.self_attn.k_proj.weight": "model-00041-of-00047.safetensors", + "model.layers.54.self_attn.o_proj.biases": "model-00041-of-00047.safetensors", + "model.layers.54.self_attn.o_proj.scales": "model-00041-of-00047.safetensors", + "model.layers.54.self_attn.o_proj.weight": "model-00041-of-00047.safetensors", + "model.layers.54.self_attn.q_norm.weight": "model-00041-of-00047.safetensors", + "model.layers.54.self_attn.q_proj.biases": "model-00041-of-00047.safetensors", + "model.layers.54.self_attn.q_proj.scales": "model-00041-of-00047.safetensors", + "model.layers.54.self_attn.q_proj.weight": "model-00041-of-00047.safetensors", + "model.layers.54.self_attn.v_proj.biases": "model-00041-of-00047.safetensors", + "model.layers.54.self_attn.v_proj.scales": "model-00041-of-00047.safetensors", + "model.layers.54.self_attn.v_proj.weight": "model-00041-of-00047.safetensors", + "model.layers.55.block_sparse_moe.e_score_correction_bias": "model-00042-of-00047.safetensors", + "model.layers.55.block_sparse_moe.gate.biases": "model-00042-of-00047.safetensors", + "model.layers.55.block_sparse_moe.gate.scales": "model-00042-of-00047.safetensors", + "model.layers.55.block_sparse_moe.gate.weight": "model-00042-of-00047.safetensors", + "model.layers.55.block_sparse_moe.switch_mlp.down_proj.biases": "model-00042-of-00047.safetensors", + "model.layers.55.block_sparse_moe.switch_mlp.down_proj.scales": "model-00042-of-00047.safetensors", + "model.layers.55.block_sparse_moe.switch_mlp.down_proj.weight": "model-00042-of-00047.safetensors", + "model.layers.55.block_sparse_moe.switch_mlp.gate_proj.biases": "model-00042-of-00047.safetensors", + "model.layers.55.block_sparse_moe.switch_mlp.gate_proj.scales": "model-00042-of-00047.safetensors", + "model.layers.55.block_sparse_moe.switch_mlp.gate_proj.weight": "model-00042-of-00047.safetensors", + "model.layers.55.block_sparse_moe.switch_mlp.up_proj.biases": "model-00042-of-00047.safetensors", + "model.layers.55.block_sparse_moe.switch_mlp.up_proj.scales": "model-00042-of-00047.safetensors", + "model.layers.55.block_sparse_moe.switch_mlp.up_proj.weight": "model-00042-of-00047.safetensors", + "model.layers.55.input_layernorm.weight": "model-00042-of-00047.safetensors", + "model.layers.55.post_attention_layernorm.weight": "model-00042-of-00047.safetensors", + "model.layers.55.self_attn.k_norm.weight": "model-00042-of-00047.safetensors", + "model.layers.55.self_attn.k_proj.biases": "model-00042-of-00047.safetensors", + "model.layers.55.self_attn.k_proj.scales": "model-00042-of-00047.safetensors", + "model.layers.55.self_attn.k_proj.weight": "model-00042-of-00047.safetensors", + "model.layers.55.self_attn.o_proj.biases": "model-00042-of-00047.safetensors", + "model.layers.55.self_attn.o_proj.scales": "model-00042-of-00047.safetensors", + "model.layers.55.self_attn.o_proj.weight": "model-00042-of-00047.safetensors", + "model.layers.55.self_attn.q_norm.weight": "model-00042-of-00047.safetensors", + "model.layers.55.self_attn.q_proj.biases": "model-00042-of-00047.safetensors", + "model.layers.55.self_attn.q_proj.scales": "model-00042-of-00047.safetensors", + "model.layers.55.self_attn.q_proj.weight": "model-00042-of-00047.safetensors", + "model.layers.55.self_attn.v_proj.biases": "model-00042-of-00047.safetensors", + "model.layers.55.self_attn.v_proj.scales": "model-00042-of-00047.safetensors", + "model.layers.55.self_attn.v_proj.weight": "model-00042-of-00047.safetensors", + "model.layers.56.block_sparse_moe.e_score_correction_bias": "model-00043-of-00047.safetensors", + "model.layers.56.block_sparse_moe.gate.biases": "model-00042-of-00047.safetensors", + "model.layers.56.block_sparse_moe.gate.scales": "model-00042-of-00047.safetensors", + "model.layers.56.block_sparse_moe.gate.weight": "model-00042-of-00047.safetensors", + "model.layers.56.block_sparse_moe.switch_mlp.down_proj.biases": "model-00043-of-00047.safetensors", + "model.layers.56.block_sparse_moe.switch_mlp.down_proj.scales": "model-00043-of-00047.safetensors", + "model.layers.56.block_sparse_moe.switch_mlp.down_proj.weight": "model-00043-of-00047.safetensors", + "model.layers.56.block_sparse_moe.switch_mlp.gate_proj.biases": "model-00043-of-00047.safetensors", + "model.layers.56.block_sparse_moe.switch_mlp.gate_proj.scales": "model-00043-of-00047.safetensors", + "model.layers.56.block_sparse_moe.switch_mlp.gate_proj.weight": "model-00043-of-00047.safetensors", + "model.layers.56.block_sparse_moe.switch_mlp.up_proj.biases": "model-00043-of-00047.safetensors", + "model.layers.56.block_sparse_moe.switch_mlp.up_proj.scales": "model-00043-of-00047.safetensors", + "model.layers.56.block_sparse_moe.switch_mlp.up_proj.weight": "model-00043-of-00047.safetensors", + "model.layers.56.input_layernorm.weight": "model-00043-of-00047.safetensors", + "model.layers.56.post_attention_layernorm.weight": "model-00043-of-00047.safetensors", + "model.layers.56.self_attn.k_norm.weight": "model-00042-of-00047.safetensors", + "model.layers.56.self_attn.k_proj.biases": "model-00042-of-00047.safetensors", + "model.layers.56.self_attn.k_proj.scales": "model-00042-of-00047.safetensors", + "model.layers.56.self_attn.k_proj.weight": "model-00042-of-00047.safetensors", + "model.layers.56.self_attn.o_proj.biases": "model-00042-of-00047.safetensors", + "model.layers.56.self_attn.o_proj.scales": "model-00042-of-00047.safetensors", + "model.layers.56.self_attn.o_proj.weight": "model-00042-of-00047.safetensors", + "model.layers.56.self_attn.q_norm.weight": "model-00042-of-00047.safetensors", + "model.layers.56.self_attn.q_proj.biases": "model-00042-of-00047.safetensors", + "model.layers.56.self_attn.q_proj.scales": "model-00042-of-00047.safetensors", + "model.layers.56.self_attn.q_proj.weight": "model-00042-of-00047.safetensors", + "model.layers.56.self_attn.v_proj.biases": "model-00042-of-00047.safetensors", + "model.layers.56.self_attn.v_proj.scales": "model-00042-of-00047.safetensors", + "model.layers.56.self_attn.v_proj.weight": "model-00042-of-00047.safetensors", + "model.layers.57.block_sparse_moe.e_score_correction_bias": "model-00044-of-00047.safetensors", + "model.layers.57.block_sparse_moe.gate.biases": "model-00043-of-00047.safetensors", + "model.layers.57.block_sparse_moe.gate.scales": "model-00043-of-00047.safetensors", + "model.layers.57.block_sparse_moe.gate.weight": "model-00043-of-00047.safetensors", + "model.layers.57.block_sparse_moe.switch_mlp.down_proj.biases": "model-00044-of-00047.safetensors", + "model.layers.57.block_sparse_moe.switch_mlp.down_proj.scales": "model-00044-of-00047.safetensors", + "model.layers.57.block_sparse_moe.switch_mlp.down_proj.weight": "model-00044-of-00047.safetensors", + "model.layers.57.block_sparse_moe.switch_mlp.gate_proj.biases": "model-00043-of-00047.safetensors", + "model.layers.57.block_sparse_moe.switch_mlp.gate_proj.scales": "model-00043-of-00047.safetensors", + "model.layers.57.block_sparse_moe.switch_mlp.gate_proj.weight": "model-00043-of-00047.safetensors", + "model.layers.57.block_sparse_moe.switch_mlp.up_proj.biases": "model-00044-of-00047.safetensors", + "model.layers.57.block_sparse_moe.switch_mlp.up_proj.scales": "model-00044-of-00047.safetensors", + "model.layers.57.block_sparse_moe.switch_mlp.up_proj.weight": "model-00044-of-00047.safetensors", + "model.layers.57.input_layernorm.weight": "model-00044-of-00047.safetensors", + "model.layers.57.post_attention_layernorm.weight": "model-00044-of-00047.safetensors", + "model.layers.57.self_attn.k_norm.weight": "model-00043-of-00047.safetensors", + "model.layers.57.self_attn.k_proj.biases": "model-00043-of-00047.safetensors", + "model.layers.57.self_attn.k_proj.scales": "model-00043-of-00047.safetensors", + "model.layers.57.self_attn.k_proj.weight": "model-00043-of-00047.safetensors", + "model.layers.57.self_attn.o_proj.biases": "model-00043-of-00047.safetensors", + "model.layers.57.self_attn.o_proj.scales": "model-00043-of-00047.safetensors", + "model.layers.57.self_attn.o_proj.weight": "model-00043-of-00047.safetensors", + "model.layers.57.self_attn.q_norm.weight": "model-00043-of-00047.safetensors", + "model.layers.57.self_attn.q_proj.biases": "model-00043-of-00047.safetensors", + "model.layers.57.self_attn.q_proj.scales": "model-00043-of-00047.safetensors", + "model.layers.57.self_attn.q_proj.weight": "model-00043-of-00047.safetensors", + "model.layers.57.self_attn.v_proj.biases": "model-00043-of-00047.safetensors", + "model.layers.57.self_attn.v_proj.scales": "model-00043-of-00047.safetensors", + "model.layers.57.self_attn.v_proj.weight": "model-00043-of-00047.safetensors", + "model.layers.58.block_sparse_moe.e_score_correction_bias": "model-00045-of-00047.safetensors", + "model.layers.58.block_sparse_moe.gate.biases": "model-00044-of-00047.safetensors", + "model.layers.58.block_sparse_moe.gate.scales": "model-00044-of-00047.safetensors", + "model.layers.58.block_sparse_moe.gate.weight": "model-00044-of-00047.safetensors", + "model.layers.58.block_sparse_moe.switch_mlp.down_proj.biases": "model-00045-of-00047.safetensors", + "model.layers.58.block_sparse_moe.switch_mlp.down_proj.scales": "model-00045-of-00047.safetensors", + "model.layers.58.block_sparse_moe.switch_mlp.down_proj.weight": "model-00045-of-00047.safetensors", + "model.layers.58.block_sparse_moe.switch_mlp.gate_proj.biases": "model-00044-of-00047.safetensors", + "model.layers.58.block_sparse_moe.switch_mlp.gate_proj.scales": "model-00044-of-00047.safetensors", + "model.layers.58.block_sparse_moe.switch_mlp.gate_proj.weight": "model-00044-of-00047.safetensors", + "model.layers.58.block_sparse_moe.switch_mlp.up_proj.biases": "model-00044-of-00047.safetensors", + "model.layers.58.block_sparse_moe.switch_mlp.up_proj.scales": "model-00044-of-00047.safetensors", + "model.layers.58.block_sparse_moe.switch_mlp.up_proj.weight": "model-00044-of-00047.safetensors", + "model.layers.58.input_layernorm.weight": "model-00045-of-00047.safetensors", + "model.layers.58.post_attention_layernorm.weight": "model-00045-of-00047.safetensors", + "model.layers.58.self_attn.k_norm.weight": "model-00044-of-00047.safetensors", + "model.layers.58.self_attn.k_proj.biases": "model-00044-of-00047.safetensors", + "model.layers.58.self_attn.k_proj.scales": "model-00044-of-00047.safetensors", + "model.layers.58.self_attn.k_proj.weight": "model-00044-of-00047.safetensors", + "model.layers.58.self_attn.o_proj.biases": "model-00044-of-00047.safetensors", + "model.layers.58.self_attn.o_proj.scales": "model-00044-of-00047.safetensors", + "model.layers.58.self_attn.o_proj.weight": "model-00044-of-00047.safetensors", + "model.layers.58.self_attn.q_norm.weight": "model-00044-of-00047.safetensors", + "model.layers.58.self_attn.q_proj.biases": "model-00044-of-00047.safetensors", + "model.layers.58.self_attn.q_proj.scales": "model-00044-of-00047.safetensors", + "model.layers.58.self_attn.q_proj.weight": "model-00044-of-00047.safetensors", + "model.layers.58.self_attn.v_proj.biases": "model-00044-of-00047.safetensors", + "model.layers.58.self_attn.v_proj.scales": "model-00044-of-00047.safetensors", + "model.layers.58.self_attn.v_proj.weight": "model-00044-of-00047.safetensors", + "model.layers.59.block_sparse_moe.e_score_correction_bias": "model-00045-of-00047.safetensors", + "model.layers.59.block_sparse_moe.gate.biases": "model-00045-of-00047.safetensors", + "model.layers.59.block_sparse_moe.gate.scales": "model-00045-of-00047.safetensors", + "model.layers.59.block_sparse_moe.gate.weight": "model-00045-of-00047.safetensors", + "model.layers.59.block_sparse_moe.switch_mlp.down_proj.biases": "model-00045-of-00047.safetensors", + "model.layers.59.block_sparse_moe.switch_mlp.down_proj.scales": "model-00045-of-00047.safetensors", + "model.layers.59.block_sparse_moe.switch_mlp.down_proj.weight": "model-00045-of-00047.safetensors", + "model.layers.59.block_sparse_moe.switch_mlp.gate_proj.biases": "model-00045-of-00047.safetensors", + "model.layers.59.block_sparse_moe.switch_mlp.gate_proj.scales": "model-00045-of-00047.safetensors", + "model.layers.59.block_sparse_moe.switch_mlp.gate_proj.weight": "model-00045-of-00047.safetensors", + "model.layers.59.block_sparse_moe.switch_mlp.up_proj.biases": "model-00045-of-00047.safetensors", + "model.layers.59.block_sparse_moe.switch_mlp.up_proj.scales": "model-00045-of-00047.safetensors", + "model.layers.59.block_sparse_moe.switch_mlp.up_proj.weight": "model-00045-of-00047.safetensors", + "model.layers.59.input_layernorm.weight": "model-00045-of-00047.safetensors", + "model.layers.59.post_attention_layernorm.weight": "model-00045-of-00047.safetensors", + "model.layers.59.self_attn.k_norm.weight": "model-00045-of-00047.safetensors", + "model.layers.59.self_attn.k_proj.biases": "model-00045-of-00047.safetensors", + "model.layers.59.self_attn.k_proj.scales": "model-00045-of-00047.safetensors", + "model.layers.59.self_attn.k_proj.weight": "model-00045-of-00047.safetensors", + "model.layers.59.self_attn.o_proj.biases": "model-00045-of-00047.safetensors", + "model.layers.59.self_attn.o_proj.scales": "model-00045-of-00047.safetensors", + "model.layers.59.self_attn.o_proj.weight": "model-00045-of-00047.safetensors", + "model.layers.59.self_attn.q_norm.weight": "model-00045-of-00047.safetensors", + "model.layers.59.self_attn.q_proj.biases": "model-00045-of-00047.safetensors", + "model.layers.59.self_attn.q_proj.scales": "model-00045-of-00047.safetensors", + "model.layers.59.self_attn.q_proj.weight": "model-00045-of-00047.safetensors", + "model.layers.59.self_attn.v_proj.biases": "model-00045-of-00047.safetensors", + "model.layers.59.self_attn.v_proj.scales": "model-00045-of-00047.safetensors", + "model.layers.59.self_attn.v_proj.weight": "model-00045-of-00047.safetensors", + "model.layers.6.block_sparse_moe.e_score_correction_bias": "model-00006-of-00047.safetensors", + "model.layers.6.block_sparse_moe.gate.biases": "model-00005-of-00047.safetensors", + "model.layers.6.block_sparse_moe.gate.scales": "model-00005-of-00047.safetensors", + "model.layers.6.block_sparse_moe.gate.weight": "model-00005-of-00047.safetensors", + "model.layers.6.block_sparse_moe.switch_mlp.down_proj.biases": "model-00006-of-00047.safetensors", + "model.layers.6.block_sparse_moe.switch_mlp.down_proj.scales": "model-00006-of-00047.safetensors", + "model.layers.6.block_sparse_moe.switch_mlp.down_proj.weight": "model-00006-of-00047.safetensors", + "model.layers.6.block_sparse_moe.switch_mlp.gate_proj.biases": "model-00005-of-00047.safetensors", + "model.layers.6.block_sparse_moe.switch_mlp.gate_proj.scales": "model-00005-of-00047.safetensors", + "model.layers.6.block_sparse_moe.switch_mlp.gate_proj.weight": "model-00005-of-00047.safetensors", + "model.layers.6.block_sparse_moe.switch_mlp.up_proj.biases": "model-00005-of-00047.safetensors", + "model.layers.6.block_sparse_moe.switch_mlp.up_proj.scales": "model-00005-of-00047.safetensors", + "model.layers.6.block_sparse_moe.switch_mlp.up_proj.weight": "model-00005-of-00047.safetensors", + "model.layers.6.input_layernorm.weight": "model-00006-of-00047.safetensors", + "model.layers.6.post_attention_layernorm.weight": "model-00006-of-00047.safetensors", + "model.layers.6.self_attn.k_norm.weight": "model-00005-of-00047.safetensors", + "model.layers.6.self_attn.k_proj.biases": "model-00005-of-00047.safetensors", + "model.layers.6.self_attn.k_proj.scales": "model-00005-of-00047.safetensors", + "model.layers.6.self_attn.k_proj.weight": "model-00005-of-00047.safetensors", + "model.layers.6.self_attn.o_proj.biases": "model-00005-of-00047.safetensors", + "model.layers.6.self_attn.o_proj.scales": "model-00005-of-00047.safetensors", + "model.layers.6.self_attn.o_proj.weight": "model-00005-of-00047.safetensors", + "model.layers.6.self_attn.q_norm.weight": "model-00005-of-00047.safetensors", + "model.layers.6.self_attn.q_proj.biases": "model-00005-of-00047.safetensors", + "model.layers.6.self_attn.q_proj.scales": "model-00005-of-00047.safetensors", + "model.layers.6.self_attn.q_proj.weight": "model-00005-of-00047.safetensors", + "model.layers.6.self_attn.v_proj.biases": "model-00005-of-00047.safetensors", + "model.layers.6.self_attn.v_proj.scales": "model-00005-of-00047.safetensors", + "model.layers.6.self_attn.v_proj.weight": "model-00005-of-00047.safetensors", + "model.layers.60.block_sparse_moe.e_score_correction_bias": "model-00046-of-00047.safetensors", + "model.layers.60.block_sparse_moe.gate.biases": "model-00045-of-00047.safetensors", + "model.layers.60.block_sparse_moe.gate.scales": "model-00045-of-00047.safetensors", + "model.layers.60.block_sparse_moe.gate.weight": "model-00045-of-00047.safetensors", + "model.layers.60.block_sparse_moe.switch_mlp.down_proj.biases": "model-00046-of-00047.safetensors", + "model.layers.60.block_sparse_moe.switch_mlp.down_proj.scales": "model-00046-of-00047.safetensors", + "model.layers.60.block_sparse_moe.switch_mlp.down_proj.weight": "model-00046-of-00047.safetensors", + "model.layers.60.block_sparse_moe.switch_mlp.gate_proj.biases": "model-00046-of-00047.safetensors", + "model.layers.60.block_sparse_moe.switch_mlp.gate_proj.scales": "model-00046-of-00047.safetensors", + "model.layers.60.block_sparse_moe.switch_mlp.gate_proj.weight": "model-00046-of-00047.safetensors", + "model.layers.60.block_sparse_moe.switch_mlp.up_proj.biases": "model-00046-of-00047.safetensors", + "model.layers.60.block_sparse_moe.switch_mlp.up_proj.scales": "model-00046-of-00047.safetensors", + "model.layers.60.block_sparse_moe.switch_mlp.up_proj.weight": "model-00046-of-00047.safetensors", + "model.layers.60.input_layernorm.weight": "model-00046-of-00047.safetensors", + "model.layers.60.post_attention_layernorm.weight": "model-00046-of-00047.safetensors", + "model.layers.60.self_attn.k_norm.weight": "model-00045-of-00047.safetensors", + "model.layers.60.self_attn.k_proj.biases": "model-00045-of-00047.safetensors", + "model.layers.60.self_attn.k_proj.scales": "model-00045-of-00047.safetensors", + "model.layers.60.self_attn.k_proj.weight": "model-00045-of-00047.safetensors", + "model.layers.60.self_attn.o_proj.biases": "model-00045-of-00047.safetensors", + "model.layers.60.self_attn.o_proj.scales": "model-00045-of-00047.safetensors", + "model.layers.60.self_attn.o_proj.weight": "model-00045-of-00047.safetensors", + "model.layers.60.self_attn.q_norm.weight": "model-00045-of-00047.safetensors", + "model.layers.60.self_attn.q_proj.biases": "model-00045-of-00047.safetensors", + "model.layers.60.self_attn.q_proj.scales": "model-00045-of-00047.safetensors", + "model.layers.60.self_attn.q_proj.weight": "model-00045-of-00047.safetensors", + "model.layers.60.self_attn.v_proj.biases": "model-00045-of-00047.safetensors", + "model.layers.60.self_attn.v_proj.scales": "model-00045-of-00047.safetensors", + "model.layers.60.self_attn.v_proj.weight": "model-00045-of-00047.safetensors", + "model.layers.61.block_sparse_moe.e_score_correction_bias": "model-00047-of-00047.safetensors", + "model.layers.61.block_sparse_moe.gate.biases": "model-00046-of-00047.safetensors", + "model.layers.61.block_sparse_moe.gate.scales": "model-00046-of-00047.safetensors", + "model.layers.61.block_sparse_moe.gate.weight": "model-00046-of-00047.safetensors", + "model.layers.61.block_sparse_moe.switch_mlp.down_proj.biases": "model-00047-of-00047.safetensors", + "model.layers.61.block_sparse_moe.switch_mlp.down_proj.scales": "model-00047-of-00047.safetensors", + "model.layers.61.block_sparse_moe.switch_mlp.down_proj.weight": "model-00047-of-00047.safetensors", + "model.layers.61.block_sparse_moe.switch_mlp.gate_proj.biases": "model-00046-of-00047.safetensors", + "model.layers.61.block_sparse_moe.switch_mlp.gate_proj.scales": "model-00046-of-00047.safetensors", + "model.layers.61.block_sparse_moe.switch_mlp.gate_proj.weight": "model-00046-of-00047.safetensors", + "model.layers.61.block_sparse_moe.switch_mlp.up_proj.biases": "model-00047-of-00047.safetensors", + "model.layers.61.block_sparse_moe.switch_mlp.up_proj.scales": "model-00047-of-00047.safetensors", + "model.layers.61.block_sparse_moe.switch_mlp.up_proj.weight": "model-00047-of-00047.safetensors", + "model.layers.61.input_layernorm.weight": "model-00047-of-00047.safetensors", + "model.layers.61.post_attention_layernorm.weight": "model-00047-of-00047.safetensors", + "model.layers.61.self_attn.k_norm.weight": "model-00046-of-00047.safetensors", + "model.layers.61.self_attn.k_proj.biases": "model-00046-of-00047.safetensors", + "model.layers.61.self_attn.k_proj.scales": "model-00046-of-00047.safetensors", + "model.layers.61.self_attn.k_proj.weight": "model-00046-of-00047.safetensors", + "model.layers.61.self_attn.o_proj.biases": "model-00046-of-00047.safetensors", + "model.layers.61.self_attn.o_proj.scales": "model-00046-of-00047.safetensors", + "model.layers.61.self_attn.o_proj.weight": "model-00046-of-00047.safetensors", + "model.layers.61.self_attn.q_norm.weight": "model-00046-of-00047.safetensors", + "model.layers.61.self_attn.q_proj.biases": "model-00046-of-00047.safetensors", + "model.layers.61.self_attn.q_proj.scales": "model-00046-of-00047.safetensors", + "model.layers.61.self_attn.q_proj.weight": "model-00046-of-00047.safetensors", + "model.layers.61.self_attn.v_proj.biases": "model-00046-of-00047.safetensors", + "model.layers.61.self_attn.v_proj.scales": "model-00046-of-00047.safetensors", + "model.layers.61.self_attn.v_proj.weight": "model-00046-of-00047.safetensors", + "model.layers.7.block_sparse_moe.e_score_correction_bias": "model-00006-of-00047.safetensors", + "model.layers.7.block_sparse_moe.gate.biases": "model-00006-of-00047.safetensors", + "model.layers.7.block_sparse_moe.gate.scales": "model-00006-of-00047.safetensors", + "model.layers.7.block_sparse_moe.gate.weight": "model-00006-of-00047.safetensors", + "model.layers.7.block_sparse_moe.switch_mlp.down_proj.biases": "model-00006-of-00047.safetensors", + "model.layers.7.block_sparse_moe.switch_mlp.down_proj.scales": "model-00006-of-00047.safetensors", + "model.layers.7.block_sparse_moe.switch_mlp.down_proj.weight": "model-00006-of-00047.safetensors", + "model.layers.7.block_sparse_moe.switch_mlp.gate_proj.biases": "model-00006-of-00047.safetensors", + "model.layers.7.block_sparse_moe.switch_mlp.gate_proj.scales": "model-00006-of-00047.safetensors", + "model.layers.7.block_sparse_moe.switch_mlp.gate_proj.weight": "model-00006-of-00047.safetensors", + "model.layers.7.block_sparse_moe.switch_mlp.up_proj.biases": "model-00006-of-00047.safetensors", + "model.layers.7.block_sparse_moe.switch_mlp.up_proj.scales": "model-00006-of-00047.safetensors", + "model.layers.7.block_sparse_moe.switch_mlp.up_proj.weight": "model-00006-of-00047.safetensors", + "model.layers.7.input_layernorm.weight": "model-00006-of-00047.safetensors", + "model.layers.7.post_attention_layernorm.weight": "model-00006-of-00047.safetensors", + "model.layers.7.self_attn.k_norm.weight": "model-00006-of-00047.safetensors", + "model.layers.7.self_attn.k_proj.biases": "model-00006-of-00047.safetensors", + "model.layers.7.self_attn.k_proj.scales": "model-00006-of-00047.safetensors", + "model.layers.7.self_attn.k_proj.weight": "model-00006-of-00047.safetensors", + "model.layers.7.self_attn.o_proj.biases": "model-00006-of-00047.safetensors", + "model.layers.7.self_attn.o_proj.scales": "model-00006-of-00047.safetensors", + "model.layers.7.self_attn.o_proj.weight": "model-00006-of-00047.safetensors", + "model.layers.7.self_attn.q_norm.weight": "model-00006-of-00047.safetensors", + "model.layers.7.self_attn.q_proj.biases": "model-00006-of-00047.safetensors", + "model.layers.7.self_attn.q_proj.scales": "model-00006-of-00047.safetensors", + "model.layers.7.self_attn.q_proj.weight": "model-00006-of-00047.safetensors", + "model.layers.7.self_attn.v_proj.biases": "model-00006-of-00047.safetensors", + "model.layers.7.self_attn.v_proj.scales": "model-00006-of-00047.safetensors", + "model.layers.7.self_attn.v_proj.weight": "model-00006-of-00047.safetensors", + "model.layers.8.block_sparse_moe.e_score_correction_bias": "model-00007-of-00047.safetensors", + "model.layers.8.block_sparse_moe.gate.biases": "model-00006-of-00047.safetensors", + "model.layers.8.block_sparse_moe.gate.scales": "model-00006-of-00047.safetensors", + "model.layers.8.block_sparse_moe.gate.weight": "model-00006-of-00047.safetensors", + "model.layers.8.block_sparse_moe.switch_mlp.down_proj.biases": "model-00007-of-00047.safetensors", + "model.layers.8.block_sparse_moe.switch_mlp.down_proj.scales": "model-00007-of-00047.safetensors", + "model.layers.8.block_sparse_moe.switch_mlp.down_proj.weight": "model-00007-of-00047.safetensors", + "model.layers.8.block_sparse_moe.switch_mlp.gate_proj.biases": "model-00007-of-00047.safetensors", + "model.layers.8.block_sparse_moe.switch_mlp.gate_proj.scales": "model-00007-of-00047.safetensors", + "model.layers.8.block_sparse_moe.switch_mlp.gate_proj.weight": "model-00007-of-00047.safetensors", + "model.layers.8.block_sparse_moe.switch_mlp.up_proj.biases": "model-00007-of-00047.safetensors", + "model.layers.8.block_sparse_moe.switch_mlp.up_proj.scales": "model-00007-of-00047.safetensors", + "model.layers.8.block_sparse_moe.switch_mlp.up_proj.weight": "model-00007-of-00047.safetensors", + "model.layers.8.input_layernorm.weight": "model-00007-of-00047.safetensors", + "model.layers.8.post_attention_layernorm.weight": "model-00007-of-00047.safetensors", + "model.layers.8.self_attn.k_norm.weight": "model-00006-of-00047.safetensors", + "model.layers.8.self_attn.k_proj.biases": "model-00006-of-00047.safetensors", + "model.layers.8.self_attn.k_proj.scales": "model-00006-of-00047.safetensors", + "model.layers.8.self_attn.k_proj.weight": "model-00006-of-00047.safetensors", + "model.layers.8.self_attn.o_proj.biases": "model-00006-of-00047.safetensors", + "model.layers.8.self_attn.o_proj.scales": "model-00006-of-00047.safetensors", + "model.layers.8.self_attn.o_proj.weight": "model-00006-of-00047.safetensors", + "model.layers.8.self_attn.q_norm.weight": "model-00006-of-00047.safetensors", + "model.layers.8.self_attn.q_proj.biases": "model-00006-of-00047.safetensors", + "model.layers.8.self_attn.q_proj.scales": "model-00006-of-00047.safetensors", + "model.layers.8.self_attn.q_proj.weight": "model-00006-of-00047.safetensors", + "model.layers.8.self_attn.v_proj.biases": "model-00006-of-00047.safetensors", + "model.layers.8.self_attn.v_proj.scales": "model-00006-of-00047.safetensors", + "model.layers.8.self_attn.v_proj.weight": "model-00006-of-00047.safetensors", + "model.layers.9.block_sparse_moe.e_score_correction_bias": "model-00008-of-00047.safetensors", + "model.layers.9.block_sparse_moe.gate.biases": "model-00007-of-00047.safetensors", + "model.layers.9.block_sparse_moe.gate.scales": "model-00007-of-00047.safetensors", + "model.layers.9.block_sparse_moe.gate.weight": "model-00007-of-00047.safetensors", + "model.layers.9.block_sparse_moe.switch_mlp.down_proj.biases": "model-00008-of-00047.safetensors", + "model.layers.9.block_sparse_moe.switch_mlp.down_proj.scales": "model-00008-of-00047.safetensors", + "model.layers.9.block_sparse_moe.switch_mlp.down_proj.weight": "model-00008-of-00047.safetensors", + "model.layers.9.block_sparse_moe.switch_mlp.gate_proj.biases": "model-00007-of-00047.safetensors", + "model.layers.9.block_sparse_moe.switch_mlp.gate_proj.scales": "model-00007-of-00047.safetensors", + "model.layers.9.block_sparse_moe.switch_mlp.gate_proj.weight": "model-00007-of-00047.safetensors", + "model.layers.9.block_sparse_moe.switch_mlp.up_proj.biases": "model-00008-of-00047.safetensors", + "model.layers.9.block_sparse_moe.switch_mlp.up_proj.scales": "model-00008-of-00047.safetensors", + "model.layers.9.block_sparse_moe.switch_mlp.up_proj.weight": "model-00008-of-00047.safetensors", + "model.layers.9.input_layernorm.weight": "model-00008-of-00047.safetensors", + "model.layers.9.post_attention_layernorm.weight": "model-00008-of-00047.safetensors", + "model.layers.9.self_attn.k_norm.weight": "model-00007-of-00047.safetensors", + "model.layers.9.self_attn.k_proj.biases": "model-00007-of-00047.safetensors", + "model.layers.9.self_attn.k_proj.scales": "model-00007-of-00047.safetensors", + "model.layers.9.self_attn.k_proj.weight": "model-00007-of-00047.safetensors", + "model.layers.9.self_attn.o_proj.biases": "model-00007-of-00047.safetensors", + "model.layers.9.self_attn.o_proj.scales": "model-00007-of-00047.safetensors", + "model.layers.9.self_attn.o_proj.weight": "model-00007-of-00047.safetensors", + "model.layers.9.self_attn.q_norm.weight": "model-00007-of-00047.safetensors", + "model.layers.9.self_attn.q_proj.biases": "model-00007-of-00047.safetensors", + "model.layers.9.self_attn.q_proj.scales": "model-00007-of-00047.safetensors", + "model.layers.9.self_attn.q_proj.weight": "model-00007-of-00047.safetensors", + "model.layers.9.self_attn.v_proj.biases": "model-00007-of-00047.safetensors", + "model.layers.9.self_attn.v_proj.scales": "model-00007-of-00047.safetensors", + "model.layers.9.self_attn.v_proj.weight": "model-00007-of-00047.safetensors", + "model.norm.weight": "model-00047-of-00047.safetensors" + } +} \ No newline at end of file diff --git a/modeling_minimax_m2.py b/modeling_minimax_m2.py new file mode 100644 index 0000000000000000000000000000000000000000..8846d38acc932d1dcb0302bb719296313f5225a8 --- /dev/null +++ b/modeling_minimax_m2.py @@ -0,0 +1,706 @@ +# 🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨 +# This file was automatically generated from src/transformers/models/minimax_m2/modular_minimax_m2.py. +# Do NOT edit this file manually as any edits will be overwritten by the generation of +# the file from the modular. If any change should be done, please apply the change to the +# modular_minimax_m2.py file directly. One of our CI enforces this. +# 🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨 +# coding=utf-8 +# Copyright 2025 the HuggingFace Team. All rights reserved. +# +# Licensed under the Apache License, Version 2.0 (the "License"); +# you may not use this file except in compliance with the License. +# You may obtain a copy of the License at +# +# http://www.apache.org/licenses/LICENSE-2.0 +# +# Unless required by applicable law or agreed to in writing, software +# distributed under the License is distributed on an "AS IS" BASIS, +# WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. +# See the License for the specific language governing permissions and +# limitations under the License. + + +from collections.abc import Callable +from typing import Optional, Union, Unpack + +import torch +from torch import nn + +from transformers.activations import ACT2FN +from transformers.cache_utils import Cache, DynamicCache +from transformers.generation import GenerationMixin +from transformers.integrations import use_kernel_forward_from_hub +from transformers.masking_utils import create_causal_mask, create_sliding_window_causal_mask +from transformers.modeling_flash_attention_utils import FlashAttentionKwargs +from transformers.modeling_layers import ( + GenericForQuestionAnswering, + GenericForSequenceClassification, + GenericForTokenClassification, + GradientCheckpointingLayer, +) +from transformers.modeling_outputs import MoeCausalLMOutputWithPast, MoeModelOutputWithPast +from transformers.modeling_rope_utils import ROPE_INIT_FUNCTIONS, dynamic_rope_update +from transformers.modeling_utils import ALL_ATTENTION_FUNCTIONS, PreTrainedModel +from transformers.utils import TransformersKwargs, auto_docstring, can_return_tuple +from transformers.utils.deprecation import deprecate_kwarg +from transformers.utils.generic import OutputRecorder, check_model_inputs +from .configuration_minimax_m2 import MiniMaxM2Config + + +class MiniMaxM2MLP(nn.Module): + def __init__(self, config: MiniMaxM2Config): + super().__init__() + self.ffn_dim = config.intermediate_size + self.hidden_dim = config.hidden_size + + self.w1 = nn.Linear(self.hidden_dim, self.ffn_dim, bias=False) + self.w2 = nn.Linear(self.ffn_dim, self.hidden_dim, bias=False) + self.w3 = nn.Linear(self.hidden_dim, self.ffn_dim, bias=False) + + self.act_fn = ACT2FN[config.hidden_act] + + def forward(self, hidden_states): + current_hidden_states = self.act_fn(self.w1(hidden_states)) * self.w3(hidden_states) + current_hidden_states = self.w2(current_hidden_states) + return current_hidden_states + + +class MiniMaxM2Experts(nn.ModuleList): + """ + ModuleList of experts. + """ + + def __init__(self, config: MiniMaxM2Config): + super().__init__() + self.top_k = config.num_experts_per_tok + self.num_experts = config.num_local_experts + for _ in range(self.num_experts): + self.append(MiniMaxM2MLP(config)) + + def forward( + self, hidden_states: torch.Tensor, top_k_index: torch.Tensor, top_k_weights: torch.Tensor + ) -> torch.Tensor: + """ + Args: + hidden_states: (batch_size * sequence_length, hidden_dim) + selected_experts: (batch_size * sequence_length, top_k) + routing_weights: (batch_size * sequence_length, top_k) + Returns: + (batch_size * sequence_length, hidden_dim) + """ + final_hidden_states = torch.zeros_like(hidden_states) + expert_mask = torch.nn.functional.one_hot(top_k_index, num_classes=self.num_experts).permute(2, 1, 0) + + expert_hit = torch.greater(expert_mask.sum(dim=(-1, -2)), 0).nonzero() + for expert_idx in expert_hit: + idx, top_x = torch.where(expert_mask[expert_idx].squeeze(0)) + current_state = hidden_states[None, top_x].reshape(-1, hidden_states.shape[-1]) + current_hidden_states = self[expert_idx](current_state) * top_k_weights[top_x, idx, None] + final_hidden_states.index_add_(0, top_x, current_hidden_states.to(hidden_states.dtype)) + return final_hidden_states + + +class MiniMaxM2SparseMoeBlock(nn.Module): + def __init__(self, config): + super().__init__() + self.top_k = config.num_experts_per_tok + self.jitter_noise = config.router_jitter_noise + self.gate = nn.Linear(config.hidden_size, config.num_local_experts, bias=False) + self.experts = MiniMaxM2Experts(config) + self.register_buffer("e_score_correction_bias", torch.zeros(config.num_local_experts)) + + def route_tokens_to_experts(self, router_logits): + routing_weights = torch.nn.functional.sigmoid(router_logits.float()) + scores_for_choice = routing_weights + self.e_score_correction_bias + _, top_k_index = torch.topk(scores_for_choice, self.top_k, dim=-1, sorted=False) + top_k_weights = routing_weights.gather(1, top_k_index) + top_k_weights /= top_k_weights.sum(dim=-1, keepdim=True) + return top_k_index, top_k_weights.to(router_logits.dtype) + + def forward(self, hidden_states: torch.Tensor) -> tuple[torch.Tensor, torch.Tensor]: + batch_size, sequence_length, hidden_dim = hidden_states.shape + if self.training and self.jitter_noise > 0: + hidden_states *= torch.empty_like(hidden_states).uniform_(1.0 - self.jitter_noise, 1.0 + self.jitter_noise) + hidden_states = hidden_states.view(-1, hidden_states.shape[-1]) + router_logits = self.gate(hidden_states) + top_k_index, top_k_weights = self.route_tokens_to_experts(router_logits) + hidden_states = self.experts(hidden_states, top_k_index, top_k_weights.to(hidden_states.dtype)) + hidden_states = hidden_states.reshape(batch_size, sequence_length, hidden_dim) + return hidden_states, router_logits + + +@use_kernel_forward_from_hub("RMSNorm") +class MiniMaxM2RMSNorm(nn.Module): + def __init__(self, hidden_size, eps=1e-6): + """ + MiniMaxM2RMSNorm is equivalent to T5LayerNorm + """ + super().__init__() + self.weight = nn.Parameter(torch.ones(hidden_size)) + self.variance_epsilon = eps + + def forward(self, hidden_states): + input_dtype = hidden_states.dtype + hidden_states = hidden_states.to(torch.float32) + variance = hidden_states.pow(2).mean(-1, keepdim=True) + hidden_states = hidden_states * torch.rsqrt(variance + self.variance_epsilon) + return self.weight * hidden_states.to(input_dtype) + + def extra_repr(self): + return f"{tuple(self.weight.shape)}, eps={self.variance_epsilon}" + + +def repeat_kv(hidden_states: torch.Tensor, n_rep: int) -> torch.Tensor: + """ + This is the equivalent of torch.repeat_interleave(x, dim=1, repeats=n_rep). The hidden states go from (batch, + num_key_value_heads, seqlen, head_dim) to (batch, num_attention_heads, seqlen, head_dim) + """ + batch, num_key_value_heads, slen, head_dim = hidden_states.shape + if n_rep == 1: + return hidden_states + hidden_states = hidden_states[:, :, None, :, :].expand(batch, num_key_value_heads, n_rep, slen, head_dim) + return hidden_states.reshape(batch, num_key_value_heads * n_rep, slen, head_dim) + + +def eager_attention_forward( + module: nn.Module, + query: torch.Tensor, + key: torch.Tensor, + value: torch.Tensor, + attention_mask: Optional[torch.Tensor], + scaling: float, + dropout: float = 0.0, + **kwargs: Unpack[TransformersKwargs], +): + key_states = repeat_kv(key, module.num_key_value_groups) + value_states = repeat_kv(value, module.num_key_value_groups) + + attn_weights = torch.matmul(query, key_states.transpose(2, 3)) * scaling + if attention_mask is not None: + causal_mask = attention_mask[:, :, :, : key_states.shape[-2]] + attn_weights = attn_weights + causal_mask + + attn_weights = nn.functional.softmax(attn_weights, dim=-1, dtype=torch.float32).to(query.dtype) + attn_weights = nn.functional.dropout(attn_weights, p=dropout, training=module.training) + attn_output = torch.matmul(attn_weights, value_states) + attn_output = attn_output.transpose(1, 2).contiguous() + + return attn_output, attn_weights + + +def rotate_half(x): + """Rotates half the hidden dims of the input.""" + x1 = x[..., : x.shape[-1] // 2] + x2 = x[..., x.shape[-1] // 2 :] + return torch.cat((-x2, x1), dim=-1) + + +def apply_rotary_pos_emb(q, k, cos, sin, position_ids=None, unsqueeze_dim=1): + """Applies Rotary Position Embedding to the query and key tensors. + + Args: + q (`torch.Tensor`): The query tensor. + k (`torch.Tensor`): The key tensor. + cos (`torch.Tensor`): The cosine part of the rotary embedding. + sin (`torch.Tensor`): The sine part of the rotary embedding. + position_ids (`torch.Tensor`, *optional*): + Deprecated and unused. + unsqueeze_dim (`int`, *optional*, defaults to 1): + The 'unsqueeze_dim' argument specifies the dimension along which to unsqueeze cos[position_ids] and + sin[position_ids] so that they can be properly broadcasted to the dimensions of q and k. For example, note + that cos[position_ids] and sin[position_ids] have the shape [batch_size, seq_len, head_dim]. Then, if q and + k have the shape [batch_size, heads, seq_len, head_dim], then setting unsqueeze_dim=1 makes + cos[position_ids] and sin[position_ids] broadcastable to the shapes of q and k. Similarly, if q and k have + the shape [batch_size, seq_len, heads, head_dim], then set unsqueeze_dim=2. + Returns: + `tuple(torch.Tensor)` comprising of the query and key tensors rotated using the Rotary Position Embedding. + """ + cos = cos.unsqueeze(unsqueeze_dim) + sin = sin.unsqueeze(unsqueeze_dim) + + # Keep half or full tensor for later concatenation + rotary_dim = cos.shape[-1] + q_rot, q_pass = q[..., :rotary_dim], q[..., rotary_dim:] + k_rot, k_pass = k[..., :rotary_dim], k[..., rotary_dim:] + + # Apply rotary embeddings on the first half or full tensor + q_embed = (q_rot * cos) + (rotate_half(q_rot) * sin) + k_embed = (k_rot * cos) + (rotate_half(k_rot) * sin) + + # Concatenate back to full shape + q_embed = torch.cat([q_embed, q_pass], dim=-1) + k_embed = torch.cat([k_embed, k_pass], dim=-1) + return q_embed, k_embed + + +class MiniMaxM2Attention(nn.Module): + """Multi-headed attention from 'Attention Is All You Need' paper""" + + def __init__(self, config: MiniMaxM2Config, layer_idx: int): + super().__init__() + self.config = config + self.layer_idx = layer_idx + self.head_dim = getattr(config, "head_dim", None) or config.hidden_size // config.num_attention_heads + self.num_key_value_groups = config.num_attention_heads // config.num_key_value_heads + self.scaling = self.head_dim**-0.5 + self.attention_dropout = config.attention_dropout + self.is_causal = True + self.q_proj = nn.Linear(config.hidden_size, config.num_attention_heads * self.head_dim, bias=False) + self.k_proj = nn.Linear(config.hidden_size, config.num_key_value_heads * self.head_dim, bias=False) + self.v_proj = nn.Linear(config.hidden_size, config.num_key_value_heads * self.head_dim, bias=False) + self.o_proj = nn.Linear(config.num_attention_heads * self.head_dim, config.hidden_size, bias=False) + + self.use_qk_norm = config.use_qk_norm + if self.use_qk_norm: + self.q_norm = MiniMaxM2RMSNorm(self.head_dim * config.num_attention_heads, eps=config.rms_norm_eps) + self.k_norm = MiniMaxM2RMSNorm(self.head_dim * config.num_key_value_heads, eps=config.rms_norm_eps) + + @deprecate_kwarg("past_key_value", new_name="past_key_values", version="4.58") + def forward( + self, + hidden_states: torch.Tensor, + position_embeddings: tuple[torch.Tensor, torch.Tensor], + attention_mask: Optional[torch.Tensor], + past_key_values: Optional[Cache] = None, + cache_position: Optional[torch.LongTensor] = None, + **kwargs: Unpack[FlashAttentionKwargs], + ) -> tuple[torch.Tensor, Optional[torch.Tensor]]: + input_shape = hidden_states.shape[:-1] + hidden_shape = (*input_shape, -1, self.head_dim) + + query_states = self.q_proj(hidden_states) + key_states = self.k_proj(hidden_states) + value_states = self.v_proj(hidden_states) + + if self.use_qk_norm: # main diff from Llama + query_states = self.q_norm(query_states) + key_states = self.k_norm(key_states) + + key_states = key_states.view(hidden_shape) + query_states = query_states.view(hidden_shape) + value_states = value_states.view(hidden_shape) + + query_states = query_states.transpose(1, 2) + key_states = key_states.transpose(1, 2) + value_states = value_states.transpose(1, 2) + + cos, sin = position_embeddings + query_states, key_states = apply_rotary_pos_emb(query_states, key_states, cos, sin) + + if past_key_values is not None: + # sin and cos are specific to RoPE models; position_ids needed for the static cache + cache_kwargs = {"sin": sin, "cos": cos, "cache_position": cache_position} + key_states, value_states = past_key_values.update(key_states, value_states, self.layer_idx, cache_kwargs) + + attention_interface: Callable = eager_attention_forward + if self.config._attn_implementation != "eager": + attention_interface = ALL_ATTENTION_FUNCTIONS[self.config._attn_implementation] + + attn_output, attn_weights = attention_interface( + self, + query_states, + key_states, + value_states, + attention_mask, + dropout=0.0 if not self.training else self.attention_dropout, + scaling=self.scaling, + **kwargs, + ) + + attn_output = attn_output.reshape(*input_shape, -1).contiguous() + attn_output = self.o_proj(attn_output) + return attn_output, attn_weights + + +class MiniMaxM2DecoderLayer(GradientCheckpointingLayer): + def __init__(self, config: MiniMaxM2Config, layer_idx: int): + super().__init__() + self.hidden_size = config.hidden_size + + self.self_attn = MiniMaxM2Attention(config, layer_idx) + + self.block_sparse_moe = MiniMaxM2SparseMoeBlock(config) + self.input_layernorm = MiniMaxM2RMSNorm(config.hidden_size, eps=config.rms_norm_eps) + self.post_attention_layernorm = MiniMaxM2RMSNorm(config.hidden_size, eps=config.rms_norm_eps) + + @deprecate_kwarg("past_key_value", new_name="past_key_values", version="4.58") + def forward( + self, + hidden_states: torch.Tensor, + position_embeddings: tuple[torch.Tensor, torch.Tensor], + attention_mask: Optional[torch.Tensor] = None, + position_ids: Optional[torch.LongTensor] = None, + past_key_values: Optional[Cache] = None, + cache_position: Optional[torch.LongTensor] = None, + **kwargs: Unpack[TransformersKwargs], + ) -> torch.FloatTensor: + residual = hidden_states + + hidden_states = self.input_layernorm(hidden_states) + + # Self Attention + hidden_states, _ = self.self_attn( + hidden_states=hidden_states, + position_embeddings=position_embeddings, + attention_mask=attention_mask, + position_ids=position_ids, + past_key_values=past_key_values, + cache_position=cache_position, + **kwargs, + ) + hidden_states = residual + hidden_states + + # Fully Connected + residual = hidden_states + hidden_states = self.post_attention_layernorm(hidden_states) + hidden_states, _ = self.block_sparse_moe(hidden_states) + hidden_states = residual + hidden_states + + return hidden_states + + +class MiniMaxM2RotaryEmbedding(nn.Module): + inv_freq: torch.Tensor # fix linting for `register_buffer` + + def __init__(self, config: MiniMaxM2Config, device=None): + super().__init__() + # BC: "rope_type" was originally "type" + if hasattr(config, "rope_scaling") and isinstance(config.rope_scaling, dict): + self.rope_type = config.rope_scaling.get("rope_type", config.rope_scaling.get("type")) + else: + self.rope_type = "default" + self.max_seq_len_cached = config.max_position_embeddings + self.original_max_seq_len = config.max_position_embeddings + + self.config = config + self.rope_init_fn = ROPE_INIT_FUNCTIONS[self.rope_type] + + inv_freq, self.attention_scaling = self.rope_init_fn(self.config, device) + self.register_buffer("inv_freq", inv_freq, persistent=False) + self.original_inv_freq = self.inv_freq + + @torch.no_grad() + @dynamic_rope_update # power user: used with advanced RoPE types (e.g. dynamic rope) + def forward(self, x, position_ids): + inv_freq_expanded = self.inv_freq[None, :, None].float().expand(position_ids.shape[0], -1, 1).to(x.device) + position_ids_expanded = position_ids[:, None, :].float() + + device_type = x.device.type if isinstance(x.device.type, str) and x.device.type != "mps" else "cpu" + with torch.autocast(device_type=device_type, enabled=False): # Force float32 + freqs = (inv_freq_expanded.float() @ position_ids_expanded.float()).transpose(1, 2) + emb = torch.cat((freqs, freqs), dim=-1) + cos = emb.cos() * self.attention_scaling + sin = emb.sin() * self.attention_scaling + + return cos.to(dtype=x.dtype), sin.to(dtype=x.dtype) + + +@auto_docstring +class MiniMaxM2PreTrainedModel(PreTrainedModel): + config: MiniMaxM2Config + base_model_prefix = "model" + supports_gradient_checkpointing = True + _no_split_modules = ["MiniMaxM2DecoderLayer"] + _skip_keys_device_placement = ["past_key_values"] + _supports_flash_attn = True + _supports_sdpa = True + _supports_flex_attn = True + _can_compile_fullgraph = False # MoE models don't work with torch.compile (`torch.where(condition)` not supported) + _supports_attention_backend = True + _can_record_outputs = { + "router_logits": OutputRecorder(MiniMaxM2SparseMoeBlock, index=1), + "hidden_states": MiniMaxM2DecoderLayer, + "attentions": MiniMaxM2Attention, + } + + +@auto_docstring +class MiniMaxM2Model(MiniMaxM2PreTrainedModel): + def __init__(self, config: MiniMaxM2Config): + super().__init__(config) + self.padding_idx = config.pad_token_id + self.vocab_size = config.vocab_size + + self.embed_tokens = nn.Embedding(config.vocab_size, config.hidden_size, self.padding_idx) + self.layers = nn.ModuleList( + [MiniMaxM2DecoderLayer(config, layer_idx) for layer_idx in range(config.num_hidden_layers)] + ) + self.norm = MiniMaxM2RMSNorm(config.hidden_size, eps=config.rms_norm_eps) + self.rotary_emb = MiniMaxM2RotaryEmbedding(config=config) + self.gradient_checkpointing = False + + # Initialize weights and apply final processing + self.post_init() + + @check_model_inputs + @auto_docstring + def forward( + self, + input_ids: Optional[torch.LongTensor] = None, + attention_mask: Optional[torch.Tensor] = None, + position_ids: Optional[torch.LongTensor] = None, + past_key_values: Optional[Cache] = None, + inputs_embeds: Optional[torch.FloatTensor] = None, + use_cache: Optional[bool] = None, + cache_position: Optional[torch.LongTensor] = None, + **kwargs: Unpack[TransformersKwargs], + ) -> MoeModelOutputWithPast: + if (input_ids is None) ^ (inputs_embeds is not None): + raise ValueError("You must specify exactly one of input_ids or inputs_embeds") + + if use_cache and past_key_values is None: + past_key_values = DynamicCache(config=self.config) + + if inputs_embeds is None: + inputs_embeds = self.embed_tokens(input_ids) + + if cache_position is None: + past_seen_tokens = past_key_values.get_seq_length() if past_key_values is not None else 0 + cache_position = torch.arange( + past_seen_tokens, past_seen_tokens + inputs_embeds.shape[1], device=inputs_embeds.device + ) + if position_ids is None: + position_ids = cache_position.unsqueeze(0) + + mask_function = create_causal_mask if self.config.sliding_window is None else create_sliding_window_causal_mask + causal_mask = mask_function( + config=self.config, + input_embeds=inputs_embeds, + attention_mask=attention_mask, + cache_position=cache_position, + past_key_values=past_key_values, + position_ids=position_ids, + ) + + hidden_states = inputs_embeds + + # create position embeddings to be shared across the decoder layers + position_embeddings = self.rotary_emb(hidden_states, position_ids) + + for decoder_layer in self.layers[: self.config.num_hidden_layers]: + hidden_states = decoder_layer( + hidden_states, + position_embeddings=position_embeddings, + attention_mask=causal_mask, + position_ids=position_ids, + past_key_values=past_key_values, + use_cache=use_cache, + cache_position=cache_position, + **kwargs, + ) + + hidden_states = self.norm(hidden_states) + + return MoeModelOutputWithPast( # only diff with Mistral is the output type, we need MoE + last_hidden_state=hidden_states, + past_key_values=past_key_values, + ) + + +def load_balancing_loss_func( + gate_logits: Union[torch.Tensor, tuple[torch.Tensor], None], + num_experts: Optional[int] = None, + top_k=2, + attention_mask: Optional[torch.Tensor] = None, +) -> Union[torch.Tensor, int]: + r""" + Computes auxiliary load balancing loss as in Switch Transformer - implemented in Pytorch. + + See Switch Transformer (https://huggingface.co/papers/2101.03961) for more details. This function implements the loss + function presented in equations (4) - (6) of the paper. It aims at penalizing cases where the routing between + experts is too unbalanced. + + Args: + gate_logits: + Logits from the `gate`, should be a tuple of model.config.num_hidden_layers tensors of + shape [batch_size X sequence_length, num_experts]. + num_experts: + Number of experts + top_k: + The number of experts to route per-token, can be also interpreted as the `top-k` routing + parameter. + attention_mask (`torch.Tensor`, *optional*): + The attention_mask used in forward function + shape [batch_size X sequence_length] if not None. + + Returns: + The auxiliary loss. + """ + if gate_logits is None or not isinstance(gate_logits, tuple): + return 0 + + if isinstance(gate_logits, tuple): + compute_device = gate_logits[0].device + concatenated_gate_logits = torch.cat([layer_gate.to(compute_device) for layer_gate in gate_logits], dim=0) + + routing_weights = torch.nn.functional.softmax(concatenated_gate_logits, dim=-1) + + _, selected_experts = torch.topk(routing_weights, top_k, dim=-1) + + expert_mask = torch.nn.functional.one_hot(selected_experts, num_experts) + + if attention_mask is None: + # Compute the percentage of tokens routed to each experts + tokens_per_expert = torch.mean(expert_mask.float(), dim=0) + + # Compute the average probability of routing to these experts + router_prob_per_expert = torch.mean(routing_weights, dim=0) + else: + batch_size, sequence_length = attention_mask.shape + num_hidden_layers = concatenated_gate_logits.shape[0] // (batch_size * sequence_length) + + # Compute the mask that masks all padding tokens as 0 with the same shape of expert_mask + expert_attention_mask = ( + attention_mask[None, :, :, None, None] + .expand((num_hidden_layers, batch_size, sequence_length, top_k, num_experts)) + .reshape(-1, top_k, num_experts) + .to(compute_device) + ) + + # Compute the percentage of tokens routed to each experts + tokens_per_expert = torch.sum(expert_mask.float() * expert_attention_mask, dim=0) / torch.sum( + expert_attention_mask, dim=0 + ) + + # Compute the mask that masks all padding tokens as 0 with the same shape of tokens_per_expert + router_per_expert_attention_mask = ( + attention_mask[None, :, :, None] + .expand((num_hidden_layers, batch_size, sequence_length, num_experts)) + .reshape(-1, num_experts) + .to(compute_device) + ) + + # Compute the average probability of routing to these experts + router_prob_per_expert = torch.sum(routing_weights * router_per_expert_attention_mask, dim=0) / torch.sum( + router_per_expert_attention_mask, dim=0 + ) + + overall_loss = torch.sum(tokens_per_expert * router_prob_per_expert.unsqueeze(0)) + return overall_loss * num_experts + + +@auto_docstring +class MiniMaxM2ForCausalLM(MiniMaxM2PreTrainedModel, GenerationMixin): + _tied_weights_keys = ["lm_head.weight"] + _tp_plan = {"lm_head": "colwise_rep"} + _pp_plan = {"lm_head": (["hidden_states"], ["logits"])} + + def __init__(self, config): + super().__init__(config) + self.model = MiniMaxM2Model(config) + self.vocab_size = config.vocab_size + self.lm_head = nn.Linear(config.hidden_size, config.vocab_size, bias=False) + self.router_aux_loss_coef = config.router_aux_loss_coef + self.num_experts = config.num_local_experts + self.num_experts_per_tok = config.num_experts_per_tok + + # Initialize weights and apply final processing + self.post_init() + + @can_return_tuple + @auto_docstring + def forward( + self, + input_ids: Optional[torch.LongTensor] = None, + attention_mask: Optional[torch.Tensor] = None, + position_ids: Optional[torch.LongTensor] = None, + past_key_values: Optional[Cache] = None, + inputs_embeds: Optional[torch.FloatTensor] = None, + labels: Optional[torch.LongTensor] = None, + use_cache: Optional[bool] = None, + output_router_logits: Optional[bool] = None, + cache_position: Optional[torch.LongTensor] = None, + logits_to_keep: Union[int, torch.Tensor] = 0, + **kwargs: Unpack[TransformersKwargs], + ) -> MoeCausalLMOutputWithPast: + r""" + labels (`torch.LongTensor` of shape `(batch_size, sequence_length)`, *optional*): + Labels for computing the masked language modeling loss. Indices should either be in `[0, ..., + config.vocab_size]` or -100 (see `input_ids` docstring). Tokens with indices set to `-100` are ignored + (masked), the loss is only computed for the tokens with labels in `[0, ..., config.vocab_size]`. + + Example: + + ```python + >>> from transformers import AutoTokenizer, MiniMaxM2ForCausalLM + + >>> model = MiniMaxM2ForCausalLM.from_pretrained("mistralai/MiniMaxM2-8x7B-v0.1") + >>> tokenizer = AutoTokenizer.from_pretrained("mistralai/MiniMaxM2-8x7B-v0.1") + + >>> prompt = "Hey, are you conscious? Can you talk to me?" + >>> inputs = tokenizer(prompt, return_tensors="pt") + + >>> # Generate + >>> generate_ids = model.generate(inputs.input_ids, max_length=30) + >>> tokenizer.batch_decode(generate_ids, skip_special_tokens=True, clean_up_tokenization_spaces=False)[0] + "Hey, are you conscious? Can you talk to me?\nI'm not conscious, but I can talk to you." + ```""" + + output_router_logits = ( + output_router_logits if output_router_logits is not None else self.config.output_router_logits + ) + + # decoder outputs consists of (dec_features, layer_state, dec_hidden, dec_attn) + outputs: MoeModelOutputWithPast = self.model( + input_ids=input_ids, + attention_mask=attention_mask, + position_ids=position_ids, + past_key_values=past_key_values, + inputs_embeds=inputs_embeds, + use_cache=use_cache, + output_router_logits=output_router_logits, + cache_position=cache_position, + **kwargs, + ) + + hidden_states = outputs.last_hidden_state + # Only compute necessary logits, and do not upcast them to float if we are not computing the loss + slice_indices = slice(-logits_to_keep, None) if isinstance(logits_to_keep, int) else logits_to_keep + logits = self.lm_head(hidden_states[:, slice_indices, :]) + + loss = None + if labels is not None: + loss = self.loss_function(logits, labels, self.vocab_size, **kwargs) + + aux_loss = None + if output_router_logits: + aux_loss = load_balancing_loss_func( + outputs.router_logits, + self.num_experts, + self.num_experts_per_tok, + attention_mask, + ) + if labels is not None: + loss += self.router_aux_loss_coef * aux_loss.to(loss.device) # make sure to reside in the same device + + return MoeCausalLMOutputWithPast( + loss=loss, + aux_loss=aux_loss, + logits=logits, + past_key_values=outputs.past_key_values, + hidden_states=outputs.hidden_states, + attentions=outputs.attentions, + router_logits=outputs.router_logits, + ) + + +class MiniMaxM2ForSequenceClassification(GenericForSequenceClassification, MiniMaxM2PreTrainedModel): + pass + + +class MiniMaxM2ForTokenClassification(GenericForTokenClassification, MiniMaxM2PreTrainedModel): + pass + + +class MiniMaxM2ForQuestionAnswering(GenericForQuestionAnswering, MiniMaxM2PreTrainedModel): + pass + + +__all__ = [ + "MiniMaxM2ForCausalLM", + "MiniMaxM2ForQuestionAnswering", + "MiniMaxM2Model", + "MiniMaxM2PreTrainedModel", + "MiniMaxM2ForSequenceClassification", + "MiniMaxM2ForTokenClassification", +] diff --git a/tokenizer.json b/tokenizer.json new file mode 100644 index 0000000000000000000000000000000000000000..27ba950155a5ce69f8f0d6a733447616fd552b9b --- /dev/null +++ b/tokenizer.json @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:7b81e5e5cba2b169e86a0771825a927e9d41b4c4484ded4a286410f41f702f17 +size 15523144 diff --git a/tokenizer_config.json b/tokenizer_config.json new file mode 100644 index 0000000000000000000000000000000000000000..b25ef13c6ec3e5a5f17e62911d892cc4ceb36cf2 --- /dev/null +++ b/tokenizer_config.json @@ -0,0 +1,12 @@ +{ + "add_prefix_space": false, + "backend": "tokenizers", + "bos_token": "]~!b[", + "clean_up_tokenization_spaces": false, + "eos_token": "[e~[", + "is_local": true, + "model_max_length": 40960000, + "tokenizer_class": "TokenizersBackend", + "tool_parser_type": "minimax_m2", + "unk_token": "]!d~[" +} diff --git a/vmlx-logo.png b/vmlx-logo.png new file mode 100644 index 0000000000000000000000000000000000000000..b98e5ef59f1df2910aa76e4b863f020c00620925 Binary files /dev/null and b/vmlx-logo.png differ