| { | |
| "model_type": "cohere", | |
| "architectures": [ | |
| "CohereForCausalLM" | |
| ], | |
| "pre_weights": [ | |
| { | |
| "name": "model.embed_tokens.weight", | |
| "is_embed": true | |
| } | |
| ], | |
| "post_weights": [ | |
| { | |
| "name": "model.norm.weight" | |
| }, | |
| { | |
| "name": "lm_head.weight", | |
| "is_embed": true, | |
| "optional": true | |
| } | |
| ], | |
| "num_layers_config_key": "num_hidden_layers", | |
| "layer_templates": { | |
| "weights": [ | |
| { | |
| "name": "model.layers.${layer_index}.input_layernorm.weight" | |
| }, | |
| { | |
| "name": "model.layers.${layer_index}.mlp.down_proj.weight" | |
| }, | |
| { | |
| "name": "model.layers.${layer_index}.mlp.gate_proj.weight" | |
| }, | |
| { | |
| "name": "model.layers.${layer_index}.mlp.up_proj.weight" | |
| }, | |
| { | |
| "name": "model.layers.${layer_index}.self_attn.q_proj.weight" | |
| }, | |
| { | |
| "name": "model.layers.${layer_index}.self_attn.k_proj.weight" | |
| }, | |
| { | |
| "name": "model.layers.${layer_index}.self_attn.v_proj.weight" | |
| }, | |
| { | |
| "name": "model.layers.${layer_index}.self_attn.o_proj.weight" | |
| } | |
| ] | |
| } | |
| } | |