| --- |
| license: apache-2.0 |
| --- |
| |
| Here is a code to create this tiny model: |
|
|
|
|
| ```python |
| |
| import os |
| import torch |
| |
| from transformers import AutoTokenizer, AutoConfig, Lfm2MoeForCausalLM |
| |
| # # === Step 1: Define tiny model config === |
| model_id = "LiquidAI/LFM2-24B-A2B" |
| config = AutoConfig.from_pretrained(model_id) |
| |
| config.num_hidden_layers = 3 |
| config.layer_types = [ |
| "full_attention", |
| "full_attention", |
| "conv", |
| ] |
| config.num_attention_heads = 4 |
| config.num_key_value_heads = 4 |
| config.hidden_size = 16 |
| config.num_dense_layers = 1 |
| config.moe_intermediate_size = 16 |
| config.intermediate_size = 16 |
| |
| # === Step 2: Create model from config === |
| model = Lfm2MoeForCausalLM(config) |
| |
| # === Step 3: Load or create tokenizer === |
| tokenizer = AutoTokenizer.from_pretrained(model_id) |
| |
| # === Step 4: Save model and tokenizer === |
| output_dir = "./lfm2_moe" |
| os.makedirs(output_dir, exist_ok=True) |
| model.save_pretrained(output_dir, safe_serialization=False) |
| tokenizer.save_pretrained(output_dir) |
| |
| ``` |