# 自定义Megatron模型 这里介绍如何在[Mcore-Bridge](https://github.com/modelscope/mcore-bridge)中注册模型,以支持新模型在Megatron-SWIFT中的训练。我们将以MiniMax-M2.7为例子介绍。 ## 下载模型 首先,你需要下载模型配置。 ```python from swift import safe_snapshot_download model_dir = safe_snapshot_download('MiniMax/MiniMax-M2.7', download_model=False) print(f'model_dir: {model_dir}') ``` 由于模型权重很大,为了加速支持模型的效率,我们采用懒下载的方式,并只下载`num_layers`层的权重,构建mini版本的模型,用于做接入测试。以MiniMax-M2.7为例,我们构建了一层的BF16版本的权重。若有些模型出现前3层为Dense,之后为MoE,则你可以构建4层的权重。若出现Attention交替的情况,例如Qwen3.5采用linear-attention和full-attention交替,你也需要更多的层数。 ```python import os import torch from modelscope.hub.file_download import model_file_download from safetensors.torch import safe_open from swift import safe_snapshot_download from mcore_bridge.utils import Fp8Dequantizer, SafetensorLazyLoader, StreamingSafetensorSaver model_id = 'MiniMax/MiniMax-M2.7' # 有些模型会出现前几层为dense,后面为moe的情况,需合理设置该值 num_layers = 1 # 只下载`num_layers`层,节约磁盘占用和运行时显存占用 model_dir = safe_snapshot_download(model_id, download_model=False) loader = SafetensorLazyLoader(model_dir) state_dict = loader.get_state_dict() saver = StreamingSafetensorSaver(save_dir=model_dir) new_state_dict = {} fp8_dequantizer = Fp8Dequantizer() # 用于将fp8权重转成bf16 def _open_file(self, filename: str): if filename not in self._file_handles: file_path = os.path.join(self.hf_model_dir, filename) tmp_dir = os.path.join(self.hf_model_dir, 'tmp') if not os.path.exists(file_path): file_path = os.path.join(tmp_dir, filename) if not os.path.exists(file_path): file_path = model_file_download( model_id=model_id, file_path=filename, local_dir=tmp_dir, ) self._file_handles[filename] = safe_open(file_path, framework='pt') return self._file_handles[filename] SafetensorLazyLoader._open_file = _open_file # monkey patch (懒下载) for k, v in state_dict.items(): if k.startswith('model.layers.'): idx = int(k[len('model.layers.'):].split('.', 1)[0]) if idx >= num_layers: continue if k.endswith('.weight_scale_inv'): continue elif k.endswith('.weight'): weight_scale_inv = k.replace('.weight', '.weight_scale_inv') if weight_scale_inv in state_dict: v = fp8_dequantizer.convert(v.load(), state_dict[weight_scale_inv].load()).to(torch.bfloat16) new_state_dict[k] = v if isinstance(v, torch.Tensor) else v.load() for k, v in new_state_dict.items(): saver.add_tensor(k, v) saver.finalize() ``` 保存完权重后,你需要修改'config.json',将`num_hidden_layers`修改为1(与上面的代码对应),并删除`quantization_config`配置(因为权重为BF16的,而不是FP8)。FP8的训练大多数模型会自动适配,但有些模型可能需要额外适配,例如:Qwen3.5的FP8的适配参考[这个PR](https://github.com/modelscope/mcore-bridge/pull/30)。 ## 注册模型 以下提供debug代码,你需要修改代码,以确保huggingface transformers库的forward与megatron的forward对齐。 ```python import os os.environ['CUDA_VISIBLE_DEVICES'] = '0,1,2,3' os.environ['SWIFT_TEST_CONVERT_PRECISION'] = '1' from swift import export_main, ExportArguments, safe_snapshot_download model_id = 'MiniMax/MiniMax-M2.7' model_dir = safe_snapshot_download(model_id, download_model=False) export_main( ExportArguments( model=model_dir, to_mcore=True, exist_ok=True, test_convert_precision=True, torch_dtype='bfloat16', )) ``` 'minimax_m2'的注册可以查看[这个文件](https://github.com/modelscope/mcore-bridge/blob/main/src/mcore_bridge/model/gpts/minimax_m2.py)。我们注册时指定了模型对应的GPTBridge类和模型加载器loader。 ```python register_model(ModelMeta( ModelType.minimax_m2, ['minimax_m2'], bridge_cls=MinimaxM2Bridge, loader=MinimaxM2Loader, )) ``` 参数的总和对齐: ``` [INFO:swift] n_parameter: 522 [INFO:swift] total_sum: 106747128.72671509 [INFO:swift] zero_count: 0 [INFO:swift] n_parameter: 780 [INFO:swift] total_sum: 106747129.32046509 [INFO:swift] zero_count: 0 ``` 模型forward的logits对齐。(当然我们还需要对模型进行训练,训练后再测试forward的精度,避免出现这里输出tokens都是同一个的情况)。 ``` mean_diff: 2.8353377274470404e-05, max_diff: 0.0015382766723632812 mean_diff (with loss): 2.1664049199898727e-05, max_diff (with loss): 0.00021076202392578125 (Please check that mean_diff (with loss) is less than 0.1). hf_tokens: [190962, 103239, 367, 367, 367, 367, 367, 367, 367, 367, 367, 367, 367, 367, 367, 367, 367, 367, 367, 367, 367, 367, 367, 367, 367, 367, 367, 367, 367, 367, 367, 367, 367, 367, 367, 367, 367, 367, 367, 367, 367, 367, 367, 367, 367, 367, 367, 367, 367, 367, 367, 367, 367, 367, 367, 367, 367, 367, 367, 367, 367, 367, 367, 367, 367, 367, 367, 367, 367, 367, 367] mg_tokens: [190962, 103239, 367, 367, 367, 367, 367, 367, 367, 367, 367, 367, 367, 367, 367, 367, 367, 367, 367, 367, 367, 367, 367, 367, 367, 367, 367, 367, 367, 367, 367, 367, 367, 367, 367, 367, 367, 367, 367, 367, 367, 367, 367, 367, 367, 367, 367, 367, 367, 367, 367, 367, 367, 367, 367, 367, 367, 367, 367, 367, 367, 367, 367, 367, 367, 367, 367, 367, 367, 367, 367] token_diff: 0 token_diff (with loss): 0 ``` 通常在参数总数对齐和输出logits对齐后,模型就基本接入成功了。此外你可能还需要适配TP/CP的情况。你可以使用以下代码debug: ```python import os os.environ['CUDA_VISIBLE_DEVICES'] = '0,1' os.environ['SWIFT_TEST_CONVERT_PRECISION'] = '1' from swift.megatron import MegatronExportArguments, megatron_export_main from swift import safe_snapshot_download model_id = 'MiniMax/MiniMax-M2.7' model_dir = safe_snapshot_download(model_id, download_model=False) if __name__ == '__main__': megatron_export_main( MegatronExportArguments( model=model_dir, to_mcore=True, # 也可以修改成 `to_hf=True` 测试 tensor_model_parallel_size=2, sequence_parallel=True, expert_model_parallel_size=2, test_convert_precision=True, )) ``` 我们需要用torchrun启动,vscode配置: ```json { "version": "0.2.0", "configurations": [ { "name": "torchrun2", "type": "debugpy", "request": "launch", "program": "-m", "console": "integratedTerminal", "justMyCode": false, "args": [ "torch.distributed.run", "--nproc_per_node", "2", "--master_port", "29501", "${file}" ] } ] } ``` 其他模型的注册例子,可以查看对应PR:[hy_v3](https://github.com/modelscope/mcore-bridge/pull/53)、[kimi_25](https://github.com/modelscope/mcore-bridge/pull/52)。在2026年4月之前的接入PR可以在ms-swift库中寻找。 ## 测试准确性 我们对mini版本的模型进行训练,我们只使用自我认知数据集,并训练到过拟合。 ```shell # 2 * 80GiB PYTORCH_CUDA_ALLOC_CONF='expandable_segments:True' \ NPROC_PER_NODE=2 \ CUDA_VISIBLE_DEVICES=0,1 \ megatron sft \ --model /root/.cache/modelscope/models/MiniMax/MiniMax-M2.7 \ --save_safetensors true \ --dataset 'swift/self-cognition#500' \ --tensor_model_parallel_size 2 \ --sequence_parallel true \ --micro_batch_size 16 \ --global_batch_size 16 \ --recompute_granularity full \ --recompute_method uniform \ --recompute_num_layers 1 \ --finetune true \ --cross_entropy_loss_fusion true \ --lr 2e-5 \ --lr_warmup_fraction 0.05 \ --min_lr 1e-5 \ --num_train_epochs 10 \ --output_dir megatron_output \ --save_steps 500 \ --max_length 2048 \ --system 'You are a helpful assistant.' \ --dataloader_num_workers 4 \ --no_save_optim true \ --no_save_rng true \ --moe_permute_fusion true \ --expert_model_parallel_size 2 \ --moe_grouped_gemm true \ --moe_shared_expert_overlap true \ --moe_aux_loss_coeff 1e-3 \ --dataset_num_proc 4 \ --model_author swift \ --model_name swift-robot ``` 进行推理,查看训练效果: ```shell CUDA_VISIBLE_DEVICES=0 \ swift infer \ --model megatron_output/v3-20260430-143926/checkpoint-310 \ --max_new_tokens 64 \ --enable_thinking false \ --temperature 0 ``` ``` <<< 你是谁 我是一个由swift开发的人工智能助手,被称为swift-robot。我主要的目的是通过文本交流为用户提供帮助、信息和娱乐。如果您有任何疑问或需要帮助,请随时提出,我会尽力协助您。 -------------------------------------------------- <<< clear <<< who are you I am a language model developed by swift, you can call me swift-robot. How can I assist you? -------------------------------------------------- ``` 再次测试forward精度对齐: ``` mean_diff: 0.0005969047779217362, max_diff: 0.013172879815101624 mean_diff (with loss): 0.0005803848034702241, max_diff (with loss): 0.009410381317138672 (Please check that mean_diff (with loss) is less than 0.1). hf_tokens: [190962, 190962, 367, 44, 46, 2362, 5129, 6415, 75827, 343, 10, 1497, 71151, 11915, 1497, 44, 3003, 44, 46, 46, 4387, 10, 32, 10, 258, 1497, 44, 46, 46, 258, 18268, 44, 692, 13268, 42047, 3764, 46, 46, 46, 94454, 46, 46, 275, 296, 3786, 46, 46, 275, 46, 46, 3786, 46, 2329, 10, 722] mg_tokens: [190962, 190962, 367, 44, 46, 2362, 5129, 6415, 75827, 343, 10, 1497, 71151, 11915, 1497, 44, 3003, 44, 46, 46, 4387, 10, 32, 10, 258, 1497, 44, 46, 46, 258, 18268, 44, 692, 13268, 42047, 3764, 46, 46, 46, 94454, 46, 46, 275, 296, 3786, 46, 46, 275, 46, 46, 3786, 46, 2329, 10, 722] token_diff: 0 token_diff (with loss): 0 ``` 至此,模型接入成功啦! ## 提交PR 如果你想给ms-swift/mcore-bridge提交PR,你需要额外运行以下命令,对代码进行整理: ```shell pip install pre-commit pre-commit run --all-files ```