|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| """
|
| Merge Ministral 14B LoRA adapter with the official Mistral base model.
|
| Run on HuggingFace Jobs with: hf jobs uv run --flavor a10g-large --timeout 2h --secrets HF_TOKEN merge_ministral_official.py
|
| """
|
|
|
| import torch
|
| import os
|
| from peft import PeftModel
|
| from transformers import Mistral3ForConditionalGeneration, AutoTokenizer
|
| from huggingface_hub import HfApi
|
|
|
| BASE_MODEL = "mistralai/Ministral-3-14B-Base-2512"
|
| LORA_ADAPTER = "RoleModel/ministral-14b-reasoning-merged"
|
| OUTPUT_REPO = os.environ.get("OUTPUT_REPO", "RoleModel/ministral-14b-merged-official")
|
|
|
| print(f"Base model: {BASE_MODEL}")
|
| print(f"LoRA adapter: {LORA_ADAPTER}")
|
| print(f"Output repo: {OUTPUT_REPO}")
|
|
|
| print("\n=== Loading base model ===")
|
|
|
| base = Mistral3ForConditionalGeneration.from_pretrained(
|
| BASE_MODEL,
|
| torch_dtype=torch.bfloat16,
|
| device_map="cuda:0",
|
| trust_remote_code=True,
|
| low_cpu_mem_usage=True,
|
| )
|
| print(f"Base model loaded: {base.__class__.__name__}")
|
|
|
| print("\n=== Loading LoRA adapter ===")
|
| model = PeftModel.from_pretrained(base, LORA_ADAPTER)
|
| print("LoRA adapter loaded")
|
|
|
| print("\n=== Merging weights ===")
|
| merged = model.merge_and_unload()
|
| print("Merge complete")
|
|
|
| print("\n=== Moving to CPU for save ===")
|
| merged = merged.cpu()
|
| torch.cuda.empty_cache()
|
|
|
| print("\n=== Saving merged model ===")
|
| merged.save_pretrained("./merged-model", safe_serialization=True)
|
| print("Model saved locally")
|
|
|
| print("\n=== Saving tokenizer ===")
|
| tok = AutoTokenizer.from_pretrained(BASE_MODEL, trust_remote_code=True)
|
| tok.save_pretrained("./merged-model")
|
| print("Tokenizer saved")
|
|
|
| print(f"\n=== Pushing to Hub: {OUTPUT_REPO} ===")
|
| api = HfApi()
|
| api.create_repo(OUTPUT_REPO, exist_ok=True, private=True, token=os.environ.get("HF_TOKEN"))
|
| api.upload_folder(
|
| folder_path="./merged-model",
|
| repo_id=OUTPUT_REPO,
|
| repo_type="model",
|
| token=os.environ.get("HF_TOKEN")
|
| )
|
|
|
| print(f"\n=== DONE ===")
|
| print(f"Merged model available at: https://huggingface.co/{OUTPUT_REPO}")
|
|
|