devsomosahub commited on
Commit
04a4a4c
·
verified ·
1 Parent(s): 9196f6f

Upload scripts/merge_7b_cloud.py with huggingface_hub

Browse files
Files changed (1) hide show
  1. scripts/merge_7b_cloud.py +29 -0
scripts/merge_7b_cloud.py ADDED
@@ -0,0 +1,29 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ """Merge 7B LoRA adapter with FP16 base (NOT quantized) and push."""
2
+ import os, torch, gc
3
+ from peft import PeftModel
4
+ from transformers import AutoModelForCausalLM, AutoTokenizer
5
+ from huggingface_hub import login
6
+
7
+ HF_TOKEN = os.environ["HF_TOKEN"]
8
+ login(token=HF_TOKEN)
9
+
10
+ print("Loading Qwen 7B FP16 on CPU...")
11
+ base = AutoModelForCausalLM.from_pretrained(
12
+ "Qwen/Qwen2.5-7B-Instruct",
13
+ torch_dtype=torch.float16,
14
+ device_map="cpu",
15
+ trust_remote_code=True,
16
+ )
17
+
18
+ print("Loading adapter...")
19
+ model = PeftModel.from_pretrained(base, "devsomosahub/agent-os-adapter-7b")
20
+
21
+ print("Merging...")
22
+ merged = model.merge_and_unload()
23
+
24
+ tok = AutoTokenizer.from_pretrained("devsomosahub/agent-os-adapter-7b", trust_remote_code=True)
25
+
26
+ print("Pushing merged 7B to Hub...")
27
+ merged.push_to_hub("devsomosahub/agent-os-7b-merged", token=HF_TOKEN, max_shard_size="2GB")
28
+ tok.push_to_hub("devsomosahub/agent-os-7b-merged", token=HF_TOKEN)
29
+ print("DONE! https://huggingface.co/devsomosahub/agent-os-7b-merged")