| | """ |
| | Export the trained model to HuggingFace-compatible format. |
| | |
| | Creates: |
| | - model.safetensors (weights) |
| | - config.json (architecture config) |
| | - generation_config.json |
| | - tokenizer.json, tokenizer_config.json, special_tokens_map.json |
| | """ |
| |
|
| | import os |
| | import sys |
| | import json |
| | import torch |
| | from collections import OrderedDict |
| | from safetensors.torch import save_file |
| |
|
| | sys.path.insert(0, os.path.dirname(os.path.abspath(__file__))) |
| | from model.config import ModelConfig |
| | from model.transformer import Transformer |
| | from model.data import get_tokenizer |
| |
|
| | CHECKPOINT = "/jfs/deepak-kumar/checkpoints_dpo/dpo_final.pt" |
| | OUTPUT_DIR = "/home/jovyan/training/hf_model" |
| |
|
| | os.makedirs(OUTPUT_DIR, exist_ok=True) |
| |
|
| | print("=" * 60) |
| | print(" EXPORTING MODEL TO HUGGING FACE FORMAT") |
| | print("=" * 60) |
| |
|
| | |
| | print("\n[1/4] Loading checkpoint...") |
| | tokenizer = get_tokenizer() |
| | special_tokens = ["<|user|>", "<|assistant|>", "<|end|>"] |
| | vocab = tokenizer.get_vocab() |
| | new_tokens = [t for t in special_tokens if t not in vocab] |
| | if new_tokens: |
| | tokenizer.add_tokens(new_tokens, special_tokens=True) |
| |
|
| | model_config = ModelConfig() |
| | model_config.vocab_size = len(tokenizer) |
| |
|
| | model = Transformer(model_config) |
| | ckpt = torch.load(CHECKPOINT, map_location="cpu", weights_only=False) |
| | model.load_state_dict(ckpt["model"]) |
| | step = ckpt.get("step", 0) |
| | del ckpt |
| | print(f" Loaded DPO model (step {step}, vocab {model_config.vocab_size})") |
| |
|
| | |
| | print("\n[2/4] Converting weights to safetensors...") |
| |
|
| | state_dict = model.state_dict() |
| | hf_state = OrderedDict() |
| |
|
| | KEY_MAP = { |
| | "tok_embeddings.weight": "model.embed_tokens.weight", |
| | "norm.weight": "model.norm.weight", |
| | "output.weight": "lm_head.weight", |
| | } |
| |
|
| | for key, tensor in state_dict.items(): |
| | if key in KEY_MAP: |
| | hf_state[KEY_MAP[key]] = tensor |
| | continue |
| |
|
| | if key.startswith("layers."): |
| | parts = key.split(".") |
| | layer_idx = parts[1] |
| | rest = ".".join(parts[2:]) |
| |
|
| | layer_map = { |
| | "attention_norm.weight": f"model.layers.{layer_idx}.input_layernorm.weight", |
| | "ffn_norm.weight": f"model.layers.{layer_idx}.post_attention_layernorm.weight", |
| | "attention.wq.weight": f"model.layers.{layer_idx}.self_attn.q_proj.weight", |
| | "attention.wk.weight": f"model.layers.{layer_idx}.self_attn.k_proj.weight", |
| | "attention.wv.weight": f"model.layers.{layer_idx}.self_attn.v_proj.weight", |
| | "attention.wo.weight": f"model.layers.{layer_idx}.self_attn.o_proj.weight", |
| | "ffn.w_gate.weight": f"model.layers.{layer_idx}.mlp.gate_proj.weight", |
| | "ffn.w_up.weight": f"model.layers.{layer_idx}.mlp.up_proj.weight", |
| | "ffn.w_down.weight": f"model.layers.{layer_idx}.mlp.down_proj.weight", |
| | } |
| |
|
| | if rest in layer_map: |
| | hf_state[layer_map[rest]] = tensor |
| | else: |
| | print(f" WARNING: unmapped key {key}") |
| | hf_state[key] = tensor |
| | elif key == "freqs_cis": |
| | continue |
| | else: |
| | print(f" WARNING: unmapped key {key}") |
| | hf_state[key] = tensor |
| |
|
| | |
| | for k in hf_state: |
| | if hf_state[k].dtype == torch.float32: |
| | hf_state[k] = hf_state[k].to(torch.bfloat16) |
| |
|
| | safetensors_path = os.path.join(OUTPUT_DIR, "model.safetensors") |
| | save_file(hf_state, safetensors_path) |
| | size_gb = os.path.getsize(safetensors_path) / 1e9 |
| | print(f" Saved {len(hf_state)} tensors -> {safetensors_path} ({size_gb:.2f} GB)") |
| |
|
| | |
| | print("\n[3/4] Writing config files...") |
| |
|
| | config_json = { |
| | "architectures": ["LlamaForCausalLM"], |
| | "model_type": "llama", |
| | "vocab_size": model_config.vocab_size, |
| | "hidden_size": model_config.hidden_dim, |
| | "intermediate_size": model_config.intermediate_dim, |
| | "num_hidden_layers": model_config.num_layers, |
| | "num_attention_heads": model_config.num_attention_heads, |
| | "num_key_value_heads": model_config.num_kv_heads, |
| | "max_position_embeddings": model_config.max_seq_len, |
| | "rope_theta": model_config.rope_theta, |
| | "rms_norm_eps": model_config.rms_norm_eps, |
| | "hidden_act": "silu", |
| | "initializer_range": 0.02, |
| | "tie_word_embeddings": False, |
| | "torch_dtype": "bfloat16", |
| | "transformers_version": "4.40.0", |
| | "use_cache": True, |
| | "bos_token_id": tokenizer.bos_token_id, |
| | "eos_token_id": tokenizer.eos_token_id, |
| | "pad_token_id": tokenizer.pad_token_id, |
| | } |
| |
|
| | with open(os.path.join(OUTPUT_DIR, "config.json"), "w") as f: |
| | json.dump(config_json, f, indent=2) |
| | print(" config.json") |
| |
|
| | gen_config = { |
| | "bos_token_id": tokenizer.bos_token_id, |
| | "eos_token_id": tokenizer.eos_token_id, |
| | "pad_token_id": tokenizer.pad_token_id, |
| | "do_sample": True, |
| | "temperature": 0.7, |
| | "top_k": 50, |
| | "top_p": 0.9, |
| | "repetition_penalty": 1.15, |
| | "max_new_tokens": 512, |
| | "transformers_version": "4.40.0", |
| | } |
| |
|
| | with open(os.path.join(OUTPUT_DIR, "generation_config.json"), "w") as f: |
| | json.dump(gen_config, f, indent=2) |
| | print(" generation_config.json") |
| |
|
| | |
| | print("\n[4/4] Exporting tokenizer...") |
| | tokenizer.save_pretrained(OUTPUT_DIR) |
| | print(" Tokenizer files saved") |
| |
|
| | print("\n" + "=" * 60) |
| | print(" EXPORT COMPLETE -> " + OUTPUT_DIR) |
| | print("=" * 60) |
| | print("\nFiles:") |
| | for f in sorted(os.listdir(OUTPUT_DIR)): |
| | size = os.path.getsize(os.path.join(OUTPUT_DIR, f)) |
| | if size > 1e6: |
| | print(f" {f:40s} {size/1e6:.1f} MB") |
| | else: |
| | print(f" {f:40s} {size/1e3:.1f} KB") |
| |
|