| """ |
| train_dpo.py — TeenEmo DPO(SFT済みLoRAへの継続学習) |
| |
| フロー: |
| 1. SFT 済み LoRA アダプタを HF Hub またはローカルからロード |
| 2. DPO 用 LoRA アダプタを追加設定 |
| 3. 選好データセットを HF Hub から取得・チャットテンプレート適用 |
| 4. DPOTrainer で学習(SFT アダプタに継続学習) |
| 5. [STEP 2/3] DPO 完了 → HF Hub に push |
| 6. [STEP 3/3] GGUF 変換 → HF Hub に push |
| |
| 実行例: |
| python train_dpo.py |
| DPO_EPOCHS=1 DPO_BETA=0.05 python train_dpo.py |
| """ |
|
|
| from __future__ import annotations |
|
|
| import os |
| import sys |
| import traceback |
| from datetime import datetime, timezone |
| from pathlib import Path |
|
|
| if not os.environ.get("HF_TOKEN"): |
| print("[ERROR] HF_TOKEN が未設定です。export HF_TOKEN='hf_...' を実行してください。") |
| sys.exit(1) |
|
|
| import torch |
| from unsloth import FastLanguageModel, PatchDPOTrainer, is_bfloat16_supported |
| from trl import DPOTrainer, DPOConfig |
|
|
| import train_config as cfg |
| from train_utils import ( |
| setup_logger, log_gpu_info, log_training_config, |
| load_pref_dataset, apply_chat_template_dpo, |
| ) |
|
|
|
|
| def main() -> None: |
| |
| PatchDPOTrainer() |
|
|
| start_time = datetime.now(timezone.utc) |
| log_dir = Path(cfg.DPO_OUTPUT_DIR) / "logs" |
| log_dir.mkdir(parents=True, exist_ok=True) |
| log_file = log_dir / f"dpo_{start_time.strftime('%Y%m%d_%H%M%S')}.log" |
|
|
| logger = setup_logger("dpo", str(log_file)) |
| logger.info(f"=== TeenEmo DPO 開始 [{start_time.isoformat()}] ===") |
|
|
| log_gpu_info(logger) |
| log_training_config(logger, "DPO") |
|
|
| |
| |
| sft_lora_dir = Path(cfg.SFT_OUTPUT_DIR) / "lora" |
| if sft_lora_dir.exists(): |
| model_path = str(sft_lora_dir) |
| logger.info(f"SFT LoRA アダプタ(ローカル)からロード: {model_path}") |
| else: |
| model_path = cfg.SFT_HF_REPO |
| logger.info(f"SFT LoRA アダプタ(HF Hub)からロード: {model_path}") |
|
|
| try: |
| model, tokenizer = FastLanguageModel.from_pretrained( |
| model_name=model_path, |
| max_seq_length=cfg.MAX_SEQ_LENGTH, |
| dtype=None, |
| load_in_4bit=False, |
| token=cfg.HF_TOKEN or None, |
| ) |
| logger.info("SFT モデルロード完了 ✅") |
| except Exception as e: |
| logger.error(f"モデルロードエラー: {e}") |
| logger.debug(traceback.format_exc()) |
| logger.info("先に train_sft.py を実行してください。") |
| raise |
|
|
| |
| |
| FastLanguageModel.for_training(model) |
| trainable = sum(p.numel() for p in model.parameters() if p.requires_grad) |
| logger.info(f" 学習可能パラメータ: {trainable:,}") |
|
|
| |
| logger.info("選好データセット準備中...") |
| try: |
| raw_ds = load_pref_dataset(logger) |
|
|
| |
| logger.info("チャットテンプレート適用中...") |
| ds = raw_ds.map( |
| lambda x: apply_chat_template_dpo(x, tokenizer, logger), |
| batched=True, |
| desc="DPO チャットテンプレート適用", |
| ) |
| logger.info(f"選好データ準備完了: {len(ds)} 件") |
|
|
| |
| logger.debug(f" prompt[0]: {ds[0]['prompt'][:100]}") |
| logger.debug(f" chosen[0]: {ds[0]['chosen'][:100]}") |
| except Exception as e: |
| logger.error(f"データセット準備エラー: {e}") |
| logger.debug(traceback.format_exc()) |
| raise |
|
|
| |
| logger.info("DPOTrainer 初期化中...") |
| try: |
| dpo_trainer = DPOTrainer( |
| model=model, |
| ref_model=None, |
| args=DPOConfig( |
| output_dir=cfg.DPO_OUTPUT_DIR, |
| per_device_train_batch_size=cfg.DPO_BATCH_SIZE, |
| gradient_accumulation_steps=cfg.DPO_GRAD_ACCUM, |
| num_train_epochs=cfg.DPO_EPOCHS, |
| learning_rate=cfg.DPO_LR, |
| warmup_ratio=cfg.DPO_WARMUP_RATIO, |
| lr_scheduler_type=cfg.DPO_LR_SCHEDULER, |
| weight_decay=cfg.DPO_WEIGHT_DECAY, |
| fp16=not is_bfloat16_supported(), |
| bf16=is_bfloat16_supported(), |
| logging_steps=cfg.DPO_LOGGING_STEPS, |
| save_steps=cfg.DPO_SAVE_STEPS, |
| save_total_limit=2, |
| optim="adamw_8bit", |
| seed=42, |
| report_to="none", |
| ), |
| beta=cfg.DPO_BETA, |
| train_dataset=ds, |
| tokenizer=tokenizer, |
| max_length=cfg.DPO_MAX_LENGTH, |
| max_prompt_length=cfg.DPO_MAX_PROMPT_LENGTH, |
| ) |
| logger.info("DPOTrainer 初期化完了 ✅") |
| except Exception as e: |
| logger.error(f"DPOTrainer 初期化エラー: {e}") |
| logger.debug(traceback.format_exc()) |
| raise |
|
|
| |
| logger.info("DPO 学習開始...") |
| try: |
| train_result = dpo_trainer.train() |
| logger.info("DPO 学習完了 ✅") |
| logger.info(f" train_loss: {train_result.training_loss:.4f}") |
| logger.info(f" train_runtime: {train_result.metrics.get('train_runtime', 0):.0f}s") |
| logger.info(f" samples/sec: {train_result.metrics.get('train_samples_per_second', 0):.2f}") |
| logger.info(f" rewards/chosen: {train_result.metrics.get('rewards/chosen', 'N/A')}") |
| logger.info(f" rewards/rejected: {train_result.metrics.get('rewards/rejected', 'N/A')}") |
| logger.info(f" rewards/margin: {train_result.metrics.get('rewards/margins', 'N/A')}") |
| except Exception as e: |
| logger.error(f"DPO 学習エラー: {e}") |
| logger.debug(traceback.format_exc()) |
| raise |
|
|
| |
| lora_dir = Path(cfg.DPO_OUTPUT_DIR) / "lora" |
| logger.info(f"LoRA アダプタ保存: {lora_dir}") |
| try: |
| model.save_pretrained(str(lora_dir)) |
| tokenizer.save_pretrained(str(lora_dir)) |
| logger.info("LoRA アダプタ保存完了 ✅") |
| except Exception as e: |
| logger.error(f"LoRA 保存エラー: {e}") |
| logger.debug(traceback.format_exc()) |
| raise |
|
|
| |
| if cfg.PUSH_TO_HUB and cfg.HF_TOKEN: |
| logger.info(f"[STEP 2/3] HF Hub に SFT+DPO チェックポイント push: {cfg.DPO_HF_REPO}") |
| try: |
| model.push_to_hub(cfg.DPO_HF_REPO, token=cfg.HF_TOKEN) |
| tokenizer.push_to_hub(cfg.DPO_HF_REPO, token=cfg.HF_TOKEN) |
| logger.info(f" ✅ https://huggingface.co/{cfg.DPO_HF_REPO}") |
| except Exception as e: |
| logger.error(f"DPO push エラー: {e}") |
| logger.debug(traceback.format_exc()) |
|
|
| |
| if cfg.SAVE_GGUF: |
| logger.info(f"[STEP 3/3] GGUF 変換中 ({cfg.GGUF_QUANTIZATION})...") |
| try: |
| gguf_dir = Path(cfg.DPO_OUTPUT_DIR) / "gguf" |
| gguf_dir.mkdir(parents=True, exist_ok=True) |
|
|
| if cfg.PUSH_TO_HUB and cfg.HF_TOKEN: |
| model.push_to_hub_gguf( |
| cfg.GGUF_HF_REPO, |
| tokenizer, |
| quantization_method=cfg.GGUF_QUANTIZATION, |
| token=cfg.HF_TOKEN, |
| ) |
| logger.info(f" ✅ https://huggingface.co/{cfg.GGUF_HF_REPO}") |
| else: |
| model.save_pretrained_gguf( |
| str(gguf_dir), |
| tokenizer, |
| quantization_method=cfg.GGUF_QUANTIZATION, |
| ) |
| logger.info(f" ✅ ローカル保存: {gguf_dir}") |
| except Exception as e: |
| logger.error(f"GGUF 変換エラー: {e}") |
| logger.debug(traceback.format_exc()) |
|
|
| elapsed = (datetime.now(timezone.utc) - start_time).total_seconds() |
| logger.info(f"=== DPO 完了 ({elapsed/60:.1f}分) ===") |
| logger.info(f"ログ: {log_file}") |
|
|
|
|
| if __name__ == "__main__": |
| main() |
|
|