--- license: mit language: - ko --- # open-llama-2-ko based model with modified DPO dataset This is an Korean Model based on * [beomi/open-llama-2-ko-7b] Dataset is modified from * [SJ-Donald/orca-dpo-pairs-ko] Parameters ``` learning_rate: float = 3e-4 lr_scheduler: str = "cosine" warmup_ratio: float = 0.1 lora_r: int = 16 lora_alpha: int = 16 lora_dropout: float = 0.05 optim='paged_adamw_32bit' bf16=True ```