kogpt2-chatbot-lora / README.md
jihun-pae's picture
Upload README.md with huggingface_hub
3c947ee verified
metadata
license: cc-by-nc-nd-4.0
base_model: skt/kogpt2-base-v2
tags:
  - gpt2
  - lora
  - korean
  - chatbot
language:
  - ko

๋ชจ๋ธ ์ด๋ฆ„

jihun-pae/kogpt2-chatbot-lora

๋ชจ๋ธ ์„ค๋ช…

  • LoRA ํ•œ๊ตญ์–ด ์ฑ—๋ด‡

๋ชจ๋ธ ์ƒ์„ธ

  • ๊ต์œก์šฉ ์‹ค์Šต ๋ชจ๋ธ์ž…๋‹ˆ๋‹ค.

LoRA ์„ค์ •

lora_config = LoraConfig( r=16, lora_alpha=32, target_modules=["c_attn", "c_proj", "c_fc"],
lora_dropout=0.05, bias="none", task_type=TaskType.CAUSAL_LM )

ํ•™์Šต ์„ค์ •

training_args = TrainingArguments( output_dir="./lora_koqpt2_chatbot", num_train_epochs=10, per_device_train_batch_size=4, per_device_eval_batch_size=8, gradient_accumulation_steps=4, learning_rate=0.0002, warmup_steps=100,
logging_steps=50, eval_strategy="epoch", eval_steps=100, save_strategy="epoch", save_steps=100, load_best_model_at_end=True, fp16=True, report_to="none", weight_decay=0.01, )

ํ•™์Šต ๊ฒฐ๊ณผ

์‚ฌ์šฉ ๋ฐฉ๋ฒ•

# ํ…Œ์ŠคํŠธ
test_questions = [
    "์•ˆ๋…•ํ•˜์„ธ์š”?",
    "์˜ค๋Š˜ ๋‚ ์”จ๊ฐ€ ์–ด๋•Œ?",
    "๋ฐฐ๊ณ ํ”ˆ๋ฐ ๋ญ ๋จน์„๊นŒ?",
    "์ฃผ๋ง์— ๋ญํ•˜์ง€?"
]

print("=== ์ฑ—๋ด‡ ํ…Œ์ŠคํŠธ ===")
for q in test_questions:
    print(f"
์งˆ๋ฌธ: ๊ฐ•์›๋„ ์„ค์•…์‚ฐ ๊ทผ์ฒ˜ ๋ง›์ง‘ ์ข€ ์ถ”์ฒœํ•ด์ฃผ์„ธ์š”.")
    print(f"๋‹ต๋ณ€: ์—ฌํ–‰๊ฐ€๋ด๋„ ์ข‹์„๊ฑฐ ๊ฐ™์•„์š”.")

๋ชจ๋ธ ์ •๋ณด