๋ชจ๋ธ ์ด๋ฆ„

  • kogpt2-chatbot-lora

๋ชจ๋ธ ์„ค๋ช…

  • ์ฑ—๋ด‡์ด ์œ„๋กœํ•œ๋‹ค๋Š” ์ทจ์ง€์˜ ๋ฐ์ดํ„ฐ์…‹์œผ๋กœ ํŒŒ์ธํŠœ๋‹๋œ ํ•œ๊ตญ์–ด ์ฑ—๋ด‡ ๋ชจ๋ธ

๋ชจ๋ธ ์ƒ์„ธ

  • ๋ฒ ์ด์Šค ๋ชจ๋ธ: skt/kogpt2-base-v2
  • ํŒŒ์ธ ํŠœ๋‹ ๋ฐฉ๋ฒ•: LoRA
  • ์–ธ์–ด: ํ•œ๊ตญ์–ด

LoRA ์„ค์ •

r=16,
lora_alpha=32,
target_modules=["c_attn", "c_proj", "c_fc"],  
lora_dropout=0.05,
bias="none",
task_type=TaskType.CAUSAL_LM

ํ•™์Šต ์„ค์ •

num_train_epochs=10,
per_device_train_batch_size=4
per_device_eval_batch_size=8,
gradient_accumulation_steps=4,
learning_rate=0.0002,
warmup_steps=100,  
logging_steps=50,
eval_strategy= "epoch",
eval_steps=100,
save_strategy= "epoch",
save_steps=100,
load_best_model_at_end=True,
fp16=True,
report_to="none",
weight_decay=0.01, 

์‚ฌ์šฉ ๋ฐฉ๋ฒ•

from peft import PeftModel

# ๋ฒ ์ด์Šค ๋ชจ๋ธ ๋กœ๋“œ (๋ถ„๋ฅ˜์šฉ)
print("๋ฒ ์ด์Šค ๋ชจ๋ธ ๋กœ๋”ฉ")
base_model_reload = AutoModelForSequenceClassification.from_pretrained(
    "klue/bert-base",
    num_labels=2
)

# ์—…๋กœ๋“œํ•œ LoRA ์–ด๋Œ‘ํ„ฐ ๋กœ๋“œ
print(f"LoRA ์–ด๋Œ‘ํ„ฐ ๋กœ๋”ฉ: propagation/kogpt2-chatbot-lora")
model_reload = PeftModel.from_pretrained(base_model_reload, model_name_upload)
tokenizer_reload = AutoTokenizer.from_pretrained(model_name_upload)

# GPU๋กœ ์ด๋™
device = torch.device("cuda" if torch.cuda.is_available() else "cpu")
model_reload = model_reload.to(device)
model_reload.eval()

print("๋ชจ๋ธ ๋กœ๋“œ ์™„๋ฃŒ!")
Downloads last month

-

Downloads are not tracked for this model. How to track
Inference Providers NEW
This model isn't deployed by any Inference Provider. ๐Ÿ™‹ Ask for provider support

Model tree for propagation/kogpt2-chatbot-lora

Adapter
(26)
this model