korean-chatbot-lora / README.md
JINIIII's picture
Upload README.md with huggingface_hub
2f0efad verified
metadata
license: mit
base_model: beomi/open-llama-2-ko-7b
tags:
  - llama
  - lora
  - korean
  - text-generation
language:
  - ko

Korean Chatbot (LoRA Fine-tuned)

์ด ๋ชจ๋ธ์€ ํ•œ๊ตญ์–ด ๋Œ€ํ™” ๋ชจ๋ธ์ž…๋‹ˆ๋‹ค.

์‚ฌ์šฉ ๋ฐฉ๋ฒ•

from transformers import AutoModelForCausalLM, AutoTokenizer
from peft import PeftModel
import torch

# ๋ฒ ์ด์Šค ๋ชจ๋ธ ๋กœ๋“œ
base_model = AutoModelForCausalLM.from_pretrained(
    "beomi/open-llama-2-ko-7b",
    device_map="auto",
    torch_dtype=torch.float16
)

# LoRA ์–ด๋Œ‘ํ„ฐ ๋กœ๋“œ
model = PeftModel.from_pretrained(base_model, "JINIIII/korean-chatbot-lora")
tokenizer = AutoTokenizer.from_pretrained("JINIIII/korean-chatbot-lora")

# ์ถ”๋ก 
prompt = "์งˆ๋ฌธ: ์ธ๊ณต์ง€๋Šฅ์ด๋ž€ ๋ฌด์—‡์ธ๊ฐ€์š”?\n๋‹ต๋ณ€:"

inputs = tokenizer(prompt, return_tensors="pt").to(model.device)
outputs = model.generate(**inputs, max_length=200, temperature=0.7)
print(tokenizer.decode(outputs[0], skip_special_tokens=True))

ํ™œ์šฉ ์˜ˆ์‹œ

  • ์˜ํ™” ๋ฆฌ๋ทฐ ๊ฐ์ • ๋ถ„์„
  • ์ƒํ’ˆ ๋ฆฌ๋ทฐ ๋ถ„์„
  • SNS ๊ฐ์ • ๋ชจ๋‹ˆํ„ฐ๋ง

๋ผ์ด์„ ์Šค

MIT License

์ž‘์„ฑ์ž

๋‚˜๋Š”์•ผ ์ง„์ˆ˜์•ผใ…

Note: ์ด ๋ชจ๋ธ์€ ๊ต์œก ๋ชฉ์ ์œผ๋กœ ๋งŒ๋“ค์–ด์กŒ์Šต๋‹ˆ๋‹ค.