big_fut_final / README.md
mintaeng's picture
Update README.md
0c7e800 verified
---
library_name: transformers
tags:
- unsloth
- trl
- sft
datasets:
- mintaeng/llm_futsaldata_yo
license: apache-2.0
language:
- ko
---
# FUT FUT CHAT BOT
- ์˜คํ”ˆ์†Œ์Šค ๋ชจ๋ธ์— LLM fine tuning ๊ณผ RAG ๋ฅผ ์ ์šฉํ•œ ์ƒ์„ฑํ˜• AI
- ํ’‹์‚ด์— ๋Œ€ํ•œ ๊ด€์‹ฌ์ด ๋†’์•„์ง€๋ฉด์„œ ์ˆ˜์š” ๋Œ€๋น„ ์ž…๋ฌธ์ž๋ฅผ ์œ„ํ•œ ์ •๋ณด ์ œ๊ณต ์„œ๋น„์Šค๊ฐ€ ํ•„์š”ํ•˜๋‹ค๊ณ  ๋А๊ปด ์ œ์ž‘ํ•˜๊ฒŒ ๋จ
- ํ’‹์‚ด ํ”Œ๋žซํผ์— ์‚ฌ์šฉ๋˜๋Š” ํ’‹์‚ด ์ •๋ณด ๋„์šฐ๋ฏธ ์ฑ—๋ด‡
- 'ํ•ด์š”์ฒด'๋กœ ๋‹ตํ•˜๋ฉฐ ๋ฌธ์žฅ ๋์— '์–ผ๋งˆ๋“ ์ง€ ๋ฌผ์–ด๋ณด์„ธ์š”~ ํ’‹ํ’‹~!' ์„ ์ถœ๋ ฅํ•จ
- train for 7h23m
## HOW TO USE
``` python
#!pip install transformers==4.40.0 accelerate
import os
import torch
from transformers import AutoTokenizer, AutoModelForCausalLM
model_id = 'Dongwookss/small_fut_final'
tokenizer = AutoTokenizer.from_pretrained(model_id)
model = AutoModelForCausalLM.from_pretrained(
model_id,
torch_dtype=torch.bfloat16,
device_map="auto",
)
model.eval()
```
**Query**
```python
from transformers import TextStreamer
PROMPT = '''Below is an instruction that describes a task. Write a response that appropriately completes the request.
์ œ์‹œํ•˜๋Š” context์—์„œ๋งŒ ๋Œ€๋‹ตํ•˜๊ณ  context์— ์—†๋Š” ๋‚ด์šฉ์€ ๋ชจ๋ฅด๊ฒ ๋‹ค๊ณ  ๋Œ€๋‹ตํ•ด'''
messages = [
{"role": "system", "content": f"{PROMPT}"},
{"role": "user", "content": f"{instruction}"}
]
input_ids = tokenizer.apply_chat_template(
messages,
add_generation_prompt=True,
return_tensors="pt"
).to(model.device)
terminators = [
tokenizer.eos_token_id,
tokenizer.convert_tokens_to_ids("<|eot_id|>")
]
text_streamer = TextStreamer(tokenizer)
_ = model.generate(
input_ids,
max_new_tokens=4096,
eos_token_id=terminators,
do_sample=True,
streamer = text_streamer,
temperature=0.6,
top_p=0.9,
repetition_penalty = 1.1
)
```
## Model Details
### Model Description
This is the model card of a ๐Ÿค— transformers model that has been pushed on the Hub. This model card has been automatically generated.
- **Developed by:** Dongwookss
- **Model type:** text generation
- **Language(s) (NLP):** Korean
- **Finetuned from model :** HuggingFaceH4/zephyr-7b-beta
### Data
https://huggingface.co/datasets/mintaeng/llm_futsaldata_yo
ํ•™์Šต ๋ฐ์ดํ„ฐ์…‹์€ beomi/KoAlpaca-v1.1a ๋ฅผ ๋ฒ ์ด์Šค๋กœ ์ถ”๊ฐ€, ๊ตฌ์ถ•, ์ „์ฒ˜๋ฆฌ ์ง„ํ–‰ํ•œ 23.5k ๋ฐ์ดํ„ฐ๋กœ ํŠœ๋‹ํ•˜์˜€์Šต๋‹ˆ๋‹ค.
๋ฐ์ดํ„ฐ์…‹์€ instruction, input, output ์œผ๋กœ ๊ตฌ์„ฑ๋˜์–ด ์žˆ์œผ๋ฉฐ tuning ๋ชฉํ‘œ์— ๋งž๊ฒŒ ๋งํˆฌ ์ˆ˜์ •ํ•˜์˜€์Šต๋‹ˆ๋‹ค.
๋„๋ฉ”์ธ ์ •๋ณด์— ๋Œ€ํ•œ ๋ฐ์ดํ„ฐ ์ถ”๊ฐ€ํ•˜์˜€์Šต๋‹ˆ๋‹ค.
## Training & Result
### Training Procedure
LoRA์™€ SFT Trainer ๋ฐฉ์‹์„ ์‚ฌ์šฉํ•˜์˜€์Šต๋‹ˆ๋‹ค.
#### Training Hyperparameters
- **Training regime:** bf16 mixed precision
```
r=32,
lora_alpha=64, # QLoRA : alpha = r/2 // LoRA : alpha =r*2
lora_dropout=0.05,
target_modules=[
"q_proj",
"k_proj",
"v_proj",
"o_proj",
"gate_proj",
"up_proj",
"down_proj",
], # ํƒ€๊ฒŸ ๋ชจ๋“ˆ
```
### Result
https://github.com/lucide99/Chatbot_FutFut
<!-- ## Bias, Risks, and Limitations -->
<!-- ## Model Examination [optional] -->
## Environment
L4 GPU
<!-- ## contributors -->