privacy-counsel-ko-8b-lora (v4-rebalanced)
๋ํ๋ฏผ๊ตญ ๊ฐ์ธ์ ๋ณด๋ณดํธ๋ฒ ์ ๋ฌธ ์๋ด ๋ชจ๋ธ โ LoRA ์ด๋ํฐ (667MB)
cywellai/privacy-counsel-ko-8b์ LoRA ์ด๋ํฐ์ ๋๋ค. ๋จธ์ง๋ ํ ๋ชจ๋ธ(16GB)์ด ํ์ํ๋ฉด ์ ๋งํฌ๋ฅผ ์ด์ฉํ์ธ์.
์ฑ๋ฅ ์์ฝ
| ์งํ | ๊ฐ |
|---|---|
| 5์ถ ์ด์ | 14.38 / 15 |
| Gold | 144/150 (Silver 2, Fail 4) |
| ๊ตฌ์กฐ | 2.96 / 3 |
| ๋ฒ์กฐํญ | 2.66 / 3 |
| ๋ด๋ถ๊ตฌ์กฐ | 2.95 / 3 |
| ์ค๋ฌด | 2.93 / 3 |
| ํํ | 2.87 / 3 |
| ์ํ๋ น ์ธ์ฉ๋ฅ | 67% |
| ๋ค๋ง ํจํด๋ฅ | 95% |
| ํ๊ท ์๋ต ๊ธธ์ด | 721์ |
150๋ฌธํญ ๊ณจ๋์ ยท 5์ถ v2.1 ์ฑ์ ยท Claude Opus 4.6 ์ฑ์ ยท 2026-02-27
์ฌ์ฉ๋ฒ
from transformers import AutoModelForCausalLM, AutoTokenizer
from peft import PeftModel
base_model_name = "Qwen/Qwen3-8B"
adapter_name = "cywellai/privacy-counsel-ko-8b-lora"
tokenizer = AutoTokenizer.from_pretrained(base_model_name, trust_remote_code=True)
base_model = AutoModelForCausalLM.from_pretrained(
base_model_name,
torch_dtype="bfloat16",
device_map="auto",
trust_remote_code=True,
)
model = PeftModel.from_pretrained(base_model, adapter_name)
messages = [
{"role": "user", "content": "๊ฐ์ธ์ ๋ณด ์ ์ถ ์ ๋์ ์ ์ฐจ๋?"},
]
text = tokenizer.apply_chat_template(messages, tokenize=False, add_generation_prompt=True)
inputs = tokenizer(text, return_tensors="pt").to(model.device)
outputs = model.generate(**inputs, max_new_tokens=1500, temperature=0.5, do_sample=True)
response = tokenizer.decode(outputs[0][inputs.input_ids.shape[1]:], skip_special_tokens=True)
print(response)
LoRA ์ค์
| ํญ๋ชฉ | ๊ฐ |
|---|---|
| PEFT Type | LoRA |
| Rank (r) | 64 |
| Alpha | 128 |
| Dropout | 0.05 |
| Target Modules | q_proj, k_proj, v_proj, o_proj, gate_proj, up_proj, down_proj |
| ํ์ต ๊ฐ๋ฅ ํ๋ผ๋ฏธํฐ | 174.6M / 8.37B (2.09%) |
ํ์ต ์ค์
| ํญ๋ชฉ | ๊ฐ |
|---|---|
| ๋ฒ ์ด์ค ๋ชจ๋ธ | Qwen/Qwen3-8B (์๋ณธ ์ฌ์ ํ์ต ๋ชจ๋ธ) |
| ํ์ต ๋ฐ์ดํฐ | 9,009๊ฑด (ํ์ง ๊ธฐ๋ฐ ๋ฆฌ๋ฐธ๋ฐ์ฑ) |
| ๊ฒ์ฆ ๋ฐ์ดํฐ | 900๊ฑด (์ธตํ ์ํ๋ง) |
| Epochs | 3 |
| Batch Size | 8 ร 4 (effective 32) |
| Learning Rate | 5e-5 (cosine, warmup 10%) |
| Max Seq Length | 2048 |
| ์ต์ข Eval Loss | 0.3737 |
| Token Accuracy | 88.82% |
| ํ์ต ์๊ฐ | ~70๋ถ (NVIDIA H200 143GB) |
| Framework | TRL 0.27.0, Transformers 4.57.6 |
๊ด๋ จ ๋ชจ๋ธ
- ํ ๋ชจ๋ธ (๋จธ์ง): cywellai/privacy-counsel-ko-8b (16GB, ๋ฐ๋ก ์ฌ์ฉ ๊ฐ๋ฅ)
- ๋ฒ ์ด์ค ๋ชจ๋ธ: Qwen/Qwen3-8B
- Downloads last month
- -
Model tree for cywellai/privacy-counsel-ko-8b-lora
Evaluation results
- Total Score (0-15, 5-axis)self-reported14.380
- Gold Pass Rate (150 questions)self-reported0.960