privacy-counsel-ko-8b-lora (v4-rebalanced)

๋Œ€ํ•œ๋ฏผ๊ตญ ๊ฐœ์ธ์ •๋ณด๋ณดํ˜ธ๋ฒ• ์ „๋ฌธ ์ƒ๋‹ด ๋ชจ๋ธ โ€” LoRA ์–ด๋Œ‘ํ„ฐ (667MB)

cywellai/privacy-counsel-ko-8b์˜ LoRA ์–ด๋Œ‘ํ„ฐ์ž…๋‹ˆ๋‹ค. ๋จธ์ง€๋œ ํ’€ ๋ชจ๋ธ(16GB)์ด ํ•„์š”ํ•˜๋ฉด ์œ„ ๋งํฌ๋ฅผ ์ด์šฉํ•˜์„ธ์š”.


์„ฑ๋Šฅ ์š”์•ฝ

์ง€ํ‘œ ๊ฐ’
5์ถ• ์ด์  14.38 / 15
Gold 144/150 (Silver 2, Fail 4)
๊ตฌ์กฐ 2.96 / 3
๋ฒ•์กฐํ•ญ 2.66 / 3
๋‚ด๋ถ€๊ตฌ์กฐ 2.95 / 3
์‹ค๋ฌด 2.93 / 3
ํ‘œํ˜„ 2.87 / 3
์‹œํ–‰๋ น ์ธ์šฉ๋ฅ  67%
๋‹ค๋งŒ ํŒจํ„ด๋ฅ  95%
ํ‰๊ท  ์‘๋‹ต ๊ธธ์ด 721์ž

150๋ฌธํ•ญ ๊ณจ๋“œ์…‹ ยท 5์ถ• v2.1 ์ฑ„์  ยท Claude Opus 4.6 ์ฑ„์  ยท 2026-02-27


์‚ฌ์šฉ๋ฒ•

from transformers import AutoModelForCausalLM, AutoTokenizer
from peft import PeftModel

base_model_name = "Qwen/Qwen3-8B"
adapter_name = "cywellai/privacy-counsel-ko-8b-lora"

tokenizer = AutoTokenizer.from_pretrained(base_model_name, trust_remote_code=True)
base_model = AutoModelForCausalLM.from_pretrained(
    base_model_name,
    torch_dtype="bfloat16",
    device_map="auto",
    trust_remote_code=True,
)
model = PeftModel.from_pretrained(base_model, adapter_name)

messages = [
    {"role": "user", "content": "๊ฐœ์ธ์ •๋ณด ์œ ์ถœ ์‹œ ๋Œ€์‘ ์ ˆ์ฐจ๋Š”?"},
]

text = tokenizer.apply_chat_template(messages, tokenize=False, add_generation_prompt=True)
inputs = tokenizer(text, return_tensors="pt").to(model.device)
outputs = model.generate(**inputs, max_new_tokens=1500, temperature=0.5, do_sample=True)
response = tokenizer.decode(outputs[0][inputs.input_ids.shape[1]:], skip_special_tokens=True)
print(response)

LoRA ์„ค์ •

ํ•ญ๋ชฉ ๊ฐ’
PEFT Type LoRA
Rank (r) 64
Alpha 128
Dropout 0.05
Target Modules q_proj, k_proj, v_proj, o_proj, gate_proj, up_proj, down_proj
ํ•™์Šต ๊ฐ€๋Šฅ ํŒŒ๋ผ๋ฏธํ„ฐ 174.6M / 8.37B (2.09%)

ํ•™์Šต ์„ค์ •

ํ•ญ๋ชฉ ๊ฐ’
๋ฒ ์ด์Šค ๋ชจ๋ธ Qwen/Qwen3-8B (์›๋ณธ ์‚ฌ์ „ํ•™์Šต ๋ชจ๋ธ)
ํ•™์Šต ๋ฐ์ดํ„ฐ 9,009๊ฑด (ํ’ˆ์งˆ ๊ธฐ๋ฐ˜ ๋ฆฌ๋ฐธ๋Ÿฐ์‹ฑ)
๊ฒ€์ฆ ๋ฐ์ดํ„ฐ 900๊ฑด (์ธตํ™” ์ƒ˜ํ”Œ๋ง)
Epochs 3
Batch Size 8 ร— 4 (effective 32)
Learning Rate 5e-5 (cosine, warmup 10%)
Max Seq Length 2048
์ตœ์ข… Eval Loss 0.3737
Token Accuracy 88.82%
ํ•™์Šต ์‹œ๊ฐ„ ~70๋ถ„ (NVIDIA H200 143GB)
Framework TRL 0.27.0, Transformers 4.57.6

๊ด€๋ จ ๋ชจ๋ธ

Downloads last month
-
Inference Providers NEW
This model isn't deployed by any Inference Provider. ๐Ÿ™‹ Ask for provider support

Model tree for cywellai/privacy-counsel-ko-8b-lora

Finetuned
Qwen/Qwen3-8B
Adapter
(1098)
this model

Evaluation results