WeIN_bio_Qwen3-8B

ํ•œ๊ตญ์–ด ์˜๋ฃŒ/๋ฐ”์ด์˜ค ๋„๋ฉ”์ธ์— ํŠนํ™”๋œ Qwen3-8B ๊ธฐ๋ฐ˜ LoRA ํŒŒ์ธํŠœ๋‹ ๋ชจ๋ธ์ž…๋‹ˆ๋‹ค. KorMedMCQA ๋ฒค์น˜๋งˆํฌ์—์„œ 69.06% ์ •ํ™•๋„๋ฅผ ๋‹ฌ์„ฑํ•˜์—ฌ, ๋ชฉํ‘œ 65%๋ฅผ ์ดˆ๊ณผ ๋‹ฌ์„ฑํ–ˆ์Šต๋‹ˆ๋‹ค.

Model Details

Model Description

  • Base Model: Qwen/Qwen3-8B
  • Model Type: Causal Language Model (LoRA Adapter)
  • Fine-tuning Method: SFT (Supervised Fine-Tuning) with LoRA via TRL
  • Language: Korean (ko)
  • License: Apache-2.0
  • Domain: ์˜๋ฃŒ/๋ฐ”์ด์˜ค (Medical/Bio)
  • Adapter Size: 167MB (safetensors)

Model Sources

Training Details

Training Data

  • Dataset: ํ•œ๊ตญ์–ด ์˜๋ฃŒ SFT ๋ฐ์ดํ„ฐ (Korean Medical SFT Dataset)
  • Training Samples: 35,882
  • Format: Instruction-following format with Chain-of-Thought reasoning
  • Domain Coverage: ์น˜๊ณผ, ์˜ํ•™, ๊ฐ„ํ˜ธํ•™, ์•ฝํ•™, ์•ฝ์‚ฌ ์‹œํ—˜ ๋ฌธ์ œ ๊ธฐ๋ฐ˜
  • Data Sources: AIHub ์˜๋ฃŒ ๋ฐ์ดํ„ฐ(15,354), KorMedMCQA(3,401), PubMedQA(827), ๋กœ๋“œ๋งต ๋ฐ”์ด์˜ค(657), ๊ฑด๊ฐ•๊ธฐ๋Šฅ์‹ํ’ˆ(346) ๋“ฑ

Training Hyperparameters

Parameter Value
Base Model Qwen/Qwen3-8B
LoRA Rank (r) 16
LoRA Alpha 32
LoRA Dropout 0.1
LoRA Target Modules q_proj, k_proj, v_proj, o_proj, gate_proj, up_proj, down_proj
Learning Rate 2e-4
LR Scheduler Cosine
Warmup Ratio 0.03
Epochs 3
Per-device Batch Size 2
Gradient Accumulation Steps 16
Effective Batch Size 32
Max Sequence Length 2,048
Attention Implementation SDPA
Precision bf16
Optimizer AdamW (fused)
Seed 42
DoRA False
RSLoRA False

Training Infrastructure

  • Framework: PEFT, TRL, Transformers
  • Hardware: NVIDIA H200 (143GB VRAM)
  • Training Duration: ~4-5 hours
  • Final Train Loss: 0.2854
  • Mean Token Accuracy: 91.85%

Evaluation

Benchmark: KorMedMCQA

ํ•œ๊ตญ ์˜๋ฃŒ ์ž๊ฒฉ์‹œํ—˜ ๊ธฐ๋ฐ˜ ๊ฐ๊ด€์‹ ๋ฌธ์ œ (Multiple Choice QA) ๋ฒค์น˜๋งˆํฌ์—์„œ ํ‰๊ฐ€ํ–ˆ์Šต๋‹ˆ๋‹ค.

Overall Performance

Metric Value
Overall Accuracy 69.06%
Total Samples 3,009
Correct 2,078
Extract Fail Rate 0.00%
Evaluation Mode Direct (zero-shot)

Per-Subject Performance

Subject Correct Total Accuracy
๊ฐ„ํ˜ธ์‚ฌ (Nurse) 687 878 78.25%
์•ฝ์‚ฌ (Pharmacist) 198 271 73.06%
์•ฝํ•™ (Pharm Science) 422 614 68.73%
์˜์‚ฌ (Doctor) 297 435 68.28%
์น˜๊ณผ์˜์‚ฌ (Dentist) 474 811 58.45%

Performance History

Experiment Accuracy Note
qwen3_baseline 60.39% Qwen3-8B ๊ธฐ์ค€์„  (ํŒŒ์ธํŠœ๋‹ ์—†์Œ)
qwen3_sft_001 64.27% ์ดˆ๊ธฐ SFT (2 epochs)
qwen3_sft_002 67.50% ํ™•์žฅ ํ•™์Šต (3 epochs)
qwen3_sft_v3 69.06% ์ตœ์‹  ๋ฐ์ดํ„ฐ์…‹ v3 ์ ์šฉ (์ตœ๊ณ  ์„ฑ๋Šฅ)

SOTA Comparison on KorMedMCQA

Model Accuracy Parameters License
WeIN_bio_Qwen3-8B (๋ณธ ๋ชจ๋ธ) 69.06% 8B Apache 2.0
Qwen3-8B (baseline) 60.39% 8B Apache 2.0
EXAONE 7.8B 56.10% 7.8B Non-Commercial
Random Guess 20.00% - -

Usage

Quick Start

from transformers import AutoModelForCausalLM, AutoTokenizer
from peft import PeftModel

# Load base model and adapter
base_model_id = "Qwen/Qwen3-8B"
adapter_id = "dhkim0324/WeIN_bio_Qwen3-8B"

tokenizer = AutoTokenizer.from_pretrained(adapter_id, trust_remote_code=True)
model = AutoModelForCausalLM.from_pretrained(
    base_model_id,
    device_map="auto",
    torch_dtype="auto",
    trust_remote_code=True,
)
model = PeftModel.from_pretrained(model, adapter_id)

# Inference
prompt = "๋‹ค์Œ ์˜๋ฃŒ ๊ด€๋ จ ๊ฐ๊ด€์‹ ๋ฌธ์ œ์— ๋‹ตํ•˜์‹œ์˜ค.\n\n๋ฌธ์ œ: ์‹ฌ๊ทผ๊ฒฝ์ƒ‰์˜ ๊ฐ€์žฅ ํ”ํ•œ ์›์ธ์€?\n1. ๊ด€์ƒ๋™๋งฅ ์ฃฝ์ƒ๊ฒฝํ™”์ฆ\n2. ์‹ฌ์žฅํŒ๋ง‰์งˆํ™˜\n3. ์‹ฌ๊ทผ์—ผ\n4. ๋Œ€๋™๋งฅ๋ฐ•๋ฆฌ\n\n์ •๋‹ต:"
inputs = tokenizer(prompt, return_tensors="pt").to(model.device)
outputs = model.generate(**inputs, max_new_tokens=128, do_sample=False)
print(tokenizer.decode(outputs[0], skip_special_tokens=True))

Merge Adapter (Optional)

from transformers import AutoModelForCausalLM
from peft import PeftModel

base_model = AutoModelForCausalLM.from_pretrained("Qwen/Qwen3-8B", torch_dtype="auto")
model = PeftModel.from_pretrained(base_model, "dhkim0324/WeIN_bio_Qwen3-8B")
merged_model = model.merge_and_unload()
merged_model.save_pretrained("merged_model")

Limitations

  • ์˜๋ฃŒ ์ „๋ฌธ๊ฐ€์˜ ์ž„์ƒ ํŒ๋‹จ์„ ๋Œ€์ฒดํ•  ์ˆ˜ ์—†์œผ๋ฉฐ, ์—ฐ๊ตฌ ๋ฐ ๊ต์œก ๋ชฉ์ ์œผ๋กœ๋งŒ ์‚ฌ์šฉํ•ด์•ผ ํ•ฉ๋‹ˆ๋‹ค.
  • KorMedMCQA ๋ฒค์น˜๋งˆํฌ ๊ธฐ์ค€ ํ‰๊ฐ€์ด๋ฉฐ, ์‹ค์ œ ์ž„์ƒ ํ™˜๊ฒฝ์—์„œ์˜ ์„ฑ๋Šฅ์€ ๋‹ค๋ฅผ ์ˆ˜ ์žˆ์Šต๋‹ˆ๋‹ค.
  • ์น˜๊ณผ์˜์‚ฌ ๋„๋ฉ”์ธ์—์„œ ์ƒ๋Œ€์ ์œผ๋กœ ๋‚ฎ์€ ์„ฑ๋Šฅ(58.45%)์„ ๋ณด์ž…๋‹ˆ๋‹ค.
  • Chain-of-Thought ์ถ”๋ก  ์‹œ ์˜ํ•™์ ์œผ๋กœ ๋ถ€์ •ํ™•ํ•œ ์ถ”๋ก ์„ ์ƒ์„ฑํ•  ์ˆ˜ ์žˆ์Šต๋‹ˆ๋‹ค.
  • ํ•™์Šต ๋ฐ์ดํ„ฐ ์‹œ์  ์ดํ›„์˜ ์ตœ์‹  ์˜ํ•™์  ๋ฐœ์ „์€ ๋ฐ˜์˜๋˜์ง€ ์•Š์Šต๋‹ˆ๋‹ค.

Citation

@misc{wein_bio_qwen3_2026,
    title={WeIN_bio_Qwen3-8B: Korean Medical Domain LoRA Adapter for Qwen3-8B},
    author={dhkim0324},
    year={2026},
    publisher={Hugging Face}
}
Downloads last month
2
Inference Providers NEW
This model isn't deployed by any Inference Provider. ๐Ÿ™‹ Ask for provider support

Model tree for dhkim0324/WeIN_bio_Qwen3-8B

Finetuned
Qwen/Qwen3-8B
Adapter
(1095)
this model

Papers for dhkim0324/WeIN_bio_Qwen3-8B

Evaluation results