KoELECTRA Intent Classifier
μ 무 μλν μν¬νλ‘μ° μμ΄μ νΈ(λλ)λ₯Ό μν νκ΅μ΄ μλ λΆλ₯ λͺ¨λΈμ λλ€.
μ¬μ©μμ μμ°μ΄ μ λ ₯μ 8κ° μ 무 μλ(intent)λ‘ λΆλ₯ν©λλ€.
Model Details
| Item | Detail |
|---|---|
| Base Model | monologg/koelectra-base-v3-discriminator |
| Architecture | ElectraForSequenceClassification |
| Parameters | 112.9M |
| Language | Korean |
| Experiment | v2_stage6 |
Intent Labels (8 classes)
| ID | Intent | Description |
|---|---|---|
| 0 | judgment |
μ 무 νλ¨ μμ² (μΉμΈ/λ°λ €/κ²ν ) |
| 1 | doc_search |
λ¬Έμ κ²μ |
| 2 | doc_generate |
λ¬Έμ μμ± (νμλ‘, λ³΄κ³ μ λ±) |
| 3 | doc_summary |
λ¬Έμ μμ½ |
| 4 | schedule_add |
μΌμ μΆκ°/λ±λ‘ |
| 5 | schedule_view |
μΌμ μ‘°ν/νμΈ |
| 6 | general |
μΌλ° λν/μ§λ¬Έ |
| 7 | doc_qa |
λ¬Έμ κΈ°λ° Q&A |
Performance
| Metric | Score |
|---|---|
| Test F1 | 97.88% |
| Adversarial F1 | 87.84% |
| Inference Speed | 7.9ms / sample |
- Training Data: 2,425 sentences (2,327 base + 98 augmented)
- Test Data: 286 samples + 450 adversarial samples
- Label Smoothing: 0.1
Usage
from transformers import AutoTokenizer, AutoModelForSequenceClassification
import torch
model_name = "jiyong1110/koelectra-intent-classifier"
tokenizer = AutoTokenizer.from_pretrained(model_name)
model = AutoModelForSequenceClassification.from_pretrained(model_name)
text = "λ΄μΌ μ€ν 3μμ νμ μ‘μμ€"
inputs = tokenizer(text, return_tensors="pt", truncation=True, max_length=128)
with torch.no_grad():
outputs = model(**inputs)
pred = torch.argmax(outputs.logits, dim=-1).item()
id2label = model.config.id2label
print(f"Intent: {id2label[pred]}") # schedule_add
Training Details
7λ¨κ³ μ€νμ κ±°μ³ μ΅μ νλ λͺ¨λΈμ λλ€:
- Stage 1: Claude + GPT-4o κΈ°λ° νμ΅ λ°μ΄ν° μμ±
- Stage 2: 3κ° λͺ¨λΈ λ² μ΄μ€λΌμΈ λΉκ΅ (BERT, KoBERT, KoELECTRA)
- Stage 3: 32-point νμ΄νΌνλΌλ―Έν° 그리λ μμΉ
- Stage 4: μ΅μ’ νκ° (μ λμ ν μ€νΈ, μλ λ²€μΉλ§ν¬)
- Stage 5: μλ¬ λΆμ λ° νκ² μ¦κ°
- Stage 6: Label smoothing μ μ©
- Stage 7: μλλ¦¬μ€ ν μ€νΈ (100 samples)
Project
SKN21-FINAL-3TEAM β WorkFlow Agent (λλ)
LangGraph κΈ°λ° λ©ν° μμ΄μ νΈ μ 무 μλν μμ€ν μ Intent Classification λͺ¨λμ λλ€.
- Downloads last month
- 13
Model tree for jiyong1110/koelectra-intent-classifier
Base model
monologg/koelectra-base-v3-discriminatorEvaluation results
- Test F1self-reported0.979
- Adversarial F1self-reported0.878
- Inference Speedself-reported7.900