Korean XLM-RoBERTa Classifier
μ΄ λͺ¨λΈμ **xlm-roberta-base**λ₯Ό κΈ°λ°μΌλ‘ νμΈνλλ νκ΅μ΄/μμ΄ μ΄μ€μΈμ΄ ν
μ€νΈ λΆλ₯ λͺ¨λΈμ
λλ€.
μ΄ 66κ° λΌλ²¨ λΆλ₯κ° κ°λ₯νλ©°, λΌλ²¨ μ 보λ label_mapping.json νμΌμμ νμΈν μ μμ΅λλ€.
π Files in Repository
config.json: λͺ¨λΈ μ€μ tokenizer.json/tokenizer_config.json: ν ν¬λμ΄μ special_tokens_map.json: νΉμ ν ν° λ§€νpytorch_model.binλλmodel.safetensors(λ μ€ νλλ§ μ¬μ©,safetensorsκΆμ₯)label_mapping.json: μΈλ±μ€ β λΌλ²¨ λ§€νclassifier.pkl,label_embeddings.pkl: μΆκ° λΆλ₯κΈ° λ° μλ² λ©label_independence_analysis.py: λΆμ μ€ν¬λ¦½νΈ (λΆκ° μλ£)
π Usage
from transformers import AutoTokenizer, AutoModelForSequenceClassification
import torch
model_name = "Halfotter/home" # Hugging Face repo κ²½λ‘
# ν ν¬λμ΄μ μ λͺ¨λΈ λΆλ¬μ€κΈ°
tokenizer = AutoTokenizer.from_pretrained(model_name)
model = AutoModelForSequenceClassification.from_pretrained(model_name)
# μ
λ ₯ μμ
inputs = tokenizer("ν
μ€νΈ λ¬Έμ₯", return_tensors="pt")
outputs = model(**inputs)
# μννΈλ§₯μ€λ‘ νλ₯ λ³ν
probs = torch.nn.functional.softmax(outputs.logits, dim=-1)
label_id = torch.argmax(probs).item()
print("Predicted Label ID:", label_id)
Model tree for Halfotter/home
Base model
FacebookAI/xlm-roberta-base