Model Card for Model ID


language: ko datasets: - imdb metrics: - accuracy - f1 model_name: bert-based-uncased-imdb-sentiment license: apache-2.0 tags: - ํ…์ŠคํŠธ๋ถ„๋ฅ˜ - ๊ฐ์„ฑ๋ถ„์„ - imdb

๋ชจ๋ธ ์นด๋“œ: IMDB ๋ฐ์ดํ„ฐ๋กœ ํŒŒ์ธํŠœ๋‹๋œ BERT-base-uncased ๊ฐ์„ฑ ๋ถ„์„ ๋ชจ๋ธ

๋ชจ๋ธ ๊ฐœ์š”

์ด ๋ชจ๋ธ์€ BERT-base-uncased๋ฅผ ๊ธฐ๋ฐ˜์œผ๋กœ ํ•˜๋ฉฐ, IMDB ์˜ํ™” ๋ฆฌ๋ทฐ ๋ฐ์ดํ„ฐ์…‹์„ ์ด์šฉํ•ด ๊ฐ์„ฑ ๋ถ„์„(Sentiment Analysis) ์ž‘์—…์— ๋งž๊ฒŒ ํŒŒ์ธํŠœ๋‹๋˜์—ˆ์Šต๋‹ˆ๋‹ค.
ํ…์ŠคํŠธ๋ฅผ ๋‘ ๊ฐ€์ง€ ๊ฐ์ •์œผ๋กœ ๋ถ„๋ฅ˜ํ•ฉ๋‹ˆ๋‹ค:

  • 0: ๋ถ€์ • (Negative)
  • 1: ๊ธ์ • (Positive)

๋ชจ๋ธ ์ƒ์„ธ ์ •๋ณด

  • ๊ธฐ๋ฐ˜ ๋ชจ๋ธ: bert-base-uncased
  • ์ž‘์—…(Task): ๊ฐ์„ฑ ๋ถ„๋ฅ˜ (Sentiment Classification)
  • ๋ฐ์ดํ„ฐ์…‹: IMDB
  • ๋ผ๋ฒจ ์ˆ˜: 2 (๋ถ€์ •, ๊ธ์ •)
  • ์–ธ์–ด: ์˜์–ด

ํ•™์Šต ์ •๋ณด

  • ํŒŒ์ธํŠœ๋‹ ๋ฐ์ดํ„ฐ์…‹: IMDB (Hugging Face Datasets์˜ ๊ณต์‹ ๋ถ„ํ•  ์‚ฌ์šฉ)
  • ์—ํฌํฌ ์ˆ˜: (ํ•„์š” ์‹œ ๋ช…์‹œ)
  • ์˜ตํ‹ฐ๋งˆ์ด์ €: AdamW
  • ํ•™์Šต๋ฅ : (ํ•„์š” ์‹œ ๋ช…์‹œ)
  • ํ‰๊ฐ€ ์ง€ํ‘œ: Accuracy, F1-score

์„ฑ๋Šฅ ์ง€ํ‘œ

์ง€ํ‘œ ์ ์ˆ˜
์ •ํ™•๋„ (Accuracy) 0.88
F1 ์ ์ˆ˜ (F1 Score) 0.88

์ธก์ •์€ IMDB ํ…Œ์ŠคํŠธ ์„ธํŠธ์—์„œ ์ˆ˜ํ–‰๋˜์—ˆ์Šต๋‹ˆ๋‹ค.

์‚ฌ์šฉ ๋ฐฉ๋ฒ•

from transformers import AutoTokenizer, AutoModelForSequenceClassification

model_name = "your-username/bert-based-uncased-imdb-sentiment"
tokenizer = AutoTokenizer.from_pretrained(model_name)
model = AutoModelForSequenceClassification.from_pretrained(model_name)

text = "This movie was amazing!"
inputs = tokenizer(text, return_tensors="pt")
outputs = model(**inputs)
pred = outputs.logits.argmax(-1).item()
print("Sentiment:", "Positive" if pred == 1 else "Negative")

์‚ฌ์šฉ ์˜๋„

์ด ๋ชจ๋ธ์€ ์˜์–ด ์˜ํ™” ๋ฆฌ๋ทฐ ๊ฐ์„ฑ ๋ถ„์„ ์ž‘์—…์„ ์œ„ํ•ด ์„ค๊ณ„๋˜์—ˆ์Šต๋‹ˆ๋‹ค.
๋‹ค๋ฅธ ์˜์–ด ํ…์ŠคํŠธ ๊ฐ์„ฑ ๋ถ„์„ ๊ณผ์ œ์— ์‚ฌ์šฉํ•  ๊ฒฝ์šฐ, ์ถ”๊ฐ€ ํŒŒ์ธํŠœ๋‹์„ ํ†ตํ•ด ํ™•์žฅ ์‘์šฉํ•  ์ˆ˜ ์žˆ์Šต๋‹ˆ๋‹ค.

์ œํ•œ ์‚ฌํ•ญ

  • ์˜ํ™” ๋ฆฌ๋ทฐ ๋ฐ์ดํ„ฐ์— ํ•œ์ •๋˜์–ด ํ•™์Šต๋˜์—ˆ๊ธฐ ๋•Œ๋ฌธ์—, ๋‹ค๋ฅธ ๋„๋ฉ”์ธ(์˜ˆ: ์ œํ’ˆ ๋ฆฌ๋ทฐ, ๋‰ด์Šค ๊ธฐ์‚ฌ ๋“ฑ)์—์„œ๋Š” ์ •ํ™•๋„๊ฐ€ ๋‚ฎ์„ ์ˆ˜ ์žˆ์Šต๋‹ˆ๋‹ค.
  • ์ค‘๋ฆฝ์ ์ด๊ฑฐ๋‚˜ ๋ณตํ•ฉ์ ์ธ ๊ฐ์ •์ด ํฌํ•จ๋œ ๋ฌธ์žฅ์€ ์˜ค๋ถ„๋ฅ˜๋  ๊ฐ€๋Šฅ์„ฑ์ด ์žˆ์Šต๋‹ˆ๋‹ค.

์ธ์šฉ

์—ฐ๊ตฌ ๋˜๋Š” ์„œ๋น„์Šค์—์„œ ๋ณธ ๋ชจ๋ธ์„ ์‚ฌ์šฉํ•  ๊ฒฝ์šฐ ์•„๋ž˜์™€ ๊ฐ™์ด ์ธ์šฉํ•ด์ฃผ์„ธ์š”.

@misc{bert-imdb-finetuned,
  title={BERT-base-uncased fine-tuned on IMDB Sentiment Dataset},
  author={Your Name},
  year={2025},
  howpublished={Hugging Face Hub},
}

๋ผ์ด์„ ์Šค

๋ณธ ๋ชจ๋ธ์€ Apache 2.0 ๋ผ์ด์„ ์Šค ํ•˜์— ๋ฐฐํฌ๋ฉ๋‹ˆ๋‹ค.

์ฐธ๊ณ  ๋ฌธํ—Œ


Downloads last month
-
Safetensors
Model size
0.1B params
Tensor type
F32
ยท
Inference Providers NEW
This model isn't deployed by any Inference Provider. ๐Ÿ™‹ Ask for provider support

Paper for blockenters/bert-based-uncased-imdb2-v01