sentiment-roberta-base

Fine-tuned RoBERTa-base for binary sentiment classification on the Sentiment140 dataset (1.6M tweets).

Base model

FacebookAI/roberta-base โ€” the original RoBERTa-base from Liu et al. (2019), 125M parameters.

Training

  • Dataset: Sentiment140 (1.6M tweets, 80/20 split, seed 42)
  • Hyperparameters: learning rate 2e-5, batch size 16, 3 epochs
  • Hardware: NVIDIA A10G, AWS SageMaker (g5.2xlarge)
  • Training time: 7.5 hours
  • Trainer: Hugging Face Transformers + Trainer API; load_best_model_at_end=True

Test set performance

Metric Value
Accuracy 89.11%
Precision 0.901
Recall 0.879
F1 0.890

Intended use

Demonstration model for an academic purposes

Limitations

  • English only, binary sentiment, 2009-era Twitter language.
  • Sentiment140 labels generated automatically using emoticons (distant supervision), introducing systematic noise.
  • Does not handle sarcasm reliably (the dataset does not separate it as a phenomenon).
Downloads last month
50
Safetensors
Model size
0.1B params
Tensor type
F32
ยท
Inference Providers NEW
This model isn't deployed by any Inference Provider. ๐Ÿ™‹ Ask for provider support

Dataset used to train heican/sentiment-roberta-base

Space using heican/sentiment-roberta-base 1