--- base_model: google-bert/bert-base-uncased datasets: - stanfordnlp/sentiment140 pipeline_tag: text-classification --- # sentiment-bert-base Fine-tuned BERT-base for binary sentiment classification on the Sentiment140 dataset (1.6M tweets). ## Base model [google-bert/bert-base-uncased](https://huggingface.co/google-bert/bert-base-uncased) — the original BERT-base-uncased from Devlin et al. (2019), 110M parameters. ## Training - Dataset: Sentiment140 (1.6M tweets, 80/20 split, seed 42) - Hyperparameters: learning rate 2e-5, batch size 16, 3 epochs - Hardware: NVIDIA A10G, AWS SageMaker (g5.2xlarge) - Training time: 7.3 hours - Trainer: Hugging Face Transformers + Trainer API; load_best_model_at_end=True ## Test set performance | Metric | Value | |---|---| | Accuracy | 87.46% | | Precision | 0.880 | | Recall | 0.869 | | F1 | 0.874 | ## Intended use Demonstration model for an academic purposes ## Limitations - English only, binary sentiment, 2009-era Twitter language. - Sentiment140 labels generated automatically using emoticons (distant supervision), introducing systematic noise. - Does not handle sarcasm reliably (the dataset does not separate it as a phenomenon).