BERT: Pre-training of Deep Bidirectional Transformers for Language Understanding
Paper
•
1810.04805
•
Published
•
25
This model is a fine-tuned version of bert-base-uncased on the Stanford Sentiment Treebank v2 (SST-2) dataset. It was trained to perform binary sentiment classification (positive/negative) on movie review sentences.
The model was fine-tuned on the SST-2 dataset from the GLUE benchmark:
The SST-2 dataset consists of sentences from movie reviews with their associated binary sentiment labels.
For more information about the dataset, see the GLUE benchmark dataset card.
bert-base-uncased| Epoch | Training Loss | Validation Loss | Accuracy |
|---|---|---|---|
| 1 | 0.256300 | 0.427576 | 0.899083 |
| 2 | 0.169200 | 0.415616 | 0.903670 |
| 3 | 0.095600 | 0.426083 | 0.903670 |
Final training loss: 0.19818013534790577
This model may have inherited biases from its training data and pre-training corpus:
from transformers import pipeline
sentiment_analyzer = pipeline("sentiment-analysis", model="radubutucelea23/bert_base_uncased_sst2")
texts = ["I really enjoyed this movie, the acting was superb.",
"The plot was confusing and the characters were poorly developed."]
results = sentiment_analyzer(texts)
print(results)
If you use this model, please cite:
@inproceedings{socher-etal-2013-recursive,
title = "Recursive Deep Models for Semantic Compositionality Over a Sentiment Treebank",
author = "Socher, Richard and Perelygin, Alex and Wu, Jean and Chuang, Jason and Manning, Christopher D. and Ng, Andrew Y. and Potts, Christopher",
booktitle = "Proceedings of EMNLP",
year = "2013"
}
@article{devlin2019bert,
title={BERT: Pre-training of Deep Bidirectional Transformers for Language Understanding},
author={Devlin, Jacob and Chang, Ming-Wei and Lee, Kenton and Toutanova, Kristina},
journal={arXiv preprint arXiv:1810.04805},
year={2018}
}
For questions and feedback, please contact me through my Hugging Face profile: radubutucelea23
Base model
google-bert/bert-base-uncased