AnkitAI's picture
Update README.md
9656e51 verified
---
license: mit
base_model: cardiffnlp/twitter-roberta-base-sentiment-latest
language:
- en
library_name: transformers
tags:
- Roberta
- Sentiment Analysis
widget:
- text: This product is really great!
- text: This product is really bad!
---
# Fine-tuned RoBERTa for Sentiment Analysis on Reviews
This model is a fine-tuned version of [cardiffnlp/twitter-roberta-base-sentiment-latest](https://huggingface.co/cardiffnlp/twitter-roberta-base-sentiment-latest) on the [Amazon Reviews dataset](https://www.kaggle.com/datasets/bittlingmayer/amazonreviews) for sentiment analysis.
## Model Details
- **Model Name:** `AnkitAI/reviews-roberta-base-sentiment-analysis`
- **Base Model:** `cardiffnlp/twitter-roberta-base-sentiment-latest`
- **Dataset:** [Amazon Reviews](https://www.kaggle.com/datasets/bittlingmayer/amazonreviews)
- **Fine-tuning:** This model was fine-tuned for sentiment analysis with a classification head for binary sentiment classification (positive and negative).
## Training
The model was trained using the following parameters:
- **Learning Rate:** 2e-5
- **Batch Size:** 16
- **Weight Decay:** 0.01
- **Evaluation Strategy:** Epoch
### Training Details
- **Evaluation Loss:** 0.1049
- **Evaluation Runtime:** 3177.538 seconds
- **Evaluation Samples/Second:** 226.591
- **Evaluation Steps/Second:** 7.081
- **Training Runtime:** 110070.6349 seconds
- **Training Samples/Second:** 78.495
- **Training Steps/Second:** 2.453
- **Training Loss:** 0.0858
- **Evaluation Accuracy:** 97.19%
- **Evaluation Precision:** 97.9%
- **Evaluation Recall:** 97.18%
- **Evaluation F1 Score:** 97.19%
## Usage
You can use this model directly with the Hugging Face `transformers` library:
```python
from transformers import RobertaForSequenceClassification, RobertaTokenizer
model_name = "AnkitAI/reviews-roberta-base-sentiment-analysis"
model = RobertaForSequenceClassification.from_pretrained(model_name)
tokenizer = RobertaTokenizer.from_pretrained(model_name)
# Example usage
inputs = tokenizer("This product is great!", return_tensors="pt")
outputs = model(**inputs) # 1 for positive, 0 for negative
```
## License
This model is licensed under the [MIT License](LICENSE).