Beehzod's picture
Update README.md
c533c63 verified
---
license: apache-2.0
language: en
tags:
- sentiment-analysis
- distilbert
- transformers
datasets:
- imdb
metrics:
- accuracy
- f1
- precision
- recall
model_type: distilbert
---
# Fine-tuned DistilBERT for Sentiment Analysis
## Model Description
This model is a fine-tuned version of DistilBERT for sentiment analysis tasks. It was trained on the IMDB dataset to classify movie reviews as **positive** or **negative**. It can be used in applications where text sentiment analysis is needed, such as social media monitoring or customer feedback analysis.
- **Model Architecture**: DistilBERT (transformer-based model)
- **Task**: Sentiment Analysis
- **Labels**:
- **Positive**
- **Negative**
## Training Details
- **Dataset**: IMDB movie reviews dataset
- **Training Data Size**: 20,000 samples for training and 5,000 samples for evaluation
- **Epochs**: 3
- **Batch Size**: 16
- **Learning Rate**: 2e-5
- **Optimizer**: AdamW with weight decay
## Evaluation Metrics
The model was evaluated on a held-out test set using the following metrics:
- **Accuracy**: 0.95
- **F1 Score**: 0.94
- **Precision**: 0.93
- **Recall**: 0.92
## Usage
### Example Code
To use this sentiment analysis model with the Hugging Face Transformers library:
```python
from transformers import pipeline
# Load the model from the Hugging Face Hub
sentiment_pipeline = pipeline("sentiment-analysis", model="Beehzod/smart_sentiment_analysis")
# Example predictions
text = "This movie was fantastic! I really enjoyed it."
results = sentiment_pipeline(text)
for result in results:
print(f"Label: {result['label']}, Score: {result['score']:.4f}")