File size: 1,655 Bytes
7b13ddd
c533c63
7b13ddd
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
c533c63
 
7b13ddd
c533c63
 
 
 
 
7b13ddd
c533c63
 
 
 
 
 
 
 
 
 
 
 
 
 
7b13ddd
 
c533c63
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
38
39
40
41
42
43
44
45
46
47
48
49
50
51
52
53
54
55
56
57
58
59
60
61
---
license: apache-2.0
language: en
tags:
  - sentiment-analysis
  - distilbert
  - transformers
datasets:
  - imdb
metrics:
  - accuracy
  - f1
  - precision
  - recall
model_type: distilbert
---

# Fine-tuned DistilBERT for Sentiment Analysis

## Model Description
This model is a fine-tuned version of DistilBERT for sentiment analysis tasks. It was trained on the IMDB dataset to classify movie reviews as **positive** or **negative**. It can be used in applications where text sentiment analysis is needed, such as social media monitoring or customer feedback analysis.

- **Model Architecture**: DistilBERT (transformer-based model)
- **Task**: Sentiment Analysis
- **Labels**: 
  - **Positive**
  - **Negative**

## Training Details
- **Dataset**: IMDB movie reviews dataset
- **Training Data Size**: 20,000 samples for training and 5,000 samples for evaluation
- **Epochs**: 3
- **Batch Size**: 16
- **Learning Rate**: 2e-5
- **Optimizer**: AdamW with weight decay

## Evaluation Metrics
The model was evaluated on a held-out test set using the following metrics:
- **Accuracy**: 0.95
- **F1 Score**: 0.94
- **Precision**: 0.93
- **Recall**: 0.92

## Usage

### Example Code
To use this sentiment analysis model with the Hugging Face Transformers library:

```python
from transformers import pipeline

# Load the model from the Hugging Face Hub
sentiment_pipeline = pipeline("sentiment-analysis", model="Beehzod/smart_sentiment_analysis")

# Example predictions
text = "This movie was fantastic! I really enjoyed it."
results = sentiment_pipeline(text)

for result in results:
    print(f"Label: {result['label']}, Score: {result['score']:.4f}")