File size: 2,518 Bytes
25f05ab
 
6903c2c
 
 
 
645be16
f27a7c1
8248e1c
b0e413d
 
 
 
 
 
 
 
 
 
 
 
 
 
a97869e
 
 
25f05ab
89f0f5f
 
 
b3ced13
 
89f0f5f
 
b3ced13
562daa4
 
 
 
 
 
 
 
 
 
 
6baa627
25f05ab
 
 
 
19c3759
25f05ab
93d7843
19c3759
 
 
 
25f05ab
 
 
 
 
 
19c3759
25f05ab
 
 
19c3759
25f05ab
 
 
19c3759
25f05ab
 
 
19c3759
25f05ab
19c3759
25f05ab
19c3759
25f05ab
19c3759
25f05ab
19c3759
25f05ab
19c3759
25f05ab
19c3759
25f05ab
01b238d
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
38
39
40
41
42
43
44
45
46
47
48
49
50
51
52
53
54
55
56
57
58
59
60
61
62
63
64
65
66
67
68
69
70
71
72
73
74
75
76
77
78
79
80
81
82
83
84
85
86
87
88
89
90
91
---
library_name: transformers
datasets:
- stanfordnlp/imdb
metrics:
- accuracy
tags:
- PyTorch
model-index:
- name: distilbert-imdb
  results:
  - task:
      name: Text Classification
      type: text-classification
    dataset:
      name: imdb
      type: imdb
      args: plain_text
    metrics:
    - name: Accuracy
      type: accuracy
      value: 0.9316
pipeline_tag: text-classification
license: apache-2.0
language:
- en
---
# distilbert-imdb

This is a fine-tuned version of [distilbert-base-uncased](https://huggingface.co/distilbert-base-uncased) on imdb dataset.

## Performance
- Loss: 0.1958
- Accuracy: 0.932

## How to Get Started with the Model

Use the code below to get started with the model:

```python
from transformers import pipeline,DistilBertTokenizer

tokenizer = DistilBertTokenizer.from_pretrained("distilbert-base-uncased")
classifier = pipeline("sentiment-analysis", model="3oclock/distilbert-imdb", tokenizer=tokenizer)
result = classifier("I love this movie!")
print(result)
```
## Model Details

### Model Description

This is the model card for a fine-tuned 🤗 transformers model on the IMDb dataset.

- **Developed by:** Ge Li
- **Model type:** DistilBERT for Sequence Classification
- **Language(s) (NLP):** English
- **License:** [Specify License, e.g., Apache 2.0]
- **Finetuned from model:** `distilbert-base-uncased`


## Uses

### Direct Use

This model can be used directly for sentiment analysis on movie reviews. It is best suited for classifying English-language text that is similar in nature to movie reviews.

### Downstream Use [optional]

This model can be fine-tuned on other sentiment analysis tasks or adapted for tasks like text classification in domains similar to IMDb movie reviews.

### Out-of-Scope Use

The model may not perform well on non-English text or text that is significantly different in style and content from the IMDb dataset (e.g., technical documents, social media posts).

## Bias, Risks, and Limitations

### Bias

The IMDb dataset primarily consists of English-language movie reviews and may not generalize well to other languages or types of reviews.

### Risks

Misclassification in sentiment analysis can lead to incorrect conclusions in applications relying on this model.

### Limitations

The model was trained on a dataset of movie reviews, so it may not perform as well on other types of text data.

### Recommendations

Users (both direct and downstream) should be made aware of the risks, biases, and limitations of the model.