File size: 2,518 Bytes
25f05ab 6903c2c 645be16 f27a7c1 8248e1c b0e413d a97869e 25f05ab 89f0f5f b3ced13 89f0f5f b3ced13 562daa4 6baa627 25f05ab 19c3759 25f05ab 93d7843 19c3759 25f05ab 19c3759 25f05ab 19c3759 25f05ab 19c3759 25f05ab 19c3759 25f05ab 19c3759 25f05ab 19c3759 25f05ab 19c3759 25f05ab 19c3759 25f05ab 19c3759 25f05ab 19c3759 25f05ab 01b238d |
1 2 3 4 5 6 7 8 9 10 11 12 13 14 15 16 17 18 19 20 21 22 23 24 25 26 27 28 29 30 31 32 33 34 35 36 37 38 39 40 41 42 43 44 45 46 47 48 49 50 51 52 53 54 55 56 57 58 59 60 61 62 63 64 65 66 67 68 69 70 71 72 73 74 75 76 77 78 79 80 81 82 83 84 85 86 87 88 89 90 91 |
---
library_name: transformers
datasets:
- stanfordnlp/imdb
metrics:
- accuracy
tags:
- PyTorch
model-index:
- name: distilbert-imdb
results:
- task:
name: Text Classification
type: text-classification
dataset:
name: imdb
type: imdb
args: plain_text
metrics:
- name: Accuracy
type: accuracy
value: 0.9316
pipeline_tag: text-classification
license: apache-2.0
language:
- en
---
# distilbert-imdb
This is a fine-tuned version of [distilbert-base-uncased](https://huggingface.co/distilbert-base-uncased) on imdb dataset.
## Performance
- Loss: 0.1958
- Accuracy: 0.932
## How to Get Started with the Model
Use the code below to get started with the model:
```python
from transformers import pipeline,DistilBertTokenizer
tokenizer = DistilBertTokenizer.from_pretrained("distilbert-base-uncased")
classifier = pipeline("sentiment-analysis", model="3oclock/distilbert-imdb", tokenizer=tokenizer)
result = classifier("I love this movie!")
print(result)
```
## Model Details
### Model Description
This is the model card for a fine-tuned 🤗 transformers model on the IMDb dataset.
- **Developed by:** Ge Li
- **Model type:** DistilBERT for Sequence Classification
- **Language(s) (NLP):** English
- **License:** [Specify License, e.g., Apache 2.0]
- **Finetuned from model:** `distilbert-base-uncased`
## Uses
### Direct Use
This model can be used directly for sentiment analysis on movie reviews. It is best suited for classifying English-language text that is similar in nature to movie reviews.
### Downstream Use [optional]
This model can be fine-tuned on other sentiment analysis tasks or adapted for tasks like text classification in domains similar to IMDb movie reviews.
### Out-of-Scope Use
The model may not perform well on non-English text or text that is significantly different in style and content from the IMDb dataset (e.g., technical documents, social media posts).
## Bias, Risks, and Limitations
### Bias
The IMDb dataset primarily consists of English-language movie reviews and may not generalize well to other languages or types of reviews.
### Risks
Misclassification in sentiment analysis can lead to incorrect conclusions in applications relying on this model.
### Limitations
The model was trained on a dataset of movie reviews, so it may not perform as well on other types of text data.
### Recommendations
Users (both direct and downstream) should be made aware of the risks, biases, and limitations of the model. |