Update README.md
Browse files
README.md
CHANGED
|
@@ -13,18 +13,7 @@ pipeline_tag: text-classification
|
|
| 13 |
metrics:
|
| 14 |
- accuracy
|
| 15 |
---
|
| 16 |
-
|
| 17 |
-
- stanfordnlp/imdb
|
| 18 |
-
language:
|
| 19 |
-
- en
|
| 20 |
-
base_model:
|
| 21 |
-
- distilbert/distilbert-base-uncased
|
| 22 |
-
tags:
|
| 23 |
-
- torch
|
| 24 |
-
- code
|
| 25 |
-
library_name: transformers
|
| 26 |
-
pipeline_tag: text-classification
|
| 27 |
-
---
|
| 28 |
# Model Card for DistilBERT Fine-Tuned on IMDB Sentiment Analysis
|
| 29 |
|
| 30 |
This model is a fine-tuned version of `DistilBERT` for sentiment analysis on the IMDB movie reviews dataset. It classifies movie reviews into two categories: positive and negative sentiments.
|
|
@@ -35,8 +24,8 @@ This model is a fine-tuned version of `DistilBERT` for sentiment analysis on the
|
|
| 35 |
|
| 36 |
This model has been fine-tuned on the IMDB dataset, which contains movie reviews labeled with sentiments: `positive` or `negative`. The model is based on the `DistilBERT` architecture, which is a lighter, more efficient variant of BERT, offering faster inference without significantly sacrificing accuracy.
|
| 37 |
|
| 38 |
-
- **Developed by:**
|
| 39 |
-
- **Shared by [optional]:**
|
| 40 |
- **Model type:** Transformer-based model for text classification (sentiment analysis)
|
| 41 |
- **Language(s) (NLP):** English
|
| 42 |
- **Finetuned from model [optional]:** distilbert-base-uncased
|
|
|
|
| 13 |
metrics:
|
| 14 |
- accuracy
|
| 15 |
---
|
| 16 |
+
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 17 |
# Model Card for DistilBERT Fine-Tuned on IMDB Sentiment Analysis
|
| 18 |
|
| 19 |
This model is a fine-tuned version of `DistilBERT` for sentiment analysis on the IMDB movie reviews dataset. It classifies movie reviews into two categories: positive and negative sentiments.
|
|
|
|
| 24 |
|
| 25 |
This model has been fine-tuned on the IMDB dataset, which contains movie reviews labeled with sentiments: `positive` or `negative`. The model is based on the `DistilBERT` architecture, which is a lighter, more efficient variant of BERT, offering faster inference without significantly sacrificing accuracy.
|
| 26 |
|
| 27 |
+
- **Developed by:** Leonuraht/Scilineo
|
| 28 |
+
- **Shared by [optional]:** Leonuraht/Scilineo
|
| 29 |
- **Model type:** Transformer-based model for text classification (sentiment analysis)
|
| 30 |
- **Language(s) (NLP):** English
|
| 31 |
- **Finetuned from model [optional]:** distilbert-base-uncased
|