Update README.md
Browse files
README.md
CHANGED
|
@@ -29,7 +29,7 @@ DistilBERT is a transformers model, smaller and faster than BERT, which was pret
|
|
| 29 |
|
| 30 |
## Intended uses & limitations
|
| 31 |
|
| 32 |
-
|
| 33 |
|
| 34 |
## Training and evaluation data
|
| 35 |
|
|
@@ -37,6 +37,7 @@ More information needed
|
|
| 37 |
|
| 38 |
## Training procedure
|
| 39 |
|
|
|
|
| 40 |
### Training hyperparameters
|
| 41 |
|
| 42 |
The following hyperparameters were used during training:
|
|
|
|
| 29 |
|
| 30 |
## Intended uses & limitations
|
| 31 |
|
| 32 |
+
You can use this model for sentence classification task only as it is fine-tuned on imdb dataset. If you want to try it for mask filling task, it won't give you good results.
|
| 33 |
|
| 34 |
## Training and evaluation data
|
| 35 |
|
|
|
|
| 37 |
|
| 38 |
## Training procedure
|
| 39 |
|
| 40 |
+
- Preprocessing function was created to tokenize the text and truncate the sequences longer than DistilBERT max seq length. Datasets [map](https://huggingface.co/docs/datasets/main/en/package_reference/main_classes#datasets.Dataset.map) function was used to apply the preprocessing func over the entire dataset.
|
| 41 |
### Training hyperparameters
|
| 42 |
|
| 43 |
The following hyperparameters were used during training:
|