HugeFighter commited on
Commit
7c2a887
·
verified ·
1 Parent(s): 57f5f53

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +2 -1
README.md CHANGED
@@ -29,7 +29,7 @@ DistilBERT is a transformers model, smaller and faster than BERT, which was pret
29
 
30
  ## Intended uses & limitations
31
 
32
- More information needed
33
 
34
  ## Training and evaluation data
35
 
@@ -37,6 +37,7 @@ More information needed
37
 
38
  ## Training procedure
39
 
 
40
  ### Training hyperparameters
41
 
42
  The following hyperparameters were used during training:
 
29
 
30
  ## Intended uses & limitations
31
 
32
+ You can use this model for sentence classification task only as it is fine-tuned on imdb dataset. If you want to try it for mask filling task, it won't give you good results.
33
 
34
  ## Training and evaluation data
35
 
 
37
 
38
  ## Training procedure
39
 
40
+ - Preprocessing function was created to tokenize the text and truncate the sequences longer than DistilBERT max seq length. Datasets [map](https://huggingface.co/docs/datasets/main/en/package_reference/main_classes#datasets.Dataset.map) function was used to apply the preprocessing func over the entire dataset.
41
  ### Training hyperparameters
42
 
43
  The following hyperparameters were used during training: