Commit ·
2b3797c
1
Parent(s): 90f5a69
correct grammar
Browse files
README.md
CHANGED
|
@@ -22,7 +22,7 @@ The data was split into training and testing: model trained on 90% of the data,
|
|
| 22 |
|
| 23 |
DistilBERT has a maximum input length of 512, so with this in mind the following was performed:
|
| 24 |
|
| 25 |
-
1. I used the`distilbert-base-uncased` pretrained model to initialize an `AutoTokenizer`.
|
| 26 |
2. Setting a maximum length of 256, each entry in the training, testing and validation data was truncated if it exceeded the limit and padded if it didn't reach the limit.
|
| 27 |
|
| 28 |
### Training hyperparameters
|
|
|
|
| 22 |
|
| 23 |
DistilBERT has a maximum input length of 512, so with this in mind the following was performed:
|
| 24 |
|
| 25 |
+
1. I used the `distilbert-base-uncased` pretrained model to initialize an `AutoTokenizer`.
|
| 26 |
2. Setting a maximum length of 256, each entry in the training, testing and validation data was truncated if it exceeded the limit and padded if it didn't reach the limit.
|
| 27 |
|
| 28 |
### Training hyperparameters
|