Update README.md
Browse files
README.md
CHANGED
|
@@ -14,14 +14,17 @@ widget:
|
|
| 14 |
|
| 15 |
---
|
| 16 |
Text Multi-Label Sequence Classification model used to decode if passages contain a misfortunate event, a cause for misfortune, and/or an action to mollify or prevent some misfortune.
|
| 17 |
-
8293 passages were used for Training and split into 5 folds (~6634 for the train set, ~1659 for the validation set over 5 folds).
|
|
|
|
|
|
|
| 18 |
<br>Transformer: distilbert-base-uncased
|
| 19 |
<br>Tokenizer: distilbert-base-uncased
|
| 20 |
-
<br>learning rate: 2e-
|
| 21 |
<br>weight decay: .01
|
| 22 |
<br>Dropout: .1
|
| 23 |
<br>Batch Size: 8
|
| 24 |
<br>Epochs: 15
|
|
|
|
| 25 |
<br><br>Using epoch 13, the current F1 micro score of 2074 passages not used for training is .637. individual class f1 scores are shown below. Note that at this moment, some labels have been excluded as they are not relevant for the final use of the model.
|
| 26 |
<ul>
|
| 27 |
<li>EVENT: -
|
|
|
|
| 14 |
|
| 15 |
---
|
| 16 |
Text Multi-Label Sequence Classification model used to decode if passages contain a misfortunate event, a cause for misfortune, and/or an action to mollify or prevent some misfortune.
|
| 17 |
+
8293 passages were used for Training and split into 5 folds (~6634 for the train set, ~1659 for the validation set over 5 folds).
|
| 18 |
+
|
| 19 |
+
<br><b>Parameters</b>:
|
| 20 |
<br>Transformer: distilbert-base-uncased
|
| 21 |
<br>Tokenizer: distilbert-base-uncased
|
| 22 |
+
<br>learning rate: 2e-05
|
| 23 |
<br>weight decay: .01
|
| 24 |
<br>Dropout: .1
|
| 25 |
<br>Batch Size: 8
|
| 26 |
<br>Epochs: 15
|
| 27 |
+
<br>Metric for best model: F1 micro
|
| 28 |
<br><br>Using epoch 13, the current F1 micro score of 2074 passages not used for training is .637. individual class f1 scores are shown below. Note that at this moment, some labels have been excluded as they are not relevant for the final use of the model.
|
| 29 |
<ul>
|
| 30 |
<li>EVENT: -
|