Commit
·
9ef1847
1
Parent(s):
2ca67a2
Update README.md
Browse files
README.md
CHANGED
|
@@ -86,12 +86,13 @@ Test|2414|5577|179|2445|1097|183|2838|272|374|356
|
|
| 86 |
This model was trained using a NVIDIA RTX A6000 GPU with the following hyperparameters:
|
| 87 |
|
| 88 |
- learning rate: 2e-05
|
| 89 |
-
- train batch size:
|
| 90 |
- epochs: 10
|
| 91 |
- optimizer: AdamW with betas=(0.9,0.999) and epsilon=1e-08
|
| 92 |
- scheduler: linear scheduler with num_warmup_steps=round(len(train_dataloader)/5) and num_training_steps=len(train_dataloader)*epochs
|
| 93 |
- maximum length of data sequence: 512
|
| 94 |
- patience: 2 epochs
|
|
|
|
| 95 |
|
| 96 |
In the preprocessing stage, the input texts were split into chunks with a maximum length of 300 tokens,
|
| 97 |
in order to avoid the tokenized chunks exceeding the maximum length of 512. Tokenization was performed
|
|
@@ -106,15 +107,15 @@ Evaluation results using the test dataset are listed below:
|
|
| 106 |
|
| 107 |
||Precision|Recall|F1-score
|
| 108 |
-|-|-|-
|
| 109 |
-
PERSON|0.
|
| 110 |
-
ORG|0.
|
| 111 |
-
LOC|0.
|
| 112 |
-
GPE|0.
|
| 113 |
-
PRODUCT|0.
|
| 114 |
-
EVENT|0.
|
| 115 |
-
DATE|0.
|
| 116 |
-
JON|0.
|
| 117 |
-
FIBC|0.
|
| 118 |
NORP|0.91|0.95|0.93
|
| 119 |
|
| 120 |
The metrics were calculated using the [seqeval](https://github.com/chakki-works/seqeval) library.
|
|
|
|
| 86 |
This model was trained using a NVIDIA RTX A6000 GPU with the following hyperparameters:
|
| 87 |
|
| 88 |
- learning rate: 2e-05
|
| 89 |
+
- train batch size: 24
|
| 90 |
- epochs: 10
|
| 91 |
- optimizer: AdamW with betas=(0.9,0.999) and epsilon=1e-08
|
| 92 |
- scheduler: linear scheduler with num_warmup_steps=round(len(train_dataloader)/5) and num_training_steps=len(train_dataloader)*epochs
|
| 93 |
- maximum length of data sequence: 512
|
| 94 |
- patience: 2 epochs
|
| 95 |
+
- classifier dropout: 0.3
|
| 96 |
|
| 97 |
In the preprocessing stage, the input texts were split into chunks with a maximum length of 300 tokens,
|
| 98 |
in order to avoid the tokenized chunks exceeding the maximum length of 512. Tokenization was performed
|
|
|
|
| 107 |
|
| 108 |
||Precision|Recall|F1-score
|
| 109 |
-|-|-|-
|
| 110 |
+
PERSON|0.90|0.91|0.90
|
| 111 |
+
ORG|0.84|0.87|0.86
|
| 112 |
+
LOC|0.84|0.86|0.85
|
| 113 |
+
GPE|0.91|0.91|0.91
|
| 114 |
+
PRODUCT|0.73|0.77|0.75
|
| 115 |
+
EVENT|0.69|0.73|0.71
|
| 116 |
+
DATE|0.90|0.92|0.91
|
| 117 |
+
JON|0.83|0.95|0.89
|
| 118 |
+
FIBC|0.95|0.99|0.97
|
| 119 |
NORP|0.91|0.95|0.93
|
| 120 |
|
| 121 |
The metrics were calculated using the [seqeval](https://github.com/chakki-works/seqeval) library.
|