Commit
·
bf3d8cf
1
Parent(s):
85b23de
End of training
Browse files
README.md
CHANGED
|
@@ -16,13 +16,13 @@ should probably proofread and complete it, then remove this comment. -->
|
|
| 16 |
|
| 17 |
This model is a fine-tuned version of [SCUT-DLVCLab/lilt-roberta-en-base](https://huggingface.co/SCUT-DLVCLab/lilt-roberta-en-base) on the mydata dataset.
|
| 18 |
It achieves the following results on the evaluation set:
|
| 19 |
-
- Loss: 0.
|
| 20 |
-
- In: {'precision':
|
| 21 |
-
- Ear: {'precision':
|
| 22 |
-
- Overall Precision:
|
| 23 |
-
- Overall Recall:
|
| 24 |
-
- Overall F1:
|
| 25 |
-
- Overall Accuracy:
|
| 26 |
|
| 27 |
## Model description
|
| 28 |
|
|
@@ -47,63 +47,17 @@ The following hyperparameters were used during training:
|
|
| 47 |
- seed: 42
|
| 48 |
- optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
|
| 49 |
- lr_scheduler_type: linear
|
| 50 |
-
- training_steps:
|
| 51 |
- mixed_precision_training: Native AMP
|
| 52 |
|
| 53 |
### Training results
|
| 54 |
|
| 55 |
-
| Training Loss | Epoch
|
| 56 |
-
|
| 57 |
-
| 0.
|
| 58 |
-
| 0.
|
| 59 |
-
| 0.
|
| 60 |
-
| 0.0 | 100.0
|
| 61 |
-
| 0.0 | 125.0 | 250 | 0.0000 | {'precision': 1.0, 'recall': 1.0, 'f1': 1.0, 'number': 2} | {'precision': 1.0, 'recall': 1.0, 'f1': 1.0, 'number': 2} | 1.0 | 1.0 | 1.0 | 1.0 |
|
| 62 |
-
| 0.0 | 150.0 | 300 | 0.0000 | {'precision': 1.0, 'recall': 1.0, 'f1': 1.0, 'number': 2} | {'precision': 1.0, 'recall': 1.0, 'f1': 1.0, 'number': 2} | 1.0 | 1.0 | 1.0 | 1.0 |
|
| 63 |
-
| 0.0 | 175.0 | 350 | 0.0000 | {'precision': 1.0, 'recall': 1.0, 'f1': 1.0, 'number': 2} | {'precision': 1.0, 'recall': 1.0, 'f1': 1.0, 'number': 2} | 1.0 | 1.0 | 1.0 | 1.0 |
|
| 64 |
-
| 0.0 | 200.0 | 400 | 0.0000 | {'precision': 1.0, 'recall': 1.0, 'f1': 1.0, 'number': 2} | {'precision': 1.0, 'recall': 1.0, 'f1': 1.0, 'number': 2} | 1.0 | 1.0 | 1.0 | 1.0 |
|
| 65 |
-
| 0.0 | 225.0 | 450 | 0.0000 | {'precision': 1.0, 'recall': 1.0, 'f1': 1.0, 'number': 2} | {'precision': 1.0, 'recall': 1.0, 'f1': 1.0, 'number': 2} | 1.0 | 1.0 | 1.0 | 1.0 |
|
| 66 |
-
| 0.0 | 250.0 | 500 | 0.0000 | {'precision': 1.0, 'recall': 1.0, 'f1': 1.0, 'number': 2} | {'precision': 1.0, 'recall': 1.0, 'f1': 1.0, 'number': 2} | 1.0 | 1.0 | 1.0 | 1.0 |
|
| 67 |
-
| 0.0 | 275.0 | 550 | 0.0000 | {'precision': 1.0, 'recall': 1.0, 'f1': 1.0, 'number': 2} | {'precision': 1.0, 'recall': 1.0, 'f1': 1.0, 'number': 2} | 1.0 | 1.0 | 1.0 | 1.0 |
|
| 68 |
-
| 0.0 | 300.0 | 600 | 0.0000 | {'precision': 1.0, 'recall': 1.0, 'f1': 1.0, 'number': 2} | {'precision': 1.0, 'recall': 1.0, 'f1': 1.0, 'number': 2} | 1.0 | 1.0 | 1.0 | 1.0 |
|
| 69 |
-
| 0.0 | 325.0 | 650 | 0.0000 | {'precision': 1.0, 'recall': 1.0, 'f1': 1.0, 'number': 2} | {'precision': 1.0, 'recall': 1.0, 'f1': 1.0, 'number': 2} | 1.0 | 1.0 | 1.0 | 1.0 |
|
| 70 |
-
| 0.0 | 350.0 | 700 | 0.0000 | {'precision': 1.0, 'recall': 1.0, 'f1': 1.0, 'number': 2} | {'precision': 1.0, 'recall': 1.0, 'f1': 1.0, 'number': 2} | 1.0 | 1.0 | 1.0 | 1.0 |
|
| 71 |
-
| 0.0 | 375.0 | 750 | 0.0000 | {'precision': 1.0, 'recall': 1.0, 'f1': 1.0, 'number': 2} | {'precision': 1.0, 'recall': 1.0, 'f1': 1.0, 'number': 2} | 1.0 | 1.0 | 1.0 | 1.0 |
|
| 72 |
-
| 0.0 | 400.0 | 800 | 0.0000 | {'precision': 1.0, 'recall': 1.0, 'f1': 1.0, 'number': 2} | {'precision': 1.0, 'recall': 1.0, 'f1': 1.0, 'number': 2} | 1.0 | 1.0 | 1.0 | 1.0 |
|
| 73 |
-
| 0.0 | 425.0 | 850 | 0.0000 | {'precision': 1.0, 'recall': 1.0, 'f1': 1.0, 'number': 2} | {'precision': 1.0, 'recall': 1.0, 'f1': 1.0, 'number': 2} | 1.0 | 1.0 | 1.0 | 1.0 |
|
| 74 |
-
| 0.0 | 450.0 | 900 | 0.0000 | {'precision': 1.0, 'recall': 1.0, 'f1': 1.0, 'number': 2} | {'precision': 1.0, 'recall': 1.0, 'f1': 1.0, 'number': 2} | 1.0 | 1.0 | 1.0 | 1.0 |
|
| 75 |
-
| 0.0 | 475.0 | 950 | 0.0000 | {'precision': 1.0, 'recall': 1.0, 'f1': 1.0, 'number': 2} | {'precision': 1.0, 'recall': 1.0, 'f1': 1.0, 'number': 2} | 1.0 | 1.0 | 1.0 | 1.0 |
|
| 76 |
-
| 0.0 | 500.0 | 1000 | 0.0000 | {'precision': 1.0, 'recall': 1.0, 'f1': 1.0, 'number': 2} | {'precision': 1.0, 'recall': 1.0, 'f1': 1.0, 'number': 2} | 1.0 | 1.0 | 1.0 | 1.0 |
|
| 77 |
-
| 0.0 | 525.0 | 1050 | 0.0000 | {'precision': 1.0, 'recall': 1.0, 'f1': 1.0, 'number': 2} | {'precision': 1.0, 'recall': 1.0, 'f1': 1.0, 'number': 2} | 1.0 | 1.0 | 1.0 | 1.0 |
|
| 78 |
-
| 0.0 | 550.0 | 1100 | 0.0000 | {'precision': 1.0, 'recall': 1.0, 'f1': 1.0, 'number': 2} | {'precision': 1.0, 'recall': 1.0, 'f1': 1.0, 'number': 2} | 1.0 | 1.0 | 1.0 | 1.0 |
|
| 79 |
-
| 0.0 | 575.0 | 1150 | 0.0000 | {'precision': 1.0, 'recall': 1.0, 'f1': 1.0, 'number': 2} | {'precision': 1.0, 'recall': 1.0, 'f1': 1.0, 'number': 2} | 1.0 | 1.0 | 1.0 | 1.0 |
|
| 80 |
-
| 0.0 | 600.0 | 1200 | 0.0000 | {'precision': 1.0, 'recall': 1.0, 'f1': 1.0, 'number': 2} | {'precision': 1.0, 'recall': 1.0, 'f1': 1.0, 'number': 2} | 1.0 | 1.0 | 1.0 | 1.0 |
|
| 81 |
-
| 0.0 | 625.0 | 1250 | 0.0000 | {'precision': 1.0, 'recall': 1.0, 'f1': 1.0, 'number': 2} | {'precision': 1.0, 'recall': 1.0, 'f1': 1.0, 'number': 2} | 1.0 | 1.0 | 1.0 | 1.0 |
|
| 82 |
-
| 0.0 | 650.0 | 1300 | 0.0000 | {'precision': 1.0, 'recall': 1.0, 'f1': 1.0, 'number': 2} | {'precision': 1.0, 'recall': 1.0, 'f1': 1.0, 'number': 2} | 1.0 | 1.0 | 1.0 | 1.0 |
|
| 83 |
-
| 0.0 | 675.0 | 1350 | 0.0000 | {'precision': 1.0, 'recall': 1.0, 'f1': 1.0, 'number': 2} | {'precision': 1.0, 'recall': 1.0, 'f1': 1.0, 'number': 2} | 1.0 | 1.0 | 1.0 | 1.0 |
|
| 84 |
-
| 0.0 | 700.0 | 1400 | 0.0000 | {'precision': 1.0, 'recall': 1.0, 'f1': 1.0, 'number': 2} | {'precision': 1.0, 'recall': 1.0, 'f1': 1.0, 'number': 2} | 1.0 | 1.0 | 1.0 | 1.0 |
|
| 85 |
-
| 0.0 | 725.0 | 1450 | 0.0000 | {'precision': 1.0, 'recall': 1.0, 'f1': 1.0, 'number': 2} | {'precision': 1.0, 'recall': 1.0, 'f1': 1.0, 'number': 2} | 1.0 | 1.0 | 1.0 | 1.0 |
|
| 86 |
-
| 0.0 | 750.0 | 1500 | 0.0000 | {'precision': 1.0, 'recall': 1.0, 'f1': 1.0, 'number': 2} | {'precision': 1.0, 'recall': 1.0, 'f1': 1.0, 'number': 2} | 1.0 | 1.0 | 1.0 | 1.0 |
|
| 87 |
-
| 0.0 | 775.0 | 1550 | 0.0000 | {'precision': 1.0, 'recall': 1.0, 'f1': 1.0, 'number': 2} | {'precision': 1.0, 'recall': 1.0, 'f1': 1.0, 'number': 2} | 1.0 | 1.0 | 1.0 | 1.0 |
|
| 88 |
-
| 0.0 | 800.0 | 1600 | 0.0000 | {'precision': 1.0, 'recall': 1.0, 'f1': 1.0, 'number': 2} | {'precision': 1.0, 'recall': 1.0, 'f1': 1.0, 'number': 2} | 1.0 | 1.0 | 1.0 | 1.0 |
|
| 89 |
-
| 0.0 | 825.0 | 1650 | 0.0000 | {'precision': 1.0, 'recall': 1.0, 'f1': 1.0, 'number': 2} | {'precision': 1.0, 'recall': 1.0, 'f1': 1.0, 'number': 2} | 1.0 | 1.0 | 1.0 | 1.0 |
|
| 90 |
-
| 0.0 | 850.0 | 1700 | 0.0000 | {'precision': 1.0, 'recall': 1.0, 'f1': 1.0, 'number': 2} | {'precision': 1.0, 'recall': 1.0, 'f1': 1.0, 'number': 2} | 1.0 | 1.0 | 1.0 | 1.0 |
|
| 91 |
-
| 0.0 | 875.0 | 1750 | 0.0000 | {'precision': 1.0, 'recall': 1.0, 'f1': 1.0, 'number': 2} | {'precision': 1.0, 'recall': 1.0, 'f1': 1.0, 'number': 2} | 1.0 | 1.0 | 1.0 | 1.0 |
|
| 92 |
-
| 0.0 | 900.0 | 1800 | 0.0000 | {'precision': 1.0, 'recall': 1.0, 'f1': 1.0, 'number': 2} | {'precision': 1.0, 'recall': 1.0, 'f1': 1.0, 'number': 2} | 1.0 | 1.0 | 1.0 | 1.0 |
|
| 93 |
-
| 0.0 | 925.0 | 1850 | 0.0000 | {'precision': 1.0, 'recall': 1.0, 'f1': 1.0, 'number': 2} | {'precision': 1.0, 'recall': 1.0, 'f1': 1.0, 'number': 2} | 1.0 | 1.0 | 1.0 | 1.0 |
|
| 94 |
-
| 0.0 | 950.0 | 1900 | 0.0000 | {'precision': 1.0, 'recall': 1.0, 'f1': 1.0, 'number': 2} | {'precision': 1.0, 'recall': 1.0, 'f1': 1.0, 'number': 2} | 1.0 | 1.0 | 1.0 | 1.0 |
|
| 95 |
-
| 0.0 | 975.0 | 1950 | 0.0000 | {'precision': 1.0, 'recall': 1.0, 'f1': 1.0, 'number': 2} | {'precision': 1.0, 'recall': 1.0, 'f1': 1.0, 'number': 2} | 1.0 | 1.0 | 1.0 | 1.0 |
|
| 96 |
-
| 0.0 | 1000.0 | 2000 | 0.0000 | {'precision': 1.0, 'recall': 1.0, 'f1': 1.0, 'number': 2} | {'precision': 1.0, 'recall': 1.0, 'f1': 1.0, 'number': 2} | 1.0 | 1.0 | 1.0 | 1.0 |
|
| 97 |
-
| 0.0 | 1025.0 | 2050 | 0.0000 | {'precision': 1.0, 'recall': 1.0, 'f1': 1.0, 'number': 2} | {'precision': 1.0, 'recall': 1.0, 'f1': 1.0, 'number': 2} | 1.0 | 1.0 | 1.0 | 1.0 |
|
| 98 |
-
| 0.0 | 1050.0 | 2100 | 0.0000 | {'precision': 1.0, 'recall': 1.0, 'f1': 1.0, 'number': 2} | {'precision': 1.0, 'recall': 1.0, 'f1': 1.0, 'number': 2} | 1.0 | 1.0 | 1.0 | 1.0 |
|
| 99 |
-
| 0.0 | 1075.0 | 2150 | 0.0000 | {'precision': 1.0, 'recall': 1.0, 'f1': 1.0, 'number': 2} | {'precision': 1.0, 'recall': 1.0, 'f1': 1.0, 'number': 2} | 1.0 | 1.0 | 1.0 | 1.0 |
|
| 100 |
-
| 0.0 | 1100.0 | 2200 | 0.0000 | {'precision': 1.0, 'recall': 1.0, 'f1': 1.0, 'number': 2} | {'precision': 1.0, 'recall': 1.0, 'f1': 1.0, 'number': 2} | 1.0 | 1.0 | 1.0 | 1.0 |
|
| 101 |
-
| 0.0 | 1125.0 | 2250 | 0.0000 | {'precision': 1.0, 'recall': 1.0, 'f1': 1.0, 'number': 2} | {'precision': 1.0, 'recall': 1.0, 'f1': 1.0, 'number': 2} | 1.0 | 1.0 | 1.0 | 1.0 |
|
| 102 |
-
| 0.0 | 1150.0 | 2300 | 0.0000 | {'precision': 1.0, 'recall': 1.0, 'f1': 1.0, 'number': 2} | {'precision': 1.0, 'recall': 1.0, 'f1': 1.0, 'number': 2} | 1.0 | 1.0 | 1.0 | 1.0 |
|
| 103 |
-
| 0.0 | 1175.0 | 2350 | 0.0000 | {'precision': 1.0, 'recall': 1.0, 'f1': 1.0, 'number': 2} | {'precision': 1.0, 'recall': 1.0, 'f1': 1.0, 'number': 2} | 1.0 | 1.0 | 1.0 | 1.0 |
|
| 104 |
-
| 0.0 | 1200.0 | 2400 | 0.0000 | {'precision': 1.0, 'recall': 1.0, 'f1': 1.0, 'number': 2} | {'precision': 1.0, 'recall': 1.0, 'f1': 1.0, 'number': 2} | 1.0 | 1.0 | 1.0 | 1.0 |
|
| 105 |
-
| 0.0 | 1225.0 | 2450 | 0.0000 | {'precision': 1.0, 'recall': 1.0, 'f1': 1.0, 'number': 2} | {'precision': 1.0, 'recall': 1.0, 'f1': 1.0, 'number': 2} | 1.0 | 1.0 | 1.0 | 1.0 |
|
| 106 |
-
| 0.0 | 1250.0 | 2500 | 0.0000 | {'precision': 1.0, 'recall': 1.0, 'f1': 1.0, 'number': 2} | {'precision': 1.0, 'recall': 1.0, 'f1': 1.0, 'number': 2} | 1.0 | 1.0 | 1.0 | 1.0 |
|
| 107 |
|
| 108 |
|
| 109 |
### Framework versions
|
|
|
|
| 16 |
|
| 17 |
This model is a fine-tuned version of [SCUT-DLVCLab/lilt-roberta-en-base](https://huggingface.co/SCUT-DLVCLab/lilt-roberta-en-base) on the mydata dataset.
|
| 18 |
It achieves the following results on the evaluation set:
|
| 19 |
+
- Loss: 0.0023
|
| 20 |
+
- In: {'precision': 0.5, 'recall': 0.5, 'f1': 0.5, 'number': 2}
|
| 21 |
+
- Ear: {'precision': 0.6666666666666666, 'recall': 1.0, 'f1': 0.8, 'number': 2}
|
| 22 |
+
- Overall Precision: 0.6
|
| 23 |
+
- Overall Recall: 0.75
|
| 24 |
+
- Overall F1: 0.6667
|
| 25 |
+
- Overall Accuracy: 0.9984
|
| 26 |
|
| 27 |
## Model description
|
| 28 |
|
|
|
|
| 47 |
- seed: 42
|
| 48 |
- optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
|
| 49 |
- lr_scheduler_type: linear
|
| 50 |
+
- training_steps: 200
|
| 51 |
- mixed_precision_training: Native AMP
|
| 52 |
|
| 53 |
### Training results
|
| 54 |
|
| 55 |
+
| Training Loss | Epoch | Step | Validation Loss | In | Ear | Overall Precision | Overall Recall | Overall F1 | Overall Accuracy |
|
| 56 |
+
|:-------------:|:-----:|:----:|:---------------:|:---------------------------------------------------------:|:------------------------------------------------------------------------:|:-----------------:|:--------------:|:----------:|:----------------:|
|
| 57 |
+
| 0.1628 | 25.0 | 50 | 0.0023 | {'precision': 0.5, 'recall': 0.5, 'f1': 0.5, 'number': 2} | {'precision': 0.6666666666666666, 'recall': 1.0, 'f1': 0.8, 'number': 2} | 0.6 | 0.75 | 0.6667 | 0.9984 |
|
| 58 |
+
| 0.0002 | 50.0 | 100 | 0.0015 | {'precision': 0.5, 'recall': 0.5, 'f1': 0.5, 'number': 2} | {'precision': 0.6666666666666666, 'recall': 1.0, 'f1': 0.8, 'number': 2} | 0.6 | 0.75 | 0.6667 | 0.9984 |
|
| 59 |
+
| 0.0001 | 75.0 | 150 | 0.0020 | {'precision': 0.5, 'recall': 0.5, 'f1': 0.5, 'number': 2} | {'precision': 0.6666666666666666, 'recall': 1.0, 'f1': 0.8, 'number': 2} | 0.6 | 0.75 | 0.6667 | 0.9984 |
|
| 60 |
+
| 0.0 | 100.0 | 200 | 0.0020 | {'precision': 0.5, 'recall': 0.5, 'f1': 0.5, 'number': 2} | {'precision': 0.6666666666666666, 'recall': 1.0, 'f1': 0.8, 'number': 2} | 0.6 | 0.75 | 0.6667 | 0.9984 |
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 61 |
|
| 62 |
|
| 63 |
### Framework versions
|
logs/events.out.tfevents.1685460269.MSI.3990.0
CHANGED
|
@@ -1,3 +1,3 @@
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
-
oid sha256:
|
| 3 |
-
size
|
|
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:4d8470c3ef4bd18fae7cb85f6f89d5be4e5d07bd81cd45e6e0868198be796edd
|
| 3 |
+
size 7229
|
logs/events.out.tfevents.1685460384.MSI.3990.2
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:7db659e86ef3cce24ebcf7f5af1de065a79f9efa8701a37e9b1c1743dfdb7827
|
| 3 |
+
size 592
|