End of training
Browse files- README.md +27 -27
- logs/events.out.tfevents.1741099192.DESKTOP-HA84SVN.2309656.3 +2 -2
- model.safetensors +1 -1
- tokenizer_config.json +0 -7
README.md
CHANGED
|
@@ -1,7 +1,7 @@
|
|
| 1 |
---
|
| 2 |
library_name: transformers
|
| 3 |
license: mit
|
| 4 |
-
base_model:
|
| 5 |
tags:
|
| 6 |
- generated_from_trainer
|
| 7 |
datasets:
|
|
@@ -16,16 +16,16 @@ should probably proofread and complete it, then remove this comment. -->
|
|
| 16 |
|
| 17 |
# layoutlm-with-funsd
|
| 18 |
|
| 19 |
-
This model is a fine-tuned version of [
|
| 20 |
It achieves the following results on the evaluation set:
|
| 21 |
-
- Loss: 0.
|
| 22 |
-
- Eader: {'precision': 0.
|
| 23 |
-
- Nswer: {'precision': 0.
|
| 24 |
-
- Uestion: {'precision': 0.
|
| 25 |
-
- Overall Precision: 0.
|
| 26 |
-
- Overall Recall: 0.
|
| 27 |
-
- Overall F1: 0.
|
| 28 |
-
- Overall Accuracy: 0.
|
| 29 |
|
| 30 |
## Model description
|
| 31 |
|
|
@@ -55,23 +55,23 @@ The following hyperparameters were used during training:
|
|
| 55 |
|
| 56 |
### Training results
|
| 57 |
|
| 58 |
-
| Training Loss | Epoch | Step | Validation Loss | Eader
|
| 59 |
-
|:-------------:|:-----:|:----:|:---------------:|:--------------------------------------------------------------------------------------------:|:----------------------------------------------------------------------------------------------------------:|:----------------------------------------------------------------------------------------------------------:|:-----------------:|:--------------:|:----------:|:----------------:|
|
| 60 |
-
| 1.
|
| 61 |
-
|
|
| 62 |
-
| 0.
|
| 63 |
-
| 0.
|
| 64 |
-
| 0.
|
| 65 |
-
| 0.
|
| 66 |
-
| 0.
|
| 67 |
-
| 0.
|
| 68 |
-
| 0.
|
| 69 |
-
| 0.
|
| 70 |
-
| 0.
|
| 71 |
-
| 0.
|
| 72 |
-
| 0.
|
| 73 |
-
| 0.
|
| 74 |
-
| 0.
|
| 75 |
|
| 76 |
|
| 77 |
### Framework versions
|
|
|
|
| 1 |
---
|
| 2 |
library_name: transformers
|
| 3 |
license: mit
|
| 4 |
+
base_model: microsoft/layoutlm-base-uncased
|
| 5 |
tags:
|
| 6 |
- generated_from_trainer
|
| 7 |
datasets:
|
|
|
|
| 16 |
|
| 17 |
# layoutlm-with-funsd
|
| 18 |
|
| 19 |
+
This model is a fine-tuned version of [microsoft/layoutlm-base-uncased](https://huggingface.co/microsoft/layoutlm-base-uncased) on the funsd dataset.
|
| 20 |
It achieves the following results on the evaluation set:
|
| 21 |
+
- Loss: 0.8090
|
| 22 |
+
- Eader: {'precision': 0.3333333333333333, 'recall': 0.21875, 'f1': 0.2641509433962264, 'number': 32}
|
| 23 |
+
- Nswer: {'precision': 0.3763440860215054, 'recall': 0.5, 'f1': 0.4294478527607362, 'number': 70}
|
| 24 |
+
- Uestion: {'precision': 0.3368421052631579, 'recall': 0.41025641025641024, 'f1': 0.3699421965317919, 'number': 78}
|
| 25 |
+
- Overall Precision: 0.3541
|
| 26 |
+
- Overall Recall: 0.4111
|
| 27 |
+
- Overall F1: 0.3805
|
| 28 |
+
- Overall Accuracy: 0.7559
|
| 29 |
|
| 30 |
## Model description
|
| 31 |
|
|
|
|
| 55 |
|
| 56 |
### Training results
|
| 57 |
|
| 58 |
+
| Training Loss | Epoch | Step | Validation Loss | Eader | Nswer | Uestion | Overall Precision | Overall Recall | Overall F1 | Overall Accuracy |
|
| 59 |
+
|:-------------:|:-----:|:----:|:---------------:|:---------------------------------------------------------------------------------------------:|:----------------------------------------------------------------------------------------------------------:|:----------------------------------------------------------------------------------------------------------:|:-----------------:|:--------------:|:----------:|:----------------:|
|
| 60 |
+
| 1.2801 | 1.0 | 9 | 1.0648 | {'precision': 0.0, 'recall': 0.0, 'f1': 0.0, 'number': 32} | {'precision': 0.07246376811594203, 'recall': 0.21428571428571427, 'f1': 0.10830324909747292, 'number': 70} | {'precision': 0.08292682926829269, 'recall': 0.21794871794871795, 'f1': 0.12014134275618374, 'number': 78} | 0.0777 | 0.1778 | 0.1081 | 0.6088 |
|
| 61 |
+
| 0.9803 | 2.0 | 18 | 0.8556 | {'precision': 0.0, 'recall': 0.0, 'f1': 0.0, 'number': 32} | {'precision': 0.1875, 'recall': 0.38571428571428573, 'f1': 0.25233644859813087, 'number': 70} | {'precision': 0.1259259259259259, 'recall': 0.21794871794871795, 'f1': 0.1596244131455399, 'number': 78} | 0.1577 | 0.2444 | 0.1917 | 0.7064 |
|
| 62 |
+
| 0.769 | 3.0 | 27 | 0.6782 | {'precision': 0.0, 'recall': 0.0, 'f1': 0.0, 'number': 32} | {'precision': 0.29591836734693877, 'recall': 0.4142857142857143, 'f1': 0.34523809523809523, 'number': 70} | {'precision': 0.3333333333333333, 'recall': 0.38461538461538464, 'f1': 0.3571428571428571, 'number': 78} | 0.2995 | 0.3278 | 0.3130 | 0.7740 |
|
| 63 |
+
| 0.6082 | 4.0 | 36 | 0.6412 | {'precision': 0.2, 'recall': 0.125, 'f1': 0.15384615384615385, 'number': 32} | {'precision': 0.3333333333333333, 'recall': 0.4714285714285714, 'f1': 0.3905325443786982, 'number': 70} | {'precision': 0.367816091954023, 'recall': 0.41025641025641024, 'f1': 0.3878787878787879, 'number': 78} | 0.3350 | 0.3833 | 0.3575 | 0.7655 |
|
| 64 |
+
| 0.5047 | 5.0 | 45 | 0.7447 | {'precision': 0.42105263157894735, 'recall': 0.25, 'f1': 0.3137254901960784, 'number': 32} | {'precision': 0.32978723404255317, 'recall': 0.44285714285714284, 'f1': 0.3780487804878049, 'number': 70} | {'precision': 0.36363636363636365, 'recall': 0.41025641025641024, 'f1': 0.3855421686746988, 'number': 78} | 0.3532 | 0.3944 | 0.3727 | 0.7275 |
|
| 65 |
+
| 0.422 | 6.0 | 54 | 0.6465 | {'precision': 0.2857142857142857, 'recall': 0.1875, 'f1': 0.22641509433962265, 'number': 32} | {'precision': 0.43902439024390244, 'recall': 0.5142857142857142, 'f1': 0.4736842105263158, 'number': 70} | {'precision': 0.46153846153846156, 'recall': 0.46153846153846156, 'f1': 0.46153846153846156, 'number': 78} | 0.4309 | 0.4333 | 0.4321 | 0.7951 |
|
| 66 |
+
| 0.3607 | 7.0 | 63 | 0.7246 | {'precision': 0.3684210526315789, 'recall': 0.21875, 'f1': 0.2745098039215686, 'number': 32} | {'precision': 0.3953488372093023, 'recall': 0.4857142857142857, 'f1': 0.43589743589743585, 'number': 70} | {'precision': 0.4069767441860465, 'recall': 0.44871794871794873, 'f1': 0.4268292682926829, 'number': 78} | 0.3979 | 0.4222 | 0.4097 | 0.7559 |
|
| 67 |
+
| 0.3106 | 8.0 | 72 | 0.7467 | {'precision': 0.3181818181818182, 'recall': 0.21875, 'f1': 0.25925925925925924, 'number': 32} | {'precision': 0.38202247191011235, 'recall': 0.4857142857142857, 'f1': 0.42767295597484273, 'number': 70} | {'precision': 0.3333333333333333, 'recall': 0.41025641025641024, 'f1': 0.36781609195402293, 'number': 78} | 0.3527 | 0.4056 | 0.3773 | 0.7288 |
|
| 68 |
+
| 0.2649 | 9.0 | 81 | 0.7238 | {'precision': 0.3333333333333333, 'recall': 0.21875, 'f1': 0.2641509433962264, 'number': 32} | {'precision': 0.3953488372093023, 'recall': 0.4857142857142857, 'f1': 0.43589743589743585, 'number': 70} | {'precision': 0.3404255319148936, 'recall': 0.41025641025641024, 'f1': 0.37209302325581395, 'number': 78} | 0.3632 | 0.4056 | 0.3832 | 0.7758 |
|
| 69 |
+
| 0.239 | 10.0 | 90 | 0.8137 | {'precision': 0.30434782608695654, 'recall': 0.21875, 'f1': 0.2545454545454546, 'number': 32} | {'precision': 0.4069767441860465, 'recall': 0.5, 'f1': 0.4487179487179487, 'number': 70} | {'precision': 0.37209302325581395, 'recall': 0.41025641025641024, 'f1': 0.3902439024390244, 'number': 78} | 0.3795 | 0.4111 | 0.3947 | 0.7288 |
|
| 70 |
+
| 0.2141 | 11.0 | 99 | 0.7518 | {'precision': 0.25, 'recall': 0.1875, 'f1': 0.21428571428571427, 'number': 32} | {'precision': 0.3645833333333333, 'recall': 0.5, 'f1': 0.42168674698795183, 'number': 70} | {'precision': 0.31958762886597936, 'recall': 0.3974358974358974, 'f1': 0.3542857142857143, 'number': 78} | 0.3318 | 0.4 | 0.3627 | 0.7589 |
|
| 71 |
+
| 0.1978 | 12.0 | 108 | 0.8165 | {'precision': 0.3333333333333333, 'recall': 0.21875, 'f1': 0.2641509433962264, 'number': 32} | {'precision': 0.41975308641975306, 'recall': 0.4857142857142857, 'f1': 0.4503311258278146, 'number': 70} | {'precision': 0.37209302325581395, 'recall': 0.41025641025641024, 'f1': 0.3902439024390244, 'number': 78} | 0.3883 | 0.4056 | 0.3967 | 0.7438 |
|
| 72 |
+
| 0.1807 | 13.0 | 117 | 0.7946 | {'precision': 0.3181818181818182, 'recall': 0.21875, 'f1': 0.25925925925925924, 'number': 32} | {'precision': 0.358695652173913, 'recall': 0.4714285714285714, 'f1': 0.4074074074074074, 'number': 70} | {'precision': 0.31958762886597936, 'recall': 0.3974358974358974, 'f1': 0.3542857142857143, 'number': 78} | 0.3365 | 0.3944 | 0.3632 | 0.7565 |
|
| 73 |
+
| 0.1705 | 14.0 | 126 | 0.8007 | {'precision': 0.3684210526315789, 'recall': 0.21875, 'f1': 0.2745098039215686, 'number': 32} | {'precision': 0.3684210526315789, 'recall': 0.5, 'f1': 0.4242424242424242, 'number': 70} | {'precision': 0.32989690721649484, 'recall': 0.41025641025641024, 'f1': 0.36571428571428566, 'number': 78} | 0.3507 | 0.4111 | 0.3785 | 0.7601 |
|
| 74 |
+
| 0.1676 | 15.0 | 135 | 0.8090 | {'precision': 0.3333333333333333, 'recall': 0.21875, 'f1': 0.2641509433962264, 'number': 32} | {'precision': 0.3763440860215054, 'recall': 0.5, 'f1': 0.4294478527607362, 'number': 70} | {'precision': 0.3368421052631579, 'recall': 0.41025641025641024, 'f1': 0.3699421965317919, 'number': 78} | 0.3541 | 0.4111 | 0.3805 | 0.7559 |
|
| 75 |
|
| 76 |
|
| 77 |
### Framework versions
|
logs/events.out.tfevents.1741099192.DESKTOP-HA84SVN.2309656.3
CHANGED
|
@@ -1,3 +1,3 @@
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
-
oid sha256:
|
| 3 |
-
size
|
|
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:9429a449a6d1944ffc5aedfd2b8681b4ef6f5b31f96935acc7e5019563504d27
|
| 3 |
+
size 16102
|
model.safetensors
CHANGED
|
@@ -1,3 +1,3 @@
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
-
oid sha256:
|
| 3 |
size 450548984
|
|
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:f9bb63bb8fbcf7c09decfade85afd8c76ee77c8748bcb00d27af6986fb1f349e
|
| 3 |
size 450548984
|
tokenizer_config.json
CHANGED
|
@@ -55,11 +55,9 @@
|
|
| 55 |
"do_lower_case": true,
|
| 56 |
"extra_special_tokens": {},
|
| 57 |
"mask_token": "[MASK]",
|
| 58 |
-
"max_length": 512,
|
| 59 |
"model_max_length": 512,
|
| 60 |
"never_split": null,
|
| 61 |
"only_label_first_subword": true,
|
| 62 |
-
"pad_to_multiple_of": null,
|
| 63 |
"pad_token": "[PAD]",
|
| 64 |
"pad_token_box": [
|
| 65 |
0,
|
|
@@ -68,8 +66,6 @@
|
|
| 68 |
0
|
| 69 |
],
|
| 70 |
"pad_token_label": -100,
|
| 71 |
-
"pad_token_type_id": 0,
|
| 72 |
-
"padding_side": "right",
|
| 73 |
"processor_class": "LayoutLMv2Processor",
|
| 74 |
"sep_token": "[SEP]",
|
| 75 |
"sep_token_box": [
|
|
@@ -78,11 +74,8 @@
|
|
| 78 |
1000,
|
| 79 |
1000
|
| 80 |
],
|
| 81 |
-
"stride": 0,
|
| 82 |
"strip_accents": null,
|
| 83 |
"tokenize_chinese_chars": true,
|
| 84 |
"tokenizer_class": "LayoutLMv2Tokenizer",
|
| 85 |
-
"truncation_side": "right",
|
| 86 |
-
"truncation_strategy": "longest_first",
|
| 87 |
"unk_token": "[UNK]"
|
| 88 |
}
|
|
|
|
| 55 |
"do_lower_case": true,
|
| 56 |
"extra_special_tokens": {},
|
| 57 |
"mask_token": "[MASK]",
|
|
|
|
| 58 |
"model_max_length": 512,
|
| 59 |
"never_split": null,
|
| 60 |
"only_label_first_subword": true,
|
|
|
|
| 61 |
"pad_token": "[PAD]",
|
| 62 |
"pad_token_box": [
|
| 63 |
0,
|
|
|
|
| 66 |
0
|
| 67 |
],
|
| 68 |
"pad_token_label": -100,
|
|
|
|
|
|
|
| 69 |
"processor_class": "LayoutLMv2Processor",
|
| 70 |
"sep_token": "[SEP]",
|
| 71 |
"sep_token_box": [
|
|
|
|
| 74 |
1000,
|
| 75 |
1000
|
| 76 |
],
|
|
|
|
| 77 |
"strip_accents": null,
|
| 78 |
"tokenize_chinese_chars": true,
|
| 79 |
"tokenizer_class": "LayoutLMv2Tokenizer",
|
|
|
|
|
|
|
| 80 |
"unk_token": "[UNK]"
|
| 81 |
}
|