TomasFAV commited on
Commit
2b396ae
·
verified ·
1 Parent(s): 6e65b46

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +87 -20
README.md CHANGED
@@ -1,41 +1,105 @@
1
  ---
2
  library_name: transformers
 
 
3
  tags:
4
  - generated_from_trainer
 
 
 
 
 
 
 
5
  metrics:
6
  - precision
7
  - recall
8
  - f1
9
  - accuracy
10
  model-index:
11
- - name: LiLTInvoiceCzechV1
12
  results: []
13
  ---
14
 
15
- <!-- This model card has been generated automatically according to the information the Trainer had access to. You
16
- should probably proofread and complete it, then remove this comment. -->
17
 
18
- # LiLTInvoiceCzechV1
19
 
20
- This model was trained from scratch on an unknown dataset.
21
  It achieves the following results on the evaluation set:
22
- - Loss: 0.1907
23
- - Precision: 0.6326
24
- - Recall: 0.7491
25
- - F1: 0.6859
26
- - Accuracy: 0.9660
 
 
27
 
28
  ## Model description
29
 
30
- More information needed
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
31
 
32
- ## Intended uses & limitations
33
 
34
- More information needed
 
 
 
 
 
 
 
35
 
36
- ## Training and evaluation data
 
 
 
 
 
 
 
37
 
38
- More information needed
 
 
 
 
 
 
 
39
 
40
  ## Training procedure
41
 
@@ -52,6 +116,8 @@ The following hyperparameters were used during training:
52
  - num_epochs: 10
53
  - mixed_precision_training: Native AMP
54
 
 
 
55
  ### Training results
56
 
57
  | Training Loss | Epoch | Step | Validation Loss | Precision | Recall | F1 | Accuracy |
@@ -67,10 +133,11 @@ The following hyperparameters were used during training:
67
  | No log | 9.0 | 342 | 0.1939 | 0.6700 | 0.6962 | 0.6828 | 0.9701 |
68
  | No log | 10.0 | 380 | 0.1931 | 0.6645 | 0.6928 | 0.6784 | 0.9696 |
69
 
 
70
 
71
- ### Framework versions
72
 
73
- - Transformers 5.0.0
74
- - Pytorch 2.10.0+cu128
75
- - Datasets 4.0.0
76
- - Tokenizers 0.22.2
 
1
  ---
2
  library_name: transformers
3
+ license: mit
4
+ base_model: SCUT-DLVCLab/lilt-roberta-en-base
5
  tags:
6
  - generated_from_trainer
7
+ - invoice-processing
8
+ - information-extraction
9
+ - czech-language
10
+ - document-ai
11
+ - layout-aware-model
12
+ - synthetic-data
13
+ - layout-augmentation
14
  metrics:
15
  - precision
16
  - recall
17
  - f1
18
  - accuracy
19
  model-index:
20
+ - name: LiLTInvoiceCzech-V1
21
  results: []
22
  ---
23
 
24
+ # LiLTInvoiceCzech (V1 Synthetic + Random Layout)
 
25
 
26
+ This model is a fine-tuned version of [SCUT-DLVCLab/lilt-roberta-en-base](https://huggingface.co/SCUT-DLVCLab/lilt-roberta-en-base) for structured information extraction from Czech invoices.
27
 
 
28
  It achieves the following results on the evaluation set:
29
+ - Loss: 0.1907
30
+ - Precision: 0.6326
31
+ - Recall: 0.7491
32
+ - F1: 0.6859
33
+ - Accuracy: 0.9660
34
+
35
+ ---
36
 
37
  ## Model description
38
 
39
+ LiLTInvoiceCzech (V1) extends the baseline layout-aware model by introducing layout variability into the training data.
40
+
41
+ The model performs token-level classification using both textual and spatial (bounding box) information to extract structured invoice fields:
42
+ - supplier
43
+ - customer
44
+ - invoice number
45
+ - bank details
46
+ - totals
47
+ - dates
48
+
49
+ Compared to V0, this version is trained on synthetically generated invoices with **randomized layouts**, improving robustness to spatial variations.
50
+
51
+ ---
52
+
53
+ ## Training data
54
+
55
+ The dataset consists of:
56
+
57
+ - synthetically generated invoices based on templates
58
+ - augmented variants with randomized layout structures
59
+ - corresponding bounding box annotations
60
+
61
+ Key properties:
62
+ - variable positioning of fields
63
+ - layout perturbations (shifts, spacing, ordering)
64
+ - preserved label consistency
65
+ - fully synthetic data
66
+
67
+ This dataset introduces **layout diversity**, which is especially important for layout-aware models.
68
+
69
+ ---
70
+
71
+ ## Role in the pipeline
72
+
73
+ This model corresponds to:
74
 
75
+ **V1 Synthetic templates + randomized layouts**
76
 
77
+ It is used to:
78
+ - evaluate the effect of layout variability on LiLT
79
+ - compare against:
80
+ - V0 (fixed layouts)
81
+ - later stages with hybrid and real data (V2, V3)
82
+ - analyze how layout-aware models benefit from synthetic augmentation
83
+
84
+ ---
85
 
86
+ ## Intended uses
87
+
88
+ - Research in layout-aware document understanding
89
+ - Evaluation of spatial robustness in NLP models
90
+ - Benchmarking LiLT against text-only models (BERT)
91
+ - Czech invoice information extraction
92
+
93
+ ---
94
 
95
+ ## Limitations
96
+
97
+ - Still trained only on synthetic data
98
+ - Layout variability is artificial
99
+ - No real-world noise (OCR errors, distortions)
100
+ - May not fully generalize to real invoice distributions
101
+
102
+ ---
103
 
104
  ## Training procedure
105
 
 
116
  - num_epochs: 10
117
  - mixed_precision_training: Native AMP
118
 
119
+ ---
120
+
121
  ### Training results
122
 
123
  | Training Loss | Epoch | Step | Validation Loss | Precision | Recall | F1 | Accuracy |
 
133
  | No log | 9.0 | 342 | 0.1939 | 0.6700 | 0.6962 | 0.6828 | 0.9701 |
134
  | No log | 10.0 | 380 | 0.1931 | 0.6645 | 0.6928 | 0.6784 | 0.9696 |
135
 
136
+ ---
137
 
138
+ ## Framework versions
139
 
140
+ - Transformers 5.0.0
141
+ - PyTorch 2.10.0+cu128
142
+ - Datasets 4.0.0
143
+ - Tokenizers 0.22.2