TomasFAV commited on
Commit
4f06f81
·
verified ·
1 Parent(s): 3e39b4b

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +82 -20
README.md CHANGED
@@ -4,40 +4,99 @@ license: mit
4
  base_model: SCUT-DLVCLab/lilt-roberta-en-base
5
  tags:
6
  - generated_from_trainer
 
 
 
 
 
 
7
  metrics:
8
  - precision
9
  - recall
10
  - f1
11
  - accuracy
12
  model-index:
13
- - name: LiLTInvoiceCzech
14
  results: []
15
  ---
16
 
17
- <!-- This model card has been generated automatically according to the information the Trainer had access to. You
18
- should probably proofread and complete it, then remove this comment. -->
19
 
20
- # LiLTInvoiceCzech
21
 
22
- This model is a fine-tuned version of [SCUT-DLVCLab/lilt-roberta-en-base](https://huggingface.co/SCUT-DLVCLab/lilt-roberta-en-base) on an unknown dataset.
23
  It achieves the following results on the evaluation set:
24
- - Loss: 0.1929
25
- - Precision: 0.6036
26
- - Recall: 0.7355
27
- - F1: 0.6631
28
- - Accuracy: 0.9645
 
 
29
 
30
  ## Model description
31
 
32
- More information needed
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
33
 
34
- ## Intended uses & limitations
35
 
36
- More information needed
 
 
 
 
 
37
 
38
- ## Training and evaluation data
 
 
 
 
 
 
 
39
 
40
- More information needed
 
 
 
 
 
 
 
41
 
42
  ## Training procedure
43
 
@@ -54,6 +113,8 @@ The following hyperparameters were used during training:
54
  - num_epochs: 10
55
  - mixed_precision_training: Native AMP
56
 
 
 
57
  ### Training results
58
 
59
  | Training Loss | Epoch | Step | Validation Loss | Precision | Recall | F1 | Accuracy |
@@ -69,10 +130,11 @@ The following hyperparameters were used during training:
69
  | 0.1991 | 9.0 | 675 | 0.2133 | 0.5357 | 0.7167 | 0.6131 | 0.9583 |
70
  | 0.1991 | 10.0 | 750 | 0.2198 | 0.5235 | 0.7235 | 0.6074 | 0.9569 |
71
 
 
72
 
73
- ### Framework versions
74
 
75
- - Transformers 5.0.0
76
- - Pytorch 2.10.0+cu128
77
- - Datasets 4.0.0
78
- - Tokenizers 0.22.2
 
4
  base_model: SCUT-DLVCLab/lilt-roberta-en-base
5
  tags:
6
  - generated_from_trainer
7
+ - invoice-processing
8
+ - information-extraction
9
+ - czech-language
10
+ - document-ai
11
+ - layout-aware-model
12
+ - synthetic-data
13
  metrics:
14
  - precision
15
  - recall
16
  - f1
17
  - accuracy
18
  model-index:
19
+ - name: LiLTInvoiceCzech-V0
20
  results: []
21
  ---
22
 
23
+ # LiLTInvoiceCzech (V0 Synthetic Templates Only)
 
24
 
25
+ This model is a fine-tuned version of [SCUT-DLVCLab/lilt-roberta-en-base](https://huggingface.co/SCUT-DLVCLab/lilt-roberta-en-base) for structured information extraction from Czech invoices.
26
 
 
27
  It achieves the following results on the evaluation set:
28
+ - Loss: 0.1929
29
+ - Precision: 0.6036
30
+ - Recall: 0.7355
31
+ - F1: 0.6631
32
+ - Accuracy: 0.9645
33
+
34
+ ---
35
 
36
  ## Model description
37
 
38
+ LiLTInvoiceCzech (V0) is a layout-aware model based on the LiLT architecture, designed for document understanding tasks.
39
+
40
+ The model performs token-level classification with explicit use of layout information (bounding boxes), allowing it to better capture spatial relationships between invoice fields such as:
41
+ - supplier
42
+ - customer
43
+ - invoice number
44
+ - bank details
45
+ - totals
46
+ - dates
47
+
48
+ This version is trained exclusively on synthetically generated invoice templates.
49
+
50
+ ---
51
+
52
+ ## Training data
53
+
54
+ The dataset consists of:
55
+
56
+ - synthetically generated invoices
57
+ - fixed template layouts
58
+ - associated bounding box annotations for each token
59
+
60
+ Key properties:
61
+ - consistent spatial structure
62
+ - clean and noise-free data
63
+ - precise alignment between text and layout
64
+ - no real-world documents
65
+
66
+ This represents the **baseline dataset** for layout-aware models in the pipeline.
67
+
68
+ ---
69
+
70
+ ## Role in the pipeline
71
+
72
+ This model corresponds to:
73
 
74
+ **V0 Synthetic template-based dataset only**
75
 
76
+ It is used to:
77
+ - establish a baseline for LiLT architecture
78
+ - compare layout-aware vs text-only models (e.g., BERT)
79
+ - evaluate the benefit of spatial features in a controlled setting
80
+
81
+ ---
82
 
83
+ ## Intended uses
84
+
85
+ - Document AI research with layout-aware models
86
+ - Benchmarking LiLT on structured documents
87
+ - Comparison with other architectures (BERT, LayoutLMv3, etc.)
88
+ - Czech invoice information extraction
89
+
90
+ ---
91
 
92
+ ## Limitations
93
+
94
+ - Trained only on synthetic data with fixed layouts
95
+ - Limited robustness to layout variability
96
+ - No exposure to real-world noise (OCR errors, distortions)
97
+ - Synthetic layouts may not reflect real invoice diversity
98
+
99
+ ---
100
 
101
  ## Training procedure
102
 
 
113
  - num_epochs: 10
114
  - mixed_precision_training: Native AMP
115
 
116
+ ---
117
+
118
  ### Training results
119
 
120
  | Training Loss | Epoch | Step | Validation Loss | Precision | Recall | F1 | Accuracy |
 
130
  | 0.1991 | 9.0 | 675 | 0.2133 | 0.5357 | 0.7167 | 0.6131 | 0.9583 |
131
  | 0.1991 | 10.0 | 750 | 0.2198 | 0.5235 | 0.7235 | 0.6074 | 0.9569 |
132
 
133
+ ---
134
 
135
+ ## Framework versions
136
 
137
+ - Transformers 5.0.0
138
+ - PyTorch 2.10.0+cu128
139
+ - Datasets 4.0.0
140
+ - Tokenizers 0.22.2