TomasFAV commited on
Commit
ece841e
·
verified ·
1 Parent(s): 01bc193

Model save

Browse files
Files changed (2) hide show
  1. README.md +31 -15
  2. model.safetensors +1 -1
README.md CHANGED
@@ -4,6 +4,11 @@ license: mit
4
  base_model: SCUT-DLVCLab/lilt-roberta-en-base
5
  tags:
6
  - generated_from_trainer
 
 
 
 
 
7
  model-index:
8
  - name: LiLTInvoiceCzech
9
  results: []
@@ -16,16 +21,11 @@ should probably proofread and complete it, then remove this comment. -->
16
 
17
  This model is a fine-tuned version of [SCUT-DLVCLab/lilt-roberta-en-base](https://huggingface.co/SCUT-DLVCLab/lilt-roberta-en-base) on an unknown dataset.
18
  It achieves the following results on the evaluation set:
19
- - eval_loss: 0.0224
20
- - eval_precision: 0.9615
21
- - eval_recall: 0.9650
22
- - eval_f1: 0.9632
23
- - eval_accuracy: 0.9944
24
- - eval_runtime: 1.0854
25
- - eval_samples_per_second: 60.809
26
- - eval_steps_per_second: 30.405
27
- - epoch: 13.0
28
- - step: 559
29
 
30
  ## Model description
31
 
@@ -45,18 +45,34 @@ More information needed
45
 
46
  The following hyperparameters were used during training:
47
  - learning_rate: 1e-05
48
- - train_batch_size: 16
49
  - eval_batch_size: 2
50
  - seed: 42
51
  - optimizer: Use OptimizerNames.ADAMW_TORCH_FUSED with betas=(0.9,0.999) and epsilon=1e-08 and optimizer_args=No additional optimizer arguments
52
  - lr_scheduler_type: linear
53
- - lr_scheduler_warmup_ratio: 0.1
54
- - num_epochs: 20
55
  - mixed_precision_training: Native AMP
56
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
57
  ### Framework versions
58
 
59
- - Transformers 4.57.6
60
- - Pytorch 2.9.0+cu126
61
  - Datasets 4.0.0
62
  - Tokenizers 0.22.2
 
4
  base_model: SCUT-DLVCLab/lilt-roberta-en-base
5
  tags:
6
  - generated_from_trainer
7
+ metrics:
8
+ - precision
9
+ - recall
10
+ - f1
11
+ - accuracy
12
  model-index:
13
  - name: LiLTInvoiceCzech
14
  results: []
 
21
 
22
  This model is a fine-tuned version of [SCUT-DLVCLab/lilt-roberta-en-base](https://huggingface.co/SCUT-DLVCLab/lilt-roberta-en-base) on an unknown dataset.
23
  It achieves the following results on the evaluation set:
24
+ - Loss: 0.0399
25
+ - Precision: 0.8785
26
+ - Recall: 0.9074
27
+ - F1: 0.8927
28
+ - Accuracy: 0.9919
 
 
 
 
 
29
 
30
  ## Model description
31
 
 
45
 
46
  The following hyperparameters were used during training:
47
  - learning_rate: 1e-05
48
+ - train_batch_size: 4
49
  - eval_batch_size: 2
50
  - seed: 42
51
  - optimizer: Use OptimizerNames.ADAMW_TORCH_FUSED with betas=(0.9,0.999) and epsilon=1e-08 and optimizer_args=No additional optimizer arguments
52
  - lr_scheduler_type: linear
53
+ - lr_scheduler_warmup_steps: 0.1
54
+ - num_epochs: 10
55
  - mixed_precision_training: Native AMP
56
 
57
+ ### Training results
58
+
59
+ | Training Loss | Epoch | Step | Validation Loss | Precision | Recall | F1 | Accuracy |
60
+ |:-------------:|:-----:|:----:|:---------------:|:---------:|:------:|:------:|:--------:|
61
+ | No log | 1.0 | 46 | 0.0509 | 0.7973 | 0.8498 | 0.8227 | 0.9877 |
62
+ | No log | 2.0 | 92 | 0.0417 | 0.8521 | 0.9012 | 0.8760 | 0.9907 |
63
+ | No log | 3.0 | 138 | 0.0392 | 0.8651 | 0.8971 | 0.8808 | 0.9915 |
64
+ | No log | 4.0 | 184 | 0.0432 | 0.8261 | 0.8992 | 0.8611 | 0.9900 |
65
+ | No log | 5.0 | 230 | 0.0429 | 0.8671 | 0.8992 | 0.8828 | 0.9912 |
66
+ | No log | 6.0 | 276 | 0.0412 | 0.8661 | 0.9053 | 0.8853 | 0.9913 |
67
+ | No log | 7.0 | 322 | 0.0411 | 0.8685 | 0.8971 | 0.8826 | 0.9918 |
68
+ | No log | 8.0 | 368 | 0.0409 | 0.8661 | 0.9053 | 0.8853 | 0.9914 |
69
+ | No log | 9.0 | 414 | 0.0393 | 0.8802 | 0.9074 | 0.8936 | 0.9920 |
70
+ | No log | 10.0 | 460 | 0.0399 | 0.8785 | 0.9074 | 0.8927 | 0.9919 |
71
+
72
+
73
  ### Framework versions
74
 
75
+ - Transformers 5.0.0
76
+ - Pytorch 2.10.0+cu128
77
  - Datasets 4.0.0
78
  - Tokenizers 0.22.2
model.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:0fbf94069ba2151627e46279729b6effb239d431b3f641d72abf4d1947d30b53
3
  size 520801340
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:4c643bd8dcd0cfb0eaebe7e2e18c24a3a3144c65e7939682dd03c7ae0b716b54
3
  size 520801340