Fine-tuned Construction Receipt Model
Browse files- README.md +66 -69
- config.json +1 -1
- generation_config.json +1 -1
- model.safetensors +1 -1
README.md
CHANGED
|
@@ -16,7 +16,7 @@ should probably proofread and complete it, then remove this comment. -->
|
|
| 16 |
|
| 17 |
This model is a fine-tuned version of [DanSarm/receipt-core-model](https://huggingface.co/DanSarm/receipt-core-model) on an unknown dataset.
|
| 18 |
It achieves the following results on the evaluation set:
|
| 19 |
-
- Loss: 0.
|
| 20 |
|
| 21 |
## Model description
|
| 22 |
|
|
@@ -48,77 +48,74 @@ The following hyperparameters were used during training:
|
|
| 48 |
|
| 49 |
| Training Loss | Epoch | Step | Validation Loss |
|
| 50 |
|:-------------:|:-----:|:----:|:---------------:|
|
| 51 |
-
|
|
| 52 |
-
|
|
| 53 |
-
| 0.
|
| 54 |
-
| 0.
|
| 55 |
-
| 0.
|
| 56 |
-
| 0.
|
| 57 |
-
| 0.
|
| 58 |
-
| 0.
|
| 59 |
-
| 0.
|
| 60 |
-
| 0.
|
| 61 |
-
| 0.
|
| 62 |
-
| 0.
|
| 63 |
-
| 0.
|
| 64 |
-
| 0.
|
| 65 |
-
| 0.
|
| 66 |
-
| 0.
|
| 67 |
-
| 0.
|
| 68 |
-
| 0.
|
| 69 |
-
| 0.
|
| 70 |
-
| 0.
|
| 71 |
-
| 0.
|
| 72 |
-
| 0.
|
| 73 |
-
| 0.
|
| 74 |
-
| 0.
|
| 75 |
-
| 0.
|
| 76 |
-
| 0.
|
| 77 |
-
| 0.
|
| 78 |
-
| 0.
|
| 79 |
-
| 0.
|
| 80 |
-
| 0.
|
| 81 |
-
| 0.
|
| 82 |
-
| 0.
|
| 83 |
-
| 0.
|
| 84 |
-
| 0.
|
| 85 |
-
| 0.
|
| 86 |
-
| 0.
|
| 87 |
-
| 0.
|
| 88 |
-
| 0.
|
| 89 |
-
| 0.
|
| 90 |
-
| 0.
|
| 91 |
-
| 0.
|
| 92 |
-
| 0.
|
| 93 |
-
| 0.0185 | 43.0 |
|
| 94 |
-
| 0.
|
| 95 |
-
| 0.
|
| 96 |
-
| 0.
|
| 97 |
-
| 0.
|
| 98 |
-
| 0.
|
| 99 |
-
| 0.
|
| 100 |
-
| 0.
|
| 101 |
-
| 0.
|
| 102 |
-
| 0.
|
| 103 |
-
| 0.
|
| 104 |
-
| 0.
|
| 105 |
-
| 0.
|
| 106 |
-
| 0.
|
| 107 |
-
| 0.
|
| 108 |
-
| 0.
|
| 109 |
-
| 0.
|
| 110 |
-
| 0.
|
| 111 |
-
| 0.
|
| 112 |
-
| 0.
|
| 113 |
-
| 0.
|
| 114 |
-
| 0.012 | 64.0 | 640 | 0.5479 |
|
| 115 |
-
| 0.0106 | 65.0 | 650 | 0.5456 |
|
| 116 |
-
| 0.0097 | 66.0 | 660 | 0.5390 |
|
| 117 |
|
| 118 |
|
| 119 |
### Framework versions
|
| 120 |
|
| 121 |
-
- Transformers 4.
|
| 122 |
- Pytorch 2.6.0+cu124
|
| 123 |
-
- Datasets 3.
|
| 124 |
- Tokenizers 0.21.0
|
|
|
|
| 16 |
|
| 17 |
This model is a fine-tuned version of [DanSarm/receipt-core-model](https://huggingface.co/DanSarm/receipt-core-model) on an unknown dataset.
|
| 18 |
It achieves the following results on the evaluation set:
|
| 19 |
+
- Loss: 0.2738
|
| 20 |
|
| 21 |
## Model description
|
| 22 |
|
|
|
|
| 48 |
|
| 49 |
| Training Loss | Epoch | Step | Validation Loss |
|
| 50 |
|:-------------:|:-----:|:----:|:---------------:|
|
| 51 |
+
| 1.9458 | 1.0 | 25 | 0.6631 |
|
| 52 |
+
| 0.6783 | 2.0 | 50 | 0.4201 |
|
| 53 |
+
| 0.448 | 3.0 | 75 | 0.3256 |
|
| 54 |
+
| 0.3453 | 4.0 | 100 | 0.2814 |
|
| 55 |
+
| 0.2939 | 5.0 | 125 | 0.2581 |
|
| 56 |
+
| 0.2389 | 6.0 | 150 | 0.2528 |
|
| 57 |
+
| 0.2087 | 7.0 | 175 | 0.2462 |
|
| 58 |
+
| 0.1769 | 8.0 | 200 | 0.2311 |
|
| 59 |
+
| 0.1746 | 9.0 | 225 | 0.2286 |
|
| 60 |
+
| 0.1488 | 10.0 | 250 | 0.2306 |
|
| 61 |
+
| 0.1322 | 11.0 | 275 | 0.2275 |
|
| 62 |
+
| 0.1219 | 12.0 | 300 | 0.2243 |
|
| 63 |
+
| 0.1161 | 13.0 | 325 | 0.2069 |
|
| 64 |
+
| 0.0984 | 14.0 | 350 | 0.2317 |
|
| 65 |
+
| 0.0936 | 15.0 | 375 | 0.2312 |
|
| 66 |
+
| 0.0891 | 16.0 | 400 | 0.2274 |
|
| 67 |
+
| 0.0792 | 17.0 | 425 | 0.2311 |
|
| 68 |
+
| 0.07 | 18.0 | 450 | 0.2399 |
|
| 69 |
+
| 0.0666 | 19.0 | 475 | 0.2336 |
|
| 70 |
+
| 0.0704 | 20.0 | 500 | 0.2349 |
|
| 71 |
+
| 0.0644 | 21.0 | 525 | 0.2397 |
|
| 72 |
+
| 0.0552 | 22.0 | 550 | 0.2434 |
|
| 73 |
+
| 0.0517 | 23.0 | 575 | 0.2428 |
|
| 74 |
+
| 0.0475 | 24.0 | 600 | 0.2462 |
|
| 75 |
+
| 0.0453 | 25.0 | 625 | 0.2203 |
|
| 76 |
+
| 0.0422 | 26.0 | 650 | 0.2264 |
|
| 77 |
+
| 0.0395 | 27.0 | 675 | 0.2366 |
|
| 78 |
+
| 0.0394 | 28.0 | 700 | 0.2393 |
|
| 79 |
+
| 0.0361 | 29.0 | 725 | 0.2423 |
|
| 80 |
+
| 0.0302 | 30.0 | 750 | 0.2480 |
|
| 81 |
+
| 0.0317 | 31.0 | 775 | 0.2441 |
|
| 82 |
+
| 0.0265 | 32.0 | 800 | 0.2519 |
|
| 83 |
+
| 0.027 | 33.0 | 825 | 0.2541 |
|
| 84 |
+
| 0.027 | 34.0 | 850 | 0.2512 |
|
| 85 |
+
| 0.0266 | 35.0 | 875 | 0.2590 |
|
| 86 |
+
| 0.0246 | 36.0 | 900 | 0.2319 |
|
| 87 |
+
| 0.023 | 37.0 | 925 | 0.2419 |
|
| 88 |
+
| 0.0195 | 38.0 | 950 | 0.2473 |
|
| 89 |
+
| 0.0206 | 39.0 | 975 | 0.2471 |
|
| 90 |
+
| 0.019 | 40.0 | 1000 | 0.2485 |
|
| 91 |
+
| 0.0175 | 41.0 | 1025 | 0.2635 |
|
| 92 |
+
| 0.0163 | 42.0 | 1050 | 0.2513 |
|
| 93 |
+
| 0.0185 | 43.0 | 1075 | 0.2618 |
|
| 94 |
+
| 0.0167 | 44.0 | 1100 | 0.2549 |
|
| 95 |
+
| 0.0161 | 45.0 | 1125 | 0.2540 |
|
| 96 |
+
| 0.0163 | 46.0 | 1150 | 0.2543 |
|
| 97 |
+
| 0.0149 | 47.0 | 1175 | 0.2482 |
|
| 98 |
+
| 0.016 | 48.0 | 1200 | 0.2487 |
|
| 99 |
+
| 0.0134 | 49.0 | 1225 | 0.2572 |
|
| 100 |
+
| 0.0136 | 50.0 | 1250 | 0.2589 |
|
| 101 |
+
| 0.0141 | 51.0 | 1275 | 0.2512 |
|
| 102 |
+
| 0.0108 | 52.0 | 1300 | 0.2565 |
|
| 103 |
+
| 0.011 | 53.0 | 1325 | 0.2512 |
|
| 104 |
+
| 0.0094 | 54.0 | 1350 | 0.2588 |
|
| 105 |
+
| 0.0132 | 55.0 | 1375 | 0.2515 |
|
| 106 |
+
| 0.0125 | 56.0 | 1400 | 0.2597 |
|
| 107 |
+
| 0.0118 | 57.0 | 1425 | 0.2601 |
|
| 108 |
+
| 0.0097 | 58.0 | 1450 | 0.2579 |
|
| 109 |
+
| 0.0098 | 59.0 | 1475 | 0.2586 |
|
| 110 |
+
| 0.0083 | 60.0 | 1500 | 0.2821 |
|
| 111 |
+
| 0.0081 | 61.0 | 1525 | 0.2811 |
|
| 112 |
+
| 0.0081 | 62.0 | 1550 | 0.2633 |
|
| 113 |
+
| 0.0078 | 63.0 | 1575 | 0.2738 |
|
|
|
|
|
|
|
|
|
|
| 114 |
|
| 115 |
|
| 116 |
### Framework versions
|
| 117 |
|
| 118 |
+
- Transformers 4.49.0
|
| 119 |
- Pytorch 2.6.0+cu124
|
| 120 |
+
- Datasets 3.3.1
|
| 121 |
- Tokenizers 0.21.0
|
config.json
CHANGED
|
@@ -55,7 +55,7 @@
|
|
| 55 |
}
|
| 56 |
},
|
| 57 |
"torch_dtype": "float32",
|
| 58 |
-
"transformers_version": "4.
|
| 59 |
"use_cache": true,
|
| 60 |
"vocab_size": 32128
|
| 61 |
}
|
|
|
|
| 55 |
}
|
| 56 |
},
|
| 57 |
"torch_dtype": "float32",
|
| 58 |
+
"transformers_version": "4.49.0",
|
| 59 |
"use_cache": true,
|
| 60 |
"vocab_size": 32128
|
| 61 |
}
|
generation_config.json
CHANGED
|
@@ -3,5 +3,5 @@
|
|
| 3 |
"decoder_start_token_id": 0,
|
| 4 |
"eos_token_id": 1,
|
| 5 |
"pad_token_id": 0,
|
| 6 |
-
"transformers_version": "4.
|
| 7 |
}
|
|
|
|
| 3 |
"decoder_start_token_id": 0,
|
| 4 |
"eos_token_id": 1,
|
| 5 |
"pad_token_id": 0,
|
| 6 |
+
"transformers_version": "4.49.0"
|
| 7 |
}
|
model.safetensors
CHANGED
|
@@ -1,3 +1,3 @@
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
-
oid sha256:
|
| 3 |
size 891644712
|
|
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:22194e837198408c893d67c727c0776bb8cac42c3eb2fe6486900c4e6b45987f
|
| 3 |
size 891644712
|