come-st commited on
Commit
2948d95
·
1 Parent(s): da66f7c

End of training

Browse files
README.md CHANGED
@@ -15,6 +15,8 @@ should probably proofread and complete it, then remove this comment. -->
15
  # donut_finetuning_macif
16
 
17
  This model is a fine-tuned version of [naver-clova-ix/donut-base](https://huggingface.co/naver-clova-ix/donut-base) on the imagefolder dataset.
 
 
18
 
19
  ## Model description
20
 
@@ -43,6 +45,11 @@ The following hyperparameters were used during training:
43
 
44
  ### Training results
45
 
 
 
 
 
 
46
 
47
 
48
  ### Framework versions
 
15
  # donut_finetuning_macif
16
 
17
  This model is a fine-tuned version of [naver-clova-ix/donut-base](https://huggingface.co/naver-clova-ix/donut-base) on the imagefolder dataset.
18
+ It achieves the following results on the evaluation set:
19
+ - Loss: 0.0392
20
 
21
  ## Model description
22
 
 
45
 
46
  ### Training results
47
 
48
+ | Training Loss | Epoch | Step | Validation Loss |
49
+ |:-------------:|:-----:|:----:|:---------------:|
50
+ | 0.0537 | 1.0 | 2552 | 0.0519 |
51
+ | 0.0267 | 2.0 | 5104 | 0.0400 |
52
+ | 0.0096 | 3.0 | 7656 | 0.0392 |
53
 
54
 
55
  ### Framework versions
added_tokens.json CHANGED
@@ -1,9 +1,5 @@
1
  {
2
- "</s_gt_parse>": 57526,
3
- "</s_immatriculation>": 57528,
4
- "<s_gt_parse>": 57525,
5
  "<s_iitcdip>": 57523,
6
- "<s_immatriculation>": 57527,
7
  "<s_synthdog>": 57524,
8
  "<sep/>": 57522
9
  }
 
1
  {
 
 
 
2
  "<s_iitcdip>": 57523,
 
3
  "<s_synthdog>": 57524,
4
  "<sep/>": 57522
5
  }
runs/Jul07_14-52-23_datascience-1-0-ml-g4dn-xlarge-b3e2c7fea00ae03f4a997be929d0/events.out.tfevents.1688741556.datascience-1-0-ml-g4dn-xlarge-b3e2c7fea00ae03f4a997be929d0.15876.0 CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:a19032b4df0a75ead534bc397755e4440cd7ec09306cc5fdf3b48f597137936c
3
- size 21317
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:3cd985f965ceb36b039bc42942c1628d29d4c844d3fffcaba9c5a01bddd6bea5
3
+ size 21671
special_tokens_map.json CHANGED
@@ -1,9 +1,5 @@
1
  {
2
  "additional_special_tokens": [
3
- "<s_gt_parse>",
4
- "</s_gt_parse>",
5
- "<s_immatriculation>",
6
- "</s_immatriculation>",
7
  "<s>",
8
  "</s>"
9
  ],
 
1
  {
2
  "additional_special_tokens": [
 
 
 
 
3
  "<s>",
4
  "</s>"
5
  ],
tokenizer.json CHANGED
@@ -1,21 +1,7 @@
1
  {
2
  "version": "1.0",
3
- "truncation": {
4
- "direction": "Right",
5
- "max_length": 512,
6
- "strategy": "LongestFirst",
7
- "stride": 0
8
- },
9
- "padding": {
10
- "strategy": {
11
- "Fixed": 512
12
- },
13
- "direction": "Right",
14
- "pad_to_multiple_of": null,
15
- "pad_id": 1,
16
- "pad_type_id": 0,
17
- "pad_token": "<pad>"
18
- },
19
  "added_tokens": [
20
  {
21
  "id": 0,
@@ -88,42 +74,6 @@
88
  "rstrip": false,
89
  "normalized": false,
90
  "special": true
91
- },
92
- {
93
- "id": 57525,
94
- "content": "<s_gt_parse>",
95
- "single_word": false,
96
- "lstrip": false,
97
- "rstrip": false,
98
- "normalized": false,
99
- "special": true
100
- },
101
- {
102
- "id": 57526,
103
- "content": "</s_gt_parse>",
104
- "single_word": false,
105
- "lstrip": false,
106
- "rstrip": false,
107
- "normalized": false,
108
- "special": true
109
- },
110
- {
111
- "id": 57527,
112
- "content": "<s_immatriculation>",
113
- "single_word": false,
114
- "lstrip": false,
115
- "rstrip": false,
116
- "normalized": false,
117
- "special": true
118
- },
119
- {
120
- "id": 57528,
121
- "content": "</s_immatriculation>",
122
- "single_word": false,
123
- "lstrip": false,
124
- "rstrip": false,
125
- "normalized": false,
126
- "special": true
127
  }
128
  ],
129
  "normalizer": {
 
1
  {
2
  "version": "1.0",
3
+ "truncation": null,
4
+ "padding": null,
 
 
 
 
 
 
 
 
 
 
 
 
 
 
5
  "added_tokens": [
6
  {
7
  "id": 0,
 
74
  "rstrip": false,
75
  "normalized": false,
76
  "special": true
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
77
  }
78
  ],
79
  "normalizer": {