anarenteriare commited on
Commit
34f8923
·
1 Parent(s): 8d252db

End of training

Browse files
README.md CHANGED
@@ -12,7 +12,7 @@ should probably proofread and complete it, then remove this comment. -->
12
 
13
  # donut-trained-example-2
14
 
15
- This model is a fine-tuned version of [anarenteriare/donut-trained-example-2](https://huggingface.co/anarenteriare/donut-trained-example-2) on the None dataset.
16
 
17
  ## Model description
18
 
 
12
 
13
  # donut-trained-example-2
14
 
15
+ This model is a fine-tuned version of [naver-clova-ix/donut-base](https://huggingface.co/naver-clova-ix/donut-base) on the None dataset.
16
 
17
  ## Model description
18
 
added_tokens.json CHANGED
@@ -1,13 +1,5 @@
1
  {
2
- "</s_address>": 57532,
3
- "</s_company>": 57530,
4
- "</s_date>": 57528,
5
- "</s_total>": 57526,
6
- "<s_address>": 57531,
7
- "<s_company>": 57529,
8
- "<s_date>": 57527,
9
  "<s_iitcdip>": 57523,
10
  "<s_synthdog>": 57524,
11
- "<s_total>": 57525,
12
  "<sep/>": 57522
13
  }
 
1
  {
 
 
 
 
 
 
 
2
  "<s_iitcdip>": 57523,
3
  "<s_synthdog>": 57524,
 
4
  "<sep/>": 57522
5
  }
generation_config.json CHANGED
@@ -1,10 +1,8 @@
1
  {
2
  "_from_model_config": true,
3
  "bos_token_id": 0,
4
- "decoder_start_token_id": 0,
5
  "eos_token_id": 2,
6
  "forced_eos_token_id": 2,
7
- "max_length": 512,
8
  "pad_token_id": 1,
9
  "transformers_version": "4.28.0"
10
  }
 
1
  {
2
  "_from_model_config": true,
3
  "bos_token_id": 0,
 
4
  "eos_token_id": 2,
5
  "forced_eos_token_id": 2,
 
6
  "pad_token_id": 1,
7
  "transformers_version": "4.28.0"
8
  }
preprocessor_config.json CHANGED
@@ -19,8 +19,8 @@
19
  "processor_class": "DonutProcessor",
20
  "resample": 2,
21
  "rescale_factor": 0.00392156862745098,
22
- "size": {
23
- "height": 960,
24
- "width": 720
25
- }
26
  }
 
19
  "processor_class": "DonutProcessor",
20
  "resample": 2,
21
  "rescale_factor": 0.00392156862745098,
22
+ "size": [
23
+ 720,
24
+ 960
25
+ ]
26
  }
runs/May17_17-27-51_8c876416781e/events.out.tfevents.1684344480.8c876416781e.639.4 CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:a9772b53cd6ddeb6843a91a56f6628b2424e34bfc4583c4fc181fcfbef85ae0b
3
- size 8561
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:0c33214c286c6ba0265d358b33c94f12e666ed7d55f7aec627214ac395c787af
3
+ size 8909
special_tokens_map.json CHANGED
@@ -1,13 +1,5 @@
1
  {
2
  "additional_special_tokens": [
3
- "<s_total>",
4
- "</s_total>",
5
- "<s_date>",
6
- "</s_date>",
7
- "<s_company>",
8
- "</s_company>",
9
- "<s_address>",
10
- "</s_address>",
11
  "<s>",
12
  "</s>"
13
  ],
 
1
  {
2
  "additional_special_tokens": [
 
 
 
 
 
 
 
 
3
  "<s>",
4
  "</s>"
5
  ],
tokenizer.json CHANGED
The diff for this file is too large to render. See raw diff