Edgar404 commited on
Commit
a60bc5a
·
verified ·
1 Parent(s): ed46cca

Training done

Browse files
preprocessor_config.json CHANGED
@@ -37,8 +37,8 @@
37
  "processor_class": "DonutProcessor",
38
  "resample": 2,
39
  "rescale_factor": 0.00392156862745098,
40
- "size": {
41
- "height": 720,
42
- "width": 960
43
- }
44
  }
 
37
  "processor_class": "DonutProcessor",
38
  "resample": 2,
39
  "rescale_factor": 0.00392156862745098,
40
+ "size": [
41
+ 1280,
42
+ 720
43
+ ]
44
  }
tokenizer.json CHANGED
The diff for this file is too large to render. See raw diff
 
tokenizer_config.json CHANGED
@@ -98,18 +98,11 @@
98
  "cls_token": "<s>",
99
  "eos_token": "</s>",
100
  "mask_token": "<mask>",
101
- "max_length": 512,
102
  "model_max_length": 1000000000000000019884624838656,
103
- "pad_to_multiple_of": null,
104
  "pad_token": "<pad>",
105
- "pad_token_type_id": 0,
106
- "padding_side": "right",
107
  "processor_class": "DonutProcessor",
108
  "sep_token": "</s>",
109
  "sp_model_kwargs": {},
110
- "stride": 0,
111
  "tokenizer_class": "XLMRobertaTokenizer",
112
- "truncation_side": "right",
113
- "truncation_strategy": "longest_first",
114
  "unk_token": "<unk>"
115
  }
 
98
  "cls_token": "<s>",
99
  "eos_token": "</s>",
100
  "mask_token": "<mask>",
 
101
  "model_max_length": 1000000000000000019884624838656,
 
102
  "pad_token": "<pad>",
 
 
103
  "processor_class": "DonutProcessor",
104
  "sep_token": "</s>",
105
  "sp_model_kwargs": {},
 
106
  "tokenizer_class": "XLMRobertaTokenizer",
 
 
107
  "unk_token": "<unk>"
108
  }