Initialising processor

#7
by fangrayson - opened
Files changed (3) hide show
  1. config.json +1 -3
  2. generation_config.json +0 -1
  3. pytorch_model.bin +2 -2
config.json CHANGED
@@ -87,9 +87,8 @@
87
  "typical_p": 1.0,
88
  "use_bfloat16": false,
89
  "use_cache": true,
90
- "vocab_size": 57571
91
  },
92
- "decoder_start_token_id": 3,
93
  "encoder": {
94
  "_name_or_path": "",
95
  "add_cross_attention": false,
@@ -188,7 +187,6 @@
188
  },
189
  "is_encoder_decoder": true,
190
  "model_type": "vision-encoder-decoder",
191
- "pad_token_id": 1,
192
  "tie_word_embeddings": false,
193
  "torch_dtype": "float32",
194
  "transformers_version": null
 
87
  "typical_p": 1.0,
88
  "use_bfloat16": false,
89
  "use_cache": true,
90
+ "vocab_size": 57525
91
  },
 
92
  "encoder": {
93
  "_name_or_path": "",
94
  "add_cross_attention": false,
 
187
  },
188
  "is_encoder_decoder": true,
189
  "model_type": "vision-encoder-decoder",
 
190
  "tie_word_embeddings": false,
191
  "torch_dtype": "float32",
192
  "transformers_version": null
generation_config.json CHANGED
@@ -1,7 +1,6 @@
1
  {
2
  "_from_model_config": true,
3
  "bos_token_id": 0,
4
- "decoder_start_token_id": 3,
5
  "eos_token_id": 2,
6
  "forced_eos_token_id": 2,
7
  "max_length": 768,
 
1
  {
2
  "_from_model_config": true,
3
  "bos_token_id": 0,
 
4
  "eos_token_id": 2,
5
  "forced_eos_token_id": 2,
6
  "max_length": 768,
pytorch_model.bin CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:4eca82025c9712851c93e4339c060b46e44e9b0ad7de1cae79891b2ba41c7976
3
- size 809367321
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:1d266660496e2eaa9b7282306038140ff8cfa7622d0616b10675f5e1a42b51a6
3
+ size 809172185