bombshelll commited on
Commit
4cd1b79
·
verified ·
1 Parent(s): 81d5644

Upload model

Browse files
Files changed (3) hide show
  1. config.json +4 -4
  2. generation_config.json +2 -0
  3. model.safetensors +2 -2
config.json CHANGED
@@ -12,7 +12,7 @@
12
  "attention_probs_dropout_prob": 0.1,
13
  "bad_words_ids": null,
14
  "begin_suppress_tokens": null,
15
- "bos_token_id": null,
16
  "chunk_size_feed_forward": 0,
17
  "classifier_dropout": null,
18
  "cross_attention_hidden_size": null,
@@ -21,7 +21,7 @@
21
  "do_sample": false,
22
  "early_stopping": false,
23
  "encoder_no_repeat_ngram_size": 0,
24
- "eos_token_id": null,
25
  "exponential_decay_length_penalty": null,
26
  "finetuning_task": null,
27
  "forced_bos_token_id": null,
@@ -81,7 +81,7 @@
81
  "typical_p": 1.0,
82
  "use_bfloat16": false,
83
  "use_cache": true,
84
- "vocab_size": 30522
85
  },
86
  "decoder_start_token_id": 2,
87
  "early_stopping": true,
@@ -176,5 +176,5 @@
176
  "tie_word_embeddings": false,
177
  "torch_dtype": "float32",
178
  "transformers_version": "4.44.0",
179
- "vocab_size": 30522
180
  }
 
12
  "attention_probs_dropout_prob": 0.1,
13
  "bad_words_ids": null,
14
  "begin_suppress_tokens": null,
15
+ "bos_token_id": 101,
16
  "chunk_size_feed_forward": 0,
17
  "classifier_dropout": null,
18
  "cross_attention_hidden_size": null,
 
21
  "do_sample": false,
22
  "early_stopping": false,
23
  "encoder_no_repeat_ngram_size": 0,
24
+ "eos_token_id": 102,
25
  "exponential_decay_length_penalty": null,
26
  "finetuning_task": null,
27
  "forced_bos_token_id": null,
 
81
  "typical_p": 1.0,
82
  "use_bfloat16": false,
83
  "use_cache": true,
84
+ "vocab_size": 28895
85
  },
86
  "decoder_start_token_id": 2,
87
  "early_stopping": true,
 
176
  "tie_word_embeddings": false,
177
  "torch_dtype": "float32",
178
  "transformers_version": "4.44.0",
179
+ "vocab_size": 28895
180
  }
generation_config.json CHANGED
@@ -1,5 +1,7 @@
1
  {
2
  "_from_model_config": true,
 
 
3
  "pad_token_id": 0,
4
  "transformers_version": "4.44.0"
5
  }
 
1
  {
2
  "_from_model_config": true,
3
+ "bos_token_id": 101,
4
+ "eos_token_id": 102,
5
  "pad_token_id": 0,
6
  "transformers_version": "4.44.0"
7
  }
model.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:828ebc85b6eae6c090f7f357f58f406af89f6660428e736cf2700f2c930acce1
3
- size 897147000
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:33eec4825b6eb588debec94245b705e5d5c09b3c25aa7f59ef523b098286db25
3
+ size 892142332