Jlonge4 commited on
Commit
dcc9c8d
·
verified ·
1 Parent(s): cc10110

Upload Phi3ForCausalLM

Browse files
Files changed (2) hide show
  1. config.json +3 -2
  2. generation_config.json +3 -5
config.json CHANGED
@@ -12,7 +12,7 @@
12
  "bos_token_id": 199999,
13
  "dtype": "bfloat16",
14
  "embd_pdrop": 0.0,
15
- "eos_token_id": 199999,
16
  "full_attn_mod": 1,
17
  "hidden_act": "silu",
18
  "hidden_size": 3072,
@@ -27,7 +27,7 @@
27
  "num_hidden_layers": 32,
28
  "num_key_value_heads": 8,
29
  "original_max_position_embeddings": 4096,
30
- "pad_token_id": 199999,
31
  "partial_rotary_factor": 0.75,
32
  "resid_pdrop": 0.0,
33
  "rms_norm_eps": 1e-05,
@@ -138,6 +138,7 @@
138
  "sliding_window": 262144,
139
  "tie_word_embeddings": true,
140
  "transformers_version": "4.56.0",
 
141
  "use_cache": true,
142
  "vocab_size": 200064
143
  }
 
12
  "bos_token_id": 199999,
13
  "dtype": "bfloat16",
14
  "embd_pdrop": 0.0,
15
+ "eos_token_id": 200020,
16
  "full_attn_mod": 1,
17
  "hidden_act": "silu",
18
  "hidden_size": 3072,
 
27
  "num_hidden_layers": 32,
28
  "num_key_value_heads": 8,
29
  "original_max_position_embeddings": 4096,
30
+ "pad_token_id": 200029,
31
  "partial_rotary_factor": 0.75,
32
  "resid_pdrop": 0.0,
33
  "rms_norm_eps": 1e-05,
 
138
  "sliding_window": 262144,
139
  "tie_word_embeddings": true,
140
  "transformers_version": "4.56.0",
141
+ "unsloth_fixed": true,
142
  "use_cache": true,
143
  "vocab_size": 200064
144
  }
generation_config.json CHANGED
@@ -1,10 +1,8 @@
1
  {
2
  "_from_model_config": true,
3
  "bos_token_id": 199999,
4
- "eos_token_id": [
5
- 200020,
6
- 199999
7
- ],
8
- "pad_token_id": 199999,
9
  "transformers_version": "4.56.0"
10
  }
 
1
  {
2
  "_from_model_config": true,
3
  "bos_token_id": 199999,
4
+ "eos_token_id": 200020,
5
+ "max_length": 131072,
6
+ "pad_token_id": 200029,
 
 
7
  "transformers_version": "4.56.0"
8
  }