avankumar commited on
Commit
0446b1d
·
verified ·
1 Parent(s): 909c55f

Update config.json

Browse files
Files changed (1) hide show
  1. config.json +3 -7
config.json CHANGED
@@ -1,13 +1,9 @@
1
  {
2
  "architectures": [
3
- "Phi3ForCausalLM"
4
  ],
5
  "attention_bias": false,
6
  "attention_dropout": 0.0,
7
- "auto_map": {
8
- "AutoConfig": "configuration_phi3.Phi3Config",
9
- "AutoModelForCausalLM": "modeling_phi3.Phi3ForCausalLM"
10
- },
11
  "bos_token_id": 1,
12
  "embd_pdrop": 0.0,
13
  "eos_token_id": 32000,
@@ -16,7 +12,7 @@
16
  "initializer_range": 0.02,
17
  "intermediate_size": 8192,
18
  "max_position_embeddings": 4096,
19
- "model_type": "phi3",
20
  "num_attention_heads": 32,
21
  "num_hidden_layers": 32,
22
  "num_key_value_heads": 32,
@@ -33,4 +29,4 @@
33
  "transformers_version": "4.53.2",
34
  "use_cache": true,
35
  "vocab_size": 32064
36
- }
 
1
  {
2
  "architectures": [
3
+ "AutoModelForCausalLM"
4
  ],
5
  "attention_bias": false,
6
  "attention_dropout": 0.0,
 
 
 
 
7
  "bos_token_id": 1,
8
  "embd_pdrop": 0.0,
9
  "eos_token_id": 32000,
 
12
  "initializer_range": 0.02,
13
  "intermediate_size": 8192,
14
  "max_position_embeddings": 4096,
15
+ "model_type": "phi", // changed from "phi3" to "phi" for compatibility
16
  "num_attention_heads": 32,
17
  "num_hidden_layers": 32,
18
  "num_key_value_heads": 32,
 
29
  "transformers_version": "4.53.2",
30
  "use_cache": true,
31
  "vocab_size": 32064
32
+ }