nell123 commited on
Commit
45f6f23
·
verified ·
1 Parent(s): 4e3fa99

Update config.json

Browse files
Files changed (1) hide show
  1. config.json +2 -7
config.json CHANGED
@@ -1,13 +1,7 @@
1
  {
2
- "architectures": [
3
- "Phi3ForCausalLM"
4
- ],
5
  "attention_bias": false,
6
  "attention_dropout": 0.0,
7
- "auto_map": {
8
- "AutoConfig": "configuration_phi3.Phi3Config",
9
- "AutoModelForCausalLM": "modeling_phi3.Phi3ForCausalLM"
10
- },
11
  "bos_token_id": 1,
12
  "embd_pdrop": 0.0,
13
  "eos_token_id": 32000,
@@ -34,3 +28,4 @@
34
  "use_cache": true,
35
  "vocab_size": 32064
36
  }
 
 
1
  {
2
+ "architectures": ["Phi3ForCausalLM"],
 
 
3
  "attention_bias": false,
4
  "attention_dropout": 0.0,
 
 
 
 
5
  "bos_token_id": 1,
6
  "embd_pdrop": 0.0,
7
  "eos_token_id": 32000,
 
28
  "use_cache": true,
29
  "vocab_size": 32064
30
  }
31
+