Randomize commited on
Commit
45eebdc
·
verified ·
1 Parent(s): e9b865b

Update config.json

Browse files
Files changed (1) hide show
  1. config.json +10 -6
config.json CHANGED
@@ -1,8 +1,12 @@
1
  {
2
  "architectures": [
3
- "Step4ForCausalLM"
4
  ],
5
- "model_type": "step4",
 
 
 
 
6
  "hidden_size": 4096,
7
  "intermediate_size": 11264,
8
  "num_hidden_layers": 45,
@@ -10,9 +14,8 @@
10
  "max_position_embedding": 131072,
11
  "vocab_size": 128896,
12
  "torch_dtype": "bfloat16",
13
- "use_qk_norm": false,
14
  "moe_layers_enum": "3,4,5,6,7,8,9,10,11,12,13,14,15,16,17,18,19,20,21,22,23,24,25,26,27,28,29,30,31,32,33,34,35,36,37,38,39,40,41,42,43,44",
15
- "use_mfa": false,
16
  "num_attention_heads": 64,
17
  "num_attention_groups": 8,
18
  "num_nextn_predict_layers": 1,
@@ -285,5 +288,6 @@
285
  0.0,
286
  16,
287
  0.0
288
- ]
289
- }
 
 
1
  {
2
  "architectures": [
3
+ "Step3p5ForCausalLM"
4
  ],
5
+ "model_type": "step3p5",
6
+ "auto_map": {
7
+ "AutoConfig": "configuration_step3p5.Step3p5Config",
8
+ "AutoModelForCausalLM": "modeling_step3p5.Step3p5ForCausalLM"
9
+ },
10
  "hidden_size": 4096,
11
  "intermediate_size": 11264,
12
  "num_hidden_layers": 45,
 
14
  "max_position_embedding": 131072,
15
  "vocab_size": 128896,
16
  "torch_dtype": "bfloat16",
17
+ "use_qk_norm": true,
18
  "moe_layers_enum": "3,4,5,6,7,8,9,10,11,12,13,14,15,16,17,18,19,20,21,22,23,24,25,26,27,28,29,30,31,32,33,34,35,36,37,38,39,40,41,42,43,44",
 
19
  "num_attention_heads": 64,
20
  "num_attention_groups": 8,
21
  "num_nextn_predict_layers": 1,
 
288
  0.0,
289
  16,
290
  0.0
291
+ ],
292
+ "zero_centered": true
293
+ }