Safetensors
step3p5
custom_code

Update config.json

#4
by csy0225 - opened
Files changed (1) hide show
  1. config.json +1 -1
config.json CHANGED
@@ -21,7 +21,7 @@
21
  "max_seq_len": 262144,
22
  "vocab_size": 128896,
23
  "torch_dtype": "bfloat16",
24
- "use_qk_norm": false,
25
  "moe_layers_enum": "3,4,5,6,7,8,9,10,11,12,13,14,15,16,17,18,19,20,21,22,23,24,25,26,27,28,29,30,31,32,33,34,35,36,37,38,39,40,41,42,43,44",
26
  "use_mfa": false,
27
  "num_attention_heads": 64,
 
21
  "max_seq_len": 262144,
22
  "vocab_size": 128896,
23
  "torch_dtype": "bfloat16",
24
+ "use_qk_norm": true,
25
  "moe_layers_enum": "3,4,5,6,7,8,9,10,11,12,13,14,15,16,17,18,19,20,21,22,23,24,25,26,27,28,29,30,31,32,33,34,35,36,37,38,39,40,41,42,43,44",
26
  "use_mfa": false,
27
  "num_attention_heads": 64,