Tingdan csy0225 commited on
Commit
e2554be
·
1 Parent(s): 16e32f3

Update config.json (#4)

Browse files

- Update config.json (c15f627b1dbc2a860847b3525d0eb5385f151971)


Co-authored-by: chensiyu <csy0225@users.noreply.huggingface.co>

Files changed (1) hide show
  1. config.json +1 -1
config.json CHANGED
@@ -21,7 +21,7 @@
21
  "max_seq_len": 262144,
22
  "vocab_size": 128896,
23
  "torch_dtype": "bfloat16",
24
- "use_qk_norm": false,
25
  "moe_layers_enum": "3,4,5,6,7,8,9,10,11,12,13,14,15,16,17,18,19,20,21,22,23,24,25,26,27,28,29,30,31,32,33,34,35,36,37,38,39,40,41,42,43,44",
26
  "use_mfa": false,
27
  "num_attention_heads": 64,
 
21
  "max_seq_len": 262144,
22
  "vocab_size": 128896,
23
  "torch_dtype": "bfloat16",
24
+ "use_qk_norm": true,
25
  "moe_layers_enum": "3,4,5,6,7,8,9,10,11,12,13,14,15,16,17,18,19,20,21,22,23,24,25,26,27,28,29,30,31,32,33,34,35,36,37,38,39,40,41,42,43,44",
26
  "use_mfa": false,
27
  "num_attention_heads": 64,