xue wang commited on
Commit
8941376
·
verified ·
1 Parent(s): c447b04

Upload config.json

Browse files
Files changed (1) hide show
  1. config.json +7 -5
config.json CHANGED
@@ -1,14 +1,16 @@
1
  {
2
- "architectures": ["YingLong"],
 
 
 
3
 
4
  "auto_map": {
5
  "AutoConfig": "model_config.YingLongConfig",
6
  "AutoModelForCausalLM": "model.GPT"
7
- },
8
  "org": "Alibaba",
9
  "_mlp_class": "LLaMAMLP",
10
  "_norm_class": "FusedRMSNorm",
11
- "bias": false,
12
  "block_size": 8224,
13
  "condense_ratio": 1,
14
  "haar_trans": true,
@@ -19,6 +21,7 @@
19
  "n_head": 32,
20
  "n_layer": 18,
21
  "n_query_groups": 4,
 
22
  "parallel_residual": false,
23
  "patch_size": 32,
24
  "quantitle": true,
@@ -26,6 +29,5 @@
26
  "rotary_percentage": 1.0,
27
  "shared_attention_norm": false,
28
  "unet": true,
29
- "vocab_size": 1,
30
- "vq": false,
31
  }
 
1
  {
2
+
3
+ "architectures": [
4
+ "YingLong"
5
+ ],
6
 
7
  "auto_map": {
8
  "AutoConfig": "model_config.YingLongConfig",
9
  "AutoModelForCausalLM": "model.GPT"
10
+ },
11
  "org": "Alibaba",
12
  "_mlp_class": "LLaMAMLP",
13
  "_norm_class": "FusedRMSNorm",
 
14
  "block_size": 8224,
15
  "condense_ratio": 1,
16
  "haar_trans": true,
 
21
  "n_head": 32,
22
  "n_layer": 18,
23
  "n_query_groups": 4,
24
+ "norm_eps": 1e-05,
25
  "parallel_residual": false,
26
  "patch_size": 32,
27
  "quantitle": true,
 
29
  "rotary_percentage": 1.0,
30
  "shared_attention_norm": false,
31
  "unet": true,
32
+ "vocab_size": 1
 
33
  }