| { |
| |
| "architectures": [ |
| "YingLong" |
| ], |
|
|
| "auto_map": { |
| "AutoConfig": "model_config.YingLongConfig", |
| "AutoModelForCausalLM": "model.GPT" |
| }, |
| "org": "Alibaba", |
| "_mlp_class": "LLaMAMLP", |
| "_norm_class": "FusedRMSNorm", |
| "block_size": 8224, |
| "condense_ratio": 1, |
| "haar_trans": true, |
| "haar_trans_inv": true, |
| "haar_trans_norm": "backward", |
| "intermediate_size": 4096, |
| "n_embd": 1024, |
| "n_head": 32, |
| "n_layer": 18, |
| "n_query_groups": 4, |
| "norm_eps": 1e-05, |
| "parallel_residual": false, |
| "patch_size": 32, |
| "quantitle": true, |
| "rope_base": 10000, |
| "rotary_percentage": 1.0, |
| "shared_attention_norm": false, |
| "unet": true, |
| "vocab_size": 1 |
| } |