{ "num_heads": 12, "attention_dim": 768, "vocab_size": 50260, "num_blocks": 12, "ff_dim": 2304, "dropout_rate": 0.1, "possible_opt_path": "", "max_len": 8192, "emb_splt": 256, "attn_chunks": 8, "use_fash_attention": false, "emb_init_range": 0.02, "use_rope": true, "emb_scaling_factor": 1, "res_scale": 1 }