{ "auto_map": { "AutoConfig": "configuration_prisma.PrismaConfig", "AutoModelForCausalLM": "modeling_prisma.PrismaForCausalLM" }, "aux_skip_k": 1, "aux_skip_weight": 0.1, "dropout": 0.0, "embed_dim": 0, "head_dim": 0, "hidden_size": 1024, "max_seq_len": 1024, "model_type": "prisma", "n_middle": 1, "num_heads": 16, "num_kv_heads": 4, "num_layers": 41, "transformers_version": "4.57.3", "use_g2lu": true, "vocab_size": 32000, "word_rope_base": 10.0, "word_rope_dims": 8 }