Prisma / config.json
y3i12's picture
prepping safetensor model scripts
a2df0cc
raw
history blame contribute delete
518 Bytes
{
"auto_map": {
"AutoConfig": "configuration_prisma.PrismaConfig",
"AutoModelForCausalLM": "modeling_prisma.PrismaForCausalLM"
},
"aux_skip_k": 1,
"aux_skip_weight": 0.1,
"dropout": 0.0,
"embed_dim": 0,
"head_dim": 0,
"hidden_size": 1024,
"max_seq_len": 1024,
"model_type": "prisma",
"n_middle": 1,
"num_heads": 16,
"num_kv_heads": 4,
"num_layers": 41,
"transformers_version": "4.57.3",
"use_g2lu": true,
"vocab_size": 32000,
"word_rope_base": 10.0,
"word_rope_dims": 8
}