tlm-tasks-sft-100m / config.json
wzebrowski's picture
Upload config.json with huggingface_hub
0146d66 verified
{
"model_type": "tinygplm",
"vocab_size": 30000,
"d_model": 768,
"n_heads": 12,
"n_layers": 12,
"max_seq": 3072,
"mlp_mult": 4,
"rope_base": 10000.0,
"norm": "rmsnorm",
"activation": "silu",
"weight_tying": true,
"pad_id": 0,
"bos_id": 1,
"eos_id": 2,
"notes": "Config for tinygplm-100m (TinyGPLM ~108M variant)."
}