devops-engineer-slm-1m / config.json
sathishphdai's picture
Upload folder using huggingface_hub
d62dca3 verified
raw
history blame contribute delete
372 Bytes
{
"architectures": [
"RoleSLM"
],
"model_type": "devops_engineer-slm",
"domain": "DevOps Engineer",
"vocab_size": 2107,
"n_layer": 32,
"n_head": 20,
"n_embd": 1600,
"block_size": 512,
"dropout": 0.05,
"bias": false,
"ffn_multiplier": 2.667,
"max_position_embeddings": 100000000000,
"rope_theta": 50000000000.0,
"n_parameters": 989792000
}