nanochat-d20 / config.json
dnakov's picture
Upload folder using huggingface_hub
3a3363e verified
raw
history blame contribute delete
366 Bytes
{
"architectures": [
"NanoChatForCausalLM"
],
"model_type": "nanochat",
"hidden_size": 1280,
"num_hidden_layers": 20,
"num_attention_heads": 10,
"num_key_value_heads": 10,
"vocab_size": 65536,
"max_position_embeddings": 2048,
"intermediate_size": 5120,
"rope_theta": 10000.0,
"torch_dtype": "bfloat16",
"transformers_version": "4.0.0"
}