tiny-gpt / config.json
Marmik's picture
Upload folder using huggingface_hub
97a6b05 verified
raw
history blame contribute delete
522 Bytes
{
"architectures": [
"TinyGPTForCausalLM"
],
"auto_map": {
"AutoConfig": "modeling_tiny_gpt.TinyGPTConfig",
"AutoModelForCausalLM": "modeling_tiny_gpt.TinyGPTForCausalLM"
},
"attn_dropout": 0.0,
"attn_eps": 1e-06,
"d_head": 64,
"d_model": 768,
"device": "cpu",
"ffn_eps": 1e-06,
"max_seq_len": 1024,
"model_type": "tiny_gpt",
"n_heads": 12,
"n_layers": 5,
"norm_eps": 1e-06,
"top_k": null,
"torch_dtype": "float32",
"transformers_version": "4.53.2",
"vocab_size": 50257
}