toy-llm / config.json
1y33's picture
Upload HFWrappedLLM
b31feff verified
raw
history blame contribute delete
225 Bytes
{
"architectures": [
"HFWrappedLLM"
],
"hidden_size": 64,
"max_position_embeddings": 128,
"num_heads": 1,
"num_layers": 1,
"torch_dtype": "float32",
"transformers_version": "4.48.1",
"vocab_size": 500
}