flash-stu / config.json
windsornguyen's picture
Upload FlashSTU
93d11ff verified
raw
history blame contribute delete
524 Bytes
{
"architectures": [
"FlashSTU"
],
"auto_map": {
"AutoConfig": "config.FlashSTUConfig",
"AutoModel": "model.FlashSTU"
},
"bias": false,
"bsz": 8,
"dropout": 0.0,
"mlp_scale": 4,
"model_type": "FlashSTU",
"n_embd": 768,
"n_heads": 12,
"n_layers": 12,
"num_eigh": 16,
"seq_len": 4096,
"softcap": 50.0,
"torch_dtype": "float32",
"transformers_version": "4.44.0",
"use_approx": true,
"use_flash_fft": true,
"use_hankel_L": false,
"vocab_size": 200064,
"window_size": 64
}