{ "architectures": [ "FlashSTU" ], "auto_map": { "AutoConfig": "config.FlashSTUConfig", "AutoModel": "model.FlashSTU" }, "bias": false, "bsz": 8, "dropout": 0.0, "mlp_scale": 4, "model_type": "FlashSTU", "n_embd": 768, "n_heads": 12, "n_layers": 12, "num_eigh": 16, "seq_len": 4096, "softcap": 50.0, "torch_dtype": "float32", "transformers_version": "4.44.0", "use_approx": true, "use_flash_fft": true, "use_hankel_L": false, "vocab_size": 200064, "window_size": 64 }