Anvaya-Rabbit-2.7B / config.json
tvastr's picture
replace stale pearl_meta config with proper architecture config
f2cc715 verified
raw
history blame contribute delete
343 Bytes
{
"model_type": "rta-ssm",
"architecture": "RabbitCausalLM",
"variant": "fu-64",
"vocab_size": 50280,
"d_model": 2560,
"d_state": 2560,
"n_layers": 64,
"tokenizer_class": "GPTNeoXTokenizerFast",
"tokenizer_id": "EleutherAI/gpt-neox-20b",
"torch_dtype": "bfloat16",
"framework": "RtaForge-SSM",
"schema_version": "1.0"
}