File size: 277 Bytes
8335a6f |
1 2 3 4 5 6 7 8 9 10 11 12 13 14 |
{
"architectures": [
"ASCHForCausalLM"
],
"model_type": "asch",
"torch_dtype": "float16",
"transformers_version": "4.40.0",
"vocab_size": 151936,
"hidden_size": 2048,
"num_hidden_layers": 24,
"num_attention_heads": 16,
"max_position_embeddings": 8192
}
|