i3-22m / config.json
FlameF0X's picture
Upload 6 files
7e7ec39 verified
raw
history blame contribute delete
322 Bytes
{
"architectures": [
"i3Model"
],
"model_type": "i3",
"vocab_size": 4466,
"d_model": 512,
"n_layers": 24,
"n_heads": 16,
"max_seq_len": 256,
"rank": 128,
"d_state": 64,
"tokenizer_type": "chunk",
"chunk_strategy": "variable_2_3",
"torch_dtype": "float32",
"transformers_version": "4.36.0"
}