File size: 320 Bytes
aaa1dd7 | 1 2 3 4 5 6 7 8 9 10 11 12 13 14 15 | {
"architectures": [
"MultiLUARs"
],
"embedding_size": 512,
"hidden_size": 512,
"k_bucket_size": 1024,
"model_type": "MultiLUARs",
"q_bucket_size": 512,
"sentence_transformer_support": false,
"torch_dtype": "float32",
"transformers_version": "4.46.3",
"use_memory_efficient_attention": false
}
|