sluar / config.json
noandrews's picture
Initial commit of LUAR model
34148c4 verified
raw
history blame contribute delete
274 Bytes
{
"architectures": [
"LUAR"
],
"embedding_dim": 512,
"epsilon": 1e-09,
"gradient_checkpointing": false,
"model_name_or_path": "sentence-transformers/all-MiniLM-L6-v2",
"model_type": "luar",
"torch_dtype": "bfloat16",
"transformers_version": "4.52.3"
}