PeptideMLM_base / config.json
aaronfeller's picture
initial commit
eff0d52 verified
raw
history blame contribute delete
370 Bytes
{
"model_type": "MLM_model",
"size": "base",
"ffn_hidden_dim": 1024,
"embed_dim": 768,
"num_heads": 12,
"num_blocks": 24,
"vocab_size": 405,
"output_dim": 405,
"max_seq_len": 2048,
"auto_map": {
"AutoConfig": "config.model_config",
"AutoModel": "ChemPepMTR.MLM_model"
},
"architectures": ["MLM_model"]
}