PeptideMLM_sm / config.json
aaronfeller's picture
initial commit
e92e9f7 verified
raw
history blame contribute delete
369 Bytes
{
"model_type": "MLM_model",
"size": "small",
"ffn_hidden_dim": 768,
"embed_dim": 512,
"num_heads": 8,
"num_blocks": 14,
"vocab_size": 405,
"output_dim": 405,
"max_seq_len": 2048,
"auto_map": {
"AutoConfig": "config.model_config",
"AutoModel": "ChemPepMTR.MLM_model"
},
"architectures": ["MLM_model"]
}