esm2_t33_650M / config.json
prasadsachin's picture
Upload folder using huggingface_hub
0634f0f verified
raw
history blame contribute delete
655 Bytes
{
"module": "keras_hub.src.models.esm.esm_backbone",
"class_name": "ESMBackbone",
"config": {
"name": "esm_backbone",
"trainable": true,
"vocabulary_size": 33,
"num_layers": 33,
"num_heads": 20,
"hidden_dim": 1280,
"intermediate_dim": 5120,
"dropout": 0.0,
"max_wavelength": 10000,
"use_bias": true,
"activation": "gelu",
"layer_norm_eps": 1e-05,
"use_pre_layer_norm": false,
"position_embedding_type": "rotary",
"max_sequence_length": 1026,
"pad_token_id": 1
},
"registered_name": "keras_hub>ESMBackbone"
}