| """ Prot2Text configuration""" |
|
|
| from transformers.configuration_utils import PretrainedConfig |
| from transformers import AutoConfig |
| from transformers.utils import logging |
|
|
|
|
| logger = logging.get_logger(__name__) |
|
|
|
|
| class Prot2TextConfig(PretrainedConfig): |
| model_type = "prot2text" |
| keys_to_ignore_at_inference = ["past_key_values"] |
| _keys_to_ignore_on_load_missing = [r"transformer"] |
|
|
| def __init__( |
| self, |
| cross_esm_graph=True, |
| decoder_start_token_id=50257, |
| early_stopping=True, |
| eos_token_id=50258, |
| bos_token_id=50257, |
| esm=True, |
| esm_model_name="facebook/esm2_t6_8M_UR50D", |
| gpt_model_name="gpt2", |
| length_penalty=2.0, |
| max_new_tokens=256, |
| no_repeat_ngram_size=3, |
| pad_token_id=50256, |
| prot2text_version="1.1", |
| rgcn=True, |
| rgc_input_dim=67, |
| rgcn_n_layers=6, |
| gpt_config=None, |
| esm_config=None, |
| **kwargs, |
| ): |
| self.cross_esm_graph = cross_esm_graph |
| self.decoder_start_token_id = decoder_start_token_id |
| self.early_stopping = early_stopping |
| self.eos_token_id = eos_token_id |
| self.esm = esm |
| self.esm_model_name = esm_model_name |
| self.gpt_model_name = gpt_model_name |
| self.length_penalty = length_penalty |
| self.max_new_tokens = max_new_tokens |
| self.no_repeat_ngram_size = no_repeat_ngram_size |
| self.pad_token_id = pad_token_id |
| self.prot2text_version = prot2text_version |
| self.rgcn = rgcn |
| self.rgc_input_dim = rgc_input_dim |
| self.rgcn_n_layers = rgcn_n_layers |
| if gpt_config is None: |
| self.gpt_config = AutoConfig.from_pretrained(gpt_model_name, |
| _name_or_path= gpt_model_name, |
| is_encoder_decoder=True, |
| use_cache=False, |
| add_cross_attention=True, |
| bos_token_id=bos_token_id, |
| decoder_start_token_id=decoder_start_token_id, |
| eos_token_id=eos_token_id, |
| max_new_tokens=max_new_tokens, |
| pad_token_id=50256, |
| vocab_size=50259, |
| num_beams=1, |
| max_length=256, |
| min_length=1).to_dict() |
| else: |
| self.gpt_config = gpt_config |
| if esm_config is None: |
| self.esm_config = AutoConfig.from_pretrained(esm_model_name).to_dict() |
| self.esm_config = esm_config |
|
|
| super().__init__(bos_token_id=bos_token_id, eos_token_id=eos_token_id, **kwargs) |