BALM Paper
Collection
Models from the publication: "Improving antibody language models with native pairing", Patterns (2024)
•
4 items
•
Updated
BALM-shuffled is an antibody language model that uses a RoBERTa architecture and was pre-trained on randomly shuffled paired antibody sequences from Jaffe et al. This was a control model used to evaluate the benefits of natively paired sequences in our paper published in Patterns. Therefore, this model should not be used for real use cases; use BALM-paired instead.
Load the model and tokenizer as follows:
from transformers import RobertaTokenizer, RobertaForMaskedLM
model = RobertaForMaskedLM.from_pretrained("brineylab/BALM-shuffled")
tokenizer = RobertaTokenizer.from_pretrained("brineylab/BALM-shuffled")
The tokenizer expects sequences formatted as: HEAVY_CHAIN</s>LIGHT_CHAIN.