HELM-BERT

A language model for peptide representation learning using HELM (Hierarchical Editing Language for Macromolecules) notation.

GitHub

Model Description

HELM-BERT is built upon the DeBERTa architecture, designed for peptide sequences in HELM notation:

  • Disentangled Attention: Decomposes attention into content-content and content-position terms
  • Enhanced Mask Decoder (EMD): Injects absolute position embeddings at the decoder stage
  • Span Masking: Contiguous token masking with geometric distribution
  • nGiE: n-gram Induced Encoding layer (1D convolution, kernel size 3)

Model Specifications

Parameter Value
Parameters 54.8M
Hidden size 768
Layers 6
Attention heads 12
Vocab size 78
Max token length 512

How to Use

from transformers import AutoModel, AutoTokenizer

model = AutoModel.from_pretrained("Flansma/helm-bert", trust_remote_code=True)
tokenizer = AutoTokenizer.from_pretrained("Flansma/helm-bert", trust_remote_code=True)

# Cyclosporine A
inputs = tokenizer("PEPTIDE1{[Abu].[Sar].[meL].V.[meL].A.[dA].[meL].[meL].[meV].[Me_Bmt(E)]}$PEPTIDE1,PEPTIDE1,1:R1-11:R2$$$", return_tensors="pt")
outputs = model(**inputs)
embeddings = outputs.last_hidden_state

Training Data

Pretrained on deduplicated peptide sequences from:

  • ChEMBL: Bioactive molecules database
  • CycPeptMPDB: Cyclic peptide membrane permeability database
  • Propedia: Protein-peptide interaction database

Citation

@article{lee2025helmbert,
  title={HELM-BERT: A Transformer for Medium-sized Peptide Property Prediction},
  author={Seungeon Lee and Takuto Koyama and Itsuki Maeda and Shigeyuki Matsumoto and Yasushi Okuno},
  journal={arXiv preprint arXiv:2512.23175},
  year={2025},
  url={https://arxiv.org/abs/2512.23175}
}

License

MIT License

Downloads last month
150
Safetensors
Model size
54.8M params
Tensor type
F32
·
Inference Providers NEW
This model isn't deployed by any Inference Provider. 🙋 Ask for provider support