Upload 6 files
Browse files- README.md +61 -3
- ckpt.pt +3 -0
- model_config.json +9 -0
- tokenizer_config.json +4 -0
- trainer_config.json +23 -0
- vocab.txt +27 -0
README.md
CHANGED
|
@@ -1,3 +1,61 @@
|
|
| 1 |
-
---
|
| 2 |
-
|
| 3 |
-
-
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
---
|
| 2 |
+
tags:
|
| 3 |
+
- chemistry
|
| 4 |
+
- molecular-design
|
| 5 |
+
- transformer
|
| 6 |
+
- generative-model
|
| 7 |
+
- predictive-model
|
| 8 |
+
license: bsd-3-clause
|
| 9 |
+
datasets:
|
| 10 |
+
- GuacaMol
|
| 11 |
+
- ZINC
|
| 12 |
+
- MoleculeNet
|
| 13 |
+
gated: true
|
| 14 |
+
extra_gated_fields:
|
| 15 |
+
Organization: text
|
| 16 |
+
Intended use: text
|
| 17 |
+
Contact person: text
|
| 18 |
+
E-mail: text
|
| 19 |
+
Country: country
|
| 20 |
+
Date: date_picker
|
| 21 |
+
I agree to use this model only for purposes that are non-malicious and ethically responsible: checkbox
|
| 22 |
+
I have read and accept the BSD 3-Clause license: checkbox
|
| 23 |
+
---
|
| 24 |
+
|
| 25 |
+
# Hyformer
|
| 26 |
+
|
| 27 |
+
Hyformer is a joint transformer-based model that unifies a generative decoder with a predictive encoder. Depending on the task, Hyformer uses either a causal or a bidirectional mask, outputting token probabilities or predicted property values.
|
| 28 |
+
|
| 29 |
+
## Model Details
|
| 30 |
+
- **Paper:** [Synergistic Benefits of Joint Molecule Generation and Property Prediction](https://arxiv.org/abs/2504.16559)
|
| 31 |
+
- **Authors:** Adam Izdebski, Jan Olszewski, Pankhil Gawade, Krzysztof Koras, Serra Korkmaz, Valentin Rauscher, Jakub M. Tomczak, Ewa Szczurek
|
| 32 |
+
- **License:** BSD 3-Clause
|
| 33 |
+
- **Repository:** [https://github.com/szczurek-lab/hyformer](https://github.com/szczurek-lab/hyformer)
|
| 34 |
+
|
| 35 |
+
## Model checkpoints
|
| 36 |
+
- **[Hyformer_molecules_8M](https://huggingface.co/SzczurekLab/hyformer_molecules_8M):** Trained on GuacaMol dataset ([Brown et al., 2019](https://jcheminf.biomedcentral.com/articles/10.1186/s13321-019-0351-9))
|
| 37 |
+
- **[Hyformer_molecules_50M](https://huggingface.co/SzczurekLab/hyformer_molecules_50M):** Trained on 19M molecules from ZINC, ChEMBL, and other purchasable molecular datasets ([Zhou et al., 2023](https://openreview.net/forum?id=1pPpKc9wR0Y))
|
| 38 |
+
- **[Hyformer_peptides_34M](https://huggingface.co/SzczurekLab/hyformer_peptides_34M):** Trained on 3.5M general-purpose and antimicrobial peptides
|
| 39 |
+
- **[Hyformer_peptides_34M_MIC](https://huggingface.co/SzczurekLab/hyformer_peptides_34M_MIC):** `Hyformer_peptides_34M` jointly fine-tuned on minimal inhibitory concentration values (MIC) against E. coli bacteria
|
| 40 |
+
|
| 41 |
+
## Gated Access
|
| 42 |
+
This model is available with **gated access**. To request access, please use the Hugging Face gated request form.
|
| 43 |
+
|
| 44 |
+
## Citation
|
| 45 |
+
If you use this model, please cite:
|
| 46 |
+
|
| 47 |
+
```
|
| 48 |
+
@misc{izdebski2025synergisticbenefitsjointmolecule,
|
| 49 |
+
title={Synergistic Benefits of Joint Molecule Generation and Property Prediction},
|
| 50 |
+
author={Adam Izdebski and Jan Olszewski and Pankhil Gawade and Krzysztof Koras and Serra Korkmaz and Valentin Rauscher and Jakub M. Tomczak and Ewa Szczurek},
|
| 51 |
+
year={2025},
|
| 52 |
+
eprint={2504.16559},
|
| 53 |
+
archivePrefix={arXiv},
|
| 54 |
+
primaryClass={cs.LG},
|
| 55 |
+
url={https://arxiv.org/abs/2504.16559},
|
| 56 |
+
}
|
| 57 |
+
```
|
| 58 |
+
|
| 59 |
+
## References
|
| 60 |
+
- Brown, Nathan, et al. "GuacaMol: benchmarking models for de novo molecular design." Journal of chemical information and modeling, 2019.
|
| 61 |
+
- Zhou, Gengmo, et al. "Uni-mol: A universal 3d molecular representation learning framework." ICLR, 2023.
|
ckpt.pt
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:dd2084a900ef19cc01f8b18dbfc5be564993363ae50d8dcbbec47f3d094f5045
|
| 3 |
+
size 204798416
|
model_config.json
ADDED
|
@@ -0,0 +1,9 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
{
|
| 2 |
+
"model_type": "Hyformer",
|
| 3 |
+
"embedding_dim": 512,
|
| 4 |
+
"num_attention_heads": 8,
|
| 5 |
+
"num_transformer_layers": 8,
|
| 6 |
+
"vocab_size": 34,
|
| 7 |
+
"prediction_task_type": "regression",
|
| 8 |
+
"num_prediction_tasks": 1
|
| 9 |
+
}
|
tokenizer_config.json
ADDED
|
@@ -0,0 +1,4 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
{
|
| 2 |
+
"tokenizer_type": "AATokenizer",
|
| 3 |
+
"vocabulary_path": "data/vocabulary/aa.txt"
|
| 4 |
+
}
|
trainer_config.json
ADDED
|
@@ -0,0 +1,23 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
{
|
| 2 |
+
"batch_size": 64,
|
| 3 |
+
"learning_rate": 0.0001,
|
| 4 |
+
"weight_decay": 0.01,
|
| 5 |
+
"max_epochs": 60,
|
| 6 |
+
"tasks": {
|
| 7 |
+
"prediction": 0.6,
|
| 8 |
+
"lm": 0.4
|
| 9 |
+
},
|
| 10 |
+
"compile": true,
|
| 11 |
+
"enable_ddp": false,
|
| 12 |
+
"dtype": "float32",
|
| 13 |
+
"num_workers": 16,
|
| 14 |
+
"beta1": 0.9,
|
| 15 |
+
"beta2": 0.95,
|
| 16 |
+
"gradient_accumulation_steps": 1,
|
| 17 |
+
"grad_clip": 1.0,
|
| 18 |
+
"decay_lr": true,
|
| 19 |
+
"log_interval": 10,
|
| 20 |
+
"save_interval": 5,
|
| 21 |
+
"min_lr": 1e-06,
|
| 22 |
+
"warmup_iters": 54
|
| 23 |
+
}
|
vocab.txt
ADDED
|
@@ -0,0 +1,27 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
T
|
| 2 |
+
I
|
| 3 |
+
F
|
| 4 |
+
L
|
| 5 |
+
V
|
| 6 |
+
E
|
| 7 |
+
G
|
| 8 |
+
A
|
| 9 |
+
S
|
| 10 |
+
Y
|
| 11 |
+
Q
|
| 12 |
+
C
|
| 13 |
+
W
|
| 14 |
+
H
|
| 15 |
+
R
|
| 16 |
+
K
|
| 17 |
+
M
|
| 18 |
+
D
|
| 19 |
+
N
|
| 20 |
+
P
|
| 21 |
+
B
|
| 22 |
+
U
|
| 23 |
+
Z
|
| 24 |
+
X
|
| 25 |
+
O
|
| 26 |
+
-
|
| 27 |
+
.
|