pavm595 commited on
Commit
c151b07
·
verified ·
1 Parent(s): f28c75d

Upload 6 files

Browse files
Files changed (6) hide show
  1. README.md +61 -3
  2. ckpt.pt +3 -0
  3. model_config.json +9 -0
  4. tokenizer_config.json +4 -0
  5. trainer_config.json +23 -0
  6. vocab.txt +27 -0
README.md CHANGED
@@ -1,3 +1,61 @@
1
- ---
2
- license: bsd-3-clause
3
- ---
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ ---
2
+ tags:
3
+ - chemistry
4
+ - molecular-design
5
+ - transformer
6
+ - generative-model
7
+ - predictive-model
8
+ license: bsd-3-clause
9
+ datasets:
10
+ - GuacaMol
11
+ - ZINC
12
+ - MoleculeNet
13
+ gated: true
14
+ extra_gated_fields:
15
+ Organization: text
16
+ Intended use: text
17
+ Contact person: text
18
+ E-mail: text
19
+ Country: country
20
+ Date: date_picker
21
+ I agree to use this model only for purposes that are non-malicious and ethically responsible: checkbox
22
+ I have read and accept the BSD 3-Clause license: checkbox
23
+ ---
24
+
25
+ # Hyformer
26
+
27
+ Hyformer is a joint transformer-based model that unifies a generative decoder with a predictive encoder. Depending on the task, Hyformer uses either a causal or a bidirectional mask, outputting token probabilities or predicted property values.
28
+
29
+ ## Model Details
30
+ - **Paper:** [Synergistic Benefits of Joint Molecule Generation and Property Prediction](https://arxiv.org/abs/2504.16559)
31
+ - **Authors:** Adam Izdebski, Jan Olszewski, Pankhil Gawade, Krzysztof Koras, Serra Korkmaz, Valentin Rauscher, Jakub M. Tomczak, Ewa Szczurek
32
+ - **License:** BSD 3-Clause
33
+ - **Repository:** [https://github.com/szczurek-lab/hyformer](https://github.com/szczurek-lab/hyformer)
34
+
35
+ ## Model checkpoints
36
+ - **[Hyformer_molecules_8M](https://huggingface.co/SzczurekLab/hyformer_molecules_8M):** Trained on GuacaMol dataset ([Brown et al., 2019](https://jcheminf.biomedcentral.com/articles/10.1186/s13321-019-0351-9))
37
+ - **[Hyformer_molecules_50M](https://huggingface.co/SzczurekLab/hyformer_molecules_50M):** Trained on 19M molecules from ZINC, ChEMBL, and other purchasable molecular datasets ([Zhou et al., 2023](https://openreview.net/forum?id=1pPpKc9wR0Y))
38
+ - **[Hyformer_peptides_34M](https://huggingface.co/SzczurekLab/hyformer_peptides_34M):** Trained on 3.5M general-purpose and antimicrobial peptides
39
+ - **[Hyformer_peptides_34M_MIC](https://huggingface.co/SzczurekLab/hyformer_peptides_34M_MIC):** `Hyformer_peptides_34M` jointly fine-tuned on minimal inhibitory concentration values (MIC) against E. coli bacteria
40
+
41
+ ## Gated Access
42
+ This model is available with **gated access**. To request access, please use the Hugging Face gated request form.
43
+
44
+ ## Citation
45
+ If you use this model, please cite:
46
+
47
+ ```
48
+ @misc{izdebski2025synergisticbenefitsjointmolecule,
49
+ title={Synergistic Benefits of Joint Molecule Generation and Property Prediction},
50
+ author={Adam Izdebski and Jan Olszewski and Pankhil Gawade and Krzysztof Koras and Serra Korkmaz and Valentin Rauscher and Jakub M. Tomczak and Ewa Szczurek},
51
+ year={2025},
52
+ eprint={2504.16559},
53
+ archivePrefix={arXiv},
54
+ primaryClass={cs.LG},
55
+ url={https://arxiv.org/abs/2504.16559},
56
+ }
57
+ ```
58
+
59
+ ## References
60
+ - Brown, Nathan, et al. "GuacaMol: benchmarking models for de novo molecular design." Journal of chemical information and modeling, 2019.
61
+ - Zhou, Gengmo, et al. "Uni-mol: A universal 3d molecular representation learning framework." ICLR, 2023.
ckpt.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:dd2084a900ef19cc01f8b18dbfc5be564993363ae50d8dcbbec47f3d094f5045
3
+ size 204798416
model_config.json ADDED
@@ -0,0 +1,9 @@
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "model_type": "Hyformer",
3
+ "embedding_dim": 512,
4
+ "num_attention_heads": 8,
5
+ "num_transformer_layers": 8,
6
+ "vocab_size": 34,
7
+ "prediction_task_type": "regression",
8
+ "num_prediction_tasks": 1
9
+ }
tokenizer_config.json ADDED
@@ -0,0 +1,4 @@
 
 
 
 
 
1
+ {
2
+ "tokenizer_type": "AATokenizer",
3
+ "vocabulary_path": "data/vocabulary/aa.txt"
4
+ }
trainer_config.json ADDED
@@ -0,0 +1,23 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "batch_size": 64,
3
+ "learning_rate": 0.0001,
4
+ "weight_decay": 0.01,
5
+ "max_epochs": 60,
6
+ "tasks": {
7
+ "prediction": 0.6,
8
+ "lm": 0.4
9
+ },
10
+ "compile": true,
11
+ "enable_ddp": false,
12
+ "dtype": "float32",
13
+ "num_workers": 16,
14
+ "beta1": 0.9,
15
+ "beta2": 0.95,
16
+ "gradient_accumulation_steps": 1,
17
+ "grad_clip": 1.0,
18
+ "decay_lr": true,
19
+ "log_interval": 10,
20
+ "save_interval": 5,
21
+ "min_lr": 1e-06,
22
+ "warmup_iters": 54
23
+ }
vocab.txt ADDED
@@ -0,0 +1,27 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ T
2
+ I
3
+ F
4
+ L
5
+ V
6
+ E
7
+ G
8
+ A
9
+ S
10
+ Y
11
+ Q
12
+ C
13
+ W
14
+ H
15
+ R
16
+ K
17
+ M
18
+ D
19
+ N
20
+ P
21
+ B
22
+ U
23
+ Z
24
+ X
25
+ O
26
+ -
27
+ .