Upload README.md with huggingface_hub
Browse files
README.md
CHANGED
|
@@ -7,18 +7,18 @@ tags: []
|
|
| 7 |
The GitHub with the implementation and requirements.txt can be found [here](https://github.com/Synthyra/FastPLMs.git)
|
| 8 |
|
| 9 |
# Profluent-E1
|
| 10 |
-
[Profluent-E1](https://github.com/Synthyra/ESMplusplus) is a faithful implementation of Profluent's
|
| 11 |
|
| 12 |
|
| 13 |
## Use with 🤗 transformers
|
| 14 |
### Supported models
|
| 15 |
```python
|
| 16 |
model_dict = {
|
| 17 |
-
# Synthyra/Profluent-
|
| 18 |
'Profluent-E1-150M': 'Profluent-Bio/E1-150m',
|
| 19 |
-
# Synthyra/Profluent-
|
| 20 |
'Profluent-E1-300M': 'Profluent-Bio/E1-300m',
|
| 21 |
-
# Synthyra/Profluent-
|
| 22 |
'Profluent-E1-600M': 'Profluent-Bio/E1-600m',
|
| 23 |
}
|
| 24 |
```
|
|
@@ -127,17 +127,6 @@ for param in model.classifier.parameters():
|
|
| 127 |
For a more thourough example of fine-tuning, check out our example script [here](https://github.com/Synthyra/FastPLMs/blob/main/fine_tuning_example.py).
|
| 128 |
|
| 129 |
|
| 130 |
-
## Returning attention maps
|
| 131 |
-
Usually F.scaled_dot_product_attention is used for the attention calculations, which is much faster than native PyTorch. However, it cannot return attention maps.
|
| 132 |
-
ESM++ has the option to ```output_attentions```, which will calculate attention manually. This is much slower, so do not use unless you need the attention maps.
|
| 133 |
-
|
| 134 |
-
```python
|
| 135 |
-
output = model(**tokenized, output_attentions=True)
|
| 136 |
-
att = output.attentions
|
| 137 |
-
len(att) # 30, one for each layer, size (batch_size, num_heads, seq_len, seq_len) each
|
| 138 |
-
```
|
| 139 |
-
|
| 140 |
-
|
| 141 |
### Citation
|
| 142 |
If you use any of this implementation or work please cite the following DOI and Profluent's paper.
|
| 143 |
|
|
@@ -153,7 +142,8 @@ If you use any of this implementation or work please cite the following DOI and
|
|
| 153 |
```
|
| 154 |
|
| 155 |
```
|
| 156 |
-
@article{Jain_Beazer_Ruffolo_Bhatnagar_Madani_2025,
|
|
|
|
| 157 |
url={https://www.biorxiv.org/content/early/2025/11/13/2025.11.12.688125},
|
| 158 |
DOI={10.1101/2025.11.12.688125},
|
| 159 |
journal={bioRxiv},
|
|
|
|
| 7 |
The GitHub with the implementation and requirements.txt can be found [here](https://github.com/Synthyra/FastPLMs.git)
|
| 8 |
|
| 9 |
# Profluent-E1
|
| 10 |
+
[Profluent-E1](https://github.com/Synthyra/ESMplusplus) is a faithful implementation of Profluent's [E1](https://www.profluent.bio/showcase/e1) models ([license](https://github.com/Profluent-AI/E1/tree/main?tab=License-1-ov-file)) that integrates Huggingface AutoModel compatability and nice embedding functionality.
|
| 11 |
|
| 12 |
|
| 13 |
## Use with 🤗 transformers
|
| 14 |
### Supported models
|
| 15 |
```python
|
| 16 |
model_dict = {
|
| 17 |
+
# Synthyra/Profluent-E1-150M
|
| 18 |
'Profluent-E1-150M': 'Profluent-Bio/E1-150m',
|
| 19 |
+
# Synthyra/Profluent-E1-150M
|
| 20 |
'Profluent-E1-300M': 'Profluent-Bio/E1-300m',
|
| 21 |
+
# Synthyra/Profluent-E1-150M
|
| 22 |
'Profluent-E1-600M': 'Profluent-Bio/E1-600m',
|
| 23 |
}
|
| 24 |
```
|
|
|
|
| 127 |
For a more thourough example of fine-tuning, check out our example script [here](https://github.com/Synthyra/FastPLMs/blob/main/fine_tuning_example.py).
|
| 128 |
|
| 129 |
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 130 |
### Citation
|
| 131 |
If you use any of this implementation or work please cite the following DOI and Profluent's paper.
|
| 132 |
|
|
|
|
| 142 |
```
|
| 143 |
|
| 144 |
```
|
| 145 |
+
@article{Jain_Beazer_Ruffolo_Bhatnagar_Madani_2025,
|
| 146 |
+
title={E1: Retrieval-Augmented Protein Encoder Models},
|
| 147 |
url={https://www.biorxiv.org/content/early/2025/11/13/2025.11.12.688125},
|
| 148 |
DOI={10.1101/2025.11.12.688125},
|
| 149 |
journal={bioRxiv},
|