lhallee commited on
Commit
cc328db
·
verified ·
1 Parent(s): ca1f106

Upload README.md with huggingface_hub

Browse files
Files changed (1) hide show
  1. README.md +6 -16
README.md CHANGED
@@ -7,18 +7,18 @@ tags: []
7
  The GitHub with the implementation and requirements.txt can be found [here](https://github.com/Synthyra/FastPLMs.git)
8
 
9
  # Profluent-E1
10
- [Profluent-E1](https://github.com/Synthyra/ESMplusplus) is a faithful implementation of Profluent's models [E1](https://www.profluent.bio/showcase/e1) ([license](https://github.com/Profluent-AI/E1/tree/main?tab=License-1-ov-file)) that integrates Huggingface AutoModel compatability and nice embedding functionality built in.
11
 
12
 
13
  ## Use with 🤗 transformers
14
  ### Supported models
15
  ```python
16
  model_dict = {
17
- # Synthyra/Profluent-E1_150M
18
  'Profluent-E1-150M': 'Profluent-Bio/E1-150m',
19
- # Synthyra/Profluent-E1_150M
20
  'Profluent-E1-300M': 'Profluent-Bio/E1-300m',
21
- # Synthyra/Profluent-E1_150M
22
  'Profluent-E1-600M': 'Profluent-Bio/E1-600m',
23
  }
24
  ```
@@ -127,17 +127,6 @@ for param in model.classifier.parameters():
127
  For a more thourough example of fine-tuning, check out our example script [here](https://github.com/Synthyra/FastPLMs/blob/main/fine_tuning_example.py).
128
 
129
 
130
- ## Returning attention maps
131
- Usually F.scaled_dot_product_attention is used for the attention calculations, which is much faster than native PyTorch. However, it cannot return attention maps.
132
- ESM++ has the option to ```output_attentions```, which will calculate attention manually. This is much slower, so do not use unless you need the attention maps.
133
-
134
- ```python
135
- output = model(**tokenized, output_attentions=True)
136
- att = output.attentions
137
- len(att) # 30, one for each layer, size (batch_size, num_heads, seq_len, seq_len) each
138
- ```
139
-
140
-
141
  ### Citation
142
  If you use any of this implementation or work please cite the following DOI and Profluent's paper.
143
 
@@ -153,7 +142,8 @@ If you use any of this implementation or work please cite the following DOI and
153
  ```
154
 
155
  ```
156
- @article{Jain_Beazer_Ruffolo_Bhatnagar_Madani_2025, title={E1: Retrieval-Augmented Protein Encoder Models},
 
157
  url={https://www.biorxiv.org/content/early/2025/11/13/2025.11.12.688125},
158
  DOI={10.1101/2025.11.12.688125},
159
  journal={bioRxiv},
 
7
  The GitHub with the implementation and requirements.txt can be found [here](https://github.com/Synthyra/FastPLMs.git)
8
 
9
  # Profluent-E1
10
+ [Profluent-E1](https://github.com/Synthyra/ESMplusplus) is a faithful implementation of Profluent's [E1](https://www.profluent.bio/showcase/e1) models ([license](https://github.com/Profluent-AI/E1/tree/main?tab=License-1-ov-file)) that integrates Huggingface AutoModel compatability and nice embedding functionality.
11
 
12
 
13
  ## Use with 🤗 transformers
14
  ### Supported models
15
  ```python
16
  model_dict = {
17
+ # Synthyra/Profluent-E1-150M
18
  'Profluent-E1-150M': 'Profluent-Bio/E1-150m',
19
+ # Synthyra/Profluent-E1-150M
20
  'Profluent-E1-300M': 'Profluent-Bio/E1-300m',
21
+ # Synthyra/Profluent-E1-150M
22
  'Profluent-E1-600M': 'Profluent-Bio/E1-600m',
23
  }
24
  ```
 
127
  For a more thourough example of fine-tuning, check out our example script [here](https://github.com/Synthyra/FastPLMs/blob/main/fine_tuning_example.py).
128
 
129
 
 
 
 
 
 
 
 
 
 
 
 
130
  ### Citation
131
  If you use any of this implementation or work please cite the following DOI and Profluent's paper.
132
 
 
142
  ```
143
 
144
  ```
145
+ @article{Jain_Beazer_Ruffolo_Bhatnagar_Madani_2025,
146
+ title={E1: Retrieval-Augmented Protein Encoder Models},
147
  url={https://www.biorxiv.org/content/early/2025/11/13/2025.11.12.688125},
148
  DOI={10.1101/2025.11.12.688125},
149
  journal={bioRxiv},