lhallee commited on
Commit
5b31cb9
·
verified ·
1 Parent(s): 25fe96e

Upload README.md with huggingface_hub

Browse files
Files changed (1) hide show
  1. README.md +2 -2
README.md CHANGED
@@ -7,7 +7,7 @@ tags: []
7
  The GitHub with the implementation and requirements.txt can be found [here](https://github.com/Synthyra/FastPLMs.git)
8
 
9
  # Profluent-E1
10
- [Profluent-E1](https://github.com/Synthyra/ESMplusplus) is a faithful implementation of Profluent's [E1](https://www.profluent.bio/showcase/e1) models ([license](https://github.com/Profluent-AI/E1/tree/main?tab=License-1-ov-file)) that integrates Huggingface AutoModel compatability and nice embedding functionality.
11
 
12
 
13
  ## Use with 🤗 transformers
@@ -54,7 +54,7 @@ print(logits.shape) # (batch_size, num_labels), (2, 2)
54
  E1 weights were trained in bf16 and are in bf16 by default. You can load them in the precision of your choosing by leveraging the dtype parameter:
55
  ```python
56
  import torch
57
- model = AutoModelForMaskedLM.from_pretrained('Synthyra/ESMplusplus_small', trust_remote_code=True, dtype=torch.float) # fp32
58
  ```
59
 
60
  ## Embed entire datasets with no new code
 
7
  The GitHub with the implementation and requirements.txt can be found [here](https://github.com/Synthyra/FastPLMs.git)
8
 
9
  # Profluent-E1
10
+ [Synthyra's version of Profluent-E1](https://github.com/Synthyra/Profluent-E1-300M) is a faithful implementation of Profluent's [E1](https://www.profluent.bio/showcase/e1) models ([license](https://github.com/Profluent-AI/E1/tree/main?tab=License-1-ov-file)) that integrates Huggingface AutoModel compatability and nice embedding functionality.
11
 
12
 
13
  ## Use with 🤗 transformers
 
54
  E1 weights were trained in bf16 and are in bf16 by default. You can load them in the precision of your choosing by leveraging the dtype parameter:
55
  ```python
56
  import torch
57
+ model = AutoModelForMaskedLM.from_pretrained('Synthyra/Profluent-E1-150M', trust_remote_code=True, dtype=torch.float) # fp32
58
  ```
59
 
60
  ## Embed entire datasets with no new code