lhallee commited on
Commit
ca14f86
·
verified ·
1 Parent(s): 94df558

Upload README.md with huggingface_hub

Browse files
Files changed (1) hide show
  1. README.md +35 -9
README.md CHANGED
@@ -173,16 +173,42 @@ The plot below showcases performance normalized between the negative control (ra
173
  We look at various ESM models and their throughput on an H100. Adding efficient batching between ESMC and ESM++ significantly improves the throughput, although ESM++ is also faster than ESMC for batch size one. ESM++ small is even faster than ESM2-35M with long sequences! The most gains will be seen with PyTorch > 2.5 on linux machines.
174
  ![image/png](https://cdn-uploads.huggingface.co/production/uploads/62f2bd3bdb7cbd214b658c48/Lu6nWB9Fc-7YTql3Z1hVB.png)
175
 
176
- ### Citation
177
- If you use any of this implementation or work please cite it (as well as the ESMC preprint).
 
 
 
 
 
 
 
 
 
 
178
 
 
 
 
 
 
 
 
179
  ```
180
- @misc {FastPLMs,
181
- author = { Hallee, Logan and Bichara, David and Gleghorn, Jason P.},
182
- title = { FastPLMs: Fast, efficient, protien language model inference from Huggingface AutoModel.},
183
- year = {2024},
184
- url = { https://huggingface.co/Synthyra/ESMplusplus_small },
185
- DOI = { 10.57967/hf/3726 },
186
- publisher = { Hugging Face }
 
 
 
 
 
 
 
 
 
187
  }
188
  ```
 
173
  We look at various ESM models and their throughput on an H100. Adding efficient batching between ESMC and ESM++ significantly improves the throughput, although ESM++ is also faster than ESMC for batch size one. ESM++ small is even faster than ESM2-35M with long sequences! The most gains will be seen with PyTorch > 2.5 on linux machines.
174
  ![image/png](https://cdn-uploads.huggingface.co/production/uploads/62f2bd3bdb7cbd214b658c48/Lu6nWB9Fc-7YTql3Z1hVB.png)
175
 
176
+ ### Citations
177
+
178
+ ```bibtex
179
+ @misc{FastPLMs,
180
+ author={Hallee, Logan and Bichara, David and Gleghorn, Jason P.},
181
+ title={FastPLMs: Fast, efficient, protein language model inference from Huggingface AutoModel.},
182
+ year={2024},
183
+ url={https://huggingface.co/Synthyra/ESMplusplus_small},
184
+ DOI={10.57967/hf/3726},
185
+ publisher={Hugging Face}
186
+ }
187
+ ```
188
 
189
+ ```bibtex
190
+ @article{hayes2024simulating,
191
+ title={Simulating 500 million years of evolution with a language model},
192
+ author={Hayes, Thomas and Rao, Roshan and Akin, Halil and Sofber, Nicholas J and Achour, Divya and Moez, Irfan and Garg, Rhitu and Angelova, Rami and Babu, Manan and Alcaide, Eric and others},
193
+ journal={bioRxiv},
194
+ year={2024}
195
+ }
196
  ```
197
+
198
+ ```bibtex
199
+ @article{dong2024flexattention,
200
+ title={Flex Attention: A Programming Model for Generating Optimized Attention Kernels},
201
+ author={Dong, Juechu and Feng, Boyuan and Guessous, Driss and Liang, Yanbo and He, Horace},
202
+ journal={arXiv preprint arXiv:2412.05496},
203
+ year={2024}
204
+ }
205
+ ```
206
+
207
+ ```bibtex
208
+ @inproceedings{paszke2019pytorch,
209
+ title={PyTorch: An Imperative Style, High-Performance Deep Learning Library},
210
+ author={Paszke, Adam and Gross, Sam and Massa, Francisco and Lerer, Adam and Bradbury, James and Chanan, Gregory and Killeen, Trevor and Lin, Zeming and Gimelshein, Natalia and Antiga, Luca and Desmaison, Alban and K{\"o}pf, Andreas and Yang, Edward and DeVito, Zach and Raison, Martin and Tejani, Alykhan and Chilamkurthy, Sasank and Steiner, Benoit and Fang, Lu and Bai, Junjie and Chintala, Soumith},
211
+ booktitle={Advances in Neural Information Processing Systems 32},
212
+ year={2019}
213
  }
214
  ```