lhallee commited on
Commit
20da398
·
verified ·
1 Parent(s): 34601d3

Upload README.md with huggingface_hub

Browse files
Files changed (1) hide show
  1. README.md +40 -9
README.md CHANGED
@@ -129,15 +129,46 @@ Average MSE for BF16: 0.00004125
129
  We look at various ESM models and their throughput on an H100. FastESM is over twice as fast as ESM2-650 with longer sequences. Requires PyTorch 2.5+ for the most savings, see [SDPA](https://pytorch.org/docs/stable/generated/torch.nn.functional.scaled_dot_product_attention.html).
130
  ![image/png](https://cdn-uploads.huggingface.co/production/uploads/62f2bd3bdb7cbd214b658c48/PvaBGfuJXEW2v_WLkt63y.png)
131
 
132
- ### Citation
133
- If you use any of this implementation or work please cite it (as well as the [ESM2](https://www.science.org/doi/10.1126/science.ade2574) paper).
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
134
  ```
135
- @misc {FastPLMs,
136
- author = { Hallee, Logan and Bichara, David and Gleghorn, Jason P.},
137
- title = { FastPLMs: Fast, efficient, protien language model inference from Huggingface AutoModel.},
138
- year = {2024},
139
- url = { https://huggingface.co/Synthyra/ESMplusplus_small },
140
- DOI = { 10.57967/hf/3726 },
141
- publisher = { Hugging Face }
 
 
 
 
 
 
 
 
 
142
  }
143
  ```
 
129
  We look at various ESM models and their throughput on an H100. FastESM is over twice as fast as ESM2-650 with longer sequences. Requires PyTorch 2.5+ for the most savings, see [SDPA](https://pytorch.org/docs/stable/generated/torch.nn.functional.scaled_dot_product_attention.html).
130
  ![image/png](https://cdn-uploads.huggingface.co/production/uploads/62f2bd3bdb7cbd214b658c48/PvaBGfuJXEW2v_WLkt63y.png)
131
 
132
+ ### Citations
133
+
134
+ ```bibtex
135
+ @misc{FastPLMs,
136
+ author={Hallee, Logan and Bichara, David and Gleghorn, Jason P.},
137
+ title={FastPLMs: Fast, efficient, protein language model inference from Huggingface AutoModel.},
138
+ year={2024},
139
+ url={https://huggingface.co/Synthyra/ESMplusplus_small},
140
+ DOI={10.57967/hf/3726},
141
+ publisher={Hugging Face}
142
+ }
143
+ ```
144
+
145
+ ```bibtex
146
+ @article{lin2023esm2,
147
+ title={Evolutionary-scale prediction of atomic-level protein structure with a language model},
148
+ author={Lin, Zeming and Akin, Halil and Rao, Roshan and Hie, Brian and Zhu, Zhongkai and Lu, Wenting and Smestad, Nikita and Verkuil, Robert and Kabeli, Ori and Shmueli, Yaniv and dos Santos Costa, Allan and Fazel-Zarandi, Maryam and Sercu, Tom and Candido, Salvatore and Rives, Alexander},
149
+ journal={Science},
150
+ volume={379},
151
+ number={6637},
152
+ pages={1123--1130},
153
+ year={2023},
154
+ DOI={10.1126/science.ade2574}
155
+ }
156
  ```
157
+
158
+ ```bibtex
159
+ @article{dong2024flexattention,
160
+ title={Flex Attention: A Programming Model for Generating Optimized Attention Kernels},
161
+ author={Dong, Juechu and Feng, Boyuan and Guessous, Driss and Liang, Yanbo and He, Horace},
162
+ journal={arXiv preprint arXiv:2412.05496},
163
+ year={2024}
164
+ }
165
+ ```
166
+
167
+ ```bibtex
168
+ @inproceedings{paszke2019pytorch,
169
+ title={PyTorch: An Imperative Style, High-Performance Deep Learning Library},
170
+ author={Paszke, Adam and Gross, Sam and Massa, Francisco and Lerer, Adam and Bradbury, James and Chanan, Gregory and Killeen, Trevor and Lin, Zeming and Gimelshein, Natalia and Antiga, Luca and Desmaison, Alban and K{\"o}pf, Andreas and Yang, Edward and DeVito, Zach and Raison, Martin and Tejani, Alykhan and Chilamkurthy, Sasank and Steiner, Benoit and Fang, Lu and Bai, Junjie and Chintala, Soumith},
171
+ booktitle={Advances in Neural Information Processing Systems 32},
172
+ year={2019}
173
  }
174
  ```