esm2_t6_long / README.md
gabrielbianchin's picture
Update README.md
d787692 verified
|
raw
history blame
831 Bytes
metadata
title: ESM2 Long Models

ESM2 Long

ESM2 Long is an adapted version of the ESM2 architectures. It uses local attention instead of global attention, allowing for models with longer input sizes. ESM2 Long models have a context size of 2,050, double that of the standard ESM2 model. Several ESM2 Long models are available:

For detailed information, please refer to the paper.