ImportError: cannot import name 'is_flash_attn_available' from 'transformers.utils'
#16
by
nayeon212
- opened
ImportError: cannot import name 'is_flash_attn_available' from 'transformers.utils'
In the latest version of transformers, the function is_flash_attn_available() has been removed and replaced by is_flash_attn_2_available(). It would be great if you could update the modeling_orion.py code to use is_flash_attn_2_available() instead of is_flash_attn_available(). Alternatively, it would be helpful if you could provide a working transformers library (version transformers < 4.35) that includes the required functionality.
Thanks a lot!
fixed! thanks!