fix: import `flash_attn_varlen_func` from `flash_attn` instead of `transformers.modeling_flash_attention_utils`
#4 opened 6 days ago
by
wincentIsMe
vllm support ?
➕
7
1
#2 opened 4 months ago
by
TahirC