fix: trans_a=True also need to be modified to tl.trans() 427081b verified Taykhoom commited on 1 day ago
add attention return + support eager attention or triton FA2 via config.use_flash_attn e1354bd verified Taykhoom commited on 1 day ago