fix documentation for loading the model, since the fused attention module doesnt work here either.
#4
by mber - opened
No description provided.
Ahh thank you! Yes I need that for all 70B models.
TheBloke changed pull request status to merged
With pleasure :)