File size: 508 Bytes
1c8289d
0d2afad
1c8289d
 
 
1
2
3
4
5
Weights for the `softmax` model from the paper [2Mamba2Furious: Linear in Complexity, Competitive in Accuracy](https://huggingface.co/papers/2602.17363). 
This model variant just uses plain ol softmax (FlashAttention), used for the NIAH experiment. It was trained for 400K steps with a batch size of 32.
More details of the setup can be found in the Github repo.

Instructions on how to use this model can be found in [https://github.com/gmongaras/2Mamba2Furious](https://github.com/gmongaras/2Mamba2Furious)