Update README.md
Browse files
README.md
CHANGED
|
@@ -14,18 +14,17 @@ This repository contains remote code and weights for a **Native Sparse Attention
|
|
| 14 |
|
| 15 |
To use this model, please ensure the following dependencies are installed:
|
| 16 |
|
| 17 |
-
####
|
| 18 |
```bash
|
| 19 |
pip install git+https://github.com/fnite1604/native-sparse-attention-pytorch.git
|
| 20 |
```
|
| 21 |
|
| 22 |
-
####
|
| 23 |
-
These are handled automatically by the Transformers library and include:
|
| 24 |
```bash
|
| 25 |
-
pip install transformers torch
|
| 26 |
```
|
| 27 |
|
| 28 |
-
Note: We recommend using
|
| 29 |
|
| 30 |
### Example Usage
|
| 31 |
|
|
@@ -35,6 +34,6 @@ A `quick_start.py` script is included to help you get started with inference:
|
|
| 35 |
python quick_start.py
|
| 36 |
```
|
| 37 |
|
| 38 |
-
This will load the model and generate text based on a predefined prompt using Native Sparse Attention.
|
| 39 |
|
| 40 |
|
|
|
|
| 14 |
|
| 15 |
To use this model, please ensure the following dependencies are installed:
|
| 16 |
|
| 17 |
+
#### Install the required Native Sparse Attention library from our custom fork:
|
| 18 |
```bash
|
| 19 |
pip install git+https://github.com/fnite1604/native-sparse-attention-pytorch.git
|
| 20 |
```
|
| 21 |
|
| 22 |
+
#### Install standard dependencies:
|
|
|
|
| 23 |
```bash
|
| 24 |
+
pip install transformers torch ...
|
| 25 |
```
|
| 26 |
|
| 27 |
+
Note: We recommend using the latest stable of Pytorch (currently 2.7.0) with CUDA 12.6 and the latest available version of Transformers
|
| 28 |
|
| 29 |
### Example Usage
|
| 30 |
|
|
|
|
| 34 |
python quick_start.py
|
| 35 |
```
|
| 36 |
|
| 37 |
+
This will load the model and generate text based on a predefined prompt ("What is 1 + 1?") using our Native Sparse Attention enabled reasoning model.
|
| 38 |
|
| 39 |
|