doubleblind commited on
Commit
99c33d6
·
verified ·
1 Parent(s): 6dabb43

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +5 -6
README.md CHANGED
@@ -14,18 +14,17 @@ This repository contains remote code and weights for a **Native Sparse Attention
14
 
15
  To use this model, please ensure the following dependencies are installed:
16
 
17
- #### 1. Install the required sparse attention library from our custom fork:
18
  ```bash
19
  pip install git+https://github.com/fnite1604/native-sparse-attention-pytorch.git
20
  ```
21
 
22
- #### 2. Install other standard dependencies:
23
- These are handled automatically by the Transformers library and include:
24
  ```bash
25
- pip install transformers torch
26
  ```
27
 
28
- Note: We recommend using Python 3.8+ and PyTorch 2.0+ for compatibility.
29
 
30
  ### Example Usage
31
 
@@ -35,6 +34,6 @@ A `quick_start.py` script is included to help you get started with inference:
35
  python quick_start.py
36
  ```
37
 
38
- This will load the model and generate text based on a predefined prompt using Native Sparse Attention.
39
 
40
 
 
14
 
15
  To use this model, please ensure the following dependencies are installed:
16
 
17
+ #### Install the required Native Sparse Attention library from our custom fork:
18
  ```bash
19
  pip install git+https://github.com/fnite1604/native-sparse-attention-pytorch.git
20
  ```
21
 
22
+ #### Install standard dependencies:
 
23
  ```bash
24
+ pip install transformers torch ...
25
  ```
26
 
27
+ Note: We recommend using the latest stable of Pytorch (currently 2.7.0) with CUDA 12.6 and the latest available version of Transformers
28
 
29
  ### Example Usage
30
 
 
34
  python quick_start.py
35
  ```
36
 
37
+ This will load the model and generate text based on a predefined prompt ("What is 1 + 1?") using our Native Sparse Attention enabled reasoning model.
38
 
39