Update README.md
Browse files
README.md
CHANGED
|
@@ -6,13 +6,12 @@ Here, we provide AWQ-quantized versions of the most popular NVILA models. These
|
|
| 6 |
One-command demo to chat with quantized NVILA models via [llm-awq](https://github.com/mit-han-lab/llm-awq/tree/main/tinychat#:~:text=TinyChat%20support%20NVILA) (NVILA-8B as an example):
|
| 7 |
|
| 8 |
```bash
|
| 9 |
-
'''
|
| 10 |
cd llm-awq/tinychat
|
| 11 |
python nvila_demo.py --model-path PATH/TO/NVILA \
|
| 12 |
--quant_path NVILA-8B-w4-g128-awq-v2.pt \
|
| 13 |
--media PATH/TO/ANY/IMAGES/VIDEOS \
|
| 14 |
--act_scale_path NVILA-8B-VT-smooth-scale.pt \
|
| 15 |
--all --chunk --model_type nvila
|
| 16 |
-
|
| 17 |
|
| 18 |
This command will download the quantized NVILA model and run a chat demo. If you’ve already downloaded the files, simply set the path to your local copies.
|
|
|
|
| 6 |
One-command demo to chat with quantized NVILA models via [llm-awq](https://github.com/mit-han-lab/llm-awq/tree/main/tinychat#:~:text=TinyChat%20support%20NVILA) (NVILA-8B as an example):
|
| 7 |
|
| 8 |
```bash
|
|
|
|
| 9 |
cd llm-awq/tinychat
|
| 10 |
python nvila_demo.py --model-path PATH/TO/NVILA \
|
| 11 |
--quant_path NVILA-8B-w4-g128-awq-v2.pt \
|
| 12 |
--media PATH/TO/ANY/IMAGES/VIDEOS \
|
| 13 |
--act_scale_path NVILA-8B-VT-smooth-scale.pt \
|
| 14 |
--all --chunk --model_type nvila
|
| 15 |
+
```
|
| 16 |
|
| 17 |
This command will download the quantized NVILA model and run a chat demo. If you’ve already downloaded the files, simply set the path to your local copies.
|