Leon
commited on
Update README.md
Browse files
README.md
CHANGED
|
@@ -8,7 +8,8 @@ library_name: transformers
|
|
| 8 |
---
|
| 9 |
# Aria-sequential_mlp-FP8-dynamic
|
| 10 |
FP8-Dynamic quantization from [Aria-sequential_mlp](https://huggingface.co/rhymes-ai/Aria-sequential_mlp) made with [llm-compressor](https://github.com/vllm-project/llm-compressor), requires about 24.8 GB of VRAM.
|
| 11 |
-
|
|
|
|
| 12 |
```
|
| 13 |
pip install transformers==4.45.0 accelerate==0.34.1 sentencepiece==0.2.0 torchvision requests torch Pillow compressed-tensors
|
| 14 |
pip install flash-attn --no-build-isolation
|
|
|
|
| 8 |
---
|
| 9 |
# Aria-sequential_mlp-FP8-dynamic
|
| 10 |
FP8-Dynamic quantization from [Aria-sequential_mlp](https://huggingface.co/rhymes-ai/Aria-sequential_mlp) made with [llm-compressor](https://github.com/vllm-project/llm-compressor), requires about 24.8 GB of VRAM.
|
| 11 |
+
|
| 12 |
+
### Installation
|
| 13 |
```
|
| 14 |
pip install transformers==4.45.0 accelerate==0.34.1 sentencepiece==0.2.0 torchvision requests torch Pillow compressed-tensors
|
| 15 |
pip install flash-attn --no-build-isolation
|