|
|
--- |
|
|
license: apache-2.0 |
|
|
language: |
|
|
- en |
|
|
pipeline_tag: image-text-to-text |
|
|
tags: |
|
|
- multimodal |
|
|
library_name: transformers |
|
|
base_model: |
|
|
- Qwen/Qwen2.5-VL-7B-Instruct |
|
|
--- |
|
|
|
|
|
```bash |
|
|
# Create a dedicated python env |
|
|
python3 -m venv llmcompressor |
|
|
source llmcompressor/bin/activate |
|
|
# Install llm-compressor and additionnal needed libs |
|
|
pip install llmcompressor qwen_vl_utils torchvision |
|
|
# Download model in HF cache |
|
|
hf download Qwen/Qwen2.5-VL-7B-Instruct |
|
|
# Start quantization |
|
|
wget https://github.com/vllm-project/llm-compressor/blob/main/examples/quantization_w8a8_fp8/qwen_2_5_vl_example.py -O qwen_2_5_vl_fp8.py |
|
|
python3 qwen_2_5_vl_fp8.py |
|
|
``` |