n1ck-guo commited on
Commit
f2dad8d
·
verified ·
1 Parent(s): f24d7a2

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +0 -1
README.md CHANGED
@@ -13,7 +13,6 @@ This model is an int4 model with group_size 128 and symmetric quantization of [Q
13
 
14
  ### INT4 Inference(CPU/CUDA/INTEL GPU)
15
  ```python
16
- from auto_round import AutoRoundConfig ##must import for auto-round format if transformers <= 4.51.3
17
  from transformers import AutoModelForCausalLM,AutoTokenizer
18
  quantized_model_dir = "Intel/Qwen3-8B-int4-AutoRound-inc"
19
 
 
13
 
14
  ### INT4 Inference(CPU/CUDA/INTEL GPU)
15
  ```python
 
16
  from transformers import AutoModelForCausalLM,AutoTokenizer
17
  quantized_model_dir = "Intel/Qwen3-8B-int4-AutoRound-inc"
18