omitakahiro commited on
Commit
cb9271a
·
verified ·
1 Parent(s): 684ad10

Upload quantize_autoawq.py

Browse files
Files changed (1) hide show
  1. quantize_autoawq.py +33 -0
quantize_autoawq.py ADDED
@@ -0,0 +1,33 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ import random
2
+
3
+ import pandas as pd
4
+ from awq import AutoAWQForCausalLM
5
+ from transformers import AutoTokenizer
6
+
7
+ model_path = 'stockmark/Stockmark-2-100B-Instruct-beta'
8
+ quant_path = 'Stockmark-2-100B-Instruct-beta-AWQ'
9
+
10
+ quant_config = { "zero_point": True, "q_group_size": 128, "w_bit": 4, "version": "GEMM" }
11
+
12
+ # Load model
13
+ model = AutoAWQForCausalLM.from_pretrained(model_path)
14
+ tokenizer = AutoTokenizer.from_pretrained(model_path, trust_remote_code=True)
15
+
16
+ # load dataset
17
+ ds = pd.read_json("caliblation.jsonl", lines=True).to_dict("records")
18
+ ds = [ tokenizer.apply_chat_template(doc["messages"], tokenize=False) for doc in ds ]
19
+ random.shuffle(ds)
20
+
21
+ # Quantize
22
+ model.quantize(
23
+ tokenizer,
24
+ quant_config=quant_config,
25
+ calib_data=ds,
26
+ n_parallel_calib_samples=64,
27
+ max_calib_samples=128,
28
+ max_calib_seq_len=1024
29
+ )
30
+
31
+ # Save quantized model
32
+ model.save_quantized(quant_path)
33
+ tokenizer.save_pretrained(quant_path)