K-Compression commited on
Commit
364385b
·
verified ·
1 Parent(s): 8ecf0aa

Create README.md

Browse files
Files changed (1) hide show
  1. README.md +47 -0
README.md ADDED
@@ -0,0 +1,47 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ ---
2
+ license: other
3
+ language:
4
+ - ko
5
+ - en
6
+ base_model:
7
+ - naver-hyperclovax/HyperCLOVAX-SEED-Think-14B
8
+ pipeline_tag: text-generation
9
+ library_name: transformers
10
+ ---
11
+
12
+ # HyperCLOVAX-SEED-Think-14B-GPTQ
13
+
14
+ ## Instruction
15
+
16
+ This repo contains GPTQ model files for HyperCLOVAX-SEED-Think-14B.
17
+
18
+ HyperCLOVAX-SEED-Think-14B-GPTQ was quantized using gptqmodel v4.0.0, following the guide.
19
+
20
+ ### Model Configuration
21
+
22
+ - Original model: [naver-hyperclovax/HyperCLOVAX-SEED-Think-14B](https://huggingface.co/naver-hyperclovax/HyperCLOVAX-SEED-Think-14B/blob/main/config.json)
23
+ - Quantization: GPTQ with 4-bit group-wise weight-only quantization (W4A16g128)
24
+
25
+ ## Quickstart
26
+
27
+ ```python
28
+ from transformers import AutoModelForCausalLM, AutoTokenizer
29
+
30
+ model_name = "K-Compression/HyperCLOVAX-SEED-Think-14B-GPTQ"
31
+ model = AutoModelForCausalLM.from_pretrained(
32
+ model_name,
33
+ torch_dtype="bfloat16",
34
+ device_map="auto"
35
+ )
36
+ tokenizer = AutoTokenizer.from_pretrained(model_name)
37
+ ```
38
+
39
+ ## Performance(Non-Think)
40
+ | Model | MMLU (0-shot) | HAERAE (0-shot) |
41
+ |-----------------------------------------|--------|--------|
42
+ | HyperCLOVA X SEED 14B Think | 0.7144 | 0.8130 |
43
+ | HyperCLOVA X SEED 14B Think-GPTQ | 0.7018 | 0.8139 |
44
+
45
+ ## License
46
+
47
+ The model is licensed under [HyperCLOVA X SEED Model License Agreement](https://huggingface.co/naver-hyperclovax/HyperCLOVAX-SEED-Think-14B/blob/main/LICENSE)