|
|
--- |
|
|
license: other |
|
|
language: |
|
|
- ko |
|
|
- en |
|
|
base_model: |
|
|
- naver-hyperclovax/HyperCLOVAX-SEED-Think-14B |
|
|
pipeline_tag: text-generation |
|
|
library_name: transformers |
|
|
--- |
|
|
|
|
|
# HyperCLOVAX-SEED-Think-14B-GPTQ |
|
|
|
|
|
## Instruction |
|
|
|
|
|
This repo contains GPTQ model files for HyperCLOVAX-SEED-Think-14B. |
|
|
|
|
|
HyperCLOVAX-SEED-Think-14B-GPTQ was quantized using gptqmodel v4.0.0, following the guide. |
|
|
|
|
|
### Model Configuration |
|
|
|
|
|
- Original model: [naver-hyperclovax/HyperCLOVAX-SEED-Think-14B](https://huggingface.co/naver-hyperclovax/HyperCLOVAX-SEED-Think-14B/blob/main/config.json) |
|
|
- Quantization: GPTQ with 4-bit group-wise weight-only quantization (W4A16g128) |
|
|
|
|
|
## Quickstart |
|
|
|
|
|
```python |
|
|
from transformers import AutoModelForCausalLM, AutoTokenizer |
|
|
|
|
|
model_name = "K-Compression/HyperCLOVAX-SEED-Think-14B-GPTQ" |
|
|
model = AutoModelForCausalLM.from_pretrained( |
|
|
model_name, |
|
|
torch_dtype="bfloat16", |
|
|
device_map="auto" |
|
|
) |
|
|
tokenizer = AutoTokenizer.from_pretrained(model_name) |
|
|
``` |
|
|
|
|
|
## Performance(Non-Think) |
|
|
| Model | MMLU (0-shot) | HAERAE (0-shot) | |
|
|
|-----------------------------------------|--------|--------| |
|
|
| HyperCLOVA X SEED 14B Think | 0.7144 | 0.8130 | |
|
|
| HyperCLOVA X SEED 14B Think-GPTQ | 0.7018 | 0.8139 | |
|
|
|
|
|
## License |
|
|
|
|
|
The model is licensed under [HyperCLOVA X SEED Model License Agreement](https://huggingface.co/naver-hyperclovax/HyperCLOVAX-SEED-Think-14B/blob/main/LICENSE) |
|
|
|