Safetensors
exaone4
compressed-tensors

Quantization Details

  • Base Model: LGAI-EXAONE/EXAONE-4.0-1.2B
  • Method: GPTQ W4A16
  • Group Size: 128
  • Calibration Dataset: LGAI-EXAONE/MANTA-1M (512 samples)
  • Tool: llmcompressor

Usage

from vllm import LLM
llm = LLM(model="IBDPLab/EXAONE-4.0-1.2B-W4A16-GPTQ")
Downloads last month
24
Safetensors
Model size
2B params
Tensor type
I64
·
I32
·
BF16
·
Inference Providers NEW
This model isn't deployed by any Inference Provider. 🙋 Ask for provider support