jbochi's picture
Add instructions to quantize model
e40158b
|
raw
history blame
511 Bytes
metadata
license: apache-2.0
datasets:
  - grammarly/coedit
language:
  - en
tags:
  - text-generation-inference
  - candle

Quantized weights of coedit for inference with candle.

Conversion command, using candle:

cargo run --example tensor-tools --release -- quantize \
  --quantization q6k \
  /path/to/coedit-<version>/model.safetensors \
  --out-file model<version>.gguf