metadata
license: apache-2.0
datasets:
- grammarly/coedit
language:
- en
tags:
- text-generation-inference
- candle
Quantized weights of coedit for inference with candle.
Conversion command, using candle:
cargo run --example tensor-tools --release -- quantize \
--quantization q6k \
/path/to/coedit-<version>/model.safetensors \
--out-file model<version>.gguf