Multilingual GGUF Models

Quantized GGUF models for CPU inference using llama.cpp.

Files

  • base: base-q4_k_m.gguf
  • english: en-q4_k_m.gguf
  • korean: ko-q4_k_m.gguf
  • spanish: es-q4_k_m.gguf
Downloads last month
456
GGUF
Model size
1B params
Architecture
llama
Hardware compatibility
Log In to add your hardware

4-bit

Inference Providers NEW
This model isn't deployed by any Inference Provider. ๐Ÿ™‹ Ask for provider support