translategemma-4b-it-q5_k_m.gguf

This repo contains GGUF weights for google/translategemma-4b-it.

Quantization Details

  • Method: llama-quantize
  • Llama.cpp Version: 7770 (fe44d3557)
  • Original Model Precision: BF16

Files Provided

File Quant Method Size Description
translategemma-4b-it-q5_k_m.gguf Q5_K_M 3.2 GB High quality, recommended for most uses.

Usage

You can use these models with llama.cpp

./llama-server -m translategemma-4b-it-q5_k_m.gguf -no-mmap -ngl 99 --port 8080 -c 8192 -fa 1 --jinja
Downloads last month
13
GGUF
Model size
4B params
Architecture
gemma3
Hardware compatibility
Log In to view the estimation

5-bit

Inference Providers NEW
This model isn't deployed by any Inference Provider. 🙋 Ask for provider support

Model tree for mixer3d/translategemma-4b-it-gguf

Quantized
(15)
this model