--- language: - en license: eupl-1.2 tags: - safetensors - 4-bit - transformers - 8-bit - gguf - gemma4 - bitsandbytes base_model: - google/gemma-4-E2B-it base_model_relation: quantized pipeline_tag: any-to-any datasets: - lthn/LEM-research --- # Lemer A [Gemma 4 E2B](https://huggingface.co/google/gemma-4-E2B-it) finetune by [lthn.ai](https://lthn.ai) — EUPL-1.2 **Ollama**: `ollama run hf.co/lthn/lemer:Q4_K_M` **MLX**: [bf16](https://huggingface.co/lthn/lemer/tree/bf16), [8bit](https://huggingface.co/lthn/lemer/tree/8bit), [6bit](https://huggingface.co/lthn/lemer/tree/6bit), [5bit](https://huggingface.co/lthn/lemer/tree/5bit), [4bit](https://huggingface.co/lthn/lemer/tree/4bit), [mxfp8](https://huggingface.co/lthn/lemer/tree/mxfp8), [mxfp4](https://huggingface.co/lthn/lemer/tree/mxfp4), [nvfp4](https://huggingface.co/lthn/lemer/tree/nvfp4) **GGUF**: [BF16](https://huggingface.co/lthn/lemer/tree/bf16), [Q8_0](https://huggingface.co/lthn/lemer/tree/8bit), [Q6_K](https://huggingface.co/lthn/lemer/tree/6bit), [Q5_K_M](https://huggingface.co/lthn/lemer/tree/5bit), [Q4_K_M](https://huggingface.co/lthn/lemer/tree/4bit), [Q3_K_M](https://huggingface.co/lthn/lemer/tree/3bit-gguf) **HF Transformers**: on main (4-bit NF4 + bf16 in hf-bf16/) ## Base [google/gemma-4-E2B-it](https://huggingface.co/google/gemma-4-E2B-it) ## More - [lthn.ai](https://lthn.ai) - [Lethean Network](https://github.com/LetheanNetwork) ## Licence Training data and adapter: [EUPL-1.2](https://joinup.ec.europa.eu/collection/eupl/eupl-text-eupl-12) Base model: [Apache 2.0](https://www.apache.org/licenses/LICENSE-2.0)