Lemmy โ€” Gemma 4 26B A4B MoE โ€” MLX 8-bit

The Mixture-of-Experts member of the Lemma model family by Lethean. An EUPL-1.2 fork of Gemma 4 26B A4B with the Lethean Ethical Kernel (LEK) merged into the weights.

This repo hosts the MLX 8-bit build for native Apple Silicon inference via mlx-lm and mlx-vlm. For the GGUF playground (Ollama, llama.cpp) see lthn/lemmy. For the unmodified Google base see LetheanNetwork/lemmy.

Family

Repo Format Bits
lthn/lemmy GGUF multi-quant Q4_K_M โ†’ BF16
lthn/lemmy-mlx MLX 4-bit
lthn/lemmy-mlx-8bit MLX 8-bit
lthn/lemmy-mlx-bf16 MLX bf16

License

EUPL-1.2. See Gemma Terms of Use for upstream base model terms.

Downloads last month
298
Safetensors
Model size
25B params
Tensor type
BF16
ยท
U32
ยท
MLX
Hardware compatibility
Log In to add your hardware

8-bit

Inference Providers NEW
This model isn't deployed by any Inference Provider. ๐Ÿ™‹ Ask for provider support

Model tree for lthn/lemmy-mlx-8bit

Finetuned
lthn/lemmy
Quantized
(3)
this model

Collection including lthn/lemmy-mlx-8bit