File size: 1,399 Bytes
a583c51
 
d320bab
 
a583c51
d320bab
 
 
 
 
39a4bd4
 
d320bab
 
 
 
 
a583c51
d320bab
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
38
---
tags:
- gemma4
- lemma
- mlx
- 4bit
- apple-silicon
- multimodal
- on-device
- conversational
pipeline_tag: image-text-to-text
library_name: mlx
license: eupl-1.2
license_link: https://ai.google.dev/gemma/docs/gemma_4_license
base_model:
- lthn/lemrd
base_model_relation: quantized
---

# Lemrd — Gemma 4 31B Dense — MLX 4-bit

The largest dense member of the Lemma model family by Lethean. An EUPL-1.2 fork of Gemma 4 31B with the Lethean Ethical Kernel (LEK) merged into the weights.

This repo hosts the **MLX 4-bit** build for native Apple Silicon inference via [`mlx-lm`](https://github.com/ml-explore/mlx-lm) and [`mlx-vlm`](https://github.com/Blaizzy/mlx-vlm). For the GGUF playground (Ollama, llama.cpp) see [`lthn/lemrd`](https://huggingface.co/lthn/lemrd). For the unmodified Google base see [`LetheanNetwork/lemrd`](https://huggingface.co/LetheanNetwork/lemrd).

## Family

| Repo | Format | Bits |
|---|---|---|
| [`lthn/lemrd`](https://huggingface.co/lthn/lemrd) | GGUF multi-quant | Q4_K_M → BF16 |
| [`lthn/lemrd-mlx`](https://huggingface.co/lthn/lemrd-mlx) | MLX | 4-bit |
| [`lthn/lemrd-mlx-8bit`](https://huggingface.co/lthn/lemrd-mlx-8bit) | MLX | 8-bit |
| [`lthn/lemrd-mlx-bf16`](https://huggingface.co/lthn/lemrd-mlx-bf16) | MLX | bf16 |

## License

EUPL-1.2. See [Gemma Terms of Use](https://ai.google.dev/gemma/docs/gemma_4_license) for upstream base model terms.