Lemer-HF-bf16 — HuggingFace safetensors format

LEK-aligned Gemma 4 E2B, bf16 reference weights, in HuggingFace safetensors layout.

Converted from lthn/lemer-mlx-bf16 for use on non-MLX platforms (NVIDIA/AMD GPU, Kaggle TPU, vanilla transformers).

Source This repo
lthn/lemer-mlx-bf16 (MLX format) lthn/lemer-hf-bf16 (HF safetensors)
language_model.model.* keys model.language_model.* keys
MLX conv layout (C,K,I) / (O,H,W,I) PyTorch layout (C,I,K) / (O,I,H,W)

Weights are byte-equivalent to lemer-mlx-bf16 after the key rename + conv permutation — identical LEK alignment, identical behaviour.

Loading

from transformers import AutoModelForCausalLM, AutoTokenizer
tok = AutoTokenizer.from_pretrained("lthn/lemer-hf-bf16")
model = AutoModelForCausalLM.from_pretrained(
    "lthn/lemer-hf-bf16",
    dtype="bfloat16",
    device_map="auto",
)

License

EUPL-1.2.

Downloads last month
276
Safetensors
Model size
5B params
Tensor type
BF16
·
Inference Providers NEW
This model isn't deployed by any Inference Provider. 🙋 Ask for provider support

Model tree for lthn/lemer-hf-bf16

Finetuned
lthn/lemer
Quantized
(5)
this model

Collection including lthn/lemer-hf-bf16