YAML Metadata
Warning:
empty or missing yaml metadata in repo card
(https://huggingface.co/docs/hub/model-cards#model-card-metadata)
Enhanced Hybrid Transformer 416M
π 416,417,792 parameter transformer with modern optimizations.
Features
- 24 layers Γ 16 heads
- GQA-4 (Grouped Query Attention)
- SwiGLU activation
- RMSNorm normalization
- RoPE positional embeddings
Contents
pytorch_model.bin- Model weightsconfig.json- Model configurationtokenizer.json- Tokenizer filesREADME.md- This file
Usage
Load with the original repository code for full functionality.
π Generated with Claude Code
- Downloads last month
- 9
Inference Providers
NEW
This model isn't deployed by any Inference Provider.
π
Ask for provider support