Assistant-100M-Guardian
This is a custom 124M parameter language model built from scratch on the Llama architecture. It is just an expirement model doesnot give answer.
Model Details
- Architecture: Custom Llama-based (12 Layers, 12 Heads, 768 Dim)
- Parameter Count: 124M
- Format: Safetensors
Usage & Formatting
Because this model uses a custom architecture class (LlamaNano), you must initialize your local PyTorch class first, and then load these Safetensor weights into it.
Stop Sequence: It is critical to update your generation script to use <|endoftext|> as the stop sequence. This ensures the model cleanly ends its turn after responding.
- Downloads last month
- 60
Hardware compatibility
Log In to add your hardware
16-bit