Assistant-100M-Guardian

This is a custom 124M parameter language model built from scratch on the Llama architecture. It is just an expirement model doesnot give answer.

Model Details

  • Architecture: Custom Llama-based (12 Layers, 12 Heads, 768 Dim)
  • Parameter Count: 124M
  • Format: Safetensors

Usage & Formatting

Because this model uses a custom architecture class (LlamaNano), you must initialize your local PyTorch class first, and then load these Safetensor weights into it.

Stop Sequence: It is critical to update your generation script to use <|endoftext|> as the stop sequence. This ensures the model cleanly ends its turn after responding.

Downloads last month
60
GGUF
Model size
0.2B params
Architecture
llama
Hardware compatibility
Log In to add your hardware

16-bit

Inference Providers NEW
This model isn't deployed by any Inference Provider. ๐Ÿ™‹ Ask for provider support