LlamaLite-3B-TQ2_0 (GGUF Format)
This is a quantized version of meta-llama/Llama-3.2-3B-Instruct, using TQ2_0 quantization for optimized performance and reduced size. The model is stored in GGUF format for compatibility with llama.cpp and other lightweight inference engines.
Model Details
- Base Model: Llama-3.2-3B-Instruct
- Quantization Type:
TQ2_0 - Model Size: ~1.52GB
- Format: GGUF
- Intended Use: Text Generation, Chatbots, AI Assistants
- License: MIT
Download & Usage
1️⃣ Install Dependencies
pip install huggingface_hub
- Downloads last month
- 2
Hardware compatibility
Log In
to add your hardware
2-bit
Model tree for sagar27kumar/LlamaLite-3B-TQ2_0
Base model
meta-llama/Llama-3.2-3B-Instruct