LlamaLite-3B-TQ2_0 (GGUF Format)

This is a quantized version of meta-llama/Llama-3.2-3B-Instruct, using TQ2_0 quantization for optimized performance and reduced size. The model is stored in GGUF format for compatibility with llama.cpp and other lightweight inference engines.

Model Details

  • Base Model: Llama-3.2-3B-Instruct
  • Quantization Type: TQ2_0
  • Model Size: ~1.52GB
  • Format: GGUF
  • Intended Use: Text Generation, Chatbots, AI Assistants
  • License: MIT

Download & Usage

1️⃣ Install Dependencies

pip install huggingface_hub
Downloads last month
2
GGUF
Model size
3B params
Architecture
llama
Hardware compatibility
Log In to add your hardware

2-bit

Inference Providers NEW
This model isn't deployed by any Inference Provider. 🙋 Ask for provider support

Model tree for sagar27kumar/LlamaLite-3B-TQ2_0

Quantized
(414)
this model