Qwen2.5-0.5B - HQQ Quantized Model
This is a 4-bit HQQ quantized model.
Model Details
- Base Model: Qwen/Qwen2.5-0.5B
- Quantization: HQQ 4-bit
- Framework: transformers
Usage with transformers
from transformers import AutoModelForCausalLM, AutoTokenizer
model = AutoModelForCausalLM.from_pretrained("BondingAI/Qwen2.5-0.5B-hqq-4bit", device_map="auto")
tokenizer = AutoTokenizer.from_pretrained("BondingAI/Qwen2.5-0.5B-hqq-4bit")
License
Please refer to the original model card for licensing information.
- Downloads last month
- 17
Inference Providers NEW
This model isn't deployed by any Inference Provider. 🙋 Ask for provider support
Model tree for BondingAI/Qwen2.5-0.5B-hqq-4bit
Base model
Qwen/Qwen2.5-0.5B