Qwen2.5-0.5B - HQQ Quantized Model

This is a 4-bit HQQ quantized model.

Model Details

  • Base Model: Qwen/Qwen2.5-0.5B
  • Quantization: HQQ 4-bit
  • Framework: transformers

Usage with transformers

from transformers import AutoModelForCausalLM, AutoTokenizer

model = AutoModelForCausalLM.from_pretrained("BondingAI/Qwen2.5-0.5B-hqq-4bit", device_map="auto")
tokenizer = AutoTokenizer.from_pretrained("BondingAI/Qwen2.5-0.5B-hqq-4bit")

License

Please refer to the original model card for licensing information.

Downloads last month
17
Inference Providers NEW
This model isn't deployed by any Inference Provider. 🙋 Ask for provider support

Model tree for BondingAI/Qwen2.5-0.5B-hqq-4bit

Base model

Qwen/Qwen2.5-0.5B
Finetuned
(541)
this model