mobiuslabsgmbh/CLIP-ViT-H-14-laion2B-2bit_g16_s128-HQQ Image Classification • Updated Aug 22, 2025 • 12 • 5
dropbox-dash/Llama-3.1-8B-Instruct_nvfp4_weights_calib_demo Text Generation • Updated Jun 26, 2025 • 4 • 1
dropbox-dash/Llama-3.1-8B-Instruct_mxfp4_weights_calib_demo Text Generation • Updated Jun 26, 2025 • 14 • 1
dropbox-dash/Llama-3.1-8B-Instruct_mxfp4_weights_calib_demo Text Generation • Updated Jun 26, 2025 • 14 • 1
view post Post 2129 Releasing HQQ Llama-3.1-70b 4-bit quantized version! Check it out at https://huggingface.co/mobiuslabsgmbh/Llama-3.1-70b-instruct_4bitgs64_hqq. Achieves 99% of the base model performance across various benchmarks! Details in the model card. 🔥 8 8 + Reply
view post Post 1805 Excited to announce the release of our high-quality Llama-3.1 8B 4-bit HQQ/calibrated quantized model! Achieving an impressive 99.3% relative performance to FP16, it also delivers the fastest inference speed for transformers. https://huggingface.co/mobiuslabsgmbh/Llama-3.1-8b-instruct_4bitgs64_hqq_calib 1 reply · 🔥 9 9 + Reply
mobiuslabsgmbh/Mixtral-8x7B-Instruct-v0.1-hf-attn-4bit-moe-3bit-metaoffload-HQQ Text Generation • Updated Feb 5, 2025 • 16 • 13