-
-
-
-
-
-
Inference Providers
Active filters:
2-bit
DanyDA/Kevin-32B-exl3-2.0bpw
5B
•
Updated
•
1
BitDistiller/Llama-3.1-8B-Instruct-w2g64-gptq
8B
•
Updated
•
1
kaitchup/Qwen3-30B-A3B-autoround-2bit-gptq
31B
•
Updated
•
3
DanyDA/AM-Thinking-v1-exl3-2.0bpw
Text Generation
•
5B
•
Updated
•
1
BitDistiller/Qwen-8B-w2g64-gptq
8B
•
Updated
•
40
Erland/softpick-1.8B-4096-model-GPTQ-2bit
Text Generation
•
2B
•
Updated
Erland/vanilla-1.8B-4096-model-GPTQ-2bit
Text Generation
•
2B
•
Updated
tvpavan/sarvam-m-mlx-2Bit
Text Generation
•
2B
•
Updated
•
2
Fang77777/Llama-3.2-3B-Instruct-2bit-exllamav2
Text Generation
•
Updated
steampunque/Mistral-Small-3.1-24B-Instruct-2503-Hybrid-GGUF
24B
•
Updated
•
14
•
1
MaziyarPanahi/DeepSeek-R1-0528-Qwen3-8B-GGUF
Text Generation
•
8B
•
Updated
•
162k
•
8
Text Generation
•
1B
•
Updated
•
4
mlx-community-staging/gemma-3-1b-it-mlx-4Bit-dynamic
Text Generation
•
Updated
•
3
mlx-community-staging/gemma-3-1b-it-mlx-6Bit-dynamic
Text Generation
•
Updated
•
4
MetaphoricalCode/gemma3-27b-abliterated-dpo-exl3-2bpw-hb6
Image-Text-to-Text
•
6B
•
Updated
PepitaxX/qwen3-0.6b-gptq_2bit
Text Generation
•
0.6B
•
Updated
•
2
PepitaxX/qwen3-0.6B-openQA_prefinetune_deepseek210k_2bit
Text Generation
•
0.6B
•
Updated
•
4
TheS3b/Qwen3-EfficientQAT-w2g64
0.6B
•
Updated
irish-quant/01-ai-Yi-1.5-6B-Chat-2bit
6B
•
Updated
irish-quant/01-ai-Yi-1.5-6B-2bit
6B
•
Updated
irish-quant/01-ai-Yi-1.5-9B-Chat-2bit
9B
•
Updated
irish-quant/01-ai-Yi-1.5-9B-2bit
9B
•
Updated
irish-quant/HuggingFaceTB-SmolLM-1.7B-Instruct-2bit
2B
•
Updated
irish-quant/HuggingFaceTB-SmolLM-1.7B-2bit
2B
•
Updated
irish-quant/HuggingFaceTB-SmolLM-135M-2bit
0.1B
•
Updated
irish-quant/HuggingFaceTB-SmolLM-360M-Instruct-2bit
0.4B
•
Updated
irish-quant/HuggingFaceTB-SmolLM-360M-2bit
0.4B
•
Updated
•
1
irish-quant/meta-llama-Llama-3.1-8B-Instruct-2bit
8B
•
Updated
irish-quant/meta-llama-Llama-3.1-8B-2bit
8B
•
Updated
irish-quant/meta-llama-Llama-3.2-1B-Instruct-2bit
1B
•
Updated