Inference Providers
Active filters: 4bit
mlx-community/Qwen2.5-7B-Instruct-kowiki-qa-4bit
Text Generation
• Updated • 9
ModelCloud/Falcon3-10B-Instruct-gptqmodel-4bit-vortex-v1
Text Generation
• 10B • Updated • 9
• 3
adriabama06/SmallThinker-3B-Preview-AWQ
Text Generation
• Updated • 2
• 1
exxocism/Linkbricks-Horizon-AI-Llama-3.3-Korean-70B-sft-dpo-GGUF
Text Generation
• Updated ehristoforu/Phi4-MoE-2x14B-Instruct
Text Generation
• 14B • Updated • 5
ModelCloud/Qwen2.5-0.5B-Instruct-gptqmodel-w4a16
Text Generation
• 0.5B • Updated • 40
• 1
ModelCloud/DeepSeek-R1-Distill-Qwen-7B-gptqmodel-4bit-vortex-v1
Text Generation
• 8B • Updated • 4
• 6
ModelCloud/DeepSeek-R1-Distill-Qwen-7B-gptqmodel-4bit-vortex-v2
Text Generation
• 8B • Updated • 229
• 8
vital-ai/watt-tool-70B-awq
71B • Updated • 1
• 4
curiousmind147/microsoft-phi-4-AWQ-4bit-GEMM
Text Generation
• 15B • Updated • 363
• 1
ConfidentialMind/Mistral-Small-24B-Instruct-2501_GPTQ_G128_W4A16_MSE
Text Classification
• 24B • Updated • 5
• 1
ConfidentialMind/Virtuoso-Medium-v2_GPTQ_G128_W4A16
Text Generation
• 33B • Updated • 3
ConfidentialMind/Virtuoso-Medium-v2_GPTQ_G32_W4A16
Text Generation
• 33B • Updated • 2
ConfidentialMind/Mistral-Small-24B-Instruct-2501_GPTQ_G32_W4A16
Text Generation
• 24B • Updated • 3
• 1
ConfidentialMind/Rombos-LLM-V2.6-Qwen-14b_GPTQ_G32_4bit_MSE
Text Generation
• 15B • Updated • 1
ConfidentialMind/Arcee-Blitz-GPTQ-G32-W4A16-MSE
Text Generation
• 24B • Updated • 2
mradermacher/Phi4-MoE-2x14B-Instruct-GGUF
14B • Updated • 57
Deepak7376/DeepSeek-R1-Distill-Qwen-1.5B-bnb-4bit
Text Generation
• 2B • Updated • 3
• 1
saishshinde15/Clyrai_Vortex_GGUF
3B • Updated • 5
Text Generation
• Updated GainEnergy/ogai-8x7b-4bit
Text Generation
• 48B • Updated • 2
• 1
ModelCloud/QwQ-32B-gptqmodel-4bit-vortex-v1
Text Generation
• 33B • Updated • 29
• 12
syubraj/DietRecommender_4bit_Qwen2.5-0.5B
Text Generation
• Updated Lowkey-Loki/Viper-Coder-v1.7-Vsm6-MLX-4bit
2B • Updated • 2
Lowkey-Loki/reka-flash-3-mlx-4bit
3B • Updated • 9
• 1
Tonic/c4ai-command-a-03-2025-4bit_nf4_double
Text Generation
• 114B • Updated • 27
Tonic/c4ai-command-a-03-2025-4bit_fp4
Text Generation
• 113B • Updated • 3
Tonic/c4ai-command-a-03-2025-4bit_nf4_no_double
Text Generation
• 113B • Updated • 3
Lowkey-Loki/Mistral-Small-3.1-24B-Instruct-TextOnly-mlx-4bit
Text Generation
• 4B • Updated • 42
adriabama06/ReaderLM-v2-AWQ
Text Generation
• 2B • Updated • 3
• 1