Inference Providers
Active filters: 4bit
legraphista/Qwen2.5-1.5B-Instruct-IMat-GGUF
Text Generation
• 2B • Updated • 258
legraphista/Qwen2.5-3B-Instruct-IMat-GGUF
Text Generation
• 3B • Updated • 1.4k
legraphista/Qwen2.5-7B-Instruct-IMat-GGUF
Text Generation
• 8B • Updated • 88
legraphista/Qwen2.5-14B-Instruct-IMat-GGUF
Text Generation
• 15B • Updated • 643
ModelCloud/GRIN-MoE-gptq-4bit
42B • Updated • 5
• 6
legraphista/Qwen2.5-32B-Instruct-IMat-GGUF
Text Generation
• 33B • Updated • 1.87k
legraphista/Qwen2.5-Coder-1.5B-Instruct-IMat-GGUF
Text Generation
• 2B • Updated • 340
legraphista/Qwen2.5-Math-1.5B-Instruct-IMat-GGUF
Text Generation
• 2B • Updated • 423
legraphista/Qwen2.5-Coder-7B-Instruct-IMat-GGUF
Text Generation
• 8B • Updated • 194
legraphista/Qwen2.5-Math-7B-Instruct-IMat-GGUF
Text Generation
• 8B • Updated • 355
legraphista/Qwen2.5-72B-Instruct-IMat-GGUF
Text Generation
• 73B • Updated • 250
legraphista/Llama-3.2-1B-Instruct-IMat-GGUF
Text Generation
• 1B • Updated • 401
legraphista/Llama-3.2-3B-Instruct-IMat-GGUF
Text Generation
• 3B • Updated • 416
• 2
Narrator5000/llavanext-finetuned-stackoverflow-vqa
Updated • 9
• 1
NeoChen1024/internlm2_5-20b-chat-exl2-4.25bpw-h8
Text Generation
• Updated • 2
ussipan/SipanGPT-0.1-Llama-3.2-1B-GGUF
Text Generation
• 1B • Updated • 843
• 1
ussipan/SipanGPT-0.2-Llama-3.2-1B-GGUF
Text Generation
• 1B • Updated • 18
mcavus/glm-4v-9b-gptq-4bit-dynamo
Updated • 6
• 1
ussipan/SipanGPT-0.3-Llama-3.2-1B-GGUF
Text Generation
• 1B • Updated • 37
• 1
harishnair04/Gemma-medtr-2b-sft
Text Generation
• 3B • Updated harishnair04/Gemma-medtr-2b-sft-v2
Text Generation
• 3B • Updated • 4
mradermacher/Gemma-medtr-2b-sft-v2-GGUF
3B • Updated • 60
NaomiBTW/L3-8B-Lunaris-v1-GPTQ
Text Generation
• Updated ModelCloud/Qwen2.5-Coder-32B-Instruct-gptqmodel-4bit-vortex-v1
Text Generation
• 33B • Updated • 80
• 16
Rakushaking/llm-jp-3-13b-it
ModelCloud/QwQ-32B-Preview-gptqmodel-4bit-vortex-v1
Text Generation
• 33B • Updated • 7
• 51
nisten/qwen2.5-coder-7b-abliterated-128k-AWQ
Text Generation
• 8B • Updated • 2.4k
• 1
ModelCloud/QwQ-32B-Preview-gptqmodel-4bit-vortex-v2
Text Generation
• 33B • Updated • 11
• 16
ModelCloud/QwQ-32B-Preview-gptqmodel-4bit-vortex-v3
Text Generation
• 33B • Updated • 5
• 14
mlx-community/Qwen2.5-7B-Instruct-kowiki-qa-4bit
Text Generation
• Updated • 9