Inference Providers
Active filters: chat
AIFunOver/Qwen2.5-1.5B-Instruct-openvino-4bit
Text Generation
• Updated • 4
tensorblock/Qwen1.5-MoE-A2.7B-Chat-GGUF
Text Generation
• 14B • Updated • 164
• 2
theo77186/Qwen2.5-Coder-7B-Instruct-20241106
Text Generation
• 8B • Updated • 8
• 4
tensorblock/Qwen2-0.5B-Instruct-GGUF
Text Generation
• 0.5B • Updated • 291
tensorblock/Qwen1.5-14B-Chat-GGUF
Text Generation
• 14B • Updated • 7
mradermacher/calme-3.1-llamaloi-3b-GGUF
3B • Updated • 113
async0x42/Qwen2.5.1-Coder-7B-Instruct-exl2_5.0bpw
Text Generation
• Updated • 2
mradermacher/calme-3.2-baguette-3b-GGUF
3B • Updated • 7
mradermacher/calme-3.2-instruct-3b-GGUF
3B • Updated • 28
mradermacher/calme-3.2-llamaloi-3b-GGUF
3B • Updated • 41
mradermacher/calme-3.2-qwenloi-3b-GGUF
3B • Updated • 31
mradermacher/calme-3.1-qwenloi-3b-GGUF
3B • Updated • 34
mradermacher/calme-3.3-baguette-3b-GGUF
3B • Updated • 100
mradermacher/calme-3.3-llamaloi-3b-GGUF
3B • Updated • 80
• 1
mradermacher/calme-3.3-qwenloi-3b-GGUF
3B • Updated • 105
mradermacher/calme-3.3-instruct-3b-GGUF
3B • Updated • 65
mradermacher/calme-3.1-instruct-3b-GGUF
3B • Updated • 15
mradermacher/calme-3.1-baguette-3b-GGUF
3B • Updated • 37
tensorblock/ghost-8b-beta-1608-GGUF
Text Generation
• 8B • Updated • 345
mradermacher/LongWriter-glm4-9b-abliterated-GGUF
9B • Updated • 184
• 2
mradermacher/glm-4-9b-chat-abliterated-GGUF
9B • Updated • 195
• 1
mradermacher/LongWriter-glm4-9b-abliterated-i1-GGUF
9B • Updated • 258
• 1
mradermacher/glm-4-9b-chat-abliterated-i1-GGUF
9B • Updated • 186
Qwen/Qwen2.5-Coder-0.5B-Instruct-GPTQ-Int8
Text Generation
• 0.5B • Updated • 73
• 1
Qwen/Qwen2.5-Coder-0.5B-Instruct-GPTQ-Int4
Text Generation
• 0.5B • Updated • 20
• 1
Qwen/Qwen2.5-Coder-3B-Instruct-GPTQ-Int8
Text Generation
• 3B • Updated • 113
• 1
Qwen/Qwen2.5-Coder-3B-Instruct-GPTQ-Int4
Text Generation
• 3B • Updated • 393
• 1
Qwen/Qwen2.5-Coder-14B-Instruct-GPTQ-Int8
Text Generation
• 15B • Updated • 579
• 6
Qwen/Qwen2.5-Coder-14B-Instruct-GPTQ-Int4
Text Generation
• 15B • Updated • 15.6k
• 6
Qwen/Qwen2.5-Coder-32B-Instruct-GPTQ-Int8
Text Generation
• 33B • Updated • 986
• 21