-
-
-
-
-
-
Inference Providers
Active filters:
4bit
2imi9/Qwen3-1.7b-gptq-int4
Text Generation
•
0.9B
•
Updated
•
45
Text Generation
•
0.6B
•
Updated
•
3
•
1
sweatSmile/DeepSeek-R1-Distill-Qwen-1.5B-Alpaca-Instruct
2B
•
Updated
•
2
•
1
analystgatitu/economist_model_v3
Text Generation
•
Updated
•
1
analystgatitu/economist_model_v4
Text Generation
•
3B
•
Updated
•
1
KavinduHansaka/phi4-mini-bnb-4bit
Text Generation
•
4B
•
Updated
•
3
mlx-community/Apriel-1.5-15b-Thinker-3bit-MLX
Image-Text-to-Text
•
Updated
•
7
aciklab/kubernetes-ai-4bit
Image-to-Text
•
12B
•
Updated
•
1
•
2
Dhana8907/Llama-3.1-8B-Instruct-4bit
Text Generation
•
8B
•
Updated
•
1
SiddhJagani/gpt-oss-20b-no-think-mlx-Q4
Text Generation
•
21B
•
Updated
•
11
ellyfantina/llama3-medquad-lora
MightyOctopus/qwen3-0.6B-lora-medical
Updated
iMiW/Giga-Embeddings-instruct-4bit-nf4
Feature Extraction
•
4B
•
Updated
•
906
ModelCloud/GLM-4.6-GPTQMODEL-W4A16-v1
Text Generation
•
357B
•
Updated
•
2
ModelCloud/GLM-4.6-GPTQMODEL-W4A16-v2
Text Generation
•
357B
•
Updated
•
2
•
1
MidnightPhreaker/KAT-Dev-72B-Exp-GPTQ-INT4-gs32-0.01
13B
•
Updated
•
1
•
1
MidnightPhreaker/KAT-Dev-72B-Exp-GPTQ-INT4-gs32
13B
•
Updated
MidnightPhreaker/KAT-Dev-72B-Exp-GPTQ-INT4-gs128
12B
•
Updated
MidnightPhreaker/GLM-4.5-Air-REAP-82B-A12B-GPTQ-INT4-gs32
14B
•
Updated
•
39
•
6
sweatSmile/Gemma-2-2B-MedicalQA-Assistant
Text Generation
•
3B
•
Updated
•
26
•
1
ModelCloud/GLM-4.6-REAP-268B-A32B-GPTQMODEL-W4A16
Text Generation
•
269B
•
Updated
•
47
•
2
mradermacher/Gemma-2-2B-MedicalQA-Assistant-GGUF
3B
•
Updated
•
334
toddie314/toddric_v2_merged
Text Generation
•
8B
•
Updated
•
6
sanchezalonsodavid17/DeepSeek-OCR-MBQ-Quantized-v1
Image-Text-to-Text
•
3B
•
Updated
•
619
•
6
ModelCloud/MiniMax-M2-GPTQMODEL-W4A16
Text Generation
•
229B
•
Updated
•
22
•
3
ModelCloud/Marin-32B-Base-GPTQMODEL-W4A16
Text Generation
•
33B
•
Updated
•
6
•
1
ModelCloud/Marin-32B-Base-GPTQMODEL-AWQ-W4A16
Text Generation
•
33B
•
Updated
•
4
•
1
ModelCloud/Granite-4.0-H-1B-GPTQMODEL-W4A16
Text Generation
•
1B
•
Updated
•
4