-
-
-
-
-
-
Inference Providers
Active filters: fp8
4B • Updated
• 4
medmekk/mini-deepseek-r1-4layers
15B • Updated
• 2
huihui-ai/DeepSeek-R1-bf16
Text Generation
• Updated
• 6
• 3
RedHatAI/Qwen2.5-VL-3B-Instruct-FP8-dynamic
Image-Text-to-Text
• 4B • Updated
• 10.2k
• 3
RedHatAI/Qwen2.5-VL-7B-Instruct-FP8-Dynamic
Image-Text-to-Text
• 8B • Updated
• 868
• 5
RedHatAI/Qwen2.5-VL-72B-Instruct-FP8-dynamic
Image-Text-to-Text
• 73B • Updated
• 10.4k
• 15
8B • Updated
• 1
amd-quark/llama-tiny-fp8-quant-method
17.1M • Updated
• 4.72k
71B • Updated
nm-testing/pixtral-12b-FP8-dynamic
Image-Text-to-Text
• Updated
• 1.42k
• 1
13B • Updated
13B • Updated
• 1
13B • Updated
• 904
• 10
13B • Updated
• 48
• 10
starsy/Mistral-Small-24B-Instruct-2501-FP8-Dynamic
Text Generation
• 24B • Updated
• 25
context-labs/Meta-Llama-3.1-8B-Instruct-FP8
Text Generation
• Updated
Text Generation
• Updated
• 1
chutesai/DeepSeek-V3-NextN
12B • Updated
• 2
context-labs/neuralmagic-llama-3.1-8b-instruct-FP8
Text Generation
• 8B • Updated
• 6
context-labs/neuralmagic-mistral-nemo-12b-instruct-FP8
12B • Updated
• 1
context-labs/neuralmagic-llama-3.1-70b-instruct-FP8
Text Generation
• 71B • Updated
• 3
FlagRelease/DeepSeek-R1-FlagOS-Nvidia-BF16
684B • Updated
• 7
FlagRelease/DeepSeek-R1-FlagOS-Metax-BF16
684B • Updated
• 2
thisnick/DeepSeek-R1-Distill-Llama-70B-abliterated-FP8-Dynamic
71B • Updated
• 3
RedHatAI/Mixtral-8x22B-Instruct-v0.1-FP8
141B • Updated
• 679
Text Generation
• Updated
• 2
12B • Updated
• 2.34k
• 1
12B • Updated
• 4k
• 5