-
-
-
-
-
-
Inference Providers
Active filters: fp8
TevunahAi/NextCoder-32B-FP8
Text Generation
• 33B • Updated
• 4
Float16-cloud/typhoon-ocr-3b-fp8
Image-Text-to-Text
• 4B • Updated
• 6
TevunahAi/granite-8b-code-instruct-4k-FP8
Text Generation
• 8B • Updated
• 2
TevunahAi/granite-20b-code-instruct-8k-FP8
Text Generation
• 20B • Updated
• 1
TevunahAi/granite-34b-code-instruct-8k-FP8
Text Generation
• 34B • Updated
• 10
TevunahAi/Qwen3-Next-80B-A3B-Instruct-FP8
Text Generation
• Updated
• 15
vlnkane/DeepSeek-V3-4Layer
15B • Updated
• 4
unsloth/Qwen3-VL-2B-Instruct-FP8
Image-Text-to-Text
• 2B • Updated
• 63
• 1
KJML/typhoon2.5-qwen3-30b-a3b-FP8-Dynamic
Text Generation
• 31B • Updated
• 565
unsloth/Qwen3-VL-32B-Instruct-FP8
Image-Text-to-Text
• 33B • Updated
• 128
• 1
unsloth/Qwen3-VL-32B-Thinking-FP8
Image-Text-to-Text
• 33B • Updated
• 41
• 2
unsloth/Qwen3-VL-30B-A3B-Thinking-FP8
Image-Text-to-Text
• 31B • Updated
• 71
unsloth/Qwen3-VL-30B-A3B-Instruct-FP8
Image-Text-to-Text
• 31B • Updated
• 106
unsloth/Qwen3-VL-235B-A22B-Thinking-FP8
Image-Text-to-Text
• 236B • Updated
• 21
unsloth/Qwen3-VL-235B-A22B-Instruct-FP8
Image-Text-to-Text
• 236B • Updated
• 68
haydn-jones/Intern-S1-Qwen3-FP8
241B • Updated
TevunahAi/Apertus-8B-Instruct-2509-FP8
Text Generation
• 8B • Updated
• 21
FlagRelease/MiniMax-M2-FlagOS
229B • Updated
• 2
• 1
nm-testing/Llama-3.1-8B-Instruct-QKV-Cache-FP8-Per-Tensor
Updated
nm-testing/Llama-3.1-8B-Instruct-QKV-Cache-FP8-Per-Head
Updated
nm-testing/Llama-3.1-8B-Instruct-FP8-dynamic-QKV-Cache-FP8-Per-Tensor
Updated
nm-testing/Llama-3.1-8B-Instruct-FP8-dynamic-QKV-Cache-FP8-Per-Head
Updated
nm-testing/Qwen3-32B-QKV-Cache-FP8-Per-Tensor
Updated
nm-testing/Qwen3-32B-QKV-Cache-FP8-Per-Head
Updated
nm-testing/Qwen3-32B-FP8-dynamic-QKV-Cache-FP8-Per-Tensor
Updated
nm-testing/Qwen3-32B-FP8-dynamic-QKV-Cache-FP8-Per-Head
Updated
nm-testing/Llama-3.3-70B-Instruct-QKV-Cache-FP8-Per-Tensor
Updated
nm-testing/Llama-3.3-70B-Instruct-QKV-Cache-FP8-Per-Head
Updated
nm-testing/Llama-3.3-70B-Instruct-FP8-dynamic-QKV-Cache-FP8-Per-Tensor
Updated