Inference Providers
Active filters: fp8
anyisalin/lzlv_70b_fp16_hf-FP8-D
Text Generation
• 69B • Updated • 5
anyisalin/Meta-Llama-3-70B-Instruct-FP8-D
Text Generation
• 71B • Updated • 8
anyisalin/Mixtral-8x7B-Instruct-v0.1-FP8-D
Text Generation
• 47B • Updated • 4
pcmoritz/Mixtral-8x7B-v0.1-fp8-act-scale
Text Generation
• 47B • Updated • 3
anyisalin/Meta-Llama-3-70B-Instruct-FP8
Text Generation
• 71B • Updated • 3
RedHatAI/Meta-Llama-3-8B-Instruct-FP8-KV
Text Generation
• 8B • Updated • 12.7k
• • 8
comaniac/Meta-Llama-3-8B-Instruct-FP8-v1
Text Generation
• 8B • Updated • 5
comaniac/Mixtral-8x22B-Instruct-v0.1-FP8-v1
Text Generation
• 141B • Updated • 8
RedHatAI/Meta-Llama-3-70B-Instruct-FP8
Text Generation
• 71B • Updated • 4.47k
• • 13
comaniac/Meta-Llama-3-70B-Instruct-FP8-v1
Text Generation
• 71B • Updated • 9
comaniac/Mixtral-8x7B-Instruct-v0.1-FP8-v1
Text Generation
• 47B • Updated • 7
comaniac/Mixtral-8x7B-Instruct-v0.1-FP8-v2
Text Generation
• 47B • Updated • 7
Skywork/Skywork-MoE-Base-FP8
Text Generation
• 146B • Updated • 35
• 7
RedHatAI/Qwen2-72B-Instruct-FP8
Text Generation
• 73B • Updated • 929
• 15
comaniac/Meta-Llama-3-70B-Instruct-FP8-v2
Text Generation
• 71B • Updated • 8
comaniac/Mixtral-8x7B-Instruct-v0.1-FP8-v3
Text Generation
• 47B • Updated • 3
comaniac/Mixtral-8x22B-Instruct-v0.1-FP8-v2
Text Generation
• 141B • Updated • 6
RedHatAI/Mixtral-8x22B-Instruct-v0.1-AutoFP8
Text Generation
• 141B • Updated • 179
• 3
Text Generation
• 8B • Updated RedHatAI/Qwen2-1.5B-Instruct-FP8
Text Generation
• 2B • Updated • 21.8k
RedHatAI/Qwen2-7B-Instruct-FP8
Text Generation
• 8B • Updated • 8.12k
• • 2
anyisalin/L3-70B-Euryale-v2.1-FP8
Text Generation
• 71B • Updated • 21
yentinglin/Llama-3-Taiwan-70B-Instruct-FP8
Text Generation
• 71B • Updated • 3
kuotient/llama3-instrucTrans-enko-8b-FP8
Text Generation
• 8B • Updated • 3
• 2
FlorianJc/Hermes-2-Pro-Mistral-7B-vllm-fp8
Text Generation
• 7B • Updated FlorianJc/openchat-3.6-8b-20240522-vllm-fp8
Text Generation
• 8B • Updated • 1
FlorianJc/Llama3-ChatQA-1.5-8B-vllm-fp8
Text Generation
• 8B • Updated TechxGenus/Codestral-22B-v0.1-FP8
Text Generation
• 22B • Updated • 10
Model-SafeTensors/Meta-Llama-3-70B-FP8-Dynamic
Text Generation
• 71B • Updated • 4
Model-SafeTensors/Qwen-Qwen2-72B-FP8-Dynamic
Text Generation
• 73B • Updated • 3