| Qwen/Qwen3.5-397B-A17B Qwen3.5-397B-A17B | | $0.60 | $3.60 | 262,144 | 0.76 | 62 | Yes | No |
| Qwen/Qwen3.5-397B-A17B Qwen3.5-397B-A17B | | $0.60 | $3.60 | 262,144 | 0.74 | 73 | Yes | No |
| zai-org/GLM-5 GLM-5 | | $1.00 | $3.20 | 202,800 | 1.35 | 30 | Yes | No |
| zai-org/GLM-5 GLM-5 | | $1.00 | $3.20 | 202,752 | 0.50 | 37 | Yes | Yes |
| zai-org/GLM-5 GLM-5 | | - | - | - | 4.08 | 29 | No | No |
| MiniMaxAI/MiniMax-M2.5 MiniMax-M2.5 | | $0.30 | $1.20 | 204,800 | 1.53 | 38 | Yes | No |
| moonshotai/Kimi-K2.5 Kimi-K2.5 | | $0.60 | $3.00 | 262,144 | 1.26 | 27 | No | Yes |
| moonshotai/Kimi-K2.5 Kimi-K2.5 | | $0.50 | $2.80 | 262,144 | 0.97 | 37 | Yes | Yes |
| moonshotai/Kimi-K2.5 Kimi-K2.5 | | - | - | 262,144 | 0.71 | 130 | Yes | No |
| Qwen/Qwen3-Coder-Next Qwen3-Coder-Next | | $0.20 | $1.50 | 262,144 | 0.92 | 140 | Yes | Yes |
| zai-org/GLM-4.7-Flash GLM-4.7-Flash | | $0.07 | $0.40 | 200,000 | 0.86 | 70 | Yes | No |
| zai-org/GLM-4.7-Flash GLM-4.7-Flash | | - | - | - | - | - | - | - |
| openai/gpt-oss-20b gpt-oss-20b | | $0.10 | $0.50 | 131,072 | 0.08 | 855 | Yes | No |
| openai/gpt-oss-20b gpt-oss-20b | | $0.04 | $0.15 | 131,072 | 0.75 | 145 | No | No |
| openai/gpt-oss-20b gpt-oss-20b | | $0.05 | $0.20 | 131,072 | 0.79 | 113 | Yes | Yes |
| openai/gpt-oss-20b gpt-oss-20b | | $0.10 | $0.10 | 131,072 | 0.54 | 98 | Yes | No |
| openai/gpt-oss-20b gpt-oss-20b | | $0.05 | $0.20 | 131,072 | 0.17 | 105 | Yes | Yes |
| openai/gpt-oss-20b gpt-oss-20b | | $0.05 | $0.20 | 131,072 | 0.28 | 170 | Yes | No |
| openai/gpt-oss-20b gpt-oss-20b | | $0.05 | $0.18 | 131,072 | 0.31 | 59 | Yes | Yes |
| openai/gpt-oss-120b gpt-oss-120b | | $0.15 | $0.75 | 131,072 | 0.10 | 450 | Yes | No |
| openai/gpt-oss-120b gpt-oss-120b | | $0.05 | $0.25 | 131,072 | 0.46 | 76 | Yes | Yes |
| openai/gpt-oss-120b gpt-oss-120b | | $0.25 | $0.69 | - | 0.21 | 1,188 | Yes | No |
| openai/gpt-oss-120b gpt-oss-120b | | - | - | - | - | - | - | - |
| openai/gpt-oss-120b gpt-oss-120b | | $0.10 | $0.40 | 131,072 | 0.75 | 85 | Yes | Yes |
| openai/gpt-oss-120b gpt-oss-120b | | - | - | - | - | - | - | - |
| openai/gpt-oss-120b gpt-oss-120b | | $0.15 | $0.60 | 131,072 | 0.43 | 67 | Yes | Yes |
| openai/gpt-oss-120b gpt-oss-120b | | $0.15 | $0.60 | 131,072 | 0.28 | 100 | Yes | No |
| openai/gpt-oss-120b gpt-oss-120b | | - | - | - | 0.38 | 120 | Yes | Yes |
| openai/gpt-oss-120b gpt-oss-120b | | $0.09 | $0.47 | 131,072 | 0.35 | 88 | Yes | Yes |
| meta-llama/Llama-3.1-8B-Instruct Llama-3.1-8B-Instruct | | $0.02 | $0.05 | 16,384 | 0.64 | 75 | No | No |
| meta-llama/Llama-3.1-8B-Instruct Llama-3.1-8B-Instruct | | $0.10 | $0.10 | - | 0.25 | 1,355 | No | No |
| meta-llama/Llama-3.1-8B-Instruct Llama-3.1-8B-Instruct | | $0.10 | $0.20 | 16,384 | 0.31 | 761 | Yes | Yes |
| meta-llama/Llama-3.1-8B-Instruct Llama-3.1-8B-Instruct | | $0.06 | $0.06 | 131,072 | 0.83 | 60 | No | Yes |
| meta-llama/Llama-3.1-8B-Instruct Llama-3.1-8B-Instruct | | - | - | - | - | - | - | - |
| meta-llama/Llama-3.1-8B-Instruct Llama-3.1-8B-Instruct | | - | - | - | 0.28 | 114 | Yes | Yes |
| meta-llama/Llama-3.1-8B-Instruct Llama-3.1-8B-Instruct | | $0.11 | $0.11 | 131,072 | 0.38 | 42 | Yes | No |
| deepseek-ai/DeepSeek-V3.2 DeepSeek-V3.2 | | $0.27 | $0.40 | 163,840 | 1.23 | 33 | Yes | Yes |
| deepseek-ai/DeepSeek-V3.2 DeepSeek-V3.2 | | - | - | 163,840 | 0.94 | 67 | Yes | No |
| google/gemma-3-27b-it gemma-3-27b-it | | - | - | - | 0.45 | 47 | Yes | Yes |
| deepseek-ai/DeepSeek-OCR DeepSeek-OCR | | $0.03 | $0.03 | 8,192 | 0.68 | 156 | No | No |
| Qwen/Qwen2.5-7B-Instruct Qwen2.5-7B-Instruct | | $0.30 | $0.30 | 32,768 | 0.21 | 146 | Yes | No |
| Qwen/Qwen3-4B-Instruct-2507 Qwen3-4B-Instruct-2507 | | $0.01 | $0.03 | 262,144 | 0.73 | 87 | Yes | No |
| deepseek-ai/DeepSeek-R1 DeepSeek-R1 | | $0.70 | $2.50 | 64,000 | 1.22 | 25 | Yes | No |
| deepseek-ai/DeepSeek-R1 DeepSeek-R1 | | - | - | - | 3.76 | 98 | Yes | Yes |
| deepseek-ai/DeepSeek-R1 DeepSeek-R1 | | $2.00 | $2.00 | 163,840 | 0.80 | 39 | No | No |
| deepseek-ai/DeepSeek-R1 DeepSeek-R1 | | $3.00 | $7.00 | 163,840 | 0.61 | 40 | No | Yes |
| deepseek-ai/DeepSeek-R1 DeepSeek-R1 | | - | - | - | - | - | - | - |
| Qwen/Qwen3-8B Qwen3-8B | | $0.07 | $0.18 | 40,960 | 0.86 | 59 | Yes | No |
| Qwen/Qwen3-VL-8B-Instruct Qwen3-VL-8B-Instruct | | $0.08 | $0.50 | 131,072 | 0.58 | 65 | Yes | Yes |
| Qwen/Qwen3-VL-8B-Instruct Qwen3-VL-8B-Instruct | | $0.18 | $0.68 | 262,144 | 0.22 | 93 | Yes | Yes |
| Qwen/Qwen3-Coder-Next-FP8 Qwen3-Coder-Next-FP8 | | $0.50 | $1.20 | 262,144 | 1.39 | 117 | Yes | Yes |
| moonshotai/Kimi-K2-Thinking Kimi-K2-Thinking | | $0.60 | $2.50 | 262,144 | 0.81 | 49 | Yes | No |
| moonshotai/Kimi-K2-Thinking Kimi-K2-Thinking | | $1.20 | $4.00 | 262,144 | 2.88 | 91 | Yes | Yes |
| meta-llama/Llama-3.2-3B-Instruct Llama-3.2-3B-Instruct | | - | - | - | - | - | - | - |
| meta-llama/Llama-3.2-3B-Instruct Llama-3.2-3B-Instruct | | - | - | - | - | - | - | - |
| meta-llama/Llama-3.2-3B-Instruct Llama-3.2-3B-Instruct | | $0.10 | $0.10 | 131,072 | 2.03 | 81 | No | No |
| meta-llama/Llama-3.2-3B-Instruct Llama-3.2-3B-Instruct | | $0.06 | $0.06 | 131,072 | 1.28 | 77 | Yes | Yes |
| meta-llama/Llama-3.3-70B-Instruct Llama-3.3-70B-Instruct | | $0.59 | $0.79 | 131,072 | 0.21 | 182 | Yes | No |
| meta-llama/Llama-3.3-70B-Instruct Llama-3.3-70B-Instruct | | $0.14 | $0.40 | 131,072 | 0.63 | 46 | Yes | No |
| meta-llama/Llama-3.3-70B-Instruct Llama-3.3-70B-Instruct | | - | - | - | - | - | - | - |
| meta-llama/Llama-3.3-70B-Instruct Llama-3.3-70B-Instruct | | $0.60 | $1.20 | 131,072 | 0.53 | 292 | Yes | Yes |
| meta-llama/Llama-3.3-70B-Instruct Llama-3.3-70B-Instruct | | $0.40 | $0.40 | 131,072 | 0.98 | 17 | No | Yes |
| meta-llama/Llama-3.3-70B-Instruct Llama-3.3-70B-Instruct | | $0.40 | $0.40 | 131,072 | 0.69 | 121 | Yes | No |
| meta-llama/Llama-3.3-70B-Instruct Llama-3.3-70B-Instruct | | $0.88 | $0.88 | 131,072 | 1.07 | 140 | No | Yes |
| meta-llama/Llama-3.3-70B-Instruct Llama-3.3-70B-Instruct | | - | - | - | - | - | - | - |
| meta-llama/Llama-3.3-70B-Instruct Llama-3.3-70B-Instruct | | - | - | - | 0.29 | 78 | Yes | Yes |
| meta-llama/Llama-3.3-70B-Instruct Llama-3.3-70B-Instruct | | $0.74 | $0.74 | 131,072 | 0.41 | 24 | Yes | Yes |
| Qwen/Qwen2.5-Coder-7B-Instruct Qwen2.5-Coder-7B-Instruct | | $0.01 | $0.03 | 131,072 | 0.82 | 61 | No | Yes |
| deepseek-ai/DeepSeek-V3.2-Exp DeepSeek-V3.2-Exp | | $0.27 | $0.41 | 163,840 | 0.69 | 32 | Yes | Yes |
| zai-org/GLM-4.7 GLM-4.7 | | - | - | - | - | - | - | - |
| zai-org/GLM-4.7 GLM-4.7 | | - | - | - | - | - | - | - |
| zai-org/GLM-4.7 GLM-4.7 | | - | - | - | 2.30 | 59 | Yes | No |
| Qwen/Qwen3-4B-Thinking-2507 Qwen3-4B-Thinking-2507 | | $0.01 | $0.03 | 262,144 | 0.84 | 91 | Yes | No |
| Qwen/Qwen3-32B Qwen3-32B | | $0.29 | $0.59 | 131,072 | 0.16 | 260 | Yes | No |
| Qwen/Qwen3-32B Qwen3-32B | | $0.10 | $0.45 | 40,960 | 0.54 | 42 | No | No |
| Qwen/Qwen3-32B Qwen3-32B | | - | - | - | - | - | - | - |
| Qwen/Qwen3-32B Qwen3-32B | | $0.40 | $0.80 | 32,768 | 2.62 | 195 | Yes | Yes |
| Qwen/Qwen3-32B Qwen3-32B | | $0.08 | $0.25 | 40,960 | 0.99 | 27 | Yes | Yes |
| Qwen/Qwen3-32B Qwen3-32B | | $0.09 | $0.25 | 32,768 | 0.42 | 40 | Yes | Yes |
| moonshotai/Kimi-K2-Instruct-0905 Kimi-K2-Instruct-0905 | | - | - | 262,144 | 0.11 | 195 | Yes | No |
| moonshotai/Kimi-K2-Instruct-0905 Kimi-K2-Instruct-0905 | | $0.60 | $2.50 | 262,144 | 1.09 | 24 | Yes | Yes |
| moonshotai/Kimi-K2-Instruct-0905 Kimi-K2-Instruct-0905 | | $1.00 | $3.00 | 262,144 | 0.52 | 42 | Yes | Yes |
| meta-llama/Llama-3.2-1B-Instruct Llama-3.2-1B-Instruct | | - | - | 131,000 | 0.58 | 190 | No | No |
| meta-llama/Llama-3.2-1B-Instruct Llama-3.2-1B-Instruct | | - | - | - | - | - | - | - |
| meta-llama/Llama-3.2-1B-Instruct Llama-3.2-1B-Instruct | | - | - | - | - | - | - | - |
| XiaomiMiMo/MiMo-V2-Flash MiMo-V2-Flash | | $0.10 | $0.30 | 262,144 | 1.31 | 81 | Yes | No |
| Qwen/Qwen3-Next-80B-A3B-Instruct Qwen3-Next-80B-A3B-Instruct | | $0.15 | $1.50 | 131,072 | 1.60 | 72 | Yes | No |
| Qwen/Qwen3-Next-80B-A3B-Instruct Qwen3-Next-80B-A3B-Instruct | | $0.30 | $0.30 | 262,144 | 0.53 | 176 | Yes | No |
| Qwen/Qwen3-Next-80B-A3B-Instruct Qwen3-Next-80B-A3B-Instruct | | $0.15 | $1.50 | 262,144 | 0.40 | 120 | Yes | Yes |
| allenai/Molmo2-8B Molmo2-8B | | - | - | - | 1.78 | 89 | No | Yes |
| Qwen/Qwen3-Coder-30B-A3B-Instruct Qwen3-Coder-30B-A3B-Instruct | | - | - | - | - | - | - | - |
| Qwen/Qwen3-Coder-30B-A3B-Instruct Qwen3-Coder-30B-A3B-Instruct | | - | - | - | 0.48 | 50 | Yes | No |
| Qwen/Qwen3-Coder-30B-A3B-Instruct Qwen3-Coder-30B-A3B-Instruct | | $0.07 | $0.26 | 262,144 | 0.47 | 45 | Yes | Yes |
| google/gemma-3n-E4B-it gemma-3n-E4B-it | | $0.02 | $0.04 | 32,768 | 0.30 | 51 | No | Yes |
| deepseek-ai/DeepSeek-R1-Distill-Qwen-32B DeepSeek-R1-Distill-Qwen-32B | | - | - | - | - | - | - | - |
| deepseek-ai/DeepSeek-R1-Distill-Qwen-32B DeepSeek-R1-Distill-Qwen-32B | | $0.30 | $0.30 | 131,072 | 0.89 | 28 | No | Yes |
| nvidia/NVIDIA-Nemotron-Nano-9B-v2 NVIDIA-Nemotron-Nano-9B-v2 | | $0.06 | $0.25 | 131,072 | 0.47 | 112 | Yes | Yes |
| mistralai/Mistral-7B-Instruct-v0.2 Mistral-7B-Instruct-v0.2 | | $0.20 | $0.20 | 32,768 | 0.20 | 85 | Yes | Yes |
| Qwen/Qwen2.5-VL-7B-Instruct Qwen2.5-VL-7B-Instruct | | $0.20 | $0.20 | 32,768 | 0.52 | 74 | No | No |
| MiniMaxAI/MiniMax-M2.1 MiniMax-M2.1 | | $0.30 | $1.20 | 204,800 | 3.23 | 30 | Yes | No |
| Qwen/Qwen3-VL-32B-Instruct Qwen3-VL-32B-Instruct | | $0.50 | $1.50 | 262,144 | 0.91 | 68 | No | Yes |
| ServiceNow-AI/Apriel-1.6-15b-Thinker Apriel-1.6-15b-Thinker | | $0.00 | $0.00 | 131,072 | 0.16 | 138 | No | No |
| meta-llama/Meta-Llama-3-8B-Instruct Meta-Llama-3-8B-Instruct | | - | - | - | - | - | - | - |
| meta-llama/Meta-Llama-3-8B-Instruct Meta-Llama-3-8B-Instruct | | $0.04 | $0.04 | 8,192 | 0.60 | 77 | No | No |
| meta-llama/Meta-Llama-3-8B-Instruct Meta-Llama-3-8B-Instruct | | - | - | - | - | - | - | - |
| CohereLabs/command-a-reasoning-08-2025 command-a-reasoning-08-2025 | | - | - | - | 0.16 | 81 | Yes | No |
| Qwen/Qwen3-VL-30B-A3B-Instruct Qwen3-VL-30B-A3B-Instruct | | $0.20 | $0.70 | 131,072 | 0.72 | 120 | Yes | Yes |
| deepseek-ai/DeepSeek-R1-Distill-Llama-70B DeepSeek-R1-Distill-Llama-70B | | - | - | - | - | - | - | - |
| deepseek-ai/DeepSeek-R1-Distill-Llama-70B DeepSeek-R1-Distill-Llama-70B | | $0.80 | $0.80 | 8,192 | 0.38 | 65 | No | No |
| deepseek-ai/DeepSeek-R1-Distill-Llama-70B DeepSeek-R1-Distill-Llama-70B | | $0.70 | $1.40 | 131,072 | 0.79 | 213 | No | No |
| deepseek-ai/DeepSeek-R1-Distill-Llama-70B DeepSeek-R1-Distill-Llama-70B | | $0.75 | $0.75 | 131,072 | 0.82 | 16 | No | Yes |
| deepseek-ai/DeepSeek-R1-Distill-Llama-70B DeepSeek-R1-Distill-Llama-70B | | - | - | - | - | - | - | - |
| deepseek-ai/DeepSeek-R1-Distill-Llama-70B DeepSeek-R1-Distill-Llama-70B | | - | - | - | 0.54 | 38 | No | Yes |
| deepseek-ai/DeepSeek-R1-Distill-Llama-70B DeepSeek-R1-Distill-Llama-70B | | - | - | - | - | - | - | - |
| Qwen/Qwen2.5-Coder-32B-Instruct Qwen2.5-Coder-32B-Instruct | | $0.06 | $0.20 | 131,072 | 0.88 | 28 | No | Yes |
| Qwen/Qwen2.5-Coder-32B-Instruct Qwen2.5-Coder-32B-Instruct | | - | - | - | - | - | - | - |
| Qwen/Qwen2.5-Coder-32B-Instruct Qwen2.5-Coder-32B-Instruct | | - | - | - | - | - | - | - |
| Qwen/Qwen2.5-Coder-32B-Instruct Qwen2.5-Coder-32B-Instruct | | - | - | - | 0.38 | 57 | Yes | No |
| meta-llama/Llama-3.1-70B-Instruct Llama-3.1-70B-Instruct | | - | - | - | - | - | - | - |
| meta-llama/Llama-3.1-70B-Instruct Llama-3.1-70B-Instruct | | - | - | - | 0.37 | 80 | Yes | Yes |
| deepseek-ai/DeepSeek-R1-0528 DeepSeek-R1-0528 | | $0.70 | $2.50 | 163,840 | 1.23 | 28 | Yes | No |
| deepseek-ai/DeepSeek-R1-0528 DeepSeek-R1-0528 | | $5.00 | $7.00 | 131,072 | 1.07 | 110 | Yes | Yes |
| deepseek-ai/DeepSeek-R1-0528 DeepSeek-R1-0528 | | $3.00 | $3.00 | 163,840 | 0.82 | 42 | No | No |
| deepseek-ai/DeepSeek-R1-0528 DeepSeek-R1-0528 | | $0.00 | $0.00 | 163,840 | 0.62 | 39 | No | Yes |
| deepseek-ai/DeepSeek-R1-0528 DeepSeek-R1-0528 | | - | - | - | - | - | - | - |
| Qwen/QwQ-32B QwQ-32B | | - | - | - | - | - | - | - |
| Qwen/QwQ-32B QwQ-32B | | - | - | - | - | - | - | - |
| Qwen/QwQ-32B QwQ-32B | | $0.18 | $0.20 | 131,072 | 0.82 | 28 | Yes | Yes |
| Qwen/QwQ-32B QwQ-32B | | - | - | - | - | - | - | - |
| Qwen/QwQ-32B QwQ-32B | | - | - | - | - | - | - | - |
| Qwen/Qwen3-14B Qwen3-14B | | $0.07 | $0.20 | 40,960 | 0.78 | 51 | Yes | Yes |
| moonshotai/Kimi-K2-Instruct Kimi-K2-Instruct | | $0.57 | $2.30 | 131,072 | 1.17 | 30 | Yes | Yes |
| moonshotai/Kimi-K2-Instruct Kimi-K2-Instruct | | $1.00 | $3.00 | 131,072 | 0.53 | 45 | Yes | Yes |
| moonshotai/Kimi-K2-Instruct Kimi-K2-Instruct | | - | - | - | - | - | - | - |
| zai-org/GLM-4.6V-Flash GLM-4.6V-Flash | | $0.30 | $0.90 | 131,072 | 1.08 | 32 | Yes | No |
| zai-org/GLM-4.6V-Flash GLM-4.6V-Flash | | - | - | - | 1.54 | 81 | No | No |
| Qwen/Qwen2.5-72B-Instruct Qwen2.5-72B-Instruct | | $0.38 | $0.40 | 32,000 | 0.64 | 31 | Yes | No |
| Qwen/Qwen2.5-72B-Instruct Qwen2.5-72B-Instruct | | - | - | - | - | - | - | - |
| Qwen/Qwen2.5-72B-Instruct Qwen2.5-72B-Instruct | | - | - | - | - | - | - | - |
| Qwen/Qwen2.5-72B-Instruct Qwen2.5-72B-Instruct | | - | - | - | - | - | - | - |
| deepseek-ai/DeepSeek-R1-Distill-Qwen-7B DeepSeek-R1-Distill-Qwen-7B | | $0.15 | $0.15 | 131,072 | 0.83 | 66 | No | No |
| deepseek-ai/DeepSeek-Prover-V2-671B DeepSeek-Prover-V2-671B | | $0.70 | $2.50 | 160,000 | 0.92 | 23 | No | No |
| meta-llama/Llama-4-Scout-17B-16E-Instruct Llama-4-Scout-17B-16E-Instruct | | $0.11 | $0.34 | 131,072 | 0.17 | 381 | Yes | No |
| meta-llama/Llama-4-Scout-17B-16E-Instruct Llama-4-Scout-17B-16E-Instruct | | $0.18 | $0.59 | 131,072 | 0.37 | 55 | No | No |
| meta-llama/Llama-4-Scout-17B-16E-Instruct Llama-4-Scout-17B-16E-Instruct | | - | - | - | - | - | - | - |
| meta-llama/Llama-4-Scout-17B-16E-Instruct Llama-4-Scout-17B-16E-Instruct | | $0.09 | $0.29 | 890,000 | 0.76 | 33 | Yes | Yes |
| meta-llama/Llama-4-Scout-17B-16E-Instruct Llama-4-Scout-17B-16E-Instruct | | - | - | - | - | - | - | - |
| meta-llama/Llama-4-Scout-17B-16E-Instruct Llama-4-Scout-17B-16E-Instruct | | - | - | - | - | - | - | - |
| meta-llama/Llama-4-Maverick-17B-128E-Instruct Llama-4-Maverick-17B-128E-Instruct | | $0.20 | $0.60 | 131,072 | 1.06 | 315 | Yes | No |
| meta-llama/Llama-4-Maverick-17B-128E-Instruct Llama-4-Maverick-17B-128E-Instruct | | $0.63 | $1.80 | 131,072 | 1.89 | 305 | Yes | Yes |
| meta-llama/Llama-4-Maverick-17B-128E-Instruct Llama-4-Maverick-17B-128E-Instruct | | - | - | - | - | - | - | - |
| Qwen/Qwen2.5-VL-72B-Instruct Qwen2.5-VL-72B-Instruct | | $0.60 | $0.60 | 32,768 | 0.50 | 18 | No | No |
| Qwen/Qwen2.5-VL-72B-Instruct Qwen2.5-VL-72B-Instruct | | $1.01 | $1.01 | 32,768 | 0.33 | 27 | No | Yes |
| Qwen/Qwen3-VL-235B-A22B-Thinking Qwen3-VL-235B-A22B-Thinking | | $0.98 | $3.95 | 131,072 | 0.83 | 40 | No | No |
| HuggingFaceTB/SmolLM3-3B SmolLM3-3B | | - | - | - | 0.12 | 91 | Yes | Yes |
| meta-llama/Llama-4-Maverick-17B-128E-Instruct-FP8 Llama-4-Maverick-17B-128E-Instruct-FP8 | | $0.27 | $0.85 | 1,048,576 | 0.37 | 62 | No | No |
| meta-llama/Llama-4-Maverick-17B-128E-Instruct-FP8 Llama-4-Maverick-17B-128E-Instruct-FP8 | | $0.27 | $0.85 | 1,048,576 | 0.27 | 52 | Yes | Yes |
| deepseek-ai/DeepSeek-V3.1-Terminus DeepSeek-V3.1-Terminus | | $0.27 | $1.00 | 131,072 | 1.00 | 30 | Yes | Yes |
| meta-llama/Llama-Guard-4-12B Llama-Guard-4-12B | | $0.20 | $0.20 | 131,072 | 0.18 | 10 | No | No |
| Qwen/Qwen3-VL-235B-A22B-Instruct Qwen3-VL-235B-A22B-Instruct | | $0.30 | $1.50 | 131,072 | 1.13 | 29 | Yes | Yes |
| Qwen/Qwen3-Next-80B-A3B-Thinking Qwen3-Next-80B-A3B-Thinking | | $0.15 | $1.50 | 131,072 | 1.20 | 159 | Yes | No |
| Qwen/Qwen3-Next-80B-A3B-Thinking Qwen3-Next-80B-A3B-Thinking | | $0.30 | $0.30 | 262,144 | 0.58 | 183 | Yes | No |
| Qwen/Qwen3-Next-80B-A3B-Thinking Qwen3-Next-80B-A3B-Thinking | | $0.15 | $1.50 | 262,144 | 0.44 | 111 | Yes | Yes |
| arcee-ai/Trinity-Mini Trinity-Mini | | $0.05 | $0.15 | 128,000 | 0.81 | 183 | Yes | Yes |
| deepseek-ai/DeepSeek-V3.1 DeepSeek-V3.1 | | $0.27 | $1.00 | 131,072 | 1.19 | 31 | Yes | Yes |
| deepseek-ai/DeepSeek-V3.1 DeepSeek-V3.1 | | $0.60 | $1.70 | 131,072 | 1.70 | 36 | Yes | Yes |
| deepseek-ai/DeepSeek-V3.1 DeepSeek-V3.1 | | - | - | 163,840 | 0.35 | 55 | Yes | No |
| CohereLabs/aya-vision-32b aya-vision-32b | | - | - | - | 0.37 | 37 | No | No |
| allenai/Olmo-3.1-32B-Think Olmo-3.1-32B-Think | | $0.15 | $0.50 | - | 1.35 | 28 | No | No |
| allenai/Olmo-3.1-32B-Instruct Olmo-3.1-32B-Instruct | | $0.20 | $0.60 | - | 1.34 | 53 | Yes | Yes |
| zai-org/GLM-4.6V GLM-4.6V | | - | - | - | - | - | - | - |
| zai-org/GLM-4.6V GLM-4.6V | | - | - | - | 3.34 | 33 | Yes | No |
| deepseek-ai/DeepSeek-R1-Distill-Llama-8B DeepSeek-R1-Distill-Llama-8B | | - | - | - | - | - | - | - |
| deepseek-ai/DeepSeek-R1-Distill-Llama-8B DeepSeek-R1-Distill-Llama-8B | | $0.05 | $0.05 | 131,072 | 0.86 | 56 | No | Yes |
| CohereLabs/command-a-translate-08-2025 command-a-translate-08-2025 | | - | - | - | 0.19 | 64 | Yes | No |
| openai/gpt-oss-safeguard-20b gpt-oss-safeguard-20b | | - | - | 131,072 | 0.12 | 871 | Yes | No |
| Qwen/Qwen3-30B-A3B Qwen3-30B-A3B | | $0.09 | $0.45 | 40,960 | 0.76 | 115 | Yes | No |
| Qwen/Qwen3-30B-A3B Qwen3-30B-A3B | | - | - | - | - | - | - | - |
| zai-org/GLM-4.5-Air-FP8 GLM-4.5-Air-FP8 | | $0.20 | $1.10 | 131,072 | 1.74 | 122 | Yes | Yes |
| Qwen/Qwen3-235B-A22B-Instruct-2507 Qwen3-235B-A22B-Instruct-2507 | | $0.09 | $0.58 | 131,072 | 0.72 | 24 | Yes | Yes |
| Qwen/Qwen3-235B-A22B-Instruct-2507 Qwen3-235B-A22B-Instruct-2507 | | - | - | - | - | - | - | - |
| Qwen/Qwen3-235B-A22B-Instruct-2507 Qwen3-235B-A22B-Instruct-2507 | | $0.20 | $0.60 | 32,768 | 0.87 | 24 | Yes | Yes |
| Qwen/Qwen3-235B-A22B-Instruct-2507 Qwen3-235B-A22B-Instruct-2507 | | $2.00 | $2.00 | 262,144 | 0.74 | 56 | Yes | No |
| Qwen/Qwen3-235B-A22B-Instruct-2507 Qwen3-235B-A22B-Instruct-2507 | | - | - | - | - | - | - | - |
| Qwen/Qwen3-235B-A22B-Instruct-2507 Qwen3-235B-A22B-Instruct-2507 | | - | - | - | - | - | - | - |
| Qwen/Qwen3-235B-A22B-Instruct-2507 Qwen3-235B-A22B-Instruct-2507 | | - | - | - | 0.40 | 74 | No | Yes |
| Qwen/Qwen3-Coder-480B-A35B-Instruct Qwen3-Coder-480B-A35B-Instruct | | $0.30 | $1.30 | 262,144 | 0.94 | 49 | Yes | Yes |
| Qwen/Qwen3-Coder-480B-A35B-Instruct Qwen3-Coder-480B-A35B-Instruct | | - | - | - | - | - | - | - |
| Qwen/Qwen3-Coder-480B-A35B-Instruct Qwen3-Coder-480B-A35B-Instruct | | $2.00 | $2.00 | 262,144 | 0.72 | 59 | No | No |
| Qwen/Qwen3-Coder-480B-A35B-Instruct Qwen3-Coder-480B-A35B-Instruct | | $2.00 | $2.00 | 262,144 | 0.34 | 61 | Yes | Yes |
| Qwen/Qwen3-Coder-480B-A35B-Instruct Qwen3-Coder-480B-A35B-Instruct | | - | - | - | - | - | - | - |
| meta-llama/Meta-Llama-3-70B-Instruct Meta-Llama-3-70B-Instruct | | - | - | - | - | - | - | - |
| meta-llama/Meta-Llama-3-70B-Instruct Meta-Llama-3-70B-Instruct | | $0.51 | $0.74 | 8,192 | 0.65 | 28 | No | Yes |
| meta-llama/Meta-Llama-3-70B-Instruct Meta-Llama-3-70B-Instruct | | - | - | - | - | - | - | - |
| meta-llama/Meta-Llama-3-70B-Instruct Meta-Llama-3-70B-Instruct | | - | - | - | - | - | - | - |
| deepseek-ai/DeepSeek-V3 DeepSeek-V3 | | $0.40 | $1.30 | 64,000 | 1.26 | 31 | Yes | No |
| deepseek-ai/DeepSeek-V3 DeepSeek-V3 | | $1.25 | $1.25 | 131,072 | 1.27 | 33 | Yes | Yes |
| deepseek-ai/DeepSeek-V3 DeepSeek-V3 | | - | - | - | - | - | - | - |
| Sao10K/L3-8B-Stheno-v3.2 L3-8B-Stheno-v3.2 | | $0.05 | $0.05 | 8,192 | 0.59 | 58 | No | No |
| CohereLabs/c4ai-command-r7b-12-2024 c4ai-command-r7b-12-2024 | | - | - | - | 0.17 | 133 | Yes | No |
| CohereLabs/c4ai-command-r7b-arabic-02-2025 c4ai-command-r7b-arabic-02-2025 | | - | - | - | 0.19 | 100 | Yes | No |
| Qwen/Qwen3-VL-30B-A3B-Thinking Qwen3-VL-30B-A3B-Thinking | | $0.20 | $1.00 | 131,072 | 0.87 | 91 | Yes | No |
| zai-org/GLM-4.5V GLM-4.5V | | - | - | - | - | - | - | - |
| zai-org/GLM-4.5V GLM-4.5V | | - | - | - | - | - | - | - |
| zai-org/GLM-4.5V GLM-4.5V | | - | - | - | 2.19 | 49 | Yes | No |
| baidu/ERNIE-4.5-VL-424B-A47B-Base-PT ERNIE-4.5-VL-424B-A47B-Base-PT | | $0.42 | $1.25 | 123,000 | 1.18 | 41 | No | No |
| zai-org/AutoGLM-Phone-9B-Multilingual AutoGLM-Phone-9B-Multilingual | | $0.04 | $0.14 | 65,536 | 0.65 | 68 | No | No |
| Qwen/Qwen2.5-Coder-3B-Instruct Qwen2.5-Coder-3B-Instruct | | $0.01 | $0.03 | 32,768 | 0.68 | 82 | No | Yes |
| MiniMaxAI/MiniMax-M1-80k MiniMax-M1-80k | | $0.55 | $2.20 | 1,000,000 | 3.71 | 26 | Yes | No |
| deepseek-ai/DeepSeek-R1-Distill-Qwen-1.5B DeepSeek-R1-Distill-Qwen-1.5B | | $0.10 | $0.10 | 131,072 | 0.83 | 138 | No | No |
| allenai/Olmo-3-7B-Instruct Olmo-3-7B-Instruct | | $0.10 | $0.20 | - | - | - | Yes | No |
| Qwen/Qwen3-235B-A22B-Thinking-2507 Qwen3-235B-A22B-Thinking-2507 | | $0.30 | $3.00 | 131,072 | 0.89 | 40 | Yes | No |
| Qwen/Qwen3-235B-A22B-Thinking-2507 Qwen3-235B-A22B-Thinking-2507 | | - | - | - | - | - | - | - |
| Qwen/Qwen3-235B-A22B-Thinking-2507 Qwen3-235B-A22B-Thinking-2507 | | - | - | - | - | - | - | - |
| swiss-ai/Apertus-8B-Instruct-2509 Apertus-8B-Instruct-2509 | | $0.10 | $0.20 | - | 1.40 | 108 | No | Yes |
| zai-org/GLM-4-32B-0414 GLM-4-32B-0414 | | $0.55 | $1.66 | 32,000 | 0.94 | 39 | No | No |
| deepseek-ai/DeepSeek-V3-0324 DeepSeek-V3-0324 | | $0.27 | $1.12 | 163,840 | 1.08 | 30 | Yes | Yes |
| deepseek-ai/DeepSeek-V3-0324 DeepSeek-V3-0324 | | $3.00 | $4.50 | 131,072 | 0.85 | 191 | Yes | Yes |
| deepseek-ai/DeepSeek-V3-0324 DeepSeek-V3-0324 | | $1.25 | $1.25 | 163,840 | 1.96 | 35 | Yes | No |
| deepseek-ai/DeepSeek-V3-0324 DeepSeek-V3-0324 | | $1.25 | $1.25 | 131,072 | 2.60 | 27 | Yes | No |
| deepseek-ai/DeepSeek-V3-0324 DeepSeek-V3-0324 | | - | - | - | - | - | - | - |
| zai-org/GLM-4.6 GLM-4.6 | | $0.55 | $2.20 | 204,800 | 0.65 | 60 | Yes | No |
| zai-org/GLM-4.6 GLM-4.6 | | - | - | - | - | - | - | - |
| zai-org/GLM-4.6 GLM-4.6 | | $0.60 | $2.20 | 202,752 | 0.46 | 42 | Yes | No |
| zai-org/GLM-4.6 GLM-4.6 | | - | - | - | 1.77 | 64 | Yes | No |
| zai-org/GLM-4.6-FP8 GLM-4.6-FP8 | | - | - | - | 1.86 | 67 | Yes | No |
| marin-community/marin-8b-instruct marin-8b-instruct | | $0.18 | $0.18 | 4,096 | 0.16 | 173 | No | Yes |
| zai-org/GLM-4.6V-FP8 GLM-4.6V-FP8 | | - | - | - | 2.79 | 42 | No | No |
| CohereLabs/aya-expanse-32b aya-expanse-32b | | - | - | - | 0.19 | 41 | No | No |
| deepcogito/cogito-671b-v2.1 cogito-671b-v2.1 | | $1.25 | $1.25 | 163,840 | 0.22 | 65 | No | Yes |
| deepcogito/cogito-671b-v2.1 cogito-671b-v2.1 | | - | - | 163,840 | 0.37 | 41 | Yes | No |
| deepcogito/cogito-671b-v2.1-FP8 cogito-671b-v2.1-FP8 | | $1.25 | $1.25 | 163,840 | 0.23 | 66 | No | Yes |
| CohereLabs/c4ai-command-r-08-2024 c4ai-command-r-08-2024 | | - | - | - | 0.22 | 38 | Yes | No |
| EssentialAI/rnj-1-instruct rnj-1-instruct | | $0.15 | $0.15 | 32,768 | 0.18 | 126 | Yes | Yes |
| zai-org/GLM-4.7-FP8 GLM-4.7-FP8 | | $0.45 | $2.00 | 202,752 | 0.35 | 38 | Yes | Yes |
| zai-org/GLM-4.7-FP8 GLM-4.7-FP8 | | - | - | - | - | - | - | - |
| zai-org/GLM-4.5 GLM-4.5 | | $0.60 | $2.20 | 131,072 | 0.76 | 49 | Yes | No |
| zai-org/GLM-4.5 GLM-4.5 | | - | - | - | - | - | - | - |
| zai-org/GLM-4.5 GLM-4.5 | | - | - | - | 2.20 | 46 | No | No |
| zai-org/GLM-4.5-Air GLM-4.5-Air | | $0.13 | $0.85 | 131,072 | 0.62 | 59 | Yes | No |
| zai-org/GLM-4.5-Air GLM-4.5-Air | | - | - | - | - | - | - | - |
| zai-org/GLM-4.5-Air GLM-4.5-Air | | - | - | - | 2.19 | 67 | Yes | No |
| Qwen/Qwen3-Coder-480B-A35B-Instruct-FP8 Qwen3-Coder-480B-A35B-Instruct-FP8 | | $2.00 | $2.00 | 262,144 | 0.43 | 68 | Yes | Yes |
| zai-org/GLM-4.5V-FP8 GLM-4.5V-FP8 | | - | - | - | 3.91 | 30 | Yes | No |
| aisingapore/Qwen-SEA-LION-v4-32B-IT Qwen-SEA-LION-v4-32B-IT | | $0.25 | $0.50 | - | 1.93 | 44 | Yes | Yes |
| swiss-ai/Apertus-70B-Instruct-2509 Apertus-70B-Instruct-2509 | | $0.82 | $2.92 | - | 1.18 | 57 | No | Yes |
| aisingapore/Gemma-SEA-LION-v4-27B-IT Gemma-SEA-LION-v4-27B-IT | | $0.20 | $0.40 | - | 1.99 | 46 | No | Yes |
| Qwen/Qwen3-235B-A22B Qwen3-235B-A22B | | $0.20 | $0.80 | 40,960 | 0.89 | 10 | No | No |
| Qwen/Qwen3-235B-A22B Qwen3-235B-A22B | | $0.20 | $0.60 | 32,000 | 1.18 | 23 | Yes | Yes |
| Qwen/Qwen3-235B-A22B Qwen3-235B-A22B | | - | - | - | - | - | - | - |
| Qwen/Qwen3-235B-A22B Qwen3-235B-A22B | | - | - | - | - | - | - | - |
| baidu/ERNIE-4.5-300B-A47B-Base-PT ERNIE-4.5-300B-A47B-Base-PT | | $0.28 | $1.10 | 123,000 | 1.83 | 20 | No | Yes |
| dicta-il/DictaLM-3.0-24B-Thinking DictaLM-3.0-24B-Thinking | | $0.20 | $0.40 | - | 1.61 | 63 | Yes | Yes |
| mistralai/Mixtral-8x22B-Instruct-v0.1 Mixtral-8x22B-Instruct-v0.1 | | $1.20 | $1.20 | 65,536 | 0.96 | 20 | No | Yes |
| mistralai/Mixtral-8x22B-Instruct-v0.1 Mixtral-8x22B-Instruct-v0.1 | | - | - | - | - | - | - | - |
| mistralai/Mixtral-8x22B-Instruct-v0.1 Mixtral-8x22B-Instruct-v0.1 | | $1.20 | $1.20 | 65,536 | 0.29 | 71 | No | No |
| baidu/ERNIE-4.5-VL-28B-A3B-PT ERNIE-4.5-VL-28B-A3B-PT | | $0.14 | $0.56 | 30,000 | 1.03 | 71 | No | No |
| baidu/ERNIE-4.5-21B-A3B-PT ERNIE-4.5-21B-A3B-PT | | $0.07 | $0.28 | 120,000 | 0.94 | 75 | No | No |
| NousResearch/Hermes-2-Pro-Llama-3-8B Hermes-2-Pro-Llama-3-8B | | $0.14 | $0.14 | 8,192 | 0.65 | 103 | No | No |
| CohereLabs/c4ai-command-a-03-2025 c4ai-command-a-03-2025 | | - | - | - | 1.48 | 48 | Yes | No |
| CohereLabs/command-a-vision-07-2025 command-a-vision-07-2025 | | - | - | - | 0.15 | 41 | No | No |
| katanemo/Arch-Router-1.5B Arch-Router-1.5B | | - | - | - | 0.13 | 68 | No | Yes |
| alpindale/WizardLM-2-8x22B WizardLM-2-8x22B | | $0.62 | $0.62 | 65,535 | 0.90 | 11 | No | No |
| Sao10K/L3-70B-Euryale-v2.1 L3-70B-Euryale-v2.1 | | $1.48 | $1.48 | 8,192 | 0.68 | 38 | No | No |
| Sao10K/L3-8B-Lunaris-v1 L3-8B-Lunaris-v1 | | $0.05 | $0.05 | 8,192 | 0.71 | 42 | No | No |
| utter-project/EuroLLM-22B-Instruct-2512 EuroLLM-22B-Instruct-2512 | | $0.10 | $0.20 | - | - | - | No | No |