| Qwen/Qwen3.5-397B-A17B Qwen3.5-397B-A17B | | $0.60 | $3.60 | 262,144 | 1.09 | 74 | Yes | No |
| Qwen/Qwen3.5-397B-A17B Qwen3.5-397B-A17B | | - | - | - | - | - | - | - |
| zai-org/GLM-5 GLM-5 | | $1.00 | $3.20 | 202,800 | 1.82 | 30 | Yes | No |
| zai-org/GLM-5 GLM-5 | | $1.00 | $3.20 | 202,752 | 0.50 | 20 | Yes | Yes |
| zai-org/GLM-5 GLM-5 | | - | - | - | 3.97 | 30 | No | No |
| MiniMaxAI/MiniMax-M2.5 MiniMax-M2.5 | | $0.30 | $1.20 | 204,800 | 2.55 | 31 | Yes | No |
| moonshotai/Kimi-K2.5 Kimi-K2.5 | | $0.60 | $3.00 | 262,144 | 1.48 | 30 | Yes | No |
| moonshotai/Kimi-K2.5 Kimi-K2.5 | | $0.50 | $2.80 | 262,144 | 0.37 | 34 | Yes | Yes |
| moonshotai/Kimi-K2.5 Kimi-K2.5 | | - | - | 262,144 | 0.86 | 89 | Yes | No |
| Qwen/Qwen3-Coder-Next Qwen3-Coder-Next | | $0.20 | $1.50 | 262,144 | 0.77 | 152 | Yes | Yes |
| zai-org/GLM-4.7-Flash GLM-4.7-Flash | | $0.07 | $0.40 | 200,000 | 0.81 | 46 | No | Yes |
| zai-org/GLM-4.7-Flash GLM-4.7-Flash | | - | - | - | - | - | - | - |
| openai/gpt-oss-120b gpt-oss-120b | | $0.15 | $0.75 | 131,072 | 0.90 | 334 | Yes | No |
| openai/gpt-oss-120b gpt-oss-120b | | $0.05 | $0.25 | 131,072 | 0.57 | 81 | Yes | Yes |
| openai/gpt-oss-120b gpt-oss-120b | | $0.25 | $0.69 | - | 0.31 | 1,071 | Yes | No |
| openai/gpt-oss-120b gpt-oss-120b | | $0.22 | $0.59 | 131,072 | 0.64 | 538 | Yes | Yes |
| openai/gpt-oss-120b gpt-oss-120b | | $0.10 | $0.40 | 131,072 | 0.94 | 86 | Yes | Yes |
| openai/gpt-oss-120b gpt-oss-120b | | $0.30 | $0.30 | 131,072 | 0.68 | 279 | Yes | No |
| openai/gpt-oss-120b gpt-oss-120b | | $0.15 | $0.60 | 131,072 | 1.21 | 72 | Yes | No |
| openai/gpt-oss-120b gpt-oss-120b | | $0.15 | $0.60 | 131,072 | 0.31 | 90 | Yes | No |
| openai/gpt-oss-120b gpt-oss-120b | | - | - | - | 0.45 | 87 | Yes | Yes |
| openai/gpt-oss-120b gpt-oss-120b | | $0.09 | $0.47 | 131,072 | 0.31 | 79 | Yes | Yes |
| openai/gpt-oss-20b gpt-oss-20b | | $0.10 | $0.50 | 131,072 | 0.11 | 822 | Yes | No |
| openai/gpt-oss-20b gpt-oss-20b | | $0.04 | $0.15 | 131,072 | 0.68 | 102 | No | Yes |
| openai/gpt-oss-20b gpt-oss-20b | | $0.05 | $0.20 | 131,072 | 1.20 | 109 | Yes | Yes |
| openai/gpt-oss-20b gpt-oss-20b | | $0.10 | $0.10 | 131,072 | 0.51 | 112 | Yes | No |
| openai/gpt-oss-20b gpt-oss-20b | | $0.05 | $0.20 | 131,072 | 0.24 | 102 | Yes | Yes |
| openai/gpt-oss-20b gpt-oss-20b | | $0.05 | $0.20 | 131,072 | 0.76 | 125 | Yes | No |
| openai/gpt-oss-20b gpt-oss-20b | | $0.05 | $0.18 | 131,072 | 1.00 | 38 | Yes | Yes |
| meta-llama/Llama-3.1-8B-Instruct Llama-3.1-8B-Instruct | | $0.02 | $0.05 | 16,384 | 0.71 | 77 | No | No |
| meta-llama/Llama-3.1-8B-Instruct Llama-3.1-8B-Instruct | | $0.10 | $0.10 | - | 0.40 | 1,075 | No | No |
| meta-llama/Llama-3.1-8B-Instruct Llama-3.1-8B-Instruct | | $0.10 | $0.20 | 16,384 | 1.88 | 308 | Yes | Yes |
| meta-llama/Llama-3.1-8B-Instruct Llama-3.1-8B-Instruct | | $0.06 | $0.06 | 131,072 | 1.17 | 58 | No | Yes |
| meta-llama/Llama-3.1-8B-Instruct Llama-3.1-8B-Instruct | | - | - | - | - | - | - | - |
| meta-llama/Llama-3.1-8B-Instruct Llama-3.1-8B-Instruct | | - | - | - | 0.35 | 106 | Yes | Yes |
| meta-llama/Llama-3.1-8B-Instruct Llama-3.1-8B-Instruct | | $0.11 | $0.11 | 131,072 | 0.30 | 44 | Yes | No |
| deepseek-ai/DeepSeek-V3.2 DeepSeek-V3.2 | | $0.27 | $0.40 | 163,840 | 1.16 | 35 | Yes | Yes |
| deepseek-ai/DeepSeek-V3.2 DeepSeek-V3.2 | | - | - | 163,840 | 1.10 | 61 | Yes | No |
| google/gemma-3-27b-it gemma-3-27b-it | | - | - | - | 0.42 | 43 | Yes | Yes |
| Qwen/Qwen2.5-7B-Instruct Qwen2.5-7B-Instruct | | $0.30 | $0.30 | 32,768 | 0.23 | 150 | Yes | Yes |
| Qwen/Qwen3-4B-Instruct-2507 Qwen3-4B-Instruct-2507 | | $0.01 | $0.03 | 262,144 | 0.73 | 33 | Yes | No |
| deepseek-ai/DeepSeek-OCR DeepSeek-OCR | | $0.03 | $0.03 | 8,192 | 0.66 | 225 | No | No |
| Qwen/Qwen3-VL-8B-Instruct Qwen3-VL-8B-Instruct | | $0.08 | $0.50 | 131,072 | 0.88 | 59 | Yes | No |
| Qwen/Qwen3-VL-8B-Instruct Qwen3-VL-8B-Instruct | | $0.18 | $0.68 | 262,144 | 0.18 | 117 | Yes | Yes |
| deepseek-ai/DeepSeek-R1 DeepSeek-R1 | | $0.70 | $2.50 | 64,000 | 1.08 | 28 | Yes | No |
| deepseek-ai/DeepSeek-R1 DeepSeek-R1 | | - | - | - | 1.93 | 99 | Yes | Yes |
| deepseek-ai/DeepSeek-R1 DeepSeek-R1 | | $2.00 | $2.00 | 163,840 | 0.81 | 47 | No | No |
| deepseek-ai/DeepSeek-R1 DeepSeek-R1 | | $3.00 | $7.00 | 163,840 | 0.68 | 42 | No | Yes |
| deepseek-ai/DeepSeek-R1 DeepSeek-R1 | | - | - | - | - | - | - | - |
| Qwen/Qwen3-8B Qwen3-8B | | $0.07 | $0.18 | 40,960 | 0.80 | 52 | Yes | Yes |
| meta-llama/Llama-3.3-70B-Instruct Llama-3.3-70B-Instruct | | $0.59 | $0.79 | 131,072 | 0.10 | 317 | Yes | No |
| meta-llama/Llama-3.3-70B-Instruct Llama-3.3-70B-Instruct | | $0.14 | $0.40 | 131,072 | 0.70 | 26 | Yes | No |
| meta-llama/Llama-3.3-70B-Instruct Llama-3.3-70B-Instruct | | - | - | - | - | - | - | - |
| meta-llama/Llama-3.3-70B-Instruct Llama-3.3-70B-Instruct | | $0.60 | $1.20 | 131,072 | 0.52 | 267 | Yes | Yes |
| meta-llama/Llama-3.3-70B-Instruct Llama-3.3-70B-Instruct | | $0.40 | $0.40 | 131,072 | 0.85 | 16 | No | Yes |
| meta-llama/Llama-3.3-70B-Instruct Llama-3.3-70B-Instruct | | $0.40 | $0.40 | 131,072 | 1.03 | 93 | Yes | No |
| meta-llama/Llama-3.3-70B-Instruct Llama-3.3-70B-Instruct | | $0.88 | $0.88 | 131,072 | 0.42 | 149 | Yes | Yes |
| meta-llama/Llama-3.3-70B-Instruct Llama-3.3-70B-Instruct | | - | - | - | - | - | - | - |
| meta-llama/Llama-3.3-70B-Instruct Llama-3.3-70B-Instruct | | - | - | - | 0.32 | 79 | Yes | Yes |
| meta-llama/Llama-3.3-70B-Instruct Llama-3.3-70B-Instruct | | $0.74 | $0.74 | 131,072 | 0.45 | 35 | Yes | Yes |
| moonshotai/Kimi-K2-Thinking Kimi-K2-Thinking | | $0.60 | $2.50 | 262,144 | 0.98 | 40 | Yes | No |
| moonshotai/Kimi-K2-Thinking Kimi-K2-Thinking | | $1.20 | $4.00 | 262,144 | 2.43 | 90 | Yes | Yes |
| deepseek-ai/DeepSeek-V3.2-Exp DeepSeek-V3.2-Exp | | $0.27 | $0.41 | 163,840 | 0.85 | 35 | Yes | Yes |
| Qwen/Qwen2.5-Coder-7B-Instruct Qwen2.5-Coder-7B-Instruct | | $0.01 | $0.03 | 131,072 | 1.29 | 60 | No | Yes |
| meta-llama/Llama-3.2-3B-Instruct Llama-3.2-3B-Instruct | | - | - | - | - | - | - | - |
| meta-llama/Llama-3.2-3B-Instruct Llama-3.2-3B-Instruct | | - | - | - | - | - | - | - |
| meta-llama/Llama-3.2-3B-Instruct Llama-3.2-3B-Instruct | | $0.10 | $0.10 | 131,072 | 1.67 | 95 | Yes | No |
| meta-llama/Llama-3.2-3B-Instruct Llama-3.2-3B-Instruct | | $0.06 | $0.06 | 131,072 | 1.66 | 90 | Yes | Yes |
| Qwen/Qwen3-4B-Thinking-2507 Qwen3-4B-Thinking-2507 | | $0.01 | $0.03 | 262,144 | 0.91 | 95 | Yes | No |
| zai-org/GLM-4.7 GLM-4.7 | | $0.60 | $2.20 | 204,800 | 0.78 | 77 | Yes | No |
| zai-org/GLM-4.7 GLM-4.7 | | - | - | - | - | - | - | - |
| zai-org/GLM-4.7 GLM-4.7 | | - | - | - | 2.03 | 76 | Yes | No |
| moonshotai/Kimi-K2-Instruct-0905 Kimi-K2-Instruct-0905 | | - | - | 262,144 | 0.10 | 176 | Yes | No |
| moonshotai/Kimi-K2-Instruct-0905 Kimi-K2-Instruct-0905 | | $0.60 | $2.50 | 262,144 | 0.75 | 16 | Yes | Yes |
| moonshotai/Kimi-K2-Instruct-0905 Kimi-K2-Instruct-0905 | | $1.00 | $3.00 | 262,144 | 0.57 | 53 | Yes | Yes |
| Qwen/Qwen3-32B Qwen3-32B | | $0.29 | $0.59 | 131,072 | 0.23 | 271 | Yes | No |
| Qwen/Qwen3-32B Qwen3-32B | | $0.10 | $0.45 | 40,960 | 0.66 | 45 | No | No |
| Qwen/Qwen3-32B Qwen3-32B | | - | - | - | - | - | - | - |
| Qwen/Qwen3-32B Qwen3-32B | | $0.40 | $0.80 | 32,768 | 1.45 | 229 | Yes | Yes |
| Qwen/Qwen3-32B Qwen3-32B | | $0.08 | $0.25 | 40,960 | 1.03 | 27 | Yes | Yes |
| Qwen/Qwen3-32B Qwen3-32B | | $0.09 | $0.25 | 32,768 | 0.50 | 35 | Yes | Yes |
| XiaomiMiMo/MiMo-V2-Flash MiMo-V2-Flash | | $0.10 | $0.30 | 262,144 | 1.31 | 81 | Yes | No |
| google/gemma-3n-E4B-it gemma-3n-E4B-it | | $0.02 | $0.04 | 32,768 | 0.49 | 52 | No | Yes |
| Qwen/Qwen3-Next-80B-A3B-Instruct Qwen3-Next-80B-A3B-Instruct | | $0.15 | $1.50 | 131,072 | 1.53 | 101 | Yes | No |
| Qwen/Qwen3-Next-80B-A3B-Instruct Qwen3-Next-80B-A3B-Instruct | | $0.30 | $0.30 | 262,144 | 0.65 | 170 | Yes | No |
| Qwen/Qwen3-Next-80B-A3B-Instruct Qwen3-Next-80B-A3B-Instruct | | $0.15 | $1.50 | 262,144 | 0.41 | 110 | Yes | Yes |
| meta-llama/Llama-3.2-1B-Instruct Llama-3.2-1B-Instruct | | - | - | 131,000 | 2.68 | 148 | No | No |
| meta-llama/Llama-3.2-1B-Instruct Llama-3.2-1B-Instruct | | - | - | - | - | - | - | - |
| meta-llama/Llama-3.2-1B-Instruct Llama-3.2-1B-Instruct | | - | - | - | - | - | - | - |
| MiniMaxAI/MiniMax-M2.1 MiniMax-M2.1 | | $0.30 | $1.20 | 204,800 | 1.64 | 21 | Yes | No |
| ServiceNow-AI/Apriel-1.6-15b-Thinker Apriel-1.6-15b-Thinker | | $0.00 | $0.00 | 131,072 | 0.23 | 71 | No | No |
| Qwen/Qwen3-VL-32B-Instruct Qwen3-VL-32B-Instruct | | $0.50 | $1.50 | 262,144 | 0.37 | 105 | Yes | Yes |
| Qwen/Qwen3-Coder-30B-A3B-Instruct Qwen3-Coder-30B-A3B-Instruct | | - | - | - | - | - | - | - |
| Qwen/Qwen3-Coder-30B-A3B-Instruct Qwen3-Coder-30B-A3B-Instruct | | - | - | - | 0.46 | 64 | Yes | No |
| Qwen/Qwen3-Coder-30B-A3B-Instruct Qwen3-Coder-30B-A3B-Instruct | | $0.07 | $0.26 | 262,144 | 0.33 | 66 | Yes | Yes |
| Qwen/Qwen2.5-VL-7B-Instruct Qwen2.5-VL-7B-Instruct | | $0.20 | $0.20 | 32,768 | 0.40 | 92 | No | No |
| nvidia/NVIDIA-Nemotron-Nano-9B-v2 NVIDIA-Nemotron-Nano-9B-v2 | | $0.06 | $0.25 | 131,072 | 1.02 | 103 | Yes | Yes |
| Qwen/Qwen3-VL-30B-A3B-Instruct Qwen3-VL-30B-A3B-Instruct | | $0.20 | $0.70 | 131,072 | 0.75 | 119 | Yes | Yes |
| Qwen/Qwen3-14B Qwen3-14B | | $0.07 | $0.20 | 40,960 | 1.05 | 50 | Yes | Yes |
| allenai/Molmo2-8B Molmo2-8B | | - | - | - | 2.03 | 92 | No | Yes |
| meta-llama/Meta-Llama-3-8B-Instruct Meta-Llama-3-8B-Instruct | | - | - | - | - | - | - | - |
| meta-llama/Meta-Llama-3-8B-Instruct Meta-Llama-3-8B-Instruct | | $0.04 | $0.04 | 8,192 | 0.63 | 87 | No | No |
| meta-llama/Meta-Llama-3-8B-Instruct Meta-Llama-3-8B-Instruct | | - | - | - | - | - | - | - |
| mistralai/Mistral-7B-Instruct-v0.2 Mistral-7B-Instruct-v0.2 | | $0.20 | $0.20 | 32,768 | 0.23 | 92 | No | Yes |
| CohereLabs/command-a-reasoning-08-2025 command-a-reasoning-08-2025 | | - | - | - | 0.27 | 82 | Yes | No |
| deepseek-ai/DeepSeek-R1-Distill-Qwen-32B DeepSeek-R1-Distill-Qwen-32B | | - | - | - | - | - | - | - |
| deepseek-ai/DeepSeek-R1-Distill-Qwen-32B DeepSeek-R1-Distill-Qwen-32B | | $0.30 | $0.30 | 131,072 | 0.84 | 28 | No | Yes |
| deepseek-ai/DeepSeek-R1-Distill-Llama-70B DeepSeek-R1-Distill-Llama-70B | | - | - | - | - | - | - | - |
| deepseek-ai/DeepSeek-R1-Distill-Llama-70B DeepSeek-R1-Distill-Llama-70B | | $0.80 | $0.80 | 8,192 | 0.41 | 63 | No | No |
| deepseek-ai/DeepSeek-R1-Distill-Llama-70B DeepSeek-R1-Distill-Llama-70B | | $0.70 | $1.40 | 131,072 | 2.98 | 130 | No | No |
| deepseek-ai/DeepSeek-R1-Distill-Llama-70B DeepSeek-R1-Distill-Llama-70B | | $0.75 | $0.75 | 131,072 | 0.87 | 16 | No | Yes |
| deepseek-ai/DeepSeek-R1-Distill-Llama-70B DeepSeek-R1-Distill-Llama-70B | | - | - | - | - | - | - | - |
| deepseek-ai/DeepSeek-R1-Distill-Llama-70B DeepSeek-R1-Distill-Llama-70B | | - | - | - | 0.49 | 38 | No | Yes |
| deepseek-ai/DeepSeek-R1-Distill-Llama-70B DeepSeek-R1-Distill-Llama-70B | | - | - | - | - | - | - | - |
| Qwen/Qwen2.5-Coder-32B-Instruct Qwen2.5-Coder-32B-Instruct | | $0.06 | $0.20 | 131,072 | 0.91 | 27 | No | Yes |
| Qwen/Qwen2.5-Coder-32B-Instruct Qwen2.5-Coder-32B-Instruct | | - | - | - | - | - | - | - |
| Qwen/Qwen2.5-Coder-32B-Instruct Qwen2.5-Coder-32B-Instruct | | - | - | - | - | - | - | - |
| Qwen/Qwen2.5-Coder-32B-Instruct Qwen2.5-Coder-32B-Instruct | | - | - | - | 0.42 | 69 | Yes | No |
| deepseek-ai/DeepSeek-R1-0528 DeepSeek-R1-0528 | | $0.70 | $2.50 | 163,840 | 1.20 | 27 | Yes | No |
| deepseek-ai/DeepSeek-R1-0528 DeepSeek-R1-0528 | | $5.00 | $7.00 | 131,072 | 1.78 | 162 | Yes | Yes |
| deepseek-ai/DeepSeek-R1-0528 DeepSeek-R1-0528 | | $3.00 | $3.00 | 163,840 | 1.09 | 49 | No | No |
| deepseek-ai/DeepSeek-R1-0528 DeepSeek-R1-0528 | | $0.00 | $0.00 | 163,840 | 0.64 | 44 | No | Yes |
| deepseek-ai/DeepSeek-R1-0528 DeepSeek-R1-0528 | | - | - | - | - | - | - | - |
| moonshotai/Kimi-K2-Instruct Kimi-K2-Instruct | | $0.57 | $2.30 | 131,072 | 0.97 | 33 | Yes | Yes |
| moonshotai/Kimi-K2-Instruct Kimi-K2-Instruct | | $1.00 | $3.00 | 131,072 | 0.59 | 48 | Yes | Yes |
| moonshotai/Kimi-K2-Instruct Kimi-K2-Instruct | | - | - | - | - | - | - | - |
| Qwen/QwQ-32B QwQ-32B | | - | - | - | - | - | - | - |
| Qwen/QwQ-32B QwQ-32B | | - | - | - | - | - | - | - |
| Qwen/QwQ-32B QwQ-32B | | $0.18 | $0.20 | 131,072 | 1.16 | 28 | Yes | Yes |
| Qwen/QwQ-32B QwQ-32B | | - | - | - | - | - | - | - |
| Qwen/QwQ-32B QwQ-32B | | - | - | - | - | - | - | - |
| meta-llama/Llama-3.1-70B-Instruct Llama-3.1-70B-Instruct | | - | - | - | - | - | - | - |
| meta-llama/Llama-3.1-70B-Instruct Llama-3.1-70B-Instruct | | - | - | - | 0.34 | 77 | Yes | Yes |
| openai/gpt-oss-safeguard-20b gpt-oss-safeguard-20b | | - | - | 131,072 | 0.19 | 799 | Yes | No |
| Qwen/Qwen2.5-VL-72B-Instruct Qwen2.5-VL-72B-Instruct | | - | - | - | - | - | - | - |
| Qwen/Qwen2.5-VL-72B-Instruct Qwen2.5-VL-72B-Instruct | | $1.01 | $1.01 | 32,768 | 0.37 | 33 | No | Yes |
| deepseek-ai/DeepSeek-R1-Distill-Qwen-7B DeepSeek-R1-Distill-Qwen-7B | | $0.15 | $0.15 | 131,072 | 0.71 | 66 | No | No |
| meta-llama/Llama-4-Scout-17B-16E-Instruct Llama-4-Scout-17B-16E-Instruct | | $0.11 | $0.34 | 131,072 | 0.11 | 392 | Yes | No |
| meta-llama/Llama-4-Scout-17B-16E-Instruct Llama-4-Scout-17B-16E-Instruct | | $0.18 | $0.59 | 131,072 | 0.59 | 56 | No | No |
| meta-llama/Llama-4-Scout-17B-16E-Instruct Llama-4-Scout-17B-16E-Instruct | | - | - | - | - | - | - | - |
| meta-llama/Llama-4-Scout-17B-16E-Instruct Llama-4-Scout-17B-16E-Instruct | | $0.09 | $0.29 | 890,000 | 0.95 | 33 | Yes | Yes |
| meta-llama/Llama-4-Scout-17B-16E-Instruct Llama-4-Scout-17B-16E-Instruct | | - | - | - | - | - | - | - |
| meta-llama/Llama-4-Scout-17B-16E-Instruct Llama-4-Scout-17B-16E-Instruct | | - | - | - | - | - | - | - |
| meta-llama/Llama-4-Maverick-17B-128E-Instruct Llama-4-Maverick-17B-128E-Instruct | | $0.20 | $0.60 | 131,072 | 1.21 | 317 | Yes | No |
| meta-llama/Llama-4-Maverick-17B-128E-Instruct Llama-4-Maverick-17B-128E-Instruct | | $0.63 | $1.80 | 131,072 | 0.76 | 439 | Yes | Yes |
| meta-llama/Llama-4-Maverick-17B-128E-Instruct Llama-4-Maverick-17B-128E-Instruct | | - | - | - | - | - | - | - |
| zai-org/GLM-4.6V-Flash GLM-4.6V-Flash | | $0.30 | $0.90 | 131,072 | 0.91 | 21 | Yes | No |
| zai-org/GLM-4.6V-Flash GLM-4.6V-Flash | | - | - | - | - | - | - | - |
| Qwen/Qwen2.5-72B-Instruct Qwen2.5-72B-Instruct | | $0.38 | $0.40 | 32,000 | 0.96 | 32 | Yes | No |
| Qwen/Qwen2.5-72B-Instruct Qwen2.5-72B-Instruct | | - | - | - | - | - | - | - |
| Qwen/Qwen2.5-72B-Instruct Qwen2.5-72B-Instruct | | - | - | - | - | - | - | - |
| Qwen/Qwen2.5-72B-Instruct Qwen2.5-72B-Instruct | | - | - | - | - | - | - | - |
| deepseek-ai/DeepSeek-V3 DeepSeek-V3 | | $0.40 | $1.30 | 64,000 | 1.65 | 29 | Yes | No |
| deepseek-ai/DeepSeek-V3 DeepSeek-V3 | | $1.25 | $1.25 | 131,072 | 1.97 | 37 | Yes | Yes |
| deepseek-ai/DeepSeek-V3 DeepSeek-V3 | | - | - | - | - | - | - | - |
| CohereLabs/aya-vision-32b aya-vision-32b | | - | - | - | 0.41 | 37 | No | No |
| Qwen/Qwen3-30B-A3B Qwen3-30B-A3B | | $0.09 | $0.45 | 40,960 | 0.69 | 76 | Yes | No |
| Qwen/Qwen3-30B-A3B Qwen3-30B-A3B | | - | - | - | - | - | - | - |
| meta-llama/Llama-4-Maverick-17B-128E-Instruct-FP8 Llama-4-Maverick-17B-128E-Instruct-FP8 | | $0.27 | $0.85 | 1,048,576 | 0.31 | 50 | No | No |
| meta-llama/Llama-4-Maverick-17B-128E-Instruct-FP8 Llama-4-Maverick-17B-128E-Instruct-FP8 | | $0.27 | $0.85 | 1,048,576 | 0.34 | 53 | Yes | Yes |
| Qwen/Qwen3-Next-80B-A3B-Thinking Qwen3-Next-80B-A3B-Thinking | | $0.15 | $1.50 | 131,072 | 0.98 | 161 | Yes | No |
| Qwen/Qwen3-Next-80B-A3B-Thinking Qwen3-Next-80B-A3B-Thinking | | $0.30 | $0.30 | 262,144 | 0.51 | 196 | Yes | No |
| Qwen/Qwen3-Next-80B-A3B-Thinking Qwen3-Next-80B-A3B-Thinking | | $0.15 | $1.50 | 262,144 | 0.93 | 114 | Yes | Yes |
| deepseek-ai/DeepSeek-Prover-V2-671B DeepSeek-Prover-V2-671B | | $0.70 | $2.50 | 160,000 | 0.94 | 15 | No | No |
| deepseek-ai/DeepSeek-V3.1 DeepSeek-V3.1 | | $0.27 | $1.00 | 131,072 | 1.36 | 29 | Yes | Yes |
| deepseek-ai/DeepSeek-V3.1 DeepSeek-V3.1 | | $0.60 | $1.70 | 131,072 | 1.60 | 35 | Yes | Yes |
| deepseek-ai/DeepSeek-V3.1 DeepSeek-V3.1 | | - | - | 163,840 | 0.63 | 27 | Yes | No |
| zai-org/GLM-4.6V GLM-4.6V | | - | - | - | - | - | - | - |
| zai-org/GLM-4.6V GLM-4.6V | | - | - | - | 3.33 | 50 | Yes | No |
| allenai/Olmo-3.1-32B-Think Olmo-3.1-32B-Think | | $0.15 | $0.50 | - | 2.20 | 69 | No | No |
| allenai/Olmo-3.1-32B-Instruct Olmo-3.1-32B-Instruct | | $0.20 | $0.60 | - | 1.63 | 51 | Yes | Yes |
| CohereLabs/command-a-translate-08-2025 command-a-translate-08-2025 | | - | - | - | 0.36 | 64 | Yes | No |
| meta-llama/Llama-Guard-4-12B Llama-Guard-4-12B | | $0.20 | $0.20 | 131,072 | 0.19 | 10 | No | No |
| Qwen/Qwen3-Coder-480B-A35B-Instruct Qwen3-Coder-480B-A35B-Instruct | | $0.30 | $1.30 | 262,144 | 0.98 | 48 | Yes | Yes |
| Qwen/Qwen3-Coder-480B-A35B-Instruct Qwen3-Coder-480B-A35B-Instruct | | - | - | - | - | - | - | - |
| Qwen/Qwen3-Coder-480B-A35B-Instruct Qwen3-Coder-480B-A35B-Instruct | | $2.00 | $2.00 | 262,144 | 0.89 | 32 | Yes | No |
| Qwen/Qwen3-Coder-480B-A35B-Instruct Qwen3-Coder-480B-A35B-Instruct | | $2.00 | $2.00 | 262,144 | 0.59 | 48 | Yes | Yes |
| Qwen/Qwen3-Coder-480B-A35B-Instruct Qwen3-Coder-480B-A35B-Instruct | | - | - | - | - | - | - | - |
| arcee-ai/Trinity-Mini Trinity-Mini | | $0.05 | $0.15 | 128,000 | 0.20 | 170 | Yes | Yes |
| deepseek-ai/DeepSeek-V3.1-Terminus DeepSeek-V3.1-Terminus | | $0.27 | $1.00 | 131,072 | 1.27 | 31 | Yes | Yes |
| deepseek-ai/DeepSeek-R1-Distill-Llama-8B DeepSeek-R1-Distill-Llama-8B | | - | - | - | - | - | - | - |
| deepseek-ai/DeepSeek-R1-Distill-Llama-8B DeepSeek-R1-Distill-Llama-8B | | $0.05 | $0.05 | 131,072 | 0.91 | 56 | No | No |
| Qwen/Qwen3-VL-235B-A22B-Thinking Qwen3-VL-235B-A22B-Thinking | | $0.98 | $3.95 | 131,072 | 0.66 | 46 | No | No |
| HuggingFaceTB/SmolLM3-3B SmolLM3-3B | | - | - | - | 0.11 | 91 | Yes | Yes |
| Qwen/Qwen3-VL-235B-A22B-Instruct Qwen3-VL-235B-A22B-Instruct | | $0.30 | $1.50 | 131,072 | 0.92 | 31 | Yes | Yes |
| Qwen/Qwen3-235B-A22B-Instruct-2507 Qwen3-235B-A22B-Instruct-2507 | | $0.09 | $0.58 | 131,072 | 0.76 | 27 | Yes | Yes |
| Qwen/Qwen3-235B-A22B-Instruct-2507 Qwen3-235B-A22B-Instruct-2507 | | - | - | - | - | - | - | - |
| Qwen/Qwen3-235B-A22B-Instruct-2507 Qwen3-235B-A22B-Instruct-2507 | | $0.20 | $0.60 | 32,768 | 0.98 | 24 | Yes | Yes |
| Qwen/Qwen3-235B-A22B-Instruct-2507 Qwen3-235B-A22B-Instruct-2507 | | $2.00 | $2.00 | 262,144 | 3.52 | 13 | Yes | No |
| Qwen/Qwen3-235B-A22B-Instruct-2507 Qwen3-235B-A22B-Instruct-2507 | | $0.20 | $0.60 | 262,144 | 0.72 | 58 | Yes | Yes |
| Qwen/Qwen3-235B-A22B-Instruct-2507 Qwen3-235B-A22B-Instruct-2507 | | - | - | - | - | - | - | - |
| Qwen/Qwen3-235B-A22B-Instruct-2507 Qwen3-235B-A22B-Instruct-2507 | | - | - | - | 1.10 | 63 | Yes | No |
| marin-community/marin-8b-instruct marin-8b-instruct | | $0.18 | $0.18 | 4,096 | 0.28 | 170 | No | Yes |
| CohereLabs/c4ai-command-r7b-arabic-02-2025 c4ai-command-r7b-arabic-02-2025 | | - | - | - | 0.18 | 104 | Yes | No |
| zai-org/AutoGLM-Phone-9B-Multilingual AutoGLM-Phone-9B-Multilingual | | $0.04 | $0.14 | 65,536 | 0.66 | 63 | No | No |
| Sao10K/L3-8B-Stheno-v3.2 L3-8B-Stheno-v3.2 | | $0.05 | $0.05 | 8,192 | 0.68 | 56 | No | No |
| Qwen/Qwen3-235B-A22B-Thinking-2507 Qwen3-235B-A22B-Thinking-2507 | | $0.30 | $3.00 | 131,072 | 1.01 | 57 | Yes | No |
| Qwen/Qwen3-235B-A22B-Thinking-2507 Qwen3-235B-A22B-Thinking-2507 | | - | - | - | - | - | - | - |
| Qwen/Qwen3-235B-A22B-Thinking-2507 Qwen3-235B-A22B-Thinking-2507 | | - | - | - | - | - | - | - |
| Qwen/Qwen3-VL-30B-A3B-Thinking Qwen3-VL-30B-A3B-Thinking | | $0.20 | $1.00 | 131,072 | 0.73 | 76 | Yes | No |
| meta-llama/Meta-Llama-3-70B-Instruct Meta-Llama-3-70B-Instruct | | - | - | - | - | - | - | - |
| meta-llama/Meta-Llama-3-70B-Instruct Meta-Llama-3-70B-Instruct | | $0.51 | $0.74 | 8,192 | 0.69 | 25 | No | Yes |
| meta-llama/Meta-Llama-3-70B-Instruct Meta-Llama-3-70B-Instruct | | - | - | - | - | - | - | - |
| meta-llama/Meta-Llama-3-70B-Instruct Meta-Llama-3-70B-Instruct | | - | - | - | - | - | - | - |
| baidu/ERNIE-4.5-VL-424B-A47B-Base-PT ERNIE-4.5-VL-424B-A47B-Base-PT | | $0.42 | $1.25 | 123,000 | 1.07 | 45 | No | No |
| deepseek-ai/DeepSeek-R1-Distill-Qwen-1.5B DeepSeek-R1-Distill-Qwen-1.5B | | $0.10 | $0.10 | 131,072 | 0.82 | 140 | No | No |
| MiniMaxAI/MiniMax-M1-80k MiniMax-M1-80k | | $0.55 | $2.20 | 1,000,000 | 1.40 | 26 | Yes | No |
| zai-org/GLM-4.5V GLM-4.5V | | $0.60 | $1.80 | 65,536 | 0.82 | 50 | Yes | No |
| zai-org/GLM-4.5V GLM-4.5V | | - | - | - | - | - | - | - |
| zai-org/GLM-4.5V GLM-4.5V | | - | - | - | 2.66 | 40 | Yes | No |
| swiss-ai/Apertus-8B-Instruct-2509 Apertus-8B-Instruct-2509 | | $0.10 | $0.20 | - | 1.70 | 47 | No | Yes |
| Qwen/Qwen2.5-Coder-3B-Instruct Qwen2.5-Coder-3B-Instruct | | $0.01 | $0.03 | 32,768 | 1.13 | 81 | No | Yes |
| allenai/Olmo-3-7B-Instruct Olmo-3-7B-Instruct | | $0.10 | $0.20 | - | - | - | Yes | No |
| zai-org/GLM-4-32B-0414 GLM-4-32B-0414 | | $0.55 | $1.66 | 32,000 | 0.89 | 36 | No | No |
| CohereLabs/c4ai-command-r7b-12-2024 c4ai-command-r7b-12-2024 | | - | - | - | 0.22 | 131 | Yes | No |
| CohereLabs/c4ai-command-r-08-2024 c4ai-command-r-08-2024 | | - | - | - | 0.18 | 47 | Yes | No |
| zai-org/GLM-4.6 GLM-4.6 | | $0.55 | $2.20 | 204,800 | 0.77 | 72 | Yes | No |
| zai-org/GLM-4.6 GLM-4.6 | | - | - | - | - | - | - | - |
| zai-org/GLM-4.6 GLM-4.6 | | $0.60 | $2.20 | 202,752 | 1.36 | 37 | Yes | Yes |
| zai-org/GLM-4.6 GLM-4.6 | | - | - | - | 1.97 | 70 | Yes | No |
| deepcogito/cogito-671b-v2.1 cogito-671b-v2.1 | | $1.25 | $1.25 | 163,840 | 0.29 | 64 | No | Yes |
| deepcogito/cogito-671b-v2.1 cogito-671b-v2.1 | | - | - | 163,840 | 0.66 | 52 | Yes | No |
| EssentialAI/rnj-1-instruct rnj-1-instruct | | $0.15 | $0.15 | 32,768 | 1.81 | 98 | Yes | Yes |
| deepseek-ai/DeepSeek-V3-0324 DeepSeek-V3-0324 | | $0.27 | $1.12 | 163,840 | 0.86 | 33 | Yes | Yes |
| deepseek-ai/DeepSeek-V3-0324 DeepSeek-V3-0324 | | $3.00 | $4.50 | 131,072 | 1.59 | 108 | Yes | Yes |
| deepseek-ai/DeepSeek-V3-0324 DeepSeek-V3-0324 | | $1.25 | $1.25 | 163,840 | 2.50 | 32 | Yes | No |
| deepseek-ai/DeepSeek-V3-0324 DeepSeek-V3-0324 | | $1.25 | $1.25 | 131,072 | 1.30 | 41 | Yes | Yes |
| deepseek-ai/DeepSeek-V3-0324 DeepSeek-V3-0324 | | - | - | - | - | - | - | - |
| deepcogito/cogito-671b-v2.1-FP8 cogito-671b-v2.1-FP8 | | $1.25 | $1.25 | 163,840 | 0.25 | 60 | No | Yes |
| zai-org/GLM-4.6-FP8 GLM-4.6-FP8 | | - | - | - | 1.89 | 67 | Yes | No |
| Qwen/Qwen3-Coder-480B-A35B-Instruct-FP8 Qwen3-Coder-480B-A35B-Instruct-FP8 | | $2.00 | $2.00 | 262,144 | 0.47 | 46 | Yes | Yes |
| zai-org/GLM-4.5V-FP8 GLM-4.5V-FP8 | | - | - | - | 2.84 | 41 | Yes | No |
| zai-org/GLM-4.7-FP8 GLM-4.7-FP8 | | $0.45 | $2.00 | 202,752 | 0.94 | 38 | Yes | Yes |
| zai-org/GLM-4.7-FP8 GLM-4.7-FP8 | | - | - | - | 2.11 | 75 | No | No |
| katanemo/Arch-Router-1.5B Arch-Router-1.5B | | - | - | - | 0.09 | 80 | No | Yes |
| zai-org/GLM-4.6V-FP8 GLM-4.6V-FP8 | | - | - | - | 3.30 | 34 | Yes | No |
| zai-org/GLM-4.5-Air-FP8 GLM-4.5-Air-FP8 | | $0.20 | $1.10 | 131,072 | 2.98 | 111 | Yes | Yes |
| zai-org/GLM-4.5 GLM-4.5 | | $0.60 | $2.20 | 131,072 | 0.60 | 53 | Yes | No |
| zai-org/GLM-4.5 GLM-4.5 | | - | - | - | - | - | - | - |
| zai-org/GLM-4.5 GLM-4.5 | | - | - | - | 2.28 | 48 | Yes | No |
| CohereLabs/aya-expanse-32b aya-expanse-32b | | - | - | - | 0.24 | 42 | No | No |
| zai-org/GLM-4.5-Air GLM-4.5-Air | | $0.13 | $0.85 | 131,072 | 0.59 | 66 | Yes | No |
| zai-org/GLM-4.5-Air GLM-4.5-Air | | - | - | - | - | - | - | - |
| zai-org/GLM-4.5-Air GLM-4.5-Air | | - | - | - | 1.83 | 74 | Yes | No |
| Qwen/Qwen3-235B-A22B Qwen3-235B-A22B | | $0.20 | $0.80 | 40,960 | 0.89 | 10 | No | No |
| Qwen/Qwen3-235B-A22B Qwen3-235B-A22B | | $0.20 | $0.60 | 32,000 | 1.39 | 22 | Yes | Yes |
| Qwen/Qwen3-235B-A22B Qwen3-235B-A22B | | - | - | - | - | - | - | - |
| Qwen/Qwen3-235B-A22B Qwen3-235B-A22B | | - | - | - | - | - | - | - |
| mistralai/Mixtral-8x22B-Instruct-v0.1 Mixtral-8x22B-Instruct-v0.1 | | $1.20 | $1.20 | 65,536 | 0.89 | 20 | No | Yes |
| mistralai/Mixtral-8x22B-Instruct-v0.1 Mixtral-8x22B-Instruct-v0.1 | | - | - | - | - | - | - | - |
| mistralai/Mixtral-8x22B-Instruct-v0.1 Mixtral-8x22B-Instruct-v0.1 | | $1.20 | $1.20 | 65,536 | 0.22 | 70 | No | No |
| baidu/ERNIE-4.5-21B-A3B-PT ERNIE-4.5-21B-A3B-PT | | $0.07 | $0.28 | 120,000 | 0.92 | 80 | No | No |
| baidu/ERNIE-4.5-VL-28B-A3B-PT ERNIE-4.5-VL-28B-A3B-PT | | $0.14 | $0.56 | 30,000 | 0.88 | 76 | No | No |
| baidu/ERNIE-4.5-300B-A47B-Base-PT ERNIE-4.5-300B-A47B-Base-PT | | $0.28 | $1.10 | 123,000 | 1.15 | 23 | No | Yes |
| alpindale/WizardLM-2-8x22B WizardLM-2-8x22B | | $0.62 | $0.62 | 65,535 | 1.16 | 11 | No | No |
| NousResearch/Hermes-2-Pro-Llama-3-8B Hermes-2-Pro-Llama-3-8B | | $0.14 | $0.14 | 8,192 | 0.71 | 99 | No | No |
| Sao10K/L3-8B-Lunaris-v1 L3-8B-Lunaris-v1 | | $0.05 | $0.05 | 8,192 | 1.03 | 32 | No | No |
| Sao10K/L3-70B-Euryale-v2.1 L3-70B-Euryale-v2.1 | | $1.48 | $1.48 | 8,192 | 0.85 | 35 | No | No |
| tokyotech-llm/Llama-3.3-Swallow-70B-Instruct-v0.4 Llama-3.3-Swallow-70B-Instruct-v0.4 | | $0.60 | $1.20 | 131,072 | 2.82 | 121 | No | Yes |
| CohereLabs/c4ai-command-a-03-2025 c4ai-command-a-03-2025 | | - | - | - | 0.41 | 33 | Yes | No |
| CohereLabs/command-a-vision-07-2025 command-a-vision-07-2025 | | - | - | - | 0.15 | 41 | No | No |
| utter-project/EuroLLM-22B-Instruct-2512 EuroLLM-22B-Instruct-2512 | | $0.10 | $0.20 | - | - | - | No | No |
| dicta-il/DictaLM-3.0-24B-Thinking DictaLM-3.0-24B-Thinking | | $0.20 | $0.40 | - | 1.62 | 63 | Yes | Yes |
| aisingapore/Qwen-SEA-LION-v4-32B-IT Qwen-SEA-LION-v4-32B-IT | | $0.25 | $0.50 | - | 1.85 | 43 | Yes | Yes |
| swiss-ai/Apertus-70B-Instruct-2509 Apertus-70B-Instruct-2509 | | $0.82 | $2.92 | - | 1.41 | 54 | No | Yes |
| aisingapore/Gemma-SEA-LION-v4-27B-IT Gemma-SEA-LION-v4-27B-IT | | $0.20 | $0.40 | - | 2.02 | 46 | No | Yes |