Inference Providers

·

Metrics for top trending models

GLM-5
- - - - - - -
GLM-5
together
cheapest
$1.00 $3.20 202,752 0.63 34 Yes Yes
GLM-5
zai-org
fastest
- - - 3.57 35 No No
Kimi-K2.5
- - - - - - -
Kimi-K2.5
$0.50 $2.80 262,144 0.50 80 Yes Yes
Kimi-K2.5
- - - - - - -
GLM-4.7-Flash
- - - - - - -
GLM-4.7-Flash
- - - 2.65 60 Yes No
Llama-3.1-8B-Instruct
- - - - - - -
Llama-3.1-8B-Instruct
cerebras
cheapest fastest
$0.10 $0.10 - 0.25 1,305 No No
Llama-3.1-8B-Instruct
- - - - - - -
Llama-3.1-8B-Instruct
- - - - - - -
Llama-3.1-8B-Instruct
- - - - - - -
Llama-3.1-8B-Instruct
- - - 0.31 84 Yes Yes
Llama-3.1-8B-Instruct
$0.11 $0.11 131,072 0.32 42 Yes No
gpt-oss-20b
- - - - - - -
gpt-oss-20b
- - - - - - -
gpt-oss-20b
- - - - - - -
gpt-oss-20b
hyperbolic
cheapest fastest
$0.10 $0.10 131,072 0.57 182 Yes No
gpt-oss-20b
$0.05 $0.20 131,072 0.22 107 Yes Yes
gpt-oss-20b
- - - - - - -
gpt-oss-20b
$0.05 $0.18 131,072 0.33 110 Yes Yes
MiniMax-M2.1
$0.30 $1.20 204,800 0.97 37 Yes No
gpt-oss-120b
- - - - - - -
gpt-oss-120b
- - - - - - -
gpt-oss-120b
cerebras
fastest
$0.25 $0.69 - 0.23 1,161 Yes No
gpt-oss-120b
- - - - - - -
gpt-oss-120b
- - - - - - -
gpt-oss-120b
- - - - - - -
gpt-oss-120b
$0.15 $0.60 131,072 0.34 94 Yes Yes
gpt-oss-120b
- - - - - - -
gpt-oss-120b
- - - - - - -
gpt-oss-120b
ovhcloud
cheapest
$0.09 $0.47 131,072 0.29 119 Yes Yes
GLM-4.7
- - - - - - -
GLM-4.7
- - - - - - -
GLM-4.7
- - - 1.64 89 Yes No
Qwen3-Coder-Next-FP8
$0.50 $1.20 262,144 0.74 127 Yes Yes
DeepSeek-V3.2
- - - - - - -
DeepSeek-V3.2
- - 163,840 0.67 82 Yes No
Qwen3-VL-8B-Instruct
- - - - - - -
Qwen3-VL-8B-Instruct
$0.18 $0.68 262,144 0.17 135 Yes Yes
Llama-3.2-3B-Instruct
- - - - - - -
Llama-3.2-3B-Instruct
- - - - - - -
Llama-3.2-3B-Instruct
$0.10 $0.10 131,072 0.61 95 Yes No
Llama-3.2-3B-Instruct
together
cheapest fastest
$0.06 $0.06 131,072 0.26 122 Yes Yes
MiniMax-M2
$0.30 $1.20 204,800 1.99 42 Yes No
gemma-3-27b-it
- - - 0.53 43 Yes No
Qwen2.5-7B-Instruct
$0.30 $0.30 32,768 0.32 79 Yes Yes
DeepSeek-R1
- - - - - - -
DeepSeek-R1
- - - - - - -
DeepSeek-R1
- - - - - - -
DeepSeek-R1
$3.00 $7.00 163,840 0.48 46 No Yes
DeepSeek-R1
- - - - - - -
Qwen3-32B
- - - - - - -
Qwen3-32B
$0.10 $0.45 40,960 0.53 48 No No
Qwen3-32B
cerebras
fastest
$0.40 $0.80 - 0.20 1,121 Yes No
Qwen3-32B
- - - - - - -
Qwen3-32B
- - - - - - -
Qwen3-32B
ovhcloud
cheapest
$0.09 $0.25 32,768 0.37 41 Yes Yes
Qwen3-Coder-30B-A3B-Instruct
- - - - - - -
Qwen3-Coder-30B-A3B-Instruct
scaleway
fastest
- - - 0.36 86 Yes No
Qwen3-Coder-30B-A3B-Instruct
ovhcloud
cheapest
$0.07 $0.26 262,144 0.28 71 Yes Yes
Kimi-K2-Thinking
- - - - - - -
Kimi-K2-Thinking
$1.20 $4.00 262,144 0.78 114 Yes Yes
GLM-4.6V-Flash
novita
cheapest
$0.30 $0.90 131,072 0.85 20 Yes No
GLM-4.6V-Flash
zai-org
fastest
- - - 2.22 60 Yes No
Llama-3.3-70B-Instruct
- - - - - - -
Llama-3.3-70B-Instruct
novita
cheapest
$0.14 $0.40 131,072 0.81 42 Yes No
Llama-3.3-70B-Instruct
cerebras
fastest
$0.85 $1.20 - 0.23 1,009 Yes No
Llama-3.3-70B-Instruct
$0.60 $1.20 131,072 0.48 281 Yes Yes
Llama-3.3-70B-Instruct
- - - - - - -
Llama-3.3-70B-Instruct
- - - - - - -
Llama-3.3-70B-Instruct
$0.88 $0.88 131,072 0.34 133 Yes Yes
Llama-3.3-70B-Instruct
- - - - - - -
Llama-3.3-70B-Instruct
- - - 0.34 81 Yes Yes
Llama-3.3-70B-Instruct
$0.74 $0.74 131,072 0.34 36 Yes Yes
MiMo-V2-Flash
$0.10 $0.30 262,144 1.00 57 No No
Mistral-7B-Instruct-v0.2
$0.20 $0.20 32,768 0.23 95 Yes Yes
Apriel-1.6-15b-Thinker
$0.00 $0.00 131,072 0.14 150 No No
GLM-4.5-Air
- - - - - - -
GLM-4.5-Air
- - - - - - -
GLM-4.5-Air
- - - 1.62 85 Yes No
Qwen3-Coder-480B-A35B-Instruct
- - - - - - -
Qwen3-Coder-480B-A35B-Instruct
- - - - - - -
Qwen3-Coder-480B-A35B-Instruct
- - - - - - -
Qwen3-Coder-480B-A35B-Instruct
$2.00 $2.00 262,144 0.40 62 Yes Yes
Qwen3-Coder-480B-A35B-Instruct
- - - - - - -
Qwen3-Next-80B-A3B-Instruct
- - - - - - -
Qwen3-Next-80B-A3B-Instruct
hyperbolic
cheapest fastest
$0.30 $0.30 262,144 0.52 166 No No
Qwen3-Next-80B-A3B-Instruct
$0.15 $1.50 262,144 0.43 115 Yes Yes
SmolLM3-3B
- - - 0.11 91 Yes Yes
Qwen2.5-Coder-32B-Instruct
- - - - - - -
Qwen2.5-Coder-32B-Instruct
hyperbolic
cheapest
$0.20 $0.20 32,768 0.42 41 No No
Qwen2.5-Coder-32B-Instruct
- - - - - - -
Qwen2.5-Coder-32B-Instruct
scaleway
fastest
- - - 0.37 94 Yes No
Kimi-K2-Instruct
- - - - - - -
Kimi-K2-Instruct
$1.00 $3.00 131,072 0.55 50 Yes Yes
Kimi-K2-Instruct
- - - - - - -
Olmo-3.1-32B-Think
$0.15 $0.50 - 2.39 76 No No
GLM-4.6V
- - - - - - -
GLM-4.6V
- - - 3.47 44 Yes No
Qwen3-VL-32B-Instruct
$0.50 $1.50 262,144 1.03 98 Yes No
Kimi-K2-Instruct-0905
- - - - - - -
Kimi-K2-Instruct-0905
- - - - - - -
Kimi-K2-Instruct-0905
$1.00 $3.00 262,144 0.57 49 Yes Yes
Arch-Router-1.5B
- - - 0.19 69 No Yes
Qwen3-235B-A22B-Instruct-2507
- - - - - - -
Qwen3-235B-A22B-Instruct-2507
- - - - - - -
Qwen3-235B-A22B-Instruct-2507
- - - - - - -
Qwen3-235B-A22B-Instruct-2507
$2.00 $2.00 262,144 0.82 48 Yes No
Qwen3-235B-A22B-Instruct-2507
together
cheapest
$0.20 $0.60 262,144 0.51 44 Yes Yes
Qwen3-235B-A22B-Instruct-2507
- - - - - - -
Qwen3-235B-A22B-Instruct-2507
scaleway
fastest
- - - 0.30 87 Yes Yes
DeepSeek-V3
- - - - - - -
DeepSeek-V3
$1.25 $1.25 131,072 0.51 85 Yes Yes
DeepSeek-V3
- - - - - - -
Llama-4-Maverick-17B-128E-Instruct
- - - - - - -
Llama-4-Maverick-17B-128E-Instruct
$0.63 $1.80 131,072 2.86 240 Yes Yes
Llama-4-Maverick-17B-128E-Instruct
- - - - - - -
Molmo2-8B
- - - 2.19 127 No Yes
Olmo-3.1-32B-Instruct
$0.20 $0.60 - 2.03 50 Yes Yes
Qwen3-Next-80B-A3B-Thinking
- - - - - - -
Qwen3-Next-80B-A3B-Thinking
- - - - - - -
Qwen3-Next-80B-A3B-Thinking
$0.15 $1.50 262,144 0.86 177 Yes Yes
GLM-4.7-FP8
together
cheapest
$0.45 $2.00 202,752 0.49 82 Yes Yes
GLM-4.7-FP8
zai-org
fastest
- - - 2.34 88 Yes No
Qwen3-235B-A22B
$0.20 $0.80 40,960 0.39 16 No No
Qwen3-235B-A22B
- - - - - - -
Qwen3-235B-A22B
- - - - - - -
Qwen3-235B-A22B
- - - - - - -
GLM-4.5
- - - - - - -
GLM-4.5
- - - - - - -
GLM-4.5
- - - 2.10 56 Yes No
gemma-3n-E4B-it
$0.02 $0.04 32,768 0.35 33 No Yes
Llama-4-Maverick-17B-128E-Instruct-FP8
- - - - - - -
Llama-4-Maverick-17B-128E-Instruct-FP8
$0.27 $0.85 1,048,576 0.40 104 Yes Yes
rnj-1-instruct
$0.15 $0.15 32,768 0.22 123 Yes Yes
GLM-4.5V
- - - - - - -
GLM-4.5V
- - - - - - -
GLM-4.5V
- - - 1.73 61 Yes No
NVIDIA-Nemotron-Nano-9B-v2
$0.06 $0.25 131,072 0.93 108 Yes No
Qwen3-Coder-480B-A35B-Instruct-FP8
$2.00 $2.00 262,144 0.46 63 Yes Yes
DeepSeek-V3.1
- - - - - - -
DeepSeek-V3.1
$0.60 $1.70 131,072 0.46 116 Yes No
DeepSeek-V3.1
- - - - - - -
Llama-3.1-70B-Instruct
- - - - - - -
Llama-3.1-70B-Instruct
- - - 0.31 81 Yes Yes
GLM-4.6V-FP8
- - - 3.09 43 No No
Apertus-70B-Instruct-2509
$0.82 $2.92 - 1.96 53 No Yes
GLM-4.6
- - - - - - -
GLM-4.6
- - - - - - -
GLM-4.6
together
cheapest fastest
$0.60 $2.20 202,752 0.43 73 Yes Yes
GLM-4.6
- - - 2.07 58 Yes No
Trinity-Mini
$0.05 $0.15 128,000 0.16 185 Yes Yes
DeepSeek-V3-0324
- - - - - - -
DeepSeek-V3-0324
- - - - - - -
DeepSeek-V3-0324
hyperbolic
cheapest
$1.25 $1.25 163,840 0.90 62 Yes No
DeepSeek-V3-0324
together
fastest
$1.25 $1.25 131,072 1.15 90 Yes Yes
DeepSeek-V3-0324
- - - - - - -
aya-expanse-32b
- - - 0.18 42 No No
aya-vision-32b
- - - 0.66 36 No No
GLM-4-32B-0414
$0.55 $1.66 32,000 0.88 46 No No
Apertus-8B-Instruct-2509
$0.10 $0.20 - 1.79 95 No No
L3-8B-Lunaris-v1
$0.05 $0.05 8,192 0.71 33 No No
marin-8b-instruct
$0.18 $0.18 4,096 0.15 173 No Yes
MiniMax-M1-80k
$0.55 $2.20 1,000,000 2.53 23 Yes No
cogito-671b-v2.1
$1.25 $1.25 163,840 0.32 55 No Yes
cogito-671b-v2.1
- - - - - - -
cogito-671b-v2.1-FP8
$1.25 $1.25 163,840 0.25 57 No Yes
GLM-4.5V-FP8
- - - 2.03 62 Yes No
DeepSeek-R1-0528
- - - - - - -
DeepSeek-R1-0528
sambanova
fastest
$5.00 $7.00 131,072 0.67 194 Yes Yes
DeepSeek-R1-0528
$3.00 $3.00 163,840 0.86 53 No No
DeepSeek-R1-0528
together
cheapest
$0.00 $0.00 163,840 0.75 47 No Yes
DeepSeek-R1-0528
- - - - - - -
GLM-4.6-FP8
- - - 2.28 58 Yes No
c4ai-command-r-08-2024
- - - 0.29 33 Yes No
QwQ-32B
- - - - - - -
QwQ-32B
- - - - - - -
QwQ-32B
$0.18 $0.20 131,072 1.63 28 Yes No
QwQ-32B
- - - - - - -
QwQ-32B
- - - - - - -
DeepSeek-R1-Distill-Llama-70B
- - - - - - -
DeepSeek-R1-Distill-Llama-70B
- - - - - - -
DeepSeek-R1-Distill-Llama-70B
- - - - - - -
DeepSeek-R1-Distill-Llama-70B
- - - - - - -
DeepSeek-R1-Distill-Llama-70B
- - - - - - -
DeepSeek-R1-Distill-Llama-70B
- - - 1.16 25 No Yes
DeepSeek-R1-Distill-Llama-70B
- - - - - - -
Llama-3.3-Swallow-70B-Instruct-v0.4
$0.60 $1.20 131,072 2.94 153 No Yes
c4ai-command-r7b-12-2024
- - - 0.19 130 Yes No
command-a-translate-08-2025
- - - 0.20 65 Yes No
GLM-4.5-Air-FP8
$0.20 $1.10 131,072 0.75 130 Yes Yes
command-a-reasoning-08-2025
- - - 0.16 82 Yes No
command-a-vision-07-2025
- - - 0.16 42 No No
c4ai-command-a-03-2025
- - - 0.21 38 Yes No
c4ai-command-r7b-arabic-02-2025
- - - 0.17 139 Yes No