API Marketplace

Compare LLM API pricing by input/output token costs. Expand a card to see model options.

OpenAI logo
OpenAI
LLM API
OpenAI API · Text & multimodal models priced per token.

gpt-4o
General-purpose flagship model
Input: $2.50/1M Cached: $1.25/1M Output: $10.00/1M
varies
per 1M tokens
gpt-4o-mini
Cheaper + fast for many workloads
Input: $0.15/1M Cached: $0.07/1M Output: $0.60/1M
varies
per 1M tokens
gpt-4.1
Strong reasoning/coding; newer 4.x series
Input: $2.00/1M Cached: $0.50/1M Output: $8.00/1M
varies
per 1M tokens
gpt-4.1-mini
Budget reasoning/coding
Input: $0.40/1M Cached: $0.10/1M Output: $1.60/1M
varies
per 1M tokens
Anthropic logo
Anthropic
LLM API
Claude API · Claude models priced per token, plus caching options.

Claude Sonnet 4
Balanced performance + cost
Input: $3.00/1M Cached: $0.30/1M Output: $15.00/1M
varies
per 1M tokens
Claude Haiku 4.5
Fastest + cheapest
Input: $1.00/1M Cached: $0.10/1M Output: $5.00/1M
varies
per 1M tokens
Google logo
Google
LLM API
Gemini Developer API · Gemini models with free tier + paid per-token pricing.

Free Tier (limited)
Free-of-charge usage with limits (varies by model/quotas)
Free tier is quota-limited; see official Gemini pricing page for current limits.
$0
tier
Gemini (paid tier — text/image/video)
Paid tier pricing for text/image/video inputs
Input: $0.10/1M Output: $0.40/1M
varies
per 1M tokens
Gemini (paid tier — lower-cost tier)
Lower-cost paid tier shown on official pricing page
Input: $0.05/1M Output: $0.20/1M
varies
per 1M tokens
xAI logo
xAI
LLM API
Grok API · Grok models with per-token pricing and cached input discounts.

grok-4
Frontier Grok model
Input: $3.00/1M Cached: $0.75/1M Output: $15.00/1M
varies
per 1M tokens
DeepSeek logo
DeepSeek
LLM API
DeepSeek API · High-performance open-weight models with extremely low token pricing.

deepseek-chat
Input: $0.14/1M Cached: $0.07/1M Output: $0.28/1M
Pay-as-you-go
per 1M tokens
deepseek-coder
Input: $0.14/1M Cached: $0.07/1M Output: $0.28/1M
Pay-as-you-go
per 1M tokens
CloudMart AI
What are you building with an LLM?

Tell me your use case, quality needs, latency needs, and expected monthly tokens.

API recommendations powered by xAI