Model Cost Pages
High-intent pricing pages for users already comparing OpenAI, Claude, DeepSeek, and more.
Live pricing models
Anthropic Claude models focused on long-context reasoning and stable enterprise usage.
Input: 0.0008 | Output: 0.004
Anthropic Claude models focused on long-context reasoning and stable enterprise usage.
Input: 0.005 | Output: 0.025
Anthropic Claude models focused on long-context reasoning and stable enterprise usage.
Input: 0.005 | Output: 0.025
Anthropic Claude models focused on long-context reasoning and stable enterprise usage.
Input: 0.003 | Output: 0.015
Anthropic Claude models focused on long-context reasoning and stable enterprise usage.
Input: 0.003 | Output: 0.015
DeepSeek models known for cost-efficient reasoning and coding-focused performance.
Input: 0.00014 | Output: 0.00028
DeepSeek models known for cost-efficient reasoning and coding-focused performance.
Input: 0.002 | Output: 0.004
General-purpose model suitable for text generation and reasoning in common API workflows.
Input: 0.00004 | Output: 0.00008
General-purpose model suitable for text generation and reasoning in common API workflows.
Input: 0.0001 | Output: 0.0003
Google Gemini models for text, multimodal workloads, and high-throughput inference.
Input: 0.0003 | Output: 0.0025
Google Gemini models for text, multimodal workloads, and high-throughput inference.
Input: 0.0001 | Output: 0.0004
Google Gemini models for text, multimodal workloads, and high-throughput inference.
Input: 0.0003 | Output: 0.0025
Google Gemini models for text, multimodal workloads, and high-throughput inference.
Input: 0.00025 | Output: 0.0015
Google Gemini models for text, multimodal workloads, and high-throughput inference.
Input: 0.002 | Output: 0.01
Google Gemini models for text, multimodal workloads, and high-throughput inference.
Input: 0.002 | Output: 0.012
Google Gemini models for text, multimodal workloads, and high-throughput inference.
Input: 0.0005 | Output: 0.003
OpenAI general-purpose text and multimodal models for chat, tools, and content generation.
Input: 0.002 | Output: 0.008
OpenAI general-purpose text and multimodal models for chat, tools, and content generation.
Input: 0.0004 | Output: 0.0016
OpenAI general-purpose text and multimodal models for chat, tools, and content generation.
Input: 0.0001 | Output: 0.0004
OpenAI general-purpose text and multimodal models for chat, tools, and content generation.
Input: 0.0025 | Output: 0.01
OpenAI general-purpose text and multimodal models for chat, tools, and content generation.
Input: 0.00015 | Output: 0.0006
OpenAI general-purpose text and multimodal models for chat, tools, and content generation.
Input: 0.00175 | Output: 0.014
OpenAI general-purpose text and multimodal models for chat, tools, and content generation.
Input: 0.00175 | Output: 0.014
OpenAI general-purpose text and multimodal models for chat, tools, and content generation.
Input: 0.00175 | Output: 0.014
OpenAI general-purpose text and multimodal models for chat, tools, and content generation.
Input: 0.00175 | Output: 0.014
OpenAI general-purpose text and multimodal models for chat, tools, and content generation.
Input: 0.0025 | Output: 0.015
OpenAI general-purpose text and multimodal models for chat, tools, and content generation.
Input: 0.03 | Output: 0.18
OpenAI general-purpose text and multimodal models for chat, tools, and content generation.
Input: 0.00005 | Output: 0.0004
Moonshot Kimi models designed for long-context processing and Chinese-language Q&A.
Input: 0.00014 | Output: 0.00222
Moonshot Kimi models designed for long-context processing and Chinese-language Q&A.
Input: 0.00014 | Output: 0.00222
Moonshot Kimi models designed for long-context processing and Chinese-language Q&A.
Input: 0.0001 | Output: 0.00292
Moonshot Kimi models designed for long-context processing and Chinese-language Q&A.
Input: 0.00014 | Output: 0.00222
Moonshot Kimi models designed for long-context processing and Chinese-language Q&A.
Input: 0.00014 | Output: 0.00806
Moonshot Kimi models designed for long-context processing and Chinese-language Q&A.
Input: 0.00014 | Output: 0.00806
Alibaba Cloud Qwen models optimized for general chat and Chinese language scenarios.
Input: 0.00004 | Output: 0.00008
Alibaba Cloud Qwen models optimized for general chat and Chinese language scenarios.
Input: 0.00012 | Output: 0.00024
Alibaba Cloud Qwen models optimized for general chat and Chinese language scenarios.
Input: 0.0006 | Output: 0.0018
Guides and comparisons
A clear way to estimate Claude costs, including input vs output tokens and workflow call volume.
How to estimate DeepSeek costs and compare value across tasks, especially when you optimize prompts and reduce retries.
A decision framework for choosing between GPT-4 and Claude based on real workflow token usage.
Understand Kimi pricing with input/output tokens and workflow call volume so you can control spend.
Understand OpenAI cost drivers and how to estimate cost per token for input and output.
Estimate Qwen API cost using input/output tokens and real workflow call volume — then optimize where waste hides.
