Model Cost Pages
High-intent pricing pages for users already comparing OpenAI, Claude, DeepSeek, and more.
Live pricing models
Anthropic Claude models focused on long-context reasoning and stable enterprise usage.
Input: 0.0008 | Output: 0.004
Anthropic Claude models focused on long-context reasoning and stable enterprise usage.
Input: 0.005 | Output: 0.025
Anthropic Claude models focused on long-context reasoning and stable enterprise usage.
Input: 0.005 | Output: 0.025
Anthropic Claude models focused on long-context reasoning and stable enterprise usage.
Input: 0.003 | Output: 0.015
Anthropic Claude models focused on long-context reasoning and stable enterprise usage.
Input: 0.003 | Output: 0.015
DeepSeek models known for cost-efficient reasoning and coding-focused performance.
Input: 0.00014 | Output: 0.00028
DeepSeek models known for cost-efficient reasoning and coding-focused performance.
Input: 0.002 | Output: 0.004
General-purpose model suitable for text generation and reasoning in common API workflows.
Input: 0.00004 | Output: 0.00008
General-purpose model suitable for text generation and reasoning in common API workflows.
Input: 0.0001 | Output: 0.0003
Google Gemini models for text, multimodal workloads, and high-throughput inference.
Input: 0.0003 | Output: 0.0025
Google Gemini models for text, multimodal workloads, and high-throughput inference.
Input: 0.0001 | Output: 0.0004
Google Gemini models for text, multimodal workloads, and high-throughput inference.
Input: 0.0003 | Output: 0.0025
Google Gemini models for text, multimodal workloads, and high-throughput inference.
Input: 0.00025 | Output: 0.0015
Google Gemini models for text, multimodal workloads, and high-throughput inference.
Input: 0.002 | Output: 0.01
Google Gemini models for text, multimodal workloads, and high-throughput inference.
Input: 0.002 | Output: 0.012
Google Gemini models for text, multimodal workloads, and high-throughput inference.
Input: 0.0005 | Output: 0.003
OpenAI general-purpose text and multimodal models for chat, tools, and content generation.
Input: 0.002 | Output: 0.008
OpenAI general-purpose text and multimodal models for chat, tools, and content generation.
Input: 0.0004 | Output: 0.0016
OpenAI general-purpose text and multimodal models for chat, tools, and content generation.
Input: 0.0001 | Output: 0.0004
OpenAI general-purpose text and multimodal models for chat, tools, and content generation.
Input: 0.0025 | Output: 0.01
OpenAI general-purpose text and multimodal models for chat, tools, and content generation.
Input: 0.00015 | Output: 0.0006
OpenAI general-purpose text and multimodal models for chat, tools, and content generation.
Input: 0.00175 | Output: 0.014
OpenAI general-purpose text and multimodal models for chat, tools, and content generation.
Input: 0.00175 | Output: 0.014
OpenAI general-purpose text and multimodal models for chat, tools, and content generation.
Input: 0.00175 | Output: 0.014
OpenAI general-purpose text and multimodal models for chat, tools, and content generation.
Input: 0.00175 | Output: 0.014
OpenAI general-purpose text and multimodal models for chat, tools, and content generation.
Input: 0.0025 | Output: 0.015
OpenAI general-purpose text and multimodal models for chat, tools, and content generation.
Input: 0.03 | Output: 0.18
OpenAI general-purpose text and multimodal models for chat, tools, and content generation.
Input: 0.00005 | Output: 0.0004
Moonshot Kimi models designed for long-context processing and Chinese-language Q&A.
Input: 0.00014 | Output: 0.00222
Moonshot Kimi models designed for long-context processing and Chinese-language Q&A.
Input: 0.00014 | Output: 0.00222
Moonshot Kimi models designed for long-context processing and Chinese-language Q&A.
Input: 0.0001 | Output: 0.00292
Moonshot Kimi models designed for long-context processing and Chinese-language Q&A.
Input: 0.00014 | Output: 0.00222
Moonshot Kimi models designed for long-context processing and Chinese-language Q&A.
Input: 0.00014 | Output: 0.00806
Moonshot Kimi models designed for long-context processing and Chinese-language Q&A.
Input: 0.00014 | Output: 0.00806
Alibaba Cloud Qwen models optimized for general chat and Chinese language scenarios.
Input: 0.00004 | Output: 0.00008
Alibaba Cloud Qwen models optimized for general chat and Chinese language scenarios.
Input: 0.00012 | Output: 0.00024
Alibaba Cloud Qwen models optimized for general chat and Chinese language scenarios.
Input: 0.0006 | Output: 0.0018
Guides and comparisons
Claude-Kosten schätzen, inkl. input/output Token und Call-Volume eures Workflows.
So schätzt ihr DeepSeek-Kosten und vergleicht Nutzen, wenn ihr Prompts und Retries optimiert.
Ein Rahmen, um zwischen GPT-4 und Claude basierend auf total gebillten Tokens zu entscheiden.
Verstehe Kimi Pricing mit Input/Output Tokens und Workflow-Call-Volume — damit du Spend kontrollieren kannst.
Versteht OpenAI-Kosten und schätzt cost per token für input und output.
Schätze Qwen API Kosten mit Input/Output Tokens und echtem Workflow-Call-Volume — und optimiere dort, wo Waste steckt.
