Inception: Mercury
Other
Text
Paid
Mercury is the first diffusion large language model (dLLM). Applying a breakthrough discrete diffusion approach, the model runs 5-10x faster than even speed optimized models like GPT-4.1 Nano and Claude 3.5 Haiku while matching their performance. Mercury's speed enables developers to provide responsive user experiences, including with voice agents, search interfaces, and chatbots. Read more in the blog post here.
Parameters
-
Context Window
32,000
tokens
Input Price
$0.25
per 1M tokens
Output Price
$1
per 1M tokens
Capabilities
Model capabilities and supported modalities
Performance
Reasoning
-
Math
-
Coding
-
Knowledge
-
Modalities
Input Modalities
text
Output Modalities
text
LLM Price Calculator
Calculate the cost of using this model
$0.000375
$0.003000
Input Cost:$0.000375
Output Cost:$0.003000
Total Cost:$0.003375
Estimated usage: 4,500 tokens
Monthly Cost Estimator
Based on different usage levels
Light Usage
$0.0125
~10 requests
Moderate Usage
$0.1250
~100 requests
Heavy Usage
$1.2500
~1000 requests
Enterprise
$12.5000
~10,000 requests
Note: Estimates based on current token count settings per request.
Last Updated: 1970/01/21