Inception: Mercury
Other
Text
Paid
Mercury is the first diffusion large language model (dLLM). Applying a breakthrough discrete diffusion approach, the model runs 5-10x faster than even speed optimized models like GPT-4.1 Nano and Claude 3.5 Haiku while matching their performance. Mercury's speed enables developers to provide responsive user experiences, including with voice agents, search interfaces, and chatbots. Read more in the [blog post] (https://www.inceptionlabs.ai/blog/introducing-mercury) here.
Parameters
-
Context Window
128,000
tokens
Input Price
$0.25
per 1M tokens
Output Price
$1
per 1M tokens
Capabilities
Model capabilities and supported modalities
Performance
Reasoning
-
Math
-
Coding
-
Knowledge
-
Modalities
Input Modalities
text
Output Modalities
text
LLM Price Calculator
Calculate the cost of using this model
$0.000375
$0.003000
Input Cost:$0.000375
Output Cost:$0.003000
Total Cost:$0.003375
Estimated usage: 4,500 tokens
Monthly Cost Estimator
Based on different usage levels
Light Usage
$0.0125
~10 requests
Moderate Usage
$0.1250
~100 requests
Heavy Usage
$1.2500
~1000 requests
Enterprise
$12.5000
~10,000 requests
Note: Estimates based on current token count settings per request.
Last Updated: 1970/01/21
