Inception

Inception: Mercury

Mercury is the first diffusion large language model (dLLM). Applying a breakthrough discrete diffusion approach, the model runs 5-10x faster than even speed optimized models like GPT-4.1 Nano and Claude 3.5 Haiku while matching their performance. Mercury's speed enables developers to provide responsive user experiences, including with voice agents, search interfaces, and chatbots. Read more in the [blog post] (https://www.inceptionlabs.ai/blog/introducing-mercury) here.

Context Window
128K tokens
Parameters
N/A
Input Price
$0.25/1M
Output Price
$0.75/1M
Price Tier
budget
Provider
Inception

How to Use Inception: Mercury

With CoreAI, you can start chatting with Inception: Mercury instantly — no separate subscription needed. CoreAI bundles access to Inception: Mercury along with 300+ other AI models from Inception and other providers like OpenAI, Anthropic, Google, Meta, and more.

  1. Download the CoreAI app for iOS, Android, or use the Web App
  2. Select Inception: Mercury from the model selector
  3. Start chatting, comparing, or creating with AI

More Inception Models

Inception

Inception: Mercury 2

Mercury 2 is an extremely fast reasoning LLM, and the first reasoning diffusion LLM (dLLM). Instead of generating tokens sequentially, Mercury 2 produ
128K budget
Inception

Inception: Mercury Coder

Mercury Coder is the first diffusion large language model (dLLM). Applying a breakthrough discrete diffusion approach, the model runs 5-10x faster tha
128K budget

Try Inception: Mercury Now

Chat with Inception: Mercury and 300+ other AI models — all in one app.

Download App → Try on Web App