Inception · Single modalityReleased 2026-02-24

mercury-2

inception/mercury-2 id

A diffusion-based reasoning LLM that generates text via parallel refinement (not token-by-token), delivering real-time latency with ~1k tokens/sec plus 128K context and built-in tool/JSON support.

Tool useReasoningChatFast
Type
Use mercury-2
# Drop-in OpenAI-compatible client
$ import { generateText } from 'ai'
$
$ const { text } = await generateText({
$ model: 'inception/mercury-2',
$ baseURL: 'https://synapse.garden/api/v1',
$ apiKey: process.env.MG_KEY,
$ prompt: 'Why is the sky blue?',
$ })
128K
CONTEXT WINDOW
128K
MAX OUTPUT
$0.275/M
INPUT · PER M
$0.825/M
OUTPUT · PER M
PRICING

List prices, every modality.

RatePer million tokens · USD
Input$0.275/M
Output$0.825/M
Cache read$0.028/M
Honest list pricesHow we calculate prices
MORE FROM INCEPTION

Other Inception models

See all 2
FAQ · MERCURY-2

Frequently asked

01 / 04

How do I call mercury-2 from my code?

Use the OpenAI or Anthropic SDK and point baseURL at https://synapse.garden/api/v1. Set model: ‘inception/mercury-2 and supply your Synapse Garden API key. No code changes beyond the base URL.

02 / 04

How much does mercury-2 cost?

Input: $0.275/M per million tokens. Output: $0.825/M per million tokens. The free tier includes a million tokens every month at no cost.

03 / 04

What's the context window for mercury-2?

mercury-2 supports a context window of 128K tokens, with a maximum output of 128K tokens.

04 / 04

Do I need a separate Anthropic or OpenAI account?

No. Synapse Garden is the single API surface — one key gives you OpenAI, Anthropic, Google, Meta, Mistral, DeepSeek, xAI, Cohere, and more. Billing, rate limits, and audit logs are unified.

READY

Try mercury-2 in three minutes.

Sign up, create a key, drop our base URL into your existing client. The free tier includes a million tokens every month — no credit card.