TAAFT
Free mode
100% free
Freemium
Free Trial
Deals

Mercury

Mercury applies a coarse-to-fine diffusion process to generate and refine multiple tokens in parallel instead of token-by-token decoding, aiming to deliver high throughput while staying compatible with common LLM use cases like tool use, RAG, and agentic workflows. The launch post highlights Mercury Coder as the first publicly available model in the family, optimized for code generation and evaluated on standard coding benchmarks and developer preference tests.
New Text Gen 7
Released: February 26, 2025

Overview

Mercury is Inception Labs’ diffusion large language model family built for ultra-fast, high-quality text generation, with a coding-optimized variant called Mercury Coder.

About Inception Labs

Inception’s breakthrough diffusion-based approach to language generation enables the world’s fastest, most efficient AI models with best-in-class quality.

Location: Palo Alto, California, US
View Company Profile

Tools using Mercury

  • Inception Chat
    The fastest commercial-grade diffusion LLM
    Open
    Inception Chat — v2
    Adds an explicit “reasoning” positioning, described as the fastest reasoning language model in their lineup. Introduces tunable reasoning as a first-class feature for trading speed vs depth per request. Adds native tool use plus schema-aligned JSON output as headline production features. Sharpens the “production loops” focus (agents, RAG, extraction) as the core value proposition rather than just raw fast generation. OpenAI-API compatibility so it can slot into existing stacks without rewrites.
Last updated: February 24, 2026
0 AIs selected
Clear selection
#
Name
Task