Main Menu
  • Tools
  • Developers
  • Topics
  • Discussions
  • News
  • Blogs
  • Builds
  • Contests
  • Compare
Create
    EveryDev.ai
    Sign inSubscribe
    Home
    Tools

    1,944+ AI tools

    • New
    • Trending
    • Featured
    • Compare
    Categories
    • Agents1038
    • Coding971
    • Infrastructure415
    • Marketing398
    • Design335
    • Projects313
    • Analytics299
    • Research290
    • Testing183
    • Integration167
    • Data163
    • Security156
    • MCP145
    • Learning135
    • Communication120
    • Extensions114
    • Prompts110
    • Commerce106
    • Voice102
    • DevOps84
    • Web71
    • Finance18
    1. Home
    2. Tools
    3. Inception Labs
    Inception Labs icon

    Inception Labs

    AI Infrastructure

    Diffusion-based large language models that generate tokens in parallel, delivering 5x faster inference with best-in-class quality at lower cost.

    Visit Website

    At a Glance

    Pricing
    Paid
    Mercury 2: $0.25
    Mercury 2 Output: $0.75
    Mercury Edit: $0.25
    +2 more plans

    Engagement

    Available On

    Web
    API

    Resources

    WebsiteDocsllms.txt

    Topics

    AI InfrastructureAI Coding AssistantsLocal Inference

    Alternatives

    Arcee AITrillion LabsvLLM
    Developer
    Inception LabsPalo Alto, CAEst. 2024$50000000 raised

    Listed Feb 2026

    About Inception Labs

    Inception Labs builds and deploys next-generation large language models (LLMs) powered by diffusion rather than traditional auto-regressive generation. By using diffusion, their Mercury models produce many tokens in parallel, making them several times faster and less than half the cost of conventional LLMs. The diffusion framework provides fine-grained control over outputs, allowing adherence to specific schemas and semantic constraints while offering a unified paradigm for combining language with other data modalities.

    • Parallel Token Generation enables Mercury models to generate multiple tokens simultaneously instead of one at a time, resulting in blazing-fast inference speeds that are 5x faster than traditional LLMs.

    • Mercury 2 Reasoning Model is the fastest reasoning LLM and the first reasoning diffusion LLM, ideal for complex applications where both performance and speed are crucial.

    • Mercury Edit is a small, coding-focused diffusion LLM designed for code editing and extremely latency-sensitive components of coding workflows.

    • OpenAI API Compatible means Mercury models integrate seamlessly into existing LLM workflows as a drop-in replacement with minimal code changes.

    • Enterprise-Grade Deployment options include Inception API, AWS Bedrock, Azure Foundry, and model routers like OpenRouter, with configurable data retention, private networking, and custom SLAs.

    • Real-Time Voice Applications enable natural AI engagement in voice-powered workflows like customer support, translation, and immersive gaming experiences.

    • Lightning Fast Agents automate complex coding and business workflows with ultra-responsive AI that stays in flow without interrupting user thinking.

    • Cost-Effective Pricing at $0.25 per 1M input tokens and $0.75 per 1M output tokens makes high-performance AI accessible for production applications.

    To get started, request early access through the Inception website or access Mercury through AWS Bedrock, Azure Foundry, or model routers. The API is OpenAI-compatible, requiring only a one-line code change for integration. Documentation is available at docs.inceptionlabs.ai for detailed implementation guidance.

    Inception Labs - 1

    Community Discussions

    Be the first to start a conversation about Inception Labs

    Share your experience with Inception Labs, ask questions, or help others learn from your insights.

    Pricing

    Mercury 2

    Input pricing per 1M tokens for the fastest reasoning LLM

    $0.25
    usage based
    • Fastest reasoning LLM
    • First reasoning dLLM
    • Complex applications support
    • OpenAI API compatible

    Mercury 2 Output

    Output pricing per 1M tokens for Mercury 2

    $0.75
    usage based
    • Parallel token generation
    • Best-in-class quality
    • Enterprise-grade reliability

    Mercury Edit

    Input pricing per 1M tokens for coding-focused dLLM

    $0.25
    usage based
    • Small coding-focused model
    • Code editing optimized
    • Extremely latency-sensitive workflows

    Mercury Edit Output

    Output pricing per 1M tokens for Mercury Edit

    $0.75
    usage based
    • Fast code completions
    • Tab suggestions
    • Chat responses

    Enterprise

    Custom enterprise deployment with dedicated support

    Custom
    contact sales
    • Private networking
    • Dedicated capacity
    • Custom SLAs
    • 99.5%+ uptime
    • Priority support
    • No prompt logging options
    View official pricing

    Capabilities

    Key Features

    • Parallel token generation
    • Diffusion-based language models
    • Mercury 2 reasoning model
    • Mercury Edit coding model
    • OpenAI API compatible
    • Real-time voice applications
    • Lightning fast agents
    • Instant code editing
    • Rapid search capabilities
    • Enterprise-grade privacy
    • AWS Bedrock integration
    • Azure Foundry integration
    • Custom SLAs
    • No training on customer data
    • Configurable data retention

    Integrations

    AWS Bedrock
    Azure Foundry
    OpenRouter
    Poe
    OpenAI API
    API Available
    View Docs

    Reviews & Ratings

    No ratings yet

    Be the first to rate Inception Labs and help others make informed decisions.

    Developer

    Inception Labs Team

    Inception Labs builds diffusion-based large language models that generate tokens in parallel for dramatically faster inference. The team includes leading researchers and engineers from Stanford, UCLA, Cornell, Google DeepMind, Meta AI, Microsoft AI, and OpenAI who pioneered foundational AI technologies including diffusion models, Flash Attention, and Direct Preference Optimization. The company deploys Mercury diffusion LLMs at Fortune 500 companies through its API, AWS Bedrock, and Azure Foundry.

    Founded 2024
    Palo Alto, CA
    $50000000 raised
    40 employees

    Used by

    Early access developers and enterprises…
    Read more about Inception Labs Team
    WebsiteLinkedInX / Twitter
    1 tool in directory

    Similar Tools

    Arcee AI icon

    Arcee AI

    US-based open intelligence lab building open-weight foundation models that run anywhere - on edge, on-prem, or cloud.

    Trillion Labs icon

    Trillion Labs

    Research lab building efficient foundation models for Korean and Asian languages in mission-critical domains.

    vLLM icon

    vLLM

    An open-source, high-performance library for serving and running large language models with GPU-optimized inference and efficient memory and batch management.

    Browse all tools

    Related Topics

    AI Infrastructure

    Infrastructure designed for deploying and running AI models.

    183 tools

    AI Coding Assistants

    AI tools that help write, edit, and understand code with intelligent suggestions.

    366 tools

    Local Inference

    Tools and platforms for running AI inference locally without cloud dependence.

    67 tools
    Browse all topics
    Back to all tools
    Explore AI Tools
    • AI Coding Assistants
    • Agent Frameworks
    • MCP Servers
    • AI Prompt Tools
    • Vibe Coding Tools
    • AI Design Tools
    • AI Database Tools
    • AI Website Builders
    • AI Testing Tools
    • LLM Evaluations
    Follow Us
    • X / Twitter
    • LinkedIn
    • Reddit
    • Discord
    • Threads
    • Bluesky
    • Mastodon
    • YouTube
    • GitHub
    • Instagram
    Get Started
    • About
    • Editorial Standards
    • Corrections & Disclosures
    • Community Guidelines
    • Advertise
    • Contact Us
    • Newsletter
    • Submit a Tool
    • Start a Discussion
    • Write A Blog
    • Share A Build
    • Terms of Service
    • Privacy Policy
    Explore with AI
    • ChatGPT
    • Gemini
    • Claude
    • Grok
    • Perplexity
    Agent Experience
    • llms.txt
    Theme
    With AI, Everyone is a Dev. EveryDev.ai © 2026
    10views
    Discussions