Inception Labs presents diffusion large language models

The AI startup Inception Labs announced a new generation of large language models (LLMs) with the Mercury family. The Mercury models are diffusion LLMs (dLLMs).

Image with the writing Mercury
Generated with Grok

The details

  • The new dLLMs no longer write text sequentially from left to right. Instead, the models generate text like an image. This approach is commonly used in image and video generation models.
  • According to Inception Labs, the models are up to 10x faster than current state-of-the-art LLMs. The new models process over 1000 tokens per second on NVIDIA H100 chips.
  • You can test the code generation model, Mercury Coder, for free at chat.inceptionlabs.ai. It is the first publicly available dLLM of Inception Labs.
  • According to Inception Labs, Mercury Coder performs excellently in standard coding tests, outperforming models like GPT-4o Mini and Claude 3.5 Haiku, and is up to 10x faster.
Speed test: Claude vs. ChatGPT vs. Mercury
Speed test: Claude vs. ChatGPT vs. Mercury (Gif by Inception Labs)

Our thoughts

Diffusion large language models represent a paradigm shift in LLM development. Diffusion models constantly improve their results, allowing them to fix mistakes and reduce hallucinations. The application of diffusion to discrete data such as text and code has not been successful until this publication.

We are excited about this approach because it reduces costs and speeds up the models.

More information: 🔗 Inception Labs News


Magic AI tool

Did you know you can run advanced LLMs like Llama 3, Phi 3, or Codestral locally on your computer? The VSCode plugin continue, and the open-source tool Ollama makes it happen.

In our blog article, you learn how to create a local AI coding assistant. After reading our article, you can chat with your codebase in VSCode and can use functions like autocompletion:

Local Coding assistant in VSCode

👉🏽 Learn more in our step-by-step guide!


Hand-picked articles


😀 Do you enjoy our content? If so, why not support us with a small financial contribution? This helps us fund our work to ensure we can stick around long-term.