Inception, the company behind the first commercial diffusion large language models (dLLMs), today announced the launch of Mercury 2, the fastest reasoning LLM and first reasoning dLLM. Mercury 2 ...
The new Mercury 2 AI model uses diffusion reasoning to generate 1,000 tokens per second; it runs about 5x faster than Haiku, speed limits are ...
Mercury 2 introduces diffusion LLMs to text, delivering 10x faster speeds for AI agents and production workflows without sacrificing reasoning power.
Microsoft Corp. has developed a series of large language models that can rival algorithms from OpenAI and Anthropic PBC, multiple publications reported today. Sources told Bloomberg that the LLM ...
Google at Google Cloud Next 24 unveiled three open source projects for building and running generative AI models. The company also introduced new large language models to its MaxText project of ...
A little over a year after it upended the tech industry, DeepSeek is back with another apparent breakthrough: a means to stop current large language models (LLMs) from wasting computational depth on ...
LiteLLM allows developers to integrate a diverse range of LLM models as if they were calling OpenAI’s API, with support for fallbacks, budgets, rate limits, and real-time monitoring of API calls. The ...
Exposed endpoints quietly expand attack surfaces across LLM infrastructure. Learn why endpoint privilege management is important to AI security.