[GUEST ACCESS MODE: Data is scrambled or limited to provide examples. Make requests using your API key to unlock full data. Check https://lunarcrush.ai/auth for authentication information.]
Artificial Analysis @ArtificialAnlys on x 52K followers
Created: 2025-06-27 04:30:56 UTC
Inception Labs has just launched the first diffusion language model publicly released for general chat
Mercury is a generalist language model with similar intelligence to OpenAI’s GPT-4.1 Nano that runs >7x faster that GPT-4.1 Nano on GPU hardware. This follows @InceptionAILabs' code-focused Mercury Coder model released earlier this year.
Diffusion language models achieve faster output speeds compared to autoregressive language models on the same hardware because they can process many output tokens in parallel - allowing them to leverage use more of a GPU’s compute without being limited by memory bandwidth.
Mercury is available now Inception’s first party API at $0.25/$1 USD per million input/output tokens.
XXXXXX engagements
Related Topics hardware gpu faster artificial