trending
Zac Zuo

5h ago

Mercury 2 - Fastest reasoning LLM built for instant production AI

Mercury 2 ditches sequential decoding for parallel refinement. As the first reasoning diffusion LLM, it generates tokens simultaneously to hit 1,000+ tokens/sec. This delivers reasoning-grade quality inside tight latency budgets for your agentic loops.