Mercury 2 just launched from Inception Labs.
ChatGPT runs at 71 tokens per second.
Claude Haiku runs at 89 tokens per second.
Mercury 2 runs at over 1,000 tokens per second.
Want to turn this into income? 👇
Not a small gap. A completely different league.
It uses diffusion architecture — generates all tokens in parallel instead of one at a time.
Same way image AI works. Now applied to language.
Math. Code. Reasoning. Benchmarks are competitive with the best models available.
And it costs less per token because of how it generates.
If you run AI automation workflows the latency savings compound fast.
The ones who plug this in first run more tasks, process more data, serve more clients.
Want to turn this into income? 👇
🎬 See it in action 👇