Jochen G. Fuchs’ Post

View profile for Jochen G. Fuchs, graphic

W&V: Ressortleiter KI&Tech, Commerce | Host Marketplace Convention | a.k.a Der E-Fuchs | Journalist - Autor - Moderator - Speaker

The Token Wars of the chip startups. Why the fight for the fastest inference will influence the future of AI. „Today, Cerebras, SambaNova, and Groq are all delivering over 1,000 tokens per second, and the “token wars” have revved up considerably. At the end of August, Cerebras claimed it had launched the “world’s fastest AI inference” at 1,800 tokens per second, and last week Cerebras said it had beaten that record and become the “first hardware of any kind” to exceed 2,000 tokens per second on one of Meta’s Llama models. “ „AI models are quickly being used to power far more complex applications than just chat.“ Cerebras CEO Andrew Feldman simplifies why: „Nobody’s going to build a business on an application that makes you sit around and wait“. A piece by Sharon Goldman at Fortune. https://lnkd.in/eMs7Pxux

Cerebras hopes planned IPO will supercharge its race against Nvidia and fellow chip startups for the fastest generative AI

Cerebras hopes planned IPO will supercharge its race against Nvidia and fellow chip startups for the fastest generative AI

fortune.com

To view or add a comment, sign in

Explore topics