To run an AI model, computers must constantly shift vast amounts of data between separate memory and logic chips, a process that chokes performance. To solve this, Cerebras Systems in 2019 engineered a dinner plate–sized chip—the largest ever—that embeds both memory and logic. “People thought we were mad hatters,” says Andrew Feldman, Cerebras’s CEO and co-founder, given the huge technical hurdles. In March, the company released a third generation of the chip, the record-fast Wafer-Scale Engine 3 (WSE-3), which can train models 10 times bigger than OpenAI’s GPT-4, and will comprise the Condor Galaxy 3, a supercomputer under construction in Texas.
More Must-Reads from TIME
- Inside Elon Musk’s War on Washington
- Meet the 2025 Women of the Year
- Why Do More Young Adults Have Cancer?
- Colman Domingo Leads With Radical Love
- 11 New Books to Read in Februar
- How to Get Better at Doing Things Alone
- Cecily Strong on Goober the Clown
- Column: The Rise of America’s Broligarchy
Contact us at letters@time.com