Cerebras WSE-3 Launches with 900,000 Cores to Rival Nvidia in AI Model Training

Image Source: Cerebras
Cerebras Systems, a California-based company, has launched its third-generation Wafer Scale Engine (WSE-3), a processor built for large-scale AI model training. Announced on March 13, 2024, the WSE-3 includes 900,000 AI-optimized cores and about 4 trillion transistors. Cerebras seeks to compete in the AI hardware market, taking on companies like Nvidia with this release. The processor’s specifications and performance data reflect its focus on AI applications.
[Read More: Amazon Launches Trainium2 to Challenge Nvidia’s Dominance in AI Chip Market]
Processor Specifications Detailed
The WSE-3 is constructed on a single silicon wafer, roughly the size of a dinner plate, using TSMC’s 5nm process. It delivers 125 petaflops of peak AI performance, features 44 gigabytes of on-chip SRAM, and offers 21 petabytes per second of memory bandwidth. Compared to the WSE-2, the new chip doubles performance while keeping power consumption unchanged. Cerebras claims it can train AI models with up to 24 trillion parameters. Its design avoids inter-chip connections needed in multi-GPU systems, cutting latency. A May 2024 study by researchers from Sandia, Lawrence Livermore, Los Alamos, and the National Nuclear Security Administration found the WSE-2 outperformed the Frontier supercomputer by 179 times in millisecond-scale molecular dynamics simulations; the WSE-3’s doubled performance suggests it could match or exceed this. These features target the computational needs of modern AI workloads.
Efficiency in AI Model Training
The WSE-3 powers the CS-3 supercomputer, targeting the growing need for resources to train large language models, generative AI, and scientific simulations. Cerebras reports a 210-fold performance improvement over Nvidia’s H100 GPU in carbon capture simulations, though this lacks independent benchmarking. Cerebras states that the processor’s architecture cuts training time and energy use compared to traditional GPU clusters.
The company has expanded its infrastructure through the Condor Galaxy project, a network of supercomputers currently providing 16 exaflops of AI compute across multiple U.S. locations. Plans are in place to increase this capacity to 36 exaflops by the end of 2025. These efforts aim to support organizations managing extensive AI projects.
[Read More: AMD Unveils Powerful Ryzen AI Max, X3D Chips, and Z2 Series at CES 2025]
Applications and Industry Context
The WSE-3’s launch aligns with increased demand for specialized AI hardware. The CS-3 system, powered by the WSE-3, was named one of TIME’s “200 Best Inventions of 2024” and is being used in fields such as drug discovery and climate modelling. National laboratories and energy companies are among its early adopters. Research institutions note that its performance could accelerate projects previously constrained by hardware limitations.
Cost and integration details for the WSE-3 remain undisclosed, raising questions about its accessibility beyond large organizations. Meanwhile, competitors like AMD and Intel are also advancing their AI hardware offerings, adding pressure to the market. The WSE-3’s adoption rate will likely influence its long-term role in the industry.
[Read More: AI Chips Are Evolving to Mimic the Human Brain with New Spintronic Tech]
Source: Forbes

We are your source for AI news and insights. Join us as we explore the future of AI and its impact on humanity, offering thoughtful analysis and fostering community dialogue.