-

Cerebras Demonstrates Trillion Parameter Model Training on a Single CS-3 System

SUNNYVALE, Calif. & VANCOUVER, British Columbia--(BUSINESS WIRE)--Today at NeurIPS 2024, Cerebras Systems, the pioneer in accelerating generative AI, today announced a groundbreaking achievement in collaboration with Sandia National Laboratories: successfully demonstrating training of a 1 trillion parameter AI model on a single CS-3 system. Trillion parameter models represent the state of the art in today’s LLMs, requiring thousands of GPUs and dozens of hardware experts to perform. By leveraging Cerebras’ Wafer Scale Cluster technology, researchers at Sandia were able to initiate training on a single AI accelerator – a one-of-a-kind achievement for frontier model development.

“Traditionally, training a model of this scale would require thousands of GPUs, significant infrastructure complexity, and a team of AI infrastructure experts,” said Sandia researcher Siva Rajamanickam. “With the Cerebras CS-3, the team was able to achieve this feat on a single system with no changes to model or infrastructure code. The model was then scaled up seamlessly to 16 CS-3 systems, demonstrating a step-change in the linear scalability and performance of large AI models, thanks to the Cerebras Wafer-Scale Cluster.”

Trillion parameter models require terabytes of memory — thousands of times more than what’s available on a single GPU. Thousands of GPUs must be procured and connected before being able to run a single training step or model experiment. Cerebras Wafer Scale Cluster uses unique, terabyte-scale external memory device called MemoryX to store model weights, making trillion parameter models as easy to train as a small model on a GPU.

For Sandia’s trillion parameter training run, Cerebras configured a 55 terabyte MemoryX device. By employing commodity DDR5 memory in a 1U server format, the hardware was procured and configured in mere days. AI researchers were able to run initial training steps and observed improving loss and stable training dynamics. After completing the single system run, researchers scaled training to two and sixteen CS-3 nodes with no code changes. The cluster exhibited near linear scaling with 15.3x speedup on sixteen systems. Achieving the above typically requires thousands of GPUs, megawatts of power, and many weeks of hardware and software configuration.

This result highlights the one-of-a-kind power and flexibility of Cerebras hardware. In addition to the industry’s fastest inference performance, Cerebras Wafer Scale engine dramatically simplifies AI training and frontier model development, making it a full end-to-end solution for training, fine-tuning, and inferencing the latest AI models.

For more information, please visit www.cerebras.ai.

About Cerebras Systems

Cerebras Systems is a team of pioneering computer architects, computer scientists, deep learning researchers, and engineers of all types. We have come together to accelerate generative AI by building from the ground up a new class of AI supercomputer. Our flagship product, the CS-3 system, is powered by the world's largest and fastest commercially available AI processor, our Wafer-Scale Engine-3. CS-3s are quickly and easily clustered together to make the largest AI supercomputers in the world, and make placing models on the supercomputers dead simple by avoiding the complexity of distributed computing. Cerebras Inference delivers breakthrough inference speeds, empowering customers to create cutting-edge AI applications. Leading corporations, research institutions, and governments use Cerebras solutions for the development of pathbreaking proprietary models, and to train open-source models with millions of downloads. Cerebras solutions are available through the Cerebras Cloud and on premise. For further information, visit www.cerebras.ai or follow us on LinkedIn or X.

Contacts

Media Contact
Press Contact: PR@zmcommunications.com

Cerebras Systems


Release Versions

Contacts

Media Contact
Press Contact: PR@zmcommunications.com

More News From Cerebras Systems

Cerebras Launches Cerebras Inference Cloud Availability in AWS Marketplace

PARIS--(BUSINESS WIRE)--Today at the RAISE Summit in Paris, France, Cerebras Systems announced that Cerebras Inference Cloud is now available in AWS Marketplace bringing Cerebras’ ultra-fast AI inference to enterprise customers, and enabling the next era of high performance, interactive, and intelligent agentic AI applications. Amazon Web Services (AWS) customers will now have access to Cerebras Inference Cloud directly within AWS Marketplace. This provides AWS customers with the ability to str...

Cerebras Enables Notion to Deliver Real-Time Enterprise Search for 100+ Million Workspace Users

PARIS & SUNNYVALE, Calif.--(BUSINESS WIRE)--Cerebras Systems, the pioneer in accelerating generative AI, today announced that Notion, the all-in-one connected workspace, is using Cerebras’ industry-leading AI inference technology to power instant, enterprise-scale document search for its AI offering, Notion AI for Work. With more than 100 million users worldwide, Notion is redefining productivity for teams across the globe. Now, by running enterprise search on Cerebras, Notion delivers the spee...

Cerebras Launches Qwen3-235B: World's Fastest Frontier AI Model with Full 131K Context Support

PARIS--(BUSINESS WIRE)--Cerebras Systems today announced the launch of Qwen3-235B with full 131K context support on its inference cloud platform. This milestone represents a breakthrough in AI model performance, combining frontier-level intelligence with unprecedented speed at one-tenth the cost of closed-source models, fundamentally transforming enterprise AI deployment. Frontier Intelligence on Cerebras Alibaba’s Qwen3-235B delivers model intelligence that rivals frontier models such as Claud...
Back to Newsroom