-

Cerebras Systems Introduces Software Development Kit to Extend Breadth of Wafer-Scale Applications

SDK Unleashes Power of 850,000 Cores and Domain-Specific Programming Language for New Uses Cases

SUNNYVALE, Calif.--(BUSINESS WIRE)--Cerebras Systems Inc. (“Cerebras”), the pioneer in accelerating artificial intelligence (AI) compute, today announced at SC21 a new Cerebras Software Development Kit (SDK) that will enable the developer community to create new wafer-scale applications across computational fluid dynamics, molecular dynamics, signal processing, and more. The SDK is in Beta now and will be available year end.

The SDK will allow a wide range of AI and HPC developers to invent and test new ideas on the CS-2 system, and its industry leading Wafer Scale Engine (WSE-2), at a more flexible, lower level. The WSE-2 is the industry’s fastest and most powerful processor. With 850,000 cores, 40 Gigabytes of on-chip memory, 20 Petabytes/s of memory bandwidth, and 220 Petabits/s of fabric bandwidth, the WSE-2 has 123 times more cores, 1,000 times more on chip memory, 12,733 times more memory bandwidth, and 45,833 times more fabric bandwidth than the leading graphics processing unit. With the SDK, and a new domain-specific programming language based on C and familiar parallel programming concepts, the vast computational resources of the WSE-2 are put directly in the hands of the developer, limited only by their imaginations.

“We are excited to engage with the developer community, especially here at SC21, to help accelerate their work at the leading edge of HPC and AI,” said Andy Hock, Vice President of Product, Cerebras. "Powered by our 850,000 core, second generation wafer-scale engine with high bandwidth on-chip memory and interconnect, our CS-2 system is uniquely-suited to help users scale existing workloads and unlock new ones. We can't wait to see what we can build together."

Primitives will include a library of neural network kernels and a library of BLAS kernels. Using this, developers can program standalone applications, or they can combine new pieces with existing kernels in a modular fashion, using the same tools that engineers at Cerebras use for internal low-level development. The goal of the SDK is enable users to create things which have never been done before, to explore the boundaries of our industry, and to quickly and easily test these ideas on the WSE-2.

Also at SC21, Cerebras customer Argonne National Labs announced a new ALCF AI Testbed featuring the Cerebras CS-2 to support pioneering research at the intersection of AI and high performance compute (HPC). The National Energy Technology Laboratory (NETL) also shared details on work with Cerebras to accelerate both simulations for improved cybersecurity and computational fluid dynamics to model weather and earth systems dynamics, astrophysical simulations and electrodynamics.

Cerebras has expanded its footprint beyond the U.S. with new offices in Tokyo, Japan and Toronto, Canada. Its roster of customers has grown significantly to include Argonne National Laboratory, Lawrence Livermore National Laboratory, Pittsburgh Supercomputing Center (PSC) for its groundbreaking Neocortex AI supercomputer, EPCC, the supercomputing center at the University of Edinburgh, Tokyo Electron Device, GlaxoSmithKline, and AstraZeneca.

About Cerebras Systems
Cerebras Systems is a team of pioneering computer architects, computer scientists, deep learning researchers, and subject matter experts of all types. We have come together to build a new class of computer system, designed for the singular purpose of accelerating AI and changing the future of AI work forever. Our Cerebras CS-2 system, powered by the world’s largest processor – the WSE-2, enables customers to accelerate their deep learning by orders of magnitude over general purpose compute.

Contacts

Press contact (for media only)
Kim Ziesemer
Email: pr@zmcommunications.com

Cerebras Systems Inc.


Release Summary
Cerebras Systems Introduces Software Development Kit to Extend Breadth of Wafer-Scale Applications
Release Versions

Contacts

Press contact (for media only)
Kim Ziesemer
Email: pr@zmcommunications.com

More News From Cerebras Systems Inc.

 Cerebras Announces Six New AI Datacenters Across North America and Europe to Deliver Industry’s Largest Dedicated AI Inference Cloud

SUNNYVALE, Calif.--(BUSINESS WIRE)--Cerebras Systems, the pioneer in accelerating generative AI, today announced the launch of six new AI inference datacenters powered by Cerebras Wafer-Scale Engines. These state-of-the-art facilities, equipped with thousands of Cerebras CS-3 systems, are expected to serve over 40 million Llama 70B tokens per second, making Cerebras the world’s #1 provider of high-speed inference and the largest domestic high speed inference cloud. These new datacenters mark a...

Hugging Face Partners with Cerebras to Give Developers Access to Industry’s Fastest AI Inference for Open-Source Models

SUNNYVALE, Calif.--(BUSINESS WIRE)--Cerebras and Hugging Face today announced a new partnership to bring Cerebras Inference to the Hugging Face platform. HuggingFace has integrated Cerebras into HuggingFace Hub, bringing the world’s fastest inference to over five million developers on HuggingFace. Cerebras Inference runs the industry’s most popular models at more than 2,000 tokens/s – 70x faster than leading GPU solutions. Cerebras Inference models including Llama 3.3 70B, will be available to...

Cerebras Bolsters Leadership Team with Appointment of New CISO, EVP of Worldwide Sales, and SVP of AI Cloud and Inference

SUNNYVALE, Calif.--(BUSINESS WIRE)--Cerebras Systems, the pioneer in accelerating generative AI, today announced the appointment of Naor Penso as Chief Information Security Officer (CISO), Alex Varel as Executive Vice President of Worldwide Sales, and Hagay Lupesko as Senior Vice President of AI Cloud and Inference. These strategic hires further strengthen the company's leadership team as it continues to innovate and accelerate the adoption of its groundbreaking AI hardware solutions. "We are t...
Back to Newsroom