Categories
Podcast Season 3

The Trillion-Parameter ML Model with Cerebras Systems

Demand for AI compute is growing faster than conventional systems architecture can match, so companies like Cerebras Systems are building massive special-purpose processing units.

Demand for AI compute is growing faster than conventional systems architecture can match, so companies like Cerebras Systems are building massive special-purpose processing units. In this episode, Andy Hock, VP of Product for Cerebras Systems, joins Frederic Van Haren and Stephen Foskett to discuss this new class of hardware. The Cerebras Wafer-Scale Engine (WSE-2) has 850,000 processors on a single chip the size of a dinner plate, along with 40 GB of SRAM and supporting interconnects. But Cerebras also has a software stack that integrates with standard ML frameworks like PyTorch and TensorFlow. Although the trillion-parameter model is a real need for certain applications, platforms need to be flexible to support both massive-scale and more mainstream workloads, and this is a focus for Cerebras as well.

Three Questions

  • Frederic’s Question: How small can ML get? Will we have ML-powered household appliances? Toys? Disposable devices?
  • Stephen’s Question: Will we ever see a Hollywood-style “artificial mind” like Mr. Data or other characters?
  • Leon Adato, host of the Technically Religious Podcast: I’m curious, what responsibility do you think IT folks have to insure the things that we build are ethical?

Guests and Hosts

Andy Hock, VP of Product at Cerebras Systems . Connect with Andy on LinkedIn. Follow Cerebras Systems on Twitter at @CerebrasSystems .

Frederic Van Haren, Founder at HighFens Inc., Consultancy & Services. Connect with Frederic at Highfens.com or on Twitter at @FredericVHaren.

Stephen Foskett, Publisher of Gestalt IT and Organizer of Tech Field Day. Find Stephen’s writing at GestaltIT.com and on Twitter at @SFoskett.

One reply on “The Trillion-Parameter ML Model with Cerebras Systems”