Saturday, November 23, 2024
HomeComputer HardwareCerebras Slays GPUs, Breaks File for Largest AI Fashions Educated on a...

Cerebras Slays GPUs, Breaks File for Largest AI Fashions Educated on a Single Machine


Cerebras, the corporate behind the world’s largest accelerator chip in existence, the CS-2 Wafer Scale Engine, has simply introduced a milestone: the coaching of the world’s largest NLP (Pure Language Processing) AI mannequin in a single machine. Whereas that in itself may imply many issues (it would not be a lot of a file to interrupt if the earlier largest mannequin was educated in a smartwatch, as an illustration), the AI mannequin educated by Cerebras ascended in the direction of a staggering – and unprecedented – 20 billion parameters. All with out the workload having to be scaled throughout a number of accelerators. That is sufficient to suit the web’s newest sensation, the image-from-text-generator, OpenAI’s 12-billion parameter DALL-E (opens in new tab).

An important bit in Cerebras’ achievement is the discount in infrastructure and software program complexity necessities. Granted, a single CS-2 system is akin to a supercomputer all by itself. The Wafer Scale Engine-2 – which, just like the identify implies, is etched in a single, 7 nm wafer, often sufficient for tons of of mainstream chips – includes a staggering 2.6 trillion 7 nm transistors, 850,000 cores, and 40 GB of built-in cache in a package deal consuming round 15kW.

Cerebras wafer scale engine

Cerebras’ Wafer Scale Engine-2 in all its wafer-sized glory. (Picture credit score: Cerebras)

Preserving as much as 20 billion-parameter NLP fashions in a single chip considerably reduces the overhead in coaching prices throughout 1000’s of GPUs (and related {hardware} and scaling necessities) whereas taking out the technical difficulties of partitioning fashions throughout them. Cerebras says that is “one of the painful points of NLP workloads,” typically “taking months to finish.”

It is a bespoke downside that is distinctive not solely to every neural community being processed, the specs of every GPU, and the community that ties all of it collectively – components that should be labored out upfront earlier than the primary coaching is ever began. And it could’t be ported throughout programs.

Cerebras CS-2

Cerebras’ CS-2 is a self-contained supercomputing cluster that features not solely the Wafer Scale Engine-2, but additionally all related energy, reminiscence and storage subsystems. (Picture credit score: Cerebras)

Pure numbers could make Cerebras’ achievement look underwhelming – OpenAI’s GPT-3, an NLP mannequin that may write complete articles that could typically idiot human readers, includes a staggering 175 billion parameters. DeepMind’s Gopher, launched late final 12 months, raises that quantity to 280 billion. The brains at Google Mind have even introduced the coaching of a trillion-parameter-plus mannequin, the Swap Transformer.

RELATED ARTICLES

LEAVE A REPLY

Please enter your comment!
Please enter your name here

- Advertisment -
Google search engine

Most Popular

Recent Comments