Paul van Gerven
20 August

Throwing all conventional wisdom out of the window, Silicon Valley start-up Cerebras has developed an AI processor measuring a whopping 21.5 by 21.5 cm. Packing 1.2 trillion transistors and 400,000 cores manufactured in a 16nm process, the company claims its Wafer-Scale Engine (WSE) can train neural networks up to a thousand times faster than equivalent hardware.

There are good reasons why most chips fit in the palm of your hand. Defectivity is an important one. Chip-ruining defects will typically scatter across the wafer during manufacturing, which means the smaller the chips, the more working chips will be sawed from the wafer. Another reason is cooling: the bigger the chip, the harder it is to cool properly.

The WSE is about 56 times the size of the largest GPU in the market, Cerebras claims. Credit: Cerebras

The size of the chips obviously curbs their processing power. Supercomputers and AI systems, therefore, have many chips working together. However, shuttling data between chips is much slower than within a single chip, ultimately presenting a bottleneck in performance.

This is why Cerebras went the wafer-sized chip route after all. The company worked closely with TSMC to manufacture and package the WSE and separately developed elaborate hardware to cool the silicon slab. A couple of systems will soon be installed at a select number of customer sites for testing and evaluation.