The five technical challenges Cerebras overcame in building the first trillion transistor chip
Superlatives abound at Cerebras, the until-today stealthy next-generation silicon chip company looking to make training a deep learning model as quick as buying toothpaste from Amazon. Launching after almost three years of quiet development, Cerebras introduced its new chip today — and it is a doozy. The "Wafer Scale Engine" is 1.2 trillion transistors (the most ever), 46,225 square millimeters (the largest ever), and includes 18 gigabytes of on-chip memory (the most of any chip on the market today) and 400,000 processing cores (guess the superlative).
It's made a big splash here at Stanford University at the Hot Chips conference, one of the silicon industry's big confabs for product introductions and roadmaps, with various levels of oohs and aahs among attendees. You can read more about the chip from Tiernan Ray at Fortune and read the white paper from Cerebras itself.
Also at BBC, VentureBeat, and PCWorld.
(Score: 2) by driverless on Wednesday August 21 2019, @08:29AM (1 child)
Ah, they've finally gone public so it's OK to talk about it... yeah, it's a crazy device, even when they presented it the guy started with "every other wafer-scale project has failed", followed by endless questions about why theirs would be any different, and no clear answers. It's pretty outrageous, a single 10kW device with very special-case functionality that requires something the size of a small server rack to run, why would anyone buy this when you can use the space and power for a more conventional, and far more flexible, solution? I mean, from a geeky research-project basis it's pretty cool, but why? Their talk was mostly interruptions for questions about how this thing could be even remotely practical.
(Score: 2) by takyon on Wednesday August 21 2019, @10:16AM
What, were you at Hot Chips? Or...
Anyway, I don't think it's so crazy. Obviously this is a niche product, but it could offer great performance/$ for big companies that need it.
This thing exists because Moore Slaw Dead and there is a lot of hype money in AI/machine learning, for now and perhaps many years to come.
It's possible that some of the IP here will make its way into other products. But you could also just use lots of chiplets, stacked memory, etc. on a big ass-interposer. And 3DSoC is going to revolutionize the industry by putting logic and memory as close as possible, and it will probably stay that way.
[SIG] 10/28/2017: Soylent Upgrade v14 [soylentnews.org]