Cerebras Unveils First Installation of Its AI Supercomputer at Argonne National Labs
At Supercomputing 2019 in Denver, Colo., Cerebras Systems unveiled the computer powered by the world's biggest chip. Cerebras says the computer, the CS-1, has the equivalent machine learning capabilities of hundreds of racks worth of GPU-based computers consuming hundreds of kilowatts, but it takes up only one-third of a standard rack and consumes about 17 kW. Argonne National Labs, future home of what's expected to be the United States' first exascale supercomputer, says it has already deployed a CS-1. Argonne is one of two announced U.S. National Laboratories customers for Cerebras, the other being Lawrence Livermore National Laboratory.
The system "is the fastest AI computer," says CEO and cofounder Andrew Feldman. He compared it with Google's TPU clusters (the 2nd of three generations of that company's AI computers), noting that one of those "takes 10 racks and over 100 kilowatts to deliver a third of the performance of a single [CS-1] box."
The CS-1 is designed to speed the training of novel and large neural networks, a process that can take weeks or longer. Powered by a 400,000-core, 1-trillion-transistor wafer-scale processor chip, the CS-1 should collapse that task to minutes or even seconds. However, Cerebras did not provide data showing this performance in terms of standard AI benchmarks such as the new MLPerf standards. Instead it has been wooing potential customers by having them train their own neural network models on machines at Cerebras.
[...] The CS-1's first application is in predicting cancer drug response as part of a U.S. Department of Energy and National Cancer Institute collaboration. It is also being used to help understand the behavior of colliding black holes and the gravitational waves they produce. A previous instance of that problem required 1024 out of 4392 nodes of the Theta supercomputer.
Also at TechCrunch, VentureBeat, and Wccftech.
Previously: Cerebras "Wafer Scale Engine" Has 1.2 Trillion Transistors, 400,000 Cores
(Score: 2) by c0lo on Monday November 25 2019, @06:15AM (4 children)
1. ... but does it run Linux?
2. Imagine a Beowulf cluster of these
https://www.youtube.com/watch?v=aoFiw2jMy-0
(Score: 0) by Anonymous Coward on Monday November 25 2019, @06:19AM
It found Linux too difficult to operate.
(Score: 3, Interesting) by takyon on Monday November 25 2019, @06:20AM (2 children)
[SIG] 10/28/2017: Soylent Upgrade v14 [soylentnews.org]
(Score: 0) by Anonymous Coward on Monday November 25 2019, @07:01AM (1 child)
Imagination fail.
(Score: 2) by takyon on Monday November 25 2019, @07:06AM
There's a cluster for that.
[SIG] 10/28/2017: Soylent Upgrade v14 [soylentnews.org]
(Score: 3, Interesting) by KritonK on Monday November 25 2019, @07:51AM
I wonder if the chip's first application is going to be to mine enough bitcoins to pay the cost of its development.
(Score: 2) by crafoo on Monday November 25 2019, @08:13PM (1 child)
I don't have much to add besides this is super cool. I wonder what the defect rate is for a wafer-sized CPU. Also, what sized wafer? 4", 6", 8", 12", ...? Back when I was in the semiconductor industry, we used 4", 6,", and 12" depending on the wafer material and doping. I wonder if they have some new clean room tech too, or just accept a high defect rate.
(Score: 2) by takyon on Monday November 25 2019, @09:10PM
If you look at the previous reporting on this, the Wafer Scale Engine is defect tolerant.
It is using a 300mm wafer and TSMC "16nm" node. Defect rate should be pretty low? Since that's a mature node at this point.
[SIG] 10/28/2017: Soylent Upgrade v14 [soylentnews.org]