Stories
Slash Boxes
Comments

SoylentNews is people

posted by martyb on Tuesday August 20 2019, @09:11PM   Printer-friendly
from the not-going-to-fit-in-a-cell-phone dept.

The five technical challenges Cerebras overcame in building the first trillion transistor chip

Superlatives abound at Cerebras, the until-today stealthy next-generation silicon chip company looking to make training a deep learning model as quick as buying toothpaste from Amazon. Launching after almost three years of quiet development, Cerebras introduced its new chip today — and it is a doozy. The "Wafer Scale Engine" is 1.2 trillion transistors (the most ever), 46,225 square millimeters (the largest ever), and includes 18 gigabytes of on-chip memory (the most of any chip on the market today) and 400,000 processing cores (guess the superlative).

It's made a big splash here at Stanford University at the Hot Chips conference, one of the silicon industry's big confabs for product introductions and roadmaps, with various levels of oohs and aahs among attendees. You can read more about the chip from Tiernan Ray at Fortune and read the white paper from Cerebras itself.

Also at BBC, VentureBeat, and PCWorld.


Original Submission

 
This discussion has been archived. No new comments can be posted.
Display Options Threshold/Breakthrough Mark All as Read Mark All as Unread
The Fine Print: The following comments are owned by whoever posted them. We are not responsible for them in any way.
  • (Score: 3, Interesting) by takyon on Tuesday August 20 2019, @09:18PM (7 children)

    by takyon (881) <takyonNO@SPAMsoylentnews.org> on Tuesday August 20 2019, @09:18PM (#882787) Journal

    Why is this a good idea? Keeping everything on the same chip and on-chip memory reduces latency and increases energy efficiency.

    Power consumption will be in the ballpark of 10kW to 15 kW, not counting the cooling.

    Obviously, the wafer will have defects, but it can tolerate them and has redundant cores (ya think?).

    This is the opposite of the chiplet approach, where you want small pieces of silicon to boost yields and create a wide variety of designs (from mobile to server) with the same chiplets. In this approach, you want as much performance as possible, and using the whole wafer gets you better performance than splitting up the wafer. At least, that's the story, we don't know until someone actually benchmarks something on it.

    Apparently, it can run regular code, not just machine learning stuff. Keeping in mind that these are "small cores".

    --
    [SIG] 10/28/2017: Soylent Upgrade v14 [soylentnews.org]
    Starting Score:    1  point
    Moderation   +1  
       Interesting=1, Total=1
    Extra 'Interesting' Modifier   0  
    Karma-Bonus Modifier   +1  

    Total Score:   3  
  • (Score: 4, Insightful) by JoeMerchant on Tuesday August 20 2019, @09:28PM (1 child)

    by JoeMerchant (3937) on Tuesday August 20 2019, @09:28PM (#882792)

    It's hardly "local" - the wafer area is like a square 8.5" on a side.

    It probably needs to be maintained in a clean room its whole working life...

    --
    Україна досі не є частиною Росії Слава Україні🌻 https://news.stanford.edu/2023/02/17/will-russia-ukraine-war-end
  • (Score: 2) by takyon on Tuesday August 20 2019, @09:38PM

    by takyon (881) <takyonNO@SPAMsoylentnews.org> on Tuesday August 20 2019, @09:38PM (#882802) Journal

    I'll revise that to 15 kW, not sure about the cooling, since it seems the company is providing the cooling solution.

    --
    [SIG] 10/28/2017: Soylent Upgrade v14 [soylentnews.org]
  • (Score: 1, Funny) by Anonymous Coward on Tuesday August 20 2019, @10:47PM (1 child)

    by Anonymous Coward on Tuesday August 20 2019, @10:47PM (#882834)

    This is the opposite of the chiplet approach,

    They can put several of them side to side and call it chiplets. Imagine a beowulf cluster of... nah forget it.

  • (Score: 2) by driverless on Wednesday August 21 2019, @08:29AM (1 child)

    by driverless (4770) on Wednesday August 21 2019, @08:29AM (#883022)

    Ah, they've finally gone public so it's OK to talk about it... yeah, it's a crazy device, even when they presented it the guy started with "every other wafer-scale project has failed", followed by endless questions about why theirs would be any different, and no clear answers. It's pretty outrageous, a single 10kW device with very special-case functionality that requires something the size of a small server rack to run, why would anyone buy this when you can use the space and power for a more conventional, and far more flexible, solution? I mean, from a geeky research-project basis it's pretty cool, but why? Their talk was mostly interruptions for questions about how this thing could be even remotely practical.

    • (Score: 2) by takyon on Wednesday August 21 2019, @10:16AM

      by takyon (881) <takyonNO@SPAMsoylentnews.org> on Wednesday August 21 2019, @10:16AM (#883044) Journal

      Ah, they've finally gone public so it's OK to talk about it...

      What, were you at Hot Chips? Or...

      Anyway, I don't think it's so crazy. Obviously this is a niche product, but it could offer great performance/$ for big companies that need it.

      This thing exists because Moore Slaw Dead and there is a lot of hype money in AI/machine learning, for now and perhaps many years to come.

      It's possible that some of the IP here will make its way into other products. But you could also just use lots of chiplets, stacked memory, etc. on a big ass-interposer. And 3DSoC is going to revolutionize the industry by putting logic and memory as close as possible, and it will probably stay that way.

      --
      [SIG] 10/28/2017: Soylent Upgrade v14 [soylentnews.org]