Stories
Slash Boxes
Comments

SoylentNews is people

posted by on Monday May 01 2017, @12:01PM   Printer-friendly
from the faster-is-better dept.

SK Hynix is almost ready to produce GDDR6 memory with higher than expected per-pin bandwidth:

In a surprising move, SK Hynix has announced its first memory chips based on the yet-unpublished GDDR6 standard. The new DRAM devices for video cards have capacity of 8 Gb and run at 16 Gbps per pin data rate, which is significantly higher than both standard GDDR5 and Micron's unique GDDR5X format. SK Hynix plans to produce its GDDR6 ICs in volume by early 2018.

GDDR5 memory has been used for top-of-the-range video cards for over seven years, since summer 2008 to present. Throughout its active lifespan, GDDR5 increased its data rate by over two times, from 3.6 Gbps to 9 Gbps, whereas its per chip capacities increased by 16 times from 512 Mb to 8 Gb. In fact, numerous high-end graphics cards, such as NVIDIA's GeForce GTX 1060 and 1070, still rely on the GDDR5 technology, which is not going anywhere even after the launch of Micron's GDDR5X with up to 12 Gbps data rate per pin in 2016. As it appears, GDDR6 will be used for high-end graphics cards starting in 2018, just two years after the introduction of GDDR5X.

Previously: Samsung Announces Mass Production of HBM2 DRAM
DDR5 Standard to be Finalized by JEDEC in 2018


Original Submission

 
This discussion has been archived. No new comments can be posted.
Display Options Threshold/Breakthrough Mark All as Read Mark All as Unread
The Fine Print: The following comments are owned by whoever posted them. We are not responsible for them in any way.
(1)
  • (Score: 3, Interesting) by Scrutinizer on Monday May 01 2017, @12:58PM

    by Scrutinizer (6534) on Monday May 01 2017, @12:58PM (#502245)

    I recall that SK Hynix was one of the manufacturers whose DDR4 modules were much more resistant to Rowhammer attacks [soylentnews.org] than those from Crucial/Micron, based upon test results from thirdio.com [thirdio.com]. (G.Skill was the other manufacturer with highly resistant DDR4 modules.)

    Some of their fancy DDR6 memory might go well with the card needed to drive my future Facebook-branded face mask [oculus.com].

  • (Score: 2) by fyngyrz on Monday May 01 2017, @01:07PM (14 children)

    by fyngyrz (6567) on Monday May 01 2017, @01:07PM (#502246) Journal

    My first reaction was "graphics cards? Why can't I have this stuff on the CPU?"

    CPUs have been faster than main memory for years. Without main memory that's as fast as the instruction / data cycle time, things are slower than they otherwise could be for that CPU. The problems, I suppose, would be related to getting the info in and out of the CPU at a speed to match the new RAM, and a bus to carry it... but this would certainly be useful if it could be done and commercialized affordably.

    • (Score: 2) by Runaway1956 on Monday May 01 2017, @02:20PM (11 children)

      by Runaway1956 (2926) Subscriber Badge on Monday May 01 2017, @02:20PM (#502269) Journal

      In my experience, memory has been "fast enough" for a long time now. Given sufficient memory, an ancient AMD K6 450 mhz CPU can actually run Windows XP fairly well. That is one gig of PC-100, in this case. With limited memory, a 1 Ghz Athlon struggled to run basically the same installation.

      My computer at home has 24 gig of memory, which is basically obsolete. The computer at work is quite modern and up to date, but it drags ass all day, every day. It only has 1 gig of memory, and runs Windows 7, with a buttload of background tasks. At home, I never wait for anything, except internet. At work, waiting for anything to load is painful. Microsoft needs to shitcan the paging system, and make it clear that computers require 8 gig of memory to run today. In the not-distant future, they're going to need 16 gig minimum, because everything just grows larger and more bloated all the time. Within a decade, we may need to raise the minimum to 64 gig if you want to run a Microsoft OS. (I expect Linux to stay reasonably responsive with far less memory, just because Linux still runs well on old hardware right now.)

      Given the opportunity, yes, I'll take faster memory, faster CPU, faster everything. But, more than anything, I want a machine with ADEQUATE MEMORY! Ideally, your entire operating system, as well as most of your apps can reside in memory, and never make use of virtual memory. The slowest memory on the market today is more than fast enough to make almost any computer sing, if only enough memory is installed.

      --
      “I have become friends with many school shooters” - Tampon Tim Walz
      • (Score: 3, Interesting) by takyon on Monday May 01 2017, @02:57PM (3 children)

        by takyon (881) <{takyon} {at} {soylentnews.org}> on Monday May 01 2017, @02:57PM (#502283) Journal

        The computer at work is quite modern and up to date, but it drags ass all day, every day. It only has 1 gig of memory, and runs Windows 7, with a buttload of background tasks.

        How is a machine with only 1 GB of RAM considered "quite modern"? I would struggle to find anything under 4 GB on the market today, and they would be landfill laptops and Chromebooks with an absolute minimum of 2 GB of RAM.

        You can probably find someone with a free 4 GB DDR3 module. If not, buy it yourself for $20, put it in the work computer, and take it away when you leave. Nobody will notice and you can put it in your museum later.

        --
        [SIG] 10/28/2017: Soylent Upgrade v14 [soylentnews.org]
        • (Score: 2) by VLM on Monday May 01 2017, @03:09PM

          by VLM (445) on Monday May 01 2017, @03:09PM (#502288)

          Aside from all this desktop stuff, in the server and virtualization host market, nobody ever said their memory bus was too fast.

          From an engineering standpoint it should be possible to make optimizations such that sequential reading is kinda the default and faster at the expense of totally random access. Remember "Row and column" strobes for dram addressing in the 80s or so, you could extend that concept way beyond 2 dimensions (not physically of course?) such that sequential access would require usually 1 or sometimes 2 address segment writes per cycle for a graphics display but totally randomly smacking some random address would take like 8 address segment loads. Like imagine 128 parallel data lines and 8 bits of address and a whole bunch of segment strobes to load up the address 8 bits at a time. I wonder if there's also some weird dual porting stuff going on such that it wouldn't really be your first choice for a CPU.

        • (Score: 2) by Runaway1956 on Monday May 01 2017, @04:47PM

          by Runaway1956 (2926) Subscriber Badge on Monday May 01 2017, @04:47PM (#502347) Journal

          It belongs to contractors, who have spec'd the machines to their own needs/wants. Seriously, machines that are - ohhhh - I guess they are three years old now - with decent CPU's, smallish hard drives, no optical drives, and - only ONE gig of memory. Don't ask me why, or how. They were built cheap, and that's how they run. Our own IT people weren't smart enough to realize they were getting ripped off. The machines in the offices aren't so bad as our machines in the work spaces, but they are still pretty crappy.

          --
          “I have become friends with many school shooters” - Tampon Tim Walz
        • (Score: 2) by JoeMerchant on Monday May 01 2017, @10:15PM

          by JoeMerchant (3937) on Monday May 01 2017, @10:15PM (#502537)

          This ^^^ - my father bought me a cast-off iMac from his University for $50, the only thing really "wrong" with it was that it only had 1GB of RAM - and it was upgrade capped to 2GB by Apple... We spent another (years ago) $40 to get 2GB of RAM for it and it became downright usable - for single users doing single tasks.

          --
          🌻🌻 [google.com]
      • (Score: 2) by LoRdTAW on Monday May 01 2017, @03:15PM

        by LoRdTAW (3755) on Monday May 01 2017, @03:15PM (#502292) Journal

        In my experience, memory has been "fast enough" for a long time now.

        For the most part, it certainly is. Basic desktop useage really doesn't push memory bandwidth so much as memory usage (code bloat). We also have plenty of CPU. Even a modest dual core i3 can handle most desktop use including low end gaming.

        Microsoft needs to shitcan the paging system, and make it clear that computers require 8 gig of memory to run today.

        It's not Microsoft's fault. They have a minimum as well as an optimal configuration spec. But the problem lies in the OEM's who make tons of money upselling you on a few GB of RAM.

        Dell loves pulling this shit by offering dozens of different models which are nothing more than the same PC with multiple fixed configurations. "Oh, you want more than 8GB RAM *AND* a core i3 in your optiplex? Too bad. Buy the i5 model for another $150 AND pay another $100 for the 8GB model. Or buy that nearly $900 i3 that's about $400 over the base i3. What I wind up doing is buying the base i3 model with 4GB, look up the part number of the OEM RAM module, and buy a matching 4GB module for $30-50. Or buy a compatible 8GB kit from crucial and use the extra 4GB OEM stick in another desktop to double that one to 8GB as well. You can save around $200+ per workstation doing this. Of course that's fine for small shops like mine with less than 20 desktops. If you order by the tens or hundreds, it's less practical and you wind up going with the upsell to eliminate the labour.

        Dell (along with many others) were notorious for selling low end home PC's with barely enough RAM to boot the damn OS as was the case with my friends P4 Dell in the mid 00's with 256MB RAM and XP home. Ran like shit. Opening a web browser was an 30+ second ordeal as the disk chruned and burned shuffling stuff out of RAM into the page file. I had him buy a 1GB kit that was compatible and it was like a whole new computer. And 1GB should have been the XP minimum, not 256 like MS said.

      • (Score: 2) by kaszz on Monday May 01 2017, @05:01PM (4 children)

        by kaszz (4211) on Monday May 01 2017, @05:01PM (#502355) Journal

        The paging system is what makes the difference sometimes between can or can't use a program at all. If you install enough primary memory. This will not be a problem, so take away the options for other people.

        However if people did shitcan Microsoft software then a lot of memory problems would not occur. Besides all these requirements on gigs of primary memory is a indicator of really bad programming with some exceptions like for CAD etc.

        • (Score: 2) by Runaway1956 on Monday May 01 2017, @05:54PM (3 children)

          by Runaway1956 (2926) Subscriber Badge on Monday May 01 2017, @05:54PM (#502412) Journal

          It almost seems like an arms race. Memory gets cheaper, and more plentiful, but shoddy programmers seem to be determined to squander all of that memory.

          --
          “I have become friends with many school shooters” - Tampon Tim Walz
          • (Score: 2) by kaszz on Monday May 01 2017, @06:39PM (2 children)

            by kaszz (4211) on Monday May 01 2017, @06:39PM (#502443) Journal

            I have noticed the same trend too. But there are ways to avoid it. Open source OS, open source applications.
            And I suspect C++ and other languages has a lot to do with this. On top of people that have no business designing software being shepherded into "programming".

            The CPU trend will be interesting though.. they seem to not clock faster than circa 4.5 GHz. So programmers have to be smarter about that resource or see competition running them over.

            • (Score: 0) by Anonymous Coward on Tuesday May 02 2017, @02:36AM (1 child)

              by Anonymous Coward on Tuesday May 02 2017, @02:36AM (#502623)

              And I suspect C++ and other languages has a lot to do with [bloated programs]. On top of people that have no business designing software being shepherded into "programming".

              I'm not a programmer, so keeping that in mind: is there something wrong with C++ or the way it's being used? What are the better alternatives?

              • (Score: 2) by kaszz on Tuesday May 02 2017, @03:47AM

                by kaszz (4211) on Tuesday May 02 2017, @03:47AM (#502652) Journal

                When you write software in C++ (object oriented) it will too often implicitly suck in a lot of stuff, name space can get overloaded and some programmers like to allocate but free() is less popular.

                Depending on task, use C.

      • (Score: 2) by shortscreen on Monday May 01 2017, @08:22PM

        by shortscreen (2252) on Monday May 01 2017, @08:22PM (#502493) Journal

        Pentium 3s were terribly memory speed limited. In some cases there was hardly any difference between a P3 800MHz or 1.4GHz if they both had PC-133 SDRAM. This was when Intel started building their shitty graphics into the motherboard chipset, so refreshing the display would eat up precious memory bandwidth. Changing the screen mode to a lower resolution and color depth to reduce bus contention would speed things up measurably. Check out these benchmark results from super pi:

        Pentium 3 Coppermine 933MHz (discrete video card) - 2m17s
        Pentium 3 Tualitin 800MHz (i830 graphics) - 2m35s
        Pentium 3 Tualitin 1.33GHz (i830 graphics) - 2m20s
        1.33GHz with lower latency 3-2-2 RAM - 2m06s
        1.33GHz with screen mode set to 800x600 16-bit - 1m57s

        Athlon XPs were also somewhat limited. Although they used DDR, since speeds eventually got up to 2.3GHz or so, on a 200MHz bus the latency got to be even worse than Pentium 3s (but with 64-byte line size instead of 32)

        Athlon 64s and Pentium Ms greatly reduced this bottleneck by lowering latency and improving cache hit rates, respectively.

        The weird thing is that the MOVSD instruction has never been optimized enough for a simple block copy to acheive anything close to theoretical memory throughput. It's always limited by the CPU instead (and read-before-write memory access patterns). I guess new CPUs have a fancier way to do block copies although I have not tried it.

    • (Score: 3, Interesting) by sjames on Monday May 01 2017, @03:21PM

      by sjames (2882) on Monday May 01 2017, @03:21PM (#502296) Journal

      You can, just as soon as the lumbering behemoths get around to producing CPUs that support it. It makes sense for Hynix to target GPUs first as they're more likely to adopt it faster.

      Beyond that though, for a new memory tech, GPUs used for actual graphics tend to be far more tolerant to single bit errors than CPUs. This gives a little time to tune the manufacturing and get the kinks out.

    • (Score: 2) by kaszz on Monday May 01 2017, @04:55PM

      by kaszz (4211) on Monday May 01 2017, @04:55PM (#502352) Journal

      The problem may be related to access patterns. If the CPU pattern is random enough. Then a memory which can only deliver long sequential access will not improve anything.

  • (Score: 2) by boltronics on Monday May 01 2017, @01:15PM (1 child)

    by boltronics (580) on Monday May 01 2017, @01:15PM (#502250) Homepage Journal

    Will be interesting to compare with HBM2, or even HBM1 for that matter (as found on the Fury X). My basic understanding was that GDDR5X is slower than HBM1, but it was going to be too costly for Nvidia to re-architect for HBM1 so they went with GDDR5X anyway since it was similar to the older GDDR5?

    No surprise then that the RX400 and RX500 series only uses GDDR5 due to the mid-range high value aims, but perhaps the upcoming high-end Vega range will be using HBM2? We shouldn't have to wait long to find out.

    --
    It's GNU/Linux dammit!
    • (Score: 4, Interesting) by takyon on Monday May 01 2017, @02:21PM

      by takyon (881) <{takyon} {at} {soylentnews.org}> on Monday May 01 2017, @02:21PM (#502270) Journal

      NVIDIA did skip HBM1, which is not nearly as capable as HBM2. I hear that HBM2 (which should be HBM3 before long) is still more expensive than GDDR5X or GDDR6. GDDR6 enables rather high memory bandwidth, so the benefits of HBM aren't necessarily as apparent. HBM however does take up less die space and has lower power consumption.

      Vega will have HBM. [google.com]

      The rumor mill (just noticed) is explicitly saying that NVIDIA Volta will use GDDR6 [wccftech.com]. Compare to the AnandTech article in the summary which just says "SK Hynix is not disclosing the name of its partner among GPU developers".

      --
      [SIG] 10/28/2017: Soylent Upgrade v14 [soylentnews.org]
  • (Score: 2) by kaszz on Monday May 01 2017, @04:52PM (1 child)

    by kaszz (4211) on Monday May 01 2017, @04:52PM (#502350) Journal

    The GDDR5X [micron.com] interface seems to rely on a PLL clock, selectable data bus inversion and quad data rate interface for rates up 12 Gbit/s using single ended mode. So how does GDDR6 accomplish 16 Gbit/s ?
    (btw, good schematic on page 13)

    Some other observations is that the GDDR5X is designed to operate on 1.35 V. That is a voltage level with a very thin noise margin. And..

    Note: The operating range and AC timings of a faster speed bin are a superset of all slower speed bins. Therefore it is safe to use a faster bin device as a drop-in replacement of a slower bin device when operated within the frequency range of the slower bin device.

    (page 3)
    Already present DRAM may actually be faster than its marking tells you.

    The circuit board impedance matching, length matching, reflection avoidance, shielding etc.. pain in the %#¤.

    • (Score: 2) by takyon on Monday May 01 2017, @10:40PM

      by takyon (881) <{takyon} {at} {soylentnews.org}> on Monday May 01 2017, @10:40PM (#502547) Journal

      Some other observations is that the GDDR5X is designed to operate on 1.35 V. That is a voltage level with a very thin noise margin.

      FTA:

      The company is also stating that GDDR6 will have a 10% lower operating voltage than GDDR5, though they don't specify if that's relative to the low voltage (1.35 V), standard (1.5 V) or high frequency (1.55 V) version of GDDR5.

      Therefore voltage for GDDR6 will likely fall somewhere between 1.2 V and 1.4 V.

      --
      [SIG] 10/28/2017: Soylent Upgrade v14 [soylentnews.org]
(1)