pbnjoe writes:
Ars technica is reporting on new devlopments in GPU interconnect tech.
From the article:
Nvidia and IBM have developed an interconnect that will be integrated into future graphics processing units, letting GPUs and CPUs share data five times faster than they can now, Nvidia announced today. The fatter pipe will let data flow between the CPU and GPU at rates higher than 80GB per second, compared to 16GB per second today. NVLink, the interconnect, will be part of the newly announced Pascal GPU architecture on track for release in 2016.
GPUs have become increasingly common in supercomputing, serving as accelerators or "co-processors" to help CPUs get work done faster. In the most recent list of the world's fastest 500 supercomputers, 53 systems used co-processors and 38 of these used Nvidia chips. The second and sixth most powerful supercomputers used Nvidia chips alongside CPUs. Intel still dominates, providing processors for 82.4 percent of Top 500 systems.
"Today's GPUs are connected to x86-based CPUs through the PCI Express (PCIe) interface, which limits the GPU's ability to access the CPU memory system and is four- to five-times slower than typical CPU memory systems," Nvidia said. "PCIe is an even greater bottleneck between the GPU and IBM Power CPUs, which have more bandwidth than x86 CPUs. As the NVLink interface will match the bandwidth of typical CPU memory systems, it will enable GPUs to access CPU memory at its full bandwidth... Although future Nvidia GPUs will continue to support PCIe, NVLink technology will be used for connecting GPUs to NVLink-enabled CPUs as well as providing high-bandwidth connections directly between multiple GPUs.
(Score: 2, Interesting) by Sebastopol on Thursday March 27 2014, @03:20AM
That seems awfully slow, Intel's Xeon Phi's next installment is already claiming 500GB/s on die (to EDRAM) and 384GB/s to DDR4.
http://www.extremetech.com/extreme/171678-intel-un veils-72-core-x86-knights-landing-cpu-for-exascale -supercomputing [extremetech.com]
(Score: 3, Informative) by visaris on Thursday March 27 2014, @12:46PM
From the summary: "between the CPU and GPU".
The connection between the EDRAM and the cores on the Xeon Phi is not CPU to GPU. The memory bandwidth available to a GPU core to/from GPU memory is much higher than 80GB/s. A better comparison would be to compare the PCIe3x16 bandwidth to nVidia's stated 80GB/s.
(Score: 1) by green is the enemy on Thursday March 27 2014, @05:00PM
From the linked article:
This is actually exciting... Not necessarily that Intel is making a beast of a chip for supercomputers. Rather, if NVIDIA intends to compete with Intel, they will have to implement similar technology into their GPUs. This technology will then trickle down to the mass market of gaming GPUs. Massive computing power for all :)
(Score: 0) by Anonymous Coward on Thursday March 27 2014, @07:44PM
I've used both the existing tesla stuff and the intel phi stuff. The tesla stuff blows MIC out of the water, and the software support from nvidia is lightyears ahead of what intel currently supports. It's not even a contest.
(Score: 1) by Sebastopol on Friday March 28 2014, @02:47AM
Can you link to some data that backs this up? Not that I'm a fanboi of either one, but what I've found online Tesla has an edge or is on par, but nowhere near "water blowing". Granted, nVidia has been at this a lot longer than Intel. Unless it's just another marketing war...
http://goparallel.sourceforge.net/independent-test -xeon-phi-shocks-tesla-gpu/ [sourceforge.net]
http://blog.xcelerit.com/intel-xeon-phi-vs-nvidia- tesla-gpu/ [xcelerit.com]
https://www.google.com/url?sa=t&rct=j&q=&esrc=s&so urce=web&cd=3&cad=rja&uact=8&ved=0CD4QFjAC&url=htt p%3A%2F%2Fwww.ocw.titech.ac.jp%2Findex.php%3Fmodul e%3DGeneral%26action%3DDownLoad%26file%3D201312267 17065-477-1-45.pdf%26type%3Dcal%26JWC%3D2013122671 7065&ei=KeA0U__rH6OWyAG_p4DgDA&usg=AFQjCNHCgLfZls7 xxSFTChPU4nWH4ci1GQ&sig2=JueGHWHxAvYWcSlu5QUgSQ [google.com]
(Score: 1) by Sebastopol on Friday March 28 2014, @02:50AM
Ah. Thanks for clarifying. I didn't think there was a bottleneck from CPU to GPU, but apparently there is if they are investing in higher bandwidth! (I assumed a memory controller filled local DRAM and the GPU just pointed to it, perhaps directed by instructions from the CPU, but nothing requiring that match b/w between the two.)
(Score: 0) by Anonymous Coward on Friday March 28 2014, @04:07AM
Latency is important too. The new interconnect will be much lower latency than PCIe and it is important as the GPU can seamlessly access the CPU's RAM (or another GPU's RAM) just like CPUs in a multiple socket system.
The new Xeon Phi will have a similar interconnect. It will simply plug into a CPU socket (with Intel's QPI) rather than a PCIe slot, sharing the motherboard with Xeon EX CPUs or other Xeon Phi.
(Score: 1) by maddening milkshakes on Thursday March 27 2014, @04:00AM
I can't wait to pop one of those into my new mac pro and go to town!
(Score: 2, Interesting) by Anonymous Coward on Thursday March 27 2014, @10:24AM
Something about this smells quite similar to the VESA local bus from yesteryear
http://en.wikipedia.org/wiki/VESA_Local_Bus [wikipedia.org]