Stories
Slash Boxes
Comments

SoylentNews is people

posted by janrinok on Thursday June 07 2018, @06:53PM   Printer-friendly
from the begun-the-core-wars-have dept.

AMD released Threadripper CPUs in 2017, built on the same 14nm Zen architecture as Ryzen, but with up to 16 cores and 32 threads. Threadripper was widely believed to have pushed Intel to respond with the release of enthusiast-class Skylake-X chips with up to 18 cores. AMD also released Epyc-branded server chips with up to 32 cores.

This week at Computex 2018, Intel showed off a 28-core CPU intended for enthusiasts and high end desktop users. While the part was overclocked to 5 GHz, it required a one-horsepower water chiller to do so. The demonstration seemed to be timed to steal the thunder from AMD's own news.

Now, AMD has announced two Threadripper 2 CPUs: one with 24 cores, and another with 32 cores. They use the "12nm LP" GlobalFoundries process instead of "14nm", which could improve performance, but are currently clocked lower than previous Threadripper parts. The TDP has been pushed up to 250 W from the 180 W TDP of Threadripper 1950X. Although these new chips match the core counts of top Epyc CPUs, there are some differences:

At the AMD press event at Computex, it was revealed that these new processors would have up to 32 cores in total, mirroring the 32-core versions of EPYC. On EPYC, those processors have four active dies, with eight active cores on each die (four for each CCX). On EPYC however, there are eight memory channels, and AMD's X399 platform only has support for four channels. For the first generation this meant that each of the two active die would have two memory channels attached – in the second generation Threadripper this is still the case: the two now 'active' parts of the chip do not have direct memory access.

This also means that the number of PCIe lanes remains at 64 for Threadripper 2, rather than the 128 of Epyc.

Threadripper 1 had a "game mode" that disabled one of the two active dies, so it will be interesting to see if users of the new chips will be forced to disable even more cores in some scenarios.


Original Submission

 
This discussion has been archived. No new comments can be posted.
Display Options Threshold/Breakthrough Mark All as Read Mark All as Unread
The Fine Print: The following comments are owned by whoever posted them. We are not responsible for them in any way.
  • (Score: 2) by VLM on Thursday June 07 2018, @07:08PM (22 children)

    by VLM (445) on Thursday June 07 2018, @07:08PM (#690023)

    Serious question, is this really useful for gamers today in 2018?

    I only use a couple cores for minecraft and my minecraft server even with extensive mods, so I don't know about twitchy FPS sequels maybe those use all the cores, I donno. 28 cores seems a little far fetched. I think my server has 4 allocated but its usually not using all 4.

    Ironically for a product marketed to gamers, I have a real world use for multicore in my vmware cluster in the basement. In terms of system design where only the weakest link matters, I would think my memory BW or SSD would saturate long before all 28 cores are in use, but who knows, maybe not.

    Starting Score:    1  point
    Karma-Bonus Modifier   +1  

    Total Score:   2  
  • (Score: 0) by Anonymous Coward on Thursday June 07 2018, @07:16PM (2 children)

    by Anonymous Coward on Thursday June 07 2018, @07:16PM (#690027)

    But why?

    • (Score: 2) by VLM on Friday June 08 2018, @03:24PM

      by VLM (445) on Friday June 08 2018, @03:24PM (#690352)

      You mistyped "why not?"

      What can I say, you get addicted to the virtualization lifestyle at work, next thing you know you got a cluster in the basement. All legal if for non-commercial use and the ESX-experience VMUG deal. I spend a lot more on electricity and hardware than I do on VMUG membership, thats for sure, LOL.

    • (Score: 0) by Anonymous Coward on Sunday June 10 2018, @04:59AM

      by Anonymous Coward on Sunday June 10 2018, @04:59AM (#691037)

      I have somewhat seriously discussed placing a server in the attic and running wiring out around the house to allow thin clients or even just monitor/keyboard/mouse stattions to but placed in any room so that all people in the house have access to their own computing environment from anywhere in the house. Seems like the perfect use for a personal vmware cluster. A basement is fine, too.

  • (Score: 2) by EvilSS on Thursday June 07 2018, @07:50PM

    by EvilSS (1456) Subscriber Badge on Thursday June 07 2018, @07:50PM (#690042)
    No. Maybe if you are also live-streaming at the same time but even then it's overkill. These are for workstation loads. Video editing, running a bunch of VMs, sciencey mumbo jumbo, crap like that. For games you'd be better off with fewer, faster cores.
  • (Score: 2) by JoeMerchant on Thursday June 07 2018, @07:59PM (6 children)

    by JoeMerchant (3937) on Thursday June 07 2018, @07:59PM (#690045)

    Game authors _could_ use massively multi-threaded processing for many things (think: processing NPC logic and decisions), but game authors are driven by the mass market: what does their buying audience have access to, so, no, I doubt many game authors are going to go massively multi-threaded anytime soon. Lots of "gaming systems" are still just dual core.

    --
    🌻🌻 [google.com]
    • (Score: 5, Insightful) by takyon on Thursday June 07 2018, @08:31PM (2 children)

      by takyon (881) <takyonNO@SPAMsoylentnews.org> on Thursday June 07 2018, @08:31PM (#690056) Journal

      If they develop their games on workstations, and test them on both the workstations and dual or quad-core PCs, they should be able to design engines/games that use even as many as 64 threads. Games like Skyrim ran pretty well on low-end systems as well as high-end systems. That's the way it should work: set your minimum requirements, as low as possible if you want to ensure people can at least run it, but scale to use everything available.

      While Steam stats show many 2-4 core users, we now have 6-8 core "mainstream" chips from both AMD and Intel. The latest consoles also allow game developers to access about 6-7 of the 8 cores. The writing is on the wall, and people should design their stuff to work well with 64 or more threads even if customers don't have that yet. Even if the utilization is just running some basic parallel RadiantAI type stuff on every thread, but doing most of the work on 4 cores, so be it. The bleeding edge users should be able to enjoy a game mode that allows for thousands of NPCs.

      --
      [SIG] 10/28/2017: Soylent Upgrade v14 [soylentnews.org]
      • (Score: 2) by JoeMerchant on Thursday June 07 2018, @09:00PM

        by JoeMerchant (3937) on Thursday June 07 2018, @09:00PM (#690067)

        people should design their stuff to work well with 64 or more threads even if customers don't have that yet

        Who do you think runs EA, Activision and UbiSoft? Where's their motivation?

        I agree with you, but I don't think that they do.

        --
        🌻🌻 [google.com]
      • (Score: 2) by bob_super on Thursday June 07 2018, @09:05PM

        by bob_super (1357) on Thursday June 07 2018, @09:05PM (#690069)

        The current question tends to be : Will it run on Xbox/PS4 at reasonable settings, and how much does the extra PC eye candy cost ?

    • (Score: 0) by Anonymous Coward on Friday June 08 2018, @07:41AM (2 children)

      by Anonymous Coward on Friday June 08 2018, @07:41AM (#690237)

      massively multi-threaded processing for many things (think: processing NPC logic and decisions),

      Many games have a multiplayer option. So a popular way to reduce bandwidth is that the NPC logic and decisions are actually deterministic based on committed player input - which doesn't take that much bandwidth. The same actions by the same players at the same time will have the NPCs doing the same exact things. If the NPCs were not so deterministic the various client PCs will have to send each other zillions of detailed updates of the various independent NPCs, instead of mainly only sending the player actions to each other.

      Making it massively multithreaded AND deterministic is possible but not so simple if you actually want it to be faster... And even less simple if you want it to be seem even more intelligent...

      You could have the game work differently depending on whether it's multiplayer or single player but that adds to the complexity.

      • (Score: 0) by Anonymous Coward on Friday June 08 2018, @08:53AM (1 child)

        by Anonymous Coward on Friday June 08 2018, @08:53AM (#690249)
        It's still possible if you only had a few intelligent NPCs that sent their actions similar to the players.
        • (Score: 2) by Freeman on Friday June 08 2018, @05:41PM

          by Freeman (732) on Friday June 08 2018, @05:41PM (#690416) Journal

          Intelligent NPC, now there's an Oxymoron, if I ever saw one.

          --
          Joshua 1:9 "Be strong and of a good courage; be not afraid, neither be thou dismayed: for the Lord thy God is with thee"
  • (Score: 1) by nitehawk214 on Thursday June 07 2018, @08:16PM (3 children)

    by nitehawk214 (1304) on Thursday June 07 2018, @08:16PM (#690050)

    The AMD one is certainly not marketed to gamers. The "game mode" on Threadripper cripples the processor which is somehow supposed to improve game performance. If all you are doing is playing games, you don't want Threadripper at all.

    I am considering Threadripper for my next video / photography processing workstation. Though I don't know if any software can really make use of this many cores yet.

    --
    "Don't you ever miss the days when you used to be nostalgic?" -Loiosh
    • (Score: 2) by takyon on Thursday June 07 2018, @08:40PM (1 child)

      by takyon (881) <takyonNO@SPAMsoylentnews.org> on Thursday June 07 2018, @08:40PM (#690062) Journal

      Only certain games, like DiRT Rally, require the Threadripper "game mode". In my quick search I didn't see any lists that show which games require game mode, don't require game mode, or can use all 16 (soon 32) cores. But maybe those lists are floating around somewhere.

      Game developers know that Intel and AMD are putting out non-Xeon/Epyc CPUs with a lot more cores. The 16-core TR 1950X, 18-core Core i9-7980XE. Now shooting up to 28 or 32 cores, and maybe continuing to 48 or 64 in the near future. So newer game engines should be able to handle running on these, or even use all of the cores in some cases.

      --
      [SIG] 10/28/2017: Soylent Upgrade v14 [soylentnews.org]
      • (Score: 1) by nitehawk214 on Saturday June 09 2018, @02:20AM

        by nitehawk214 (1304) on Saturday June 09 2018, @02:20AM (#690641)

        Well, they are going to have to make better use of multi core and gpu based systems and faster memory busses. The Ghz wars have ended, and Moore's law has expanded in to multi core processors. The coming generations of ships are going to add performance with more cores, not more Ghz.

        --
        "Don't you ever miss the days when you used to be nostalgic?" -Loiosh
    • (Score: 0) by Anonymous Coward on Friday June 08 2018, @06:22PM

      by Anonymous Coward on Friday June 08 2018, @06:22PM (#690437)

      i suspect kedenlive could.

  • (Score: 2) by Apparition on Thursday June 07 2018, @09:18PM

    by Apparition (6835) on Thursday June 07 2018, @09:18PM (#690073) Journal

    It's good for playing a video game while live streaming and having Google Chrome up with a dozen tabs. Most video games don't utilize more than four cores, but over the past few years more games have come out which utilize six or even eight, such as Overwatch. I don't know of any video game that utilizes more than eight cores, but I'm sure that's only a matter of time.

  • (Score: 3, Insightful) by bobthecimmerian on Thursday June 07 2018, @10:32PM (3 children)

    by bobthecimmerian (6834) on Thursday June 07 2018, @10:32PM (#690097)

    I wouldn't get it for gaming, I would guess that a $300-$500 CPU and, say, $1500 invested in the right GPU would outdo a $1000+ 32 core machine and $500 less spent on GPU.

    I rip all of my Blu Rays and DVDs to disk and then reencode them to H.265 (since that's the most bandwidth/disk space -efficient codec my streaming media devices can handle). So I could find a use for one of these machines for a few months. But as it is, right now I have an AMD FX-8320. It's a joke against the cutting edge, really, 4 cores, 8 threads, and the dedicated GPU is even older, form 2010. But I have an SSD drive and 32GB of RAM. I have my H.265 encoding running continuously in the background on 4 threads. The other 4 threads run Firefox, Chrome, Minecraft, a web server, and 3 VMs. I lock the screen and walk away, and my kids sit down to use Chrome and Mineraft. It never slows down.

    Now granted, if I was into more modern games it wouldn't work.

    • (Score: 2) by Spamalope on Thursday June 07 2018, @11:41PM (2 children)

      by Spamalope (5233) on Thursday June 07 2018, @11:41PM (#690119) Homepage

      What software are you using to re-encode/convert to a single video file?
      I haven't done it in a very long time and the landscape has completely changed.

      • (Score: 2, Insightful) by Anonymous Coward on Friday June 08 2018, @04:17AM

        by Anonymous Coward on Friday June 08 2018, @04:17AM (#690194)

        Not the GP, but MakeMKV followed by HandBrake seems to be the choice lately.

      • (Score: 2) by bobthecimmerian on Friday June 08 2018, @11:35AM

        by bobthecimmerian (6834) on Friday June 08 2018, @11:35AM (#690273)

        I am on Linux, but everything I'm doing can be done on Windows. I use MakeMKV to rip films as-is and then ffmpeg to convert to H.265 video and AAC audio. AAC is the audio codec on DVDs and it's lower quality than the audio codecs on Blu Rays, but I can't hear any differences with my mediocre sound system. I've had problems with streaming video to PCs and Android television boxes using the Blu Ray audio codecs, that's why I do the conversion. I use .mkv files instead of .mp4 or similar because Blu Ray subtitles can't be stored in .mp4 files without some kind of conversion process.

  • (Score: 2) by tibman on Thursday June 07 2018, @10:57PM

    by tibman (134) Subscriber Badge on Thursday June 07 2018, @10:57PM (#690111)

    This processor? doubtful. Currently using the Ryzen 8 core version and it's excessive. I play a coop game called Vermintide 2 that starts a local server that other people connect to. When i host we get wrecked so bad. There is enough spare CPU for the AI director to give you a really bad (or Fun?) day.

    The nice part is being able to leave all your normal programs open when you play games. You could permanently give 4 threads to your sooper seekret VM and you wouldn't notice any game slowdown at all. Most people seem to use these high core-count CPUs for streaming their play sessions to the world.

    --
    SN won't survive on lurkers alone. Write comments.
  • (Score: 2) by Freeman on Friday June 08 2018, @05:36PM

    by Freeman (732) on Friday June 08 2018, @05:36PM (#690415) Journal

    I'm not sure what the real bottleneck for performance on my current machine is, but I'm thinking its' got a whole lot more to do with developers who can't code a stable piece of software. As opposed to them legitimately maxing out my CPU, RAM or GPU. Could be that my 8GB RX480 is having issues, but I doubt it. I also got way more RAM than I will need for the foreseeable future. VR does demand a bit more CPU, RAM, and GPU power due to the nature of the beast. No one's really taking advantage of my 8-core CPU, though.

    SSDs makes loading anything feel more snappy. Comparing a HDD to a SSD is like comparing a VHS to a DVD. There's a vast improvement in performance when upgrading to a SSD. That would be the single best thing to upgrade, if you want to feel like you're going to make a serious impact on your PC performance.

    Faster RAM is notable, but probably not very noticeable when you go from 2400 to 3200mhz. You just won't see that much of a performance gain. Going from 4GB to 8GB of RAM could on the other hand make a much bigger impact, if you don't have enough to keep the 5,000 tabs up in your Browser of choice.

    GPU upgrade from 128-bit to 256-bit, or 256-bit to 512-bit will typically be a major improvement over the previous GPU. There are other factors like how much RAM is on the card and speed of the card, but the higher bandwidth cards usually follow an upward trend regarding those things.

    Also, regarding Multiplayer performance. Sometimes you just need a faster internet connection, though latency is a key issue. Other times it's down to the software developers that can't code.

    Minor jump in CPU performance? Bottom of the list of "Maybe this will help my computer run faster".

    --
    Joshua 1:9 "Be strong and of a good courage; be not afraid, neither be thou dismayed: for the Lord thy God is with thee"