Discussion Zen 5 Speculation (EPYC Turin and Strix Point/Granite Ridge - Ryzen 9000)

Page 436 - Seeking answers? Join the AnandTech community: where nearly half-a-million members share solutions and discuss the latest tech.

itsmydamnation

Platinum Member
Feb 6, 2011
2,821
3,313
136
I'm preparing a little video (no I'm not trying to be a MLID/RGT, it's a different kind of video) about Zen 5, can we recap what we know about its internals?
- 8 wide decode
- Same or higher clocks
- SPECINT +40%
- full width AVX 512 implem

What else?
we know it isn't 8 wide decode , it does something in decode but we don't know exactly what, two fetch blocks is all that is listed. is that parallel, used for branches etc .

I think the only things we can say for certain are from this slide
 

H433x0n

Senior member
Mar 15, 2023
933
1,032
96
Turin is N4P or N4X (not sure which tbh) which is in the same family as N5. Just better/more refined.
I’ve heard from people that I consider reliable that it uses N4X. I’ve got a hard time believing it since N4X was regarded as a bit of a meme.

There’s some tidbits that support it though like the increased over Zen 4 and the higher reported power consumption for desktop parts.
 
Jul 27, 2020
17,174
11,042
106
Some distributed computing enthusiasts do have Intel p+e CPUs, but even though an e core performs roughly similar to one p HT thread, these CPUs are still awkward to handle in a distributed computing node. Just recently I heard of weird issues with Windows' CPU time accounting on these CPUs. And way before that I saw several reports of performance problems of multithreaded distributed computing applications on these CPUs, which are completely to be expected and can only be worked around by restricting the application to run on cores of same type.
Not sure if the Win11 scheduler has been improved but Linux is supposedly better at dealing with hybrid cores: https://www.phoronix.com/news/Linux-6.5-Intel-Hybrid-Sched
 
Jul 27, 2020
17,174
11,042
106
- 8 wide decode
- Same or higher clocks
- SPECINT +40%
- full width AVX 512 implem

What else?
Special instructions to accelerate AI workloads. (Source: AMD slides)

Possibly more performant SMT (due to beefier execution resources). (Source: Hopium)

May hit 6 GHz boost. (Source: Hopium)

First time when a previous gen X3D chip may not be able to touch the nextgen vanilla chip in gaming workloads. (Source: Hopium)

DDR5-6400 will possibly be the base RAM configuration. (Source: Hopium)

RDNA3 iGPU should beat current Intel Core Lake iGPU and Ryzen 7000 series desktop iGPU. May get beaten by Arrow Lake iGPU. (Source: Hopium)

May beat Zen 4 vanilla chips in ECO mode. (Source: Hopium)
 

StefanR5R

Elite Member
Dec 10, 2016
5,633
8,107
136
Now that we've smoked out the rat, I can't wait for details on Zen 5 LP.
Not just the perf, but what did they take out, power draw, etc.
Be interesting to see areal density too.
Apparently there will be only a small number of LP cores. [Purpose: to host background tasks in idle situations/ connected standby maybe — not to prop up Cinebench. ;-) ] Thus, areal density, while not unimportant, may not be a central design goal. For Zen 5LP, that is.

Given Bergamo was only a 1.33x increase in cores over Genoa and the Zen5 successor is supposed to be more like 1.5x there must be a significant difference in layout there too.
Genoa and Bergamo still have some spare room under the lid. (According to published photos, not that I'd delidded one myself.) I guess the new IOD for Turin and Turin-Dense could be a more slender rectangle than Genoa's and Bergamo's IOD, for some more "shore line" to area ratio. Both for putting the additionally needed GMI links on the chip and to facilitate their routing on the package.

The 96 core Threadripper 7995WX however is very crowded under the hood. But whether there will be a direct Zen 5 based successor to this one remains to be seen anyway.

I have 352 Genoa cores myself.
On a single socket?
4x 64c/128t and 1x 96c/192t according to Mark's signature. All 1P I think.

Are you able to utilize them properly or do you need to resort to putting your workloads in VMs for better core occupancy?
In Distributed Computing, we often run n instances of single-threaded processes. This scales without problem to so many threads (on Linux; I am not up to date with Windows). Sometimes we run fewer instances of multi-threaded processes. With some of such applications, performance suffers a lot if threads of one such process end up running on different CCXs. That has been an issue with Zen 1...4 and obviously will remain with Zen 5. Hard to say what will happen with Zen 6 with its substantially changed SOCs. (Or Strix Halo already, in fact.) The problem is two-fold: Inter-thread shared data gets onto more caches than strictly needed, and inter-thread communications beyond CCX boundaries is slow and energy costly. But we don't need VMs or even containers to solve this; we can do this with helper tools, or in case of EPYCs can use a BIOS option which (ab)uses NUMA hints to coerce NUMA aware operating system into cache-aware thread scheduling. (Neither Windows' nor Linux's kernel implement a cache-aware scheduling policy. The kernel developers probably have their reasons to leave this to userspace to handle.)

EDIT: CPUs with unified last level cache would remove this inconvenience. But the price to pay would be higher latency of the last level cache, and higher chip manufacturing costs.

EDIT 2,
Some distributed computing enthusiasts do have Intel p+e CPUs, but even though an e core performs roughly similar to one p HT thread, these CPUs are still [...troublesome...]
Not sure if the Win11 scheduler has been improved but Linux is supposedly better at dealing with hybrid cores: https://www.phoronix.com/news/Linux-6.5-Intel-Hybrid-Sched
When Inte'ls offer was 8c/16t + 8c/8t, even with scheduling like that (regardless if implemented in kernelspace or userspace), you are left with a large asymmetry. It is now better with 8c/16t + 16c/16t at the top end, but still not symmetric.
As an aside, recall how Intel solved this in their LGA1700 Xeon line: ark.intel.com (Though this line is not targeted to compute servers.)
 
Last edited:
sale-70-410-exam    | Exam-200-125-pdf    | we-sale-70-410-exam    | hot-sale-70-410-exam    | Latest-exam-700-603-Dumps    | Dumps-98-363-exams-date    | Certs-200-125-date    | Dumps-300-075-exams-date    | hot-sale-book-C8010-726-book    | Hot-Sale-200-310-Exam    | Exam-Description-200-310-dumps?    | hot-sale-book-200-125-book    | Latest-Updated-300-209-Exam    | Dumps-210-260-exams-date    | Download-200-125-Exam-PDF    | Exam-Description-300-101-dumps    | Certs-300-101-date    | Hot-Sale-300-075-Exam    | Latest-exam-200-125-Dumps    | Exam-Description-200-125-dumps    | Latest-Updated-300-075-Exam    | hot-sale-book-210-260-book    | Dumps-200-901-exams-date    | Certs-200-901-date    | Latest-exam-1Z0-062-Dumps    | Hot-Sale-1Z0-062-Exam    | Certs-CSSLP-date    | 100%-Pass-70-383-Exams    | Latest-JN0-360-real-exam-questions    | 100%-Pass-4A0-100-Real-Exam-Questions    | Dumps-300-135-exams-date    | Passed-200-105-Tech-Exams    | Latest-Updated-200-310-Exam    | Download-300-070-Exam-PDF    | Hot-Sale-JN0-360-Exam    | 100%-Pass-JN0-360-Exams    | 100%-Pass-JN0-360-Real-Exam-Questions    | Dumps-JN0-360-exams-date    | Exam-Description-1Z0-876-dumps    | Latest-exam-1Z0-876-Dumps    | Dumps-HPE0-Y53-exams-date    | 2017-Latest-HPE0-Y53-Exam    | 100%-Pass-HPE0-Y53-Real-Exam-Questions    | Pass-4A0-100-Exam    | Latest-4A0-100-Questions    | Dumps-98-365-exams-date    | 2017-Latest-98-365-Exam    | 100%-Pass-VCS-254-Exams    | 2017-Latest-VCS-273-Exam    | Dumps-200-355-exams-date    | 2017-Latest-300-320-Exam    | Pass-300-101-Exam    | 100%-Pass-300-115-Exams    |
http://www.portvapes.co.uk/    | http://www.portvapes.co.uk/    |