Discussion Zen 5 Speculation (EPYC Turin and Strix Point/Granite Ridge - Ryzen 9000)

Page 440 - Seeking answers? Join the AnandTech community: where nearly half-a-million members share solutions and discuss the latest tech.

Abwx

Lifer
Apr 2, 2011
11,103
3,780
136
Wouldn't be the first time GB reported the wrong cache size. On the other hand, it could be correct and a Zen 5 update to their embedded platform.
All other cache sizes seems to be right, so it s unlikely that GB made a miss just for the L1D, and the ST score, if at about 1.4GHz, is aligned with the late april GB submissions, this would translate as about 23% better ST score than a 7950X.
 

RnR_au

Golden Member
Jun 6, 2021
1,759
4,285
106
Memory capacity is king. Strix Halo + 64GB RAM will beat a 4090 in any AI task if you scale it up sufficiently. And for local LLMs where capacity is by far the biggest bottleneck, it won't even be close.

Strix Halo directly competes for the semi-pro market that is buying 4090s instead of RTX 6000s. Significantly slower for tasks that fit inside 24GB, but can be used for tasks a 4090 can't touch.
Just on this, I noticed this thread this morning;


From my understanding, Strix Point will also feature a souped up NPU, so Windows folks might get some inferencing love if any of the partners will field Strix Point laptops with 128GB of ram.
 
Reactions: Tlh97 and Mopetar

Abwx

Lifer
Apr 2, 2011
11,103
3,780
136

poke01

Senior member
Mar 8, 2022
998
1,096
106
Just on this, I noticed this thread this morning;


From my understanding, Strix Point will also feature a souped up NPU, so Windows folks might get some inferencing love if any of the partners will field Strix Point laptops with 128GB of ram.
Great points. I would also say the huge bandwidth helps as well. That’s before Apple is using LP5X.

MLX is also advancing as well. Honestly, the more companies try to break down Nvidias iron wall the better.

I would like to say AMD should make a larger APU with more bandwidth. Maybe, Zen6?
Looking at OpenAI’s event yesterday, the hype around AI is not dying. Lots of memory is needed.
 

branch_suggestion

Senior member
Aug 4, 2023
244
523
96
Eng samples having unoptimised firmware or straight up firmware locks makes any performance assumptions dubious.
That one Weibo post that sent people mad is a very obvious example of limited testing. If they are a Lenovo product manager then it makes it far more likely that AMD would put in firmware locks to make such lapses of judgement moot.
 

H433x0n

Senior member
Mar 15, 2023
933
1,032
96
There s forcibly some big variability, in a comparision with a linux driven 7800X3D, wich is sure to be at 5Ghz, that could be 27-29% or so, at this point the GB submission give no clue about the test plateform, given that it display 6.74GB RAM it could be single channel, or not.

For the April GB scores when compared against a 14900K w/ DDR5-4800 JEDEC it ended up being 20-22%. If that were the score I think that'd actually be quite good since a desktop chip with larger L3$ would score even better than this. If you extrapolate the small IPC bump from a desktop chip having fatter L3$ and the slight fmax increase over Zen 4 you would see ~30% 1T perf increase.

I personally don't think it's worth drawing conclusions from it though since there's so much variability in GB scores. When searching through scores, you can see the same processors get results that vary by 10% from user submitted data. About the only thing that seems constant is that the Zen 5 core doesn't do great with the AES subtests.
 

moinmoin

Diamond Member
Jun 1, 2017
4,989
7,758
136
Will be highly interesting to see how this works out and what strategy for switching the threads over they might employ.
Will they switch on a thread level, on a core level (thread pairs), what thresholds, etc. ...
If the previously discussed patent 10698472 is any indication it would work on an opcode level, only implementing opcodes that can be churned through without any real computation (like JMP or anything keeping I/O going etc.) otherwise throwing an illegal opcode exception to move the computation onto a fat core.

The real question is just how castrated Z5LP really is.
Going by above patent it could well be so castrated it couldn't be used as a full core of its own. It would be little more than a fast shortcut for the simplest of all opcodes.
 

Mahboi

Senior member
Apr 4, 2024
677
1,136
91
If the previously discussed patent 10698472 is any indication it would work on an opcode level, only implementing opcodes that can be churned through without any real computation (like JMP or anything keeping I/O going etc.) otherwise throwing an illegal opcode exception to move the computation onto a fat core.
Going by above patent it could well be so castrated it couldn't be used as a full core of its own. It would be little more than a fast shortcut for the simplest of all opcodes.
I don't know, how would that work for a LP island?
Sounds like something that would fire up the fat cores for literally any workload.
 

cherullo

Member
May 19, 2019
47
108
106
I don't know, how would that work for a LP island?
Sounds like something that would fire up the fat cores for literally any workload.
Yeah, the LP cores do need to perform proper computations to actually prevent the HP cores from powering up.
Since it's transparent to the OS, it can support any mix of instructions, regardless of extension. For example, it would be nice to have memory and string instructions from AVX512, even at reduced throughput, but no FMA at all. You can even support certain zeroing idioms without supporting the instruction itself.
You can also strip out all the legacy stuff like 16- and 32-bit modes, x87 FPU, etc. It can be really, really lean.
What moinmoin described is more like a DMA engine, which is also useful.
 

SteinFG

Senior member
Dec 29, 2021
498
591
106
How possible it is for AMD to just ditch numbered tiers? Instead of ryzen 5,7,9 it's just all "Ryzen AI" for Kraken chips, and maybe "Ryzen AI HX" for Strix chips?
 
Last edited:

LightningZ71

Golden Member
Mar 10, 2017
1,645
1,929
136
Could the LP cores just be stripped of all the "go fast" bits and anything else that can't be emulated in microcode and essentially be maximally efficient microcode engines instead? They could retain full isa compatibility with the performance cores.
 

SteinFG

Senior member
Dec 29, 2021
498
591
106
Weird hypothesis, but frankly Intel and AMD always used the numbers very poorly.
Out of 9 numbers, 4 are used: 3/5/7/9. It's a generally poor system because you don't want to have many tiers(muddies the lineup), and you don't want to have lower numbers than 5. Anyone identifies i3/R3 as "bad", to the point that some noobs actually think a 2023 i3 is worse than a 2014 i7.

Frankly I think clothes did it pretty well: S/M/L/XL?
Make 4 tiers:
- Light
- Medium
- Heavy
- Ultra

But it's kind of problematic too since it's a two sided coin, you get Bigger Number Better just like you get Lower Number Worse.
This just replaces a number with a word, same meaning. What actually should've happened is, each gen should shift the window of numbers. Example: Zen 1 would have Ryzen 1,2,3 tiers, Zen 2 will have Ryzen 2,3,4,5 tiers. Zen 3 would have Ryzen 3,4,5,6 tiers. Zen 4 would have Ryzen 5,6,7 tiers. This way, you avoid the problem you are talking about. Someone will say "I have a Ryzen 4", and you instantly know it's a chip made inbetween 2019 and 2022.
 

Mahboi

Senior member
Apr 4, 2024
677
1,136
91
This just replaces a number with a word, same meaning.
Yes, that's the point.
What actually should've happened is, each gen should shift the window of numbers. Example: Zen 1 would have Ryzen 1,2,3 tiers, Zen 2 will have Ryzen 2,3,4,5 tiers. Zen 3 would have Ryzen 3,4,5,6 tiers. Zen 4 would have Ryzen 5,6,7 tiers. This way, you avoid the problem you are talking about.
You just create a huge new slew of problems.
Someone will say "I have a Ryzen 4", and you instantly know it's a chip made inbetween 2019 and 2022.
So just the difference between a 3950x and a 7600.
Just a smidge of 31% better MT perf on one, vs 30% better ST perf on the other.

No, I don't think that's better.
 
sale-70-410-exam    | Exam-200-125-pdf    | we-sale-70-410-exam    | hot-sale-70-410-exam    | Latest-exam-700-603-Dumps    | Dumps-98-363-exams-date    | Certs-200-125-date    | Dumps-300-075-exams-date    | hot-sale-book-C8010-726-book    | Hot-Sale-200-310-Exam    | Exam-Description-200-310-dumps?    | hot-sale-book-200-125-book    | Latest-Updated-300-209-Exam    | Dumps-210-260-exams-date    | Download-200-125-Exam-PDF    | Exam-Description-300-101-dumps    | Certs-300-101-date    | Hot-Sale-300-075-Exam    | Latest-exam-200-125-Dumps    | Exam-Description-200-125-dumps    | Latest-Updated-300-075-Exam    | hot-sale-book-210-260-book    | Dumps-200-901-exams-date    | Certs-200-901-date    | Latest-exam-1Z0-062-Dumps    | Hot-Sale-1Z0-062-Exam    | Certs-CSSLP-date    | 100%-Pass-70-383-Exams    | Latest-JN0-360-real-exam-questions    | 100%-Pass-4A0-100-Real-Exam-Questions    | Dumps-300-135-exams-date    | Passed-200-105-Tech-Exams    | Latest-Updated-200-310-Exam    | Download-300-070-Exam-PDF    | Hot-Sale-JN0-360-Exam    | 100%-Pass-JN0-360-Exams    | 100%-Pass-JN0-360-Real-Exam-Questions    | Dumps-JN0-360-exams-date    | Exam-Description-1Z0-876-dumps    | Latest-exam-1Z0-876-Dumps    | Dumps-HPE0-Y53-exams-date    | 2017-Latest-HPE0-Y53-Exam    | 100%-Pass-HPE0-Y53-Real-Exam-Questions    | Pass-4A0-100-Exam    | Latest-4A0-100-Questions    | Dumps-98-365-exams-date    | 2017-Latest-98-365-Exam    | 100%-Pass-VCS-254-Exams    | 2017-Latest-VCS-273-Exam    | Dumps-200-355-exams-date    | 2017-Latest-300-320-Exam    | Pass-300-101-Exam    | 100%-Pass-300-115-Exams    |
http://www.portvapes.co.uk/    | http://www.portvapes.co.uk/    |