Question Zen 6 Speculation Thread

Page 150 - Seeking answers? Join the AnandTech community: where nearly half-a-million members share solutions and discuss the latest tech.

OneEng2

Senior member
Sep 19, 2022
612
850
106
@igor_kavinski ,

I still think that using 2 different kinds of memory invalidates the test. When using the same memory type and just changing the timing, what I have generally seen in the past is that gaming performance is minimally effected.

It seems like AMD's X3D lineup conclusively showed that avoiding main memory hits (latency) is much more important than bandwidth.

I wouldn't call it a drastic increase in memory bandwidth per core
Zen 5 Turin D: 192 cores, 252Gb/sec bandwidth (DDR6000 X 12) = 1.31Gb/s/core

Zen 6 Venice D: 256 cores, 1638Gb/sec bandwidth (MRDIMM 12800) = 6.4Gb/s/core.

600% increase seems "drastic" to me.
 
Reactions: Io Magnesso

OneEng2

Senior member
Sep 19, 2022
612
850
106

A database server application. Note the 2 Xeon systems where the only difference is the memory used. The much greater memory bandwidth didn't help the 128 core Xeon gain that much.


This one actually shows a decrease in performance when gaining more bandwidth.

Then we have this one:


Xeon trounces Turin here. This seems to be a very memory sensitive benchmark.

Anyway, this page is here: https://www.phoronix.com/review/amd-epyc-9655/3

I can only assume that AMD knows what they are doing with the massive uplift in bandwidth for the next gen EPYC's. I find it difficult to believe that there isn't a really good reason to improve the bandwidth in DC.

I do question the need in desktop though. Seems like there are precious few apps that need more than DDR8000 dual channel.... and the lions share of apps that don't even need a fraction of that.
 

Io Magnesso

Member
Jun 12, 2025
30
16
36

A database server application. Note the 2 Xeon systems where the only difference is the memory used. The much greater memory bandwidth didn't help the 128 core Xeon gain that much.


This one actually shows a decrease in performance when gaining more bandwidth.

Then we have this one:


Xeon trounces Turin here. This seems to be a very memory sensitive benchmark.

Anyway, this page is here: https://www.phoronix.com/review/amd-epyc-9655/3

I can only assume that AMD knows what they are doing with the massive uplift in bandwidth for the next gen EPYC's. I find it difficult to believe that there isn't a really good reason to improve the bandwidth in DC.

I do question the need in desktop though. Seems like there are precious few apps that need more than DDR8000 dual channel.... and the lions share of apps that don't even need a fraction of that.
The dual sockets in Xeon6 are still immature.
There should be a range of improvements, but...
Is it an immature problem of control on the software side?
 
Jul 27, 2020
25,046
17,418
146
It seems like AMD's X3D lineup conclusively showed that avoiding main memory hits (latency) is much more important than bandwidth.
It's not just latency though.


Once the cache is accessed (latency), it needs to be read from/written to really quickly and that's where the increased bandwidth helps.

I think the missing piece of the puzzle is that no one has really investigated the impact of RAM speed on an X3D CPU, going from 3600 to 6400 MT/s in 1:1 mode.
 

Io Magnesso

Member
Jun 12, 2025
30
16
36

OneEng2

Senior member
Sep 19, 2022
612
850
106
As much as the memory bandwidth helps certain HPC applications I have to feel that the mega memory bandwidth is an AI play primarily.
I wonder if that isn't the case.

Last generation of DC from Intel and AMD, the Xeon had a very large bandwidth advantage yet got beaten pretty badly (on average about 40%) which would lead one to believe that EITHER most server loads aren't that bandwidth limited OR Intel did a VERY bad job with their DC processors.

As seen in the server and workstation benchmarks, there are certainly high points for the Xeon, but not that many.

As AMD had a substantial lead last generation and did so with a pretty severe bandwidth deficit, it would seem that AMD sees a future where bandwidth in DC is much more important that it is today. Perhaps AI and LLM is that reason.
 

OneEng2

Senior member
Sep 19, 2022
612
850
106
The dual sockets in Xeon6 are still immature.
There should be a range of improvements, but...
Is it an immature problem of control on the software side?
Yea, they seem pretty bad. When the original benchmarks came out everyone just assumed it was some bug that they would fix in a month or two .... yet here we are today and no update that would lead us to believe that a dual socket Xeon is a good idea (note, there were some benchmarks that it worked really well for which is quite a delimit).
 

OneEng2

Senior member
Sep 19, 2022
612
850
106
Bandwidth impact on Epyc Zen 5 DDR5-4800 vs. DDR5-6000: https://www.phoronix.com/review/amd-epyc-9755-ddr5/9
Scaling the memory bandwidth up by 25% resulted in various degrees of performance improvement. The top 5 certainly gained greatly (with #1 scaling nearly linearly with the bandwidth improvement).

I wonder how much a new IOD and a 300% increase in per core bandwidth will effect these benchmarks?
 

Thibsie

Golden Member
Apr 25, 2017
1,068
1,236
136
I wonder if that isn't the case.

Last generation of DC from Intel and AMD, the Xeon had a very large bandwidth advantage yet got beaten pretty badly (on average about 40%) which would lead one to believe that EITHER most server loads aren't that bandwidth limited OR Intel did a VERY bad job with their DC processors.
Just means bottleneck in Xeon is somewhere else.
 
Reactions: Io Magnesso

Tigerick

Senior member
Apr 1, 2022
762
717
106
@igor_kavinski ,

I still think that using 2 different kinds of memory invalidates the test. When using the same memory type and just changing the timing, what I have generally seen in the past is that gaming performance is minimally effected.

It seems like AMD's X3D lineup conclusively showed that avoiding main memory hits (latency) is much more important than bandwidth.


Zen 5 Turin D: 192 cores, 252Gb/sec bandwidth (DDR6000 X 12) = 1.31Gb/s/core

Zen 6 Venice D: 256 cores, 1638Gb/sec bandwidth (MRDIMM 12800) = 6.4Gb/s/core.

600% increase seems "drastic" to me.
I think you miscalculated Turin Dense memory bandwidth. It should be:
  • Turin Dense 12 x DDR5-6000: 614GB/s / 192 = 3.2 GB/s per core
  • Venice SP8 8 x DDR5-8000: 512GB/s / 128 = 4 GB/s per core
  • Venice 16 x DDR5-8000: 1024GB/s / 256 = 4 GB/s per core
  • Venice 16 x MRDIMM-12800: 1638GB/s / 256 = 6.4 GB/s per core
MRDIMM essential is a 6400 double bumped to 12800 (with latency trade-off). Unless there is bump in memory capacity, DDR5-8000 should be sufficient for majority customers.
 
Last edited:
Reactions: Joe NYC and OneEng2

rainy

Senior member
Jul 17, 2013
522
452
136
Zen 5 Turin D: 192 cores, 252Gb/sec bandwidth (DDR6000 X 12) = 1.31Gb/s/core

Zen 6 Venice D: 256 cores, 1638Gb/sec bandwidth (MRDIMM 12800) = 6.4Gb/s/core.

600% increase seems "drastic" to me.

It seems drastic because you've made an obvious mistake: 12 channels of DDR5-6000 translate to 576GB/s.
 

StefanR5R

Elite Member
Dec 10, 2016
6,514
10,149
136
  • Turin Dense 12 x DDR5-6000: 614GB/s / 192 = 3.2 GB/s per core
  • Venice SP8 8 x DDR5-8000: 512GB/s / 128 = 4 GB/s per core
  • Venice 16 x DDR5-8000: 1024GB/s / 256 = 4 GB/s per core
  • Venice 16 x MRDIMM-12800: 1638GB/s / 256 = 6.4 GB/s per core
Apropos.
  • Rome 8 x DDR4-3200: 204.8 GB/s / 64 = 3.2 GB/s per core
  • Naples 8 x DDR4-2666: 170.6 GB/s / 32 = 5.3 GB/s per core :-D
It's apples to oranges of course. Naples was AMD's return to server mainly via the HPC segment as their entry door. Rome was dipping into general purpose and hyperscalers. Turin Dense is mainly cloud. Venice coincides with AMD's first rack-level AI solution.
 
Last edited:
Reactions: Io Magnesso

LightningZ71

Platinum Member
Mar 10, 2017
2,238
2,742
136
Keep in mind, when X3D chips are brought up, it's not just an average latency improvement that's present, it's also an average apparent bandwidth improvement. Each of those memory hits that are served by the 3d cache don't just get the first word to the core quicker, the following words are streamed at far higher effective bandwidth than they would have if they were coming from main memory.
 

OneEng2

Senior member
Sep 19, 2022
612
850
106
It seems drastic because you've made an obvious mistake: 12 channels of DDR5-6000 translate to 576GB/s.
Not sure how I got my previous number. You are correct.

Turin D: 3GB/sec/core
Venice D: 6.4GB/sec/core

Still more than double the bandwidth per core vs Turin.... and therefore still drastic IMO.
 

OneEng2

Senior member
Sep 19, 2022
612
850
106
Keep in mind, when X3D chips are brought up, it's not just an average latency improvement that's present, it's also an average apparent bandwidth improvement. Each of those memory hits that are served by the 3d cache don't just get the first word to the core quicker, the following words are streamed at far higher effective bandwidth than they would have if they were coming from main memory.
Absolutely. Of course, for many of the benchmarks being quoted, the transfers are artificially large in order to explicitly ensure that it will come from main memory.

In real world applications, I think that this is rarely the case and quite a bit of memory access is kept local within one of the levels of cache thus preventing a main memory access. Additionally, the specific memory elements in cache are frequently accessed repetitively.

All in all, I think that the new IOD and faster memory speeds are going to be the biggest improvements we see in Zen 6 platforms next year.

I am actually not expecting much in the way of IPC of Zen 6 over Zen 5 (10-15%) and little or no increase in clock speed. In more bandwidth limited operations (likely in DC applications / HPC /AI) much larger increases in performance will be seen than the IPC bump. At a minimum, we can expect the larger core counts will result in higher performance. The IPC and bandwidth increases would then be on top of that.

Zen 6 Venice should be a very powerful platform indeed.
 

OneEng2

Senior member
Sep 19, 2022
612
850
106
Just means bottleneck in Xeon is somewhere else.
Possibly; however, bandwidth limited benchmarks indicate that the performance scales with memory bandwidth. It just doesn't scale in all cases enough to be competitive with EPYC Turin.
 

StefanR5R

Elite Member
Dec 10, 2016
6,514
10,149
136
Turin D: 3GB/sec/core
Venice D: 6.4GB/sec/core

Still more than double the bandwidth per core vs Turin.... and therefore still drastic IMO.
...vs. Turin-D!
Also much more level 3 cache per core, and much bigger level 3 cache per core complex.

This merely shows a very different focus of the highest-core-count Venice[-D] compared to the highest-core-count Turin[-D]. (A different focus enabled by manufacturing process progress, by JEDEC's work, and more.) Turin-Dense is not an AI-AI-AI powerhouse, Turin-Classic is.

As for 6.4GB/sec/core: Good job. I have exactly this at home for more than five years now. ;-)
 
sale-70-410-exam    | Exam-200-125-pdf    | we-sale-70-410-exam    | hot-sale-70-410-exam    | Latest-exam-700-603-Dumps    | Dumps-98-363-exams-date    | Certs-200-125-date    | Dumps-300-075-exams-date    | hot-sale-book-C8010-726-book    | Hot-Sale-200-310-Exam    | Exam-Description-200-310-dumps?    | hot-sale-book-200-125-book    | Latest-Updated-300-209-Exam    | Dumps-210-260-exams-date    | Download-200-125-Exam-PDF    | Exam-Description-300-101-dumps    | Certs-300-101-date    | Hot-Sale-300-075-Exam    | Latest-exam-200-125-Dumps    | Exam-Description-200-125-dumps    | Latest-Updated-300-075-Exam    | hot-sale-book-210-260-book    | Dumps-200-901-exams-date    | Certs-200-901-date    | Latest-exam-1Z0-062-Dumps    | Hot-Sale-1Z0-062-Exam    | Certs-CSSLP-date    | 100%-Pass-70-383-Exams    | Latest-JN0-360-real-exam-questions    | 100%-Pass-4A0-100-Real-Exam-Questions    | Dumps-300-135-exams-date    | Passed-200-105-Tech-Exams    | Latest-Updated-200-310-Exam    | Download-300-070-Exam-PDF    | Hot-Sale-JN0-360-Exam    | 100%-Pass-JN0-360-Exams    | 100%-Pass-JN0-360-Real-Exam-Questions    | Dumps-JN0-360-exams-date    | Exam-Description-1Z0-876-dumps    | Latest-exam-1Z0-876-Dumps    | Dumps-HPE0-Y53-exams-date    | 2017-Latest-HPE0-Y53-Exam    | 100%-Pass-HPE0-Y53-Real-Exam-Questions    | Pass-4A0-100-Exam    | Latest-4A0-100-Questions    | Dumps-98-365-exams-date    | 2017-Latest-98-365-Exam    | 100%-Pass-VCS-254-Exams    | 2017-Latest-VCS-273-Exam    | Dumps-200-355-exams-date    | 2017-Latest-300-320-Exam    | Pass-300-101-Exam    | 100%-Pass-300-115-Exams    |
http://www.portvapes.co.uk/    | http://www.portvapes.co.uk/    |