Question Incredible Apple M4 benchmarks...

Page 7 - Seeking answers? Join the AnandTech community: where nearly half-a-million members share solutions and discuss the latest tech.

Mopetar

Diamond Member
Jan 31, 2011
7,961
6,312
136
Wait, mikeegg is senttoschool? Or am I misunderstanding?

Yeah it's the same person, apparently he changed his account name. There's a few people who have done it for whatever reason. Unless there's some other way to tell that I'm not aware of you'd have to go back and look at an old post where someone replied to them as name in any quoted text from a post doesn't update with the name change on the account.
 
Reactions: igor_kavinski

StinkyPinky

Diamond Member
Jul 6, 2002
6,782
798
126
Why do people waste their energy with these synthetic benchmarks when there are plenty of real world ones that can be used. I just don't get it, Geekbench is just a fun little tool and should never be taken that seriously. Yet people seem to take it as the holy grail of CPU performance.
 
Reactions: IGBT

SarahKerrigan

Senior member
Oct 12, 2014
475
924
136
Why do people waste their energy with these synthetic benchmarks when there are plenty of real world ones that can be used. I just don't get it, Geekbench is just a fun little tool and should never be taken that seriously. Yet people seem to take it as the holy grail of CPU performance.

Geekbench isn't bad, in a pinch. SPEC is better, and now we have SPEC numbers for M4, demonstrating small but real clock-normalized gains and a decent generational clock bump.
 

SarahKerrigan

Senior member
Oct 12, 2014
475
924
136
Re: Geekbench vs. spec. Did you ever see this 2020 article on Geekbench 5?

Performance Delivered a New Way Part 2 — Geekbench versus SPEC

Author: Ram Srinivasan, Performance Architect, NUVIA

I've seen it, and I agree with some of their points, though with reservations. The tl;dr version of my opinions on Geekbench is that it's pretty good, better than other free benchmarks available, but is not a replacement for SPEC even if they tend to correlate decently well with each other. It is also notable that their prediction for the A14 correlated much less well with Anandtech's eventual SPEC estimate for that processor.

Any benchmark primarily available in binary form is going to be limiting in terms of what you can learn from it IMO. This is doubly true in that Primate Labs openly states that it uses platform-specific intrinsics when available. I'm very wary of the latter because it has a tendency to turn benchmarks from a measurement of a core's performance to a measurement of how good a machine-specific implementation of an algorithm is.

These issues could probably be addressed by buying the source license and doing a custom build, ideally excluding the platform-specific code, but Primate doesn't advertise a list price and I've never seen its source code in use anywhere myself. SPEC is cheap, ubiquitous, contains no platform-specific optimization whatsoever, and is very well-analyzed in its behavior.

Edited to add: Furthermore, I just don't think that "two integer CPU benchmark suites correlate with each other" is really meaningful. They're both integer benchmark suites. Some CPUs have faster integer perf than others. If they don't have a reasonably strong correlation, it means that one or both is not being a very good benchmark. This is extra true if you're measuring your R2 across only seven points. For the heck of it, I tried correlating SpecFP to Specint, which measure completely and utterly different things, using seven random SPEC06 results from Anandtech. I got an R2 of approximately 0.92.
 
Last edited:

Hitman928

Diamond Member
Apr 15, 2012
5,423
8,330
136
I've seen it, and I agree with some of their points, though with reservations. The tl;dr version of my opinions on Geekbench is that it's pretty good, better than other free benchmarks available, but is not a replacement for SPEC even if they tend to correlate decently well with each other. It is also notable that their prediction for the A14 correlated much less well with Anandtech's eventual SPEC estimate for that processor.

Any benchmark primarily available in binary form is going to be limiting in terms of what you can learn from it IMO. This is doubly true in that Primate Labs openly states that it uses platform-specific intrinsics when available. I'm very wary of the latter because it has a tendency to turn benchmarks from a measurement of a core's performance to a measurement of how good a machine-specific implementation of an algorithm is.

These issues could probably be addressed by buying the source license and doing a custom build, ideally excluding the platform-specific code, but Primate doesn't advertise a list price and I've never seen its source code in use anywhere myself. SPEC is cheap, ubiquitous, contains no platform-specific optimization whatsoever, and is very well-analyzed in its behavior.

Edited to add: Furthermore, I just don't think that "two integer CPU benchmark suites correlate with each other" is really meaningful. They're both integer benchmark suites. Some CPUs have faster integer perf than others. If they don't have a reasonably strong correlation, it means that one or both is not being a very good benchmark. This is extra true if you're measuring your R2 across only seven points. For the heck of it, I tried correlating SpecFP to Specint, which measure completely and utterly different things, using seven random SPEC06 results from Anandtech. I got an R2 of approximately 0.92.

Agreed. You can use pretty much any CPU benchmark and it will show very strong correlation. I tried it with Cinebench r20 from the 2700x to the 7950x and Anandtech's SPECint numbers and got 0.998 for the correlation coefficient. You know what that tells me? Faster CPUs perform better on CPU benchmarks. Not a very surprising result, but that doesn't mean that they track closely enough that you can substitute one for the other, especially as you get to larger data sets.
 

SarahKerrigan

Senior member
Oct 12, 2014
475
924
136
Agreed. You can use pretty much any CPU benchmark and it will show very strong correlation. I tried it with Cinebench r20 from the 2700x to the 7950x and Anandtech's SPECint numbers and got 0.998 for the correlation coefficient. You know what that tells me? Faster CPUs perform better on CPU benchmarks. Not a very surprising result, but that doesn't mean that they track closely enough that you can substitute one for the other, especially as you get to larger data sets.

Amen.
 
Reactions: lightmanek

Hitman928

Diamond Member
Apr 15, 2012
5,423
8,330
136

I just realized that we can easily check their predictive accuracy with newer CPUs. If I use their prediction model with the 5950x, I get:

5950x

GB5 Integer = 1405

Predicted SPECint = 6.83

Actual SPECint (Anandtech) = 7.65

Error = -10.72%

If I repeat it for the 7950x, I get a -6.02% error. For the 13900k, I get a +4.78% error.

So the prediction based on the GB5 score would overestimate the 13900k while underestimating the Zen 3 and Zen 4 CPUs. It's not huge amounts of error, but it is significant and much higher than they show in their write-up. It is made worse when you have one line of CPUs swinging one way and the other line swinging in the opposite error direction.
 

mikegg

Golden Member
Jan 30, 2010
1,813
445
136
senttoschool is quite the expert in hysterics.
Thanks man! I don't even know you though. Glad I'm famous.
Yeah it's the same person, apparently he changed his account name. There's a few people who have done it for whatever reason. Unless there's some other way to tell that I'm not aware of you'd have to go back and look at an old post where someone replied to them as name in any quoted text from a post doesn't update with the name change on the account.
I'm. Never hiding it. I can be found in many hardware forums and reddit. But I'm not sure why your post is relevant to this thread. Let's focus on the topic instead.
 
Last edited:
Reactions: Orfosaurio

mikegg

Golden Member
Jan 30, 2010
1,813
445
136
Re: Geekbench vs. spec. Did you ever see this 2020 article on Geekbench 5?

Performance Delivered a New Way Part 2 — Geekbench versus SPEC

Author: Ram Srinivasan, Performance Architect, NUVIA

See this post 2 years ago by me on Reddit:
Andrei F of Anandtech chimed in.

Since that post, the internet has changed its tune on Geekbench and Cinebench. Finally, people are getting out of the AMD marketing fueled era of using Cinebench R23 as a proxy for general CPU performance. It has its niche but needs to remain a niche.

AMD and Intel need to stop optimizing their DIY CPUs for Cinebench. Usually, their first marketing slide is Cinebench. This is very misleading since Cinebench has very poor correlation to actual CPU performance in most applications.
 
Last edited:

mikegg

Golden Member
Jan 30, 2010
1,813
445
136
Agreed. You can use pretty much any CPU benchmark and it will show very strong correlation. I tried it with Cinebench r20 from the 2700x to the 7950x and Anandtech's SPECint numbers and got 0.998 for the correlation coefficient. You know what that tells me? Faster CPUs perform better on CPU benchmarks. Not a very surprising result, but that doesn't mean that they track closely enough that you can substitute one for the other, especially as you get to larger data sets.
I'd like to see your data on Cinebench vs SPECInt.

Cinebench R20 results are quite terrible at predicting gaming performance, for example.

 
Reactions: Orfosaurio

branch_suggestion

Senior member
Aug 4, 2023
240
501
96
Since that post, the internet has changed its tune on Geekbench and Cinebench. Finally, people are getting out of the AMD marketing fueled era of using Cinebench R23 as a proxy for general CPU performance. It has its niche but needs to remain a niche.
As is well known, Cinebench does have bottlenecks that do not represent modern CPU trends or real world software developments. But it is very accessible and consistent and it is a sustained perf bench, not a burst perf bench like Geekbench.
Neither in their current forms represent good PC performance correlation, unlike SPEC. Geekbench can be cheesed in too many ways, with OS's and the like showing rather different results.
AMD used Cinebench when it suited them (Z2/3), Intel did when it suited them (ADL/RPL). ARM companies have always leaned on Geekbench because it suits them.
SPEC is the gold standard for subtest heavy sustained perf, and for rendering Blender does utilise modern CPU instructions and other capabilities very well.
AMD and Intel need to stop optimizing their DIY CPUs for Cinebench.
They really don't, it is just easy for the GP to understand. SPEC is numero uno and then specific stuff like Javascript or gaming.
Usually, their first marketing slide is Cinebench. This is very misleading since Cinebench has very poor correlation to actual CPU performance in most applications.
Well things have changed since 2020, AMD is strong in SPEC and many real world apps but specifically weak in both Cinebench and Geekbench.
Z5 is a very different beast compared to previous Zen cores, so how it does relative to Z4 and the current comp in strong/weak apps is complete guesswork.
I'd like to see your data on Cinebench vs SPECInt.

Cinebench R20 results are quite terrible at predicting gaming performance, for example.

View attachment 99053
That is false equivalence and you know it. Comparing Cinebench results to Blender/V-Ray etc sure. But not to gaming, they are very different workloads.
Memory performance largely kept weaker Intel cores competitive in gaming vs stronger AMD cores, which is why V-Cache became a thing for client. It was originally only for technical DC but someone played around with it for client workloads and it worked out really well.
 

Nothingness

Platinum Member
Jul 3, 2013
2,553
988
136
Neither in their current forms represent good PC performance correlation, unlike SPEC. Geekbench can be cheesed in too many ways, with OS's and the like showing rather different results.
Are you sure that Geekbench still varies a lot depending on the OS? Picking some random results of Linux vs Windows, I see small differences; that was not the case for GB5 IIRC where some tests depended too much on libraries (libm mostly).

BTW SPEC is not only a CPU benchmark, but also a compiler benchmark, and there you can have huge differences (even when compilers don't cheat). SPEC also depends on some extra libraries (many use jemalloc) and OS feaures (Linux THP, rebooting the machine to have the OS in an as clean as possible state). But I agree it's a better benchmark than Geekbench (especially when one uses the same compiler to compare different CPUs).
 
Reactions: lightmanek

mikegg

Golden Member
Jan 30, 2010
1,813
445
136
As is well known, Cinebench does have bottlenecks that do not represent modern CPU trends or real world software developments. But it is very accessible and consistent and it is a sustained perf bench, not a burst perf bench like Geekbench.
Neither in their current forms represent good PC performance correlation, unlike SPEC. Geekbench can be cheesed in too many ways, with OS's and the like showing rather different results.
AMD used Cinebench when it suited them (Z2/3), Intel did when it suited them (ADL/RPL). ARM companies have always leaned on Geekbench because it suits them.
SPEC is the gold standard for subtest heavy sustained perf, and for rendering Blender does utilise modern CPU instructions and other capabilities very well.
You should read my original Reddit post. Everything you wrote here has either been argued or agreed upon in the Reddit post already.

They really don't, it is just easy for the GP to understand. SPEC is numero uno and then specific stuff like Javascript or gaming.
I'm certain that they do for their consumer CPUs. It's the #1 benchmark for AMD and Intel and unfortunately, Cinebench dominates the minshare space for CPU benchmarking. Things have turned a bit more since I made the Reddit post a few years ago. Quite often, you'll see people on the internet actually cite Nuvia's medium post (for example, @Eug cited it just yesterday) as well as bring up the fact that Cinebench R23 is optimized for x86 and not ARM when comparing between the two architectures.

That is false equivalence and you know it. Comparing Cinebench results to Blender/V-Ray etc sure. But not to gaming, they are very different workloads.
Memory performance largely kept weaker Intel cores competitive in gaming vs stronger AMD cores, which is why V-Cache became a thing for client. It was originally only for technical DC but someone played around with it for client workloads and it worked out really well.
The person I quoted claims that Cinebench R20 scores are a close predictor of SPEC scores. I'd like to see the data.

You can barely compare Cinebench results to Blender. They don't correlate as much as you think.

I'm not sure how a rendering benchmark became THE mainstream CPU benchmark over the last 6 or 7 years. Most people who use Cinema4D would use a GPU renderer. Same for Blender.

I guess people need a CPU benchmark to justify buying 16 or more consumer CPUs.
 
Last edited:
Jul 27, 2020
17,159
11,022
106
I guess people need a CPU benchmark to justify buying 16 or more consumer CPUs.
I think the percentage of people who really need more than 8C/16T cores is positively dwarfed by the number of people buying more than that due to other reasons, not limited to "I bet I might need this many cores at some point so better make sure I have them available!".
 

SarahKerrigan

Senior member
Oct 12, 2014
475
924
136
Are you sure that Geekbench still varies a lot depending on the OS? Picking some random results of Linux vs Windows, I see small differences; that was not the case for GB5 IIRC where some tests depended too much on libraries (libm mostly).

BTW SPEC is not only a CPU benchmark, but also a compiler benchmark, and there you can have huge differences (even when compilers don't cheat). SPEC also depends on some extra libraries (many use jemalloc) and OS feaures (Linux THP, rebooting the machine to have the OS in an as clean as possible state). But I agree it's a better benchmark than Geekbench (especially when one uses the same compiler to compare different CPUs).

Sure, SPEC is a compiler benchmark to a degree, but that's not a bad thing. It means you can build it in whatever configuration approximates what the software you'll be shipping or running will use - rather than having to just roll with whatever Primate Labs gives you. That's valuable! Vendor submissions are free to use trick compilers in whatever bizarre configuration they want, but then you can just run it yourself at O3 with LTO and no autopar, and see something realistic.

I've never seen significant improvements from jemalloc on SPECint in my own use (though if I recall, a couple of SPECfp tests benefit from it.) I have notes saying that omnet and leela break with jemalloc, though I haven't tested that in a few years and it's possible it's no longer the case.
 

Jan Olšan

Senior member
Jan 12, 2017
287
333
136
Where is the outrage now?

As far as Zen 4, GB6 was largely ignored/minimized when evaluating Zen 4.
That should be due to AVX-512 I suppose? That's just general SIMD though, you can compile pretty much anything with AVX-512 support and at least in theory, almost all codebases could benefit in small ways at least (standard functions like MEMCPY for example). It might not be significant in practice, or it could slow CPU down if it causes downclock like on Intel, but it's just plain SIMD - a standard feature of CPU core. As far as I understand, SME has narrower applicability. Not sure how much narrower exactly, but narrower.

(Also there was no update of the benchmark with new code specifically to make use of the feature for Zen 4, but I guess let's not go there)
 

Hitman928

Diamond Member
Apr 15, 2012
5,423
8,330
136
I'd like to see your data on Cinebench vs SPECInt.

Cinebench R20 results are quite terrible at predicting gaming performance, for example.

View attachment 99053

Here you go. I've made it in the same format as in the link. Note that I quoted the correlation coefficient in my previous post but am showing the coefficient of determination (R^2) in the graph to match the link.

CB r20 1t​
SPECint 2017 1t​
2700x​
433​
4.64​
3950x​
527​
5.78​
5950x​
644​
7.65​
7950x​
796​
9.39​


 

Hitman928

Diamond Member
Apr 15, 2012
5,423
8,330
136
Another graph with some Intel CPUs thrown in. I got rid of the connecting lines to better see the trend line. Still near perfect correlation. Just don't try to use this to predict ADL/RPL SPEC scores, it won't work out very well, just like with GB.

CB r20 1t​
SPECint 2017 1t​
2700x​
433​
4.64​
9900ks​
517​
6.1​
3950x​
527​
5.78​
11900k​
626​
7.346​
5950x​
644​
7.65​
7950x​
796​
9.39​

 

Doug S

Platinum Member
Feb 8, 2020
2,370
3,787
136
Sure, SPEC is a compiler benchmark to a degree, but that's not a bad thing. It means you can build it in whatever configuration approximates what the software you'll be shipping or running will use - rather than having to just roll with whatever Primate Labs gives you. That's valuable! Vendor submissions are free to use trick compilers in whatever bizarre configuration they want, but then you can just run it yourself at O3 with LTO and no autopar, and see something realistic.

I've never seen significant improvements from jemalloc on SPECint in my own use (though if I recall, a couple of SPECfp tests benefit from it.) I have notes saying that omnet and leela break with jemalloc, though I haven't tested that in a few years and it's possible it's no longer the case.


The thing I like about SPEC involving the compiler is that everyone starts from the same source code. If you have fancy SIMD instructions, fancy SME type instructions, whatever, the compiler has to be able to on its own translate that source code into those instructions. Yes that has been abused in the past with compilers written to essentially "recognize" SPEC like code from a specific test. But that's again why I go back to looking specifically at gcc/llvm/clang compiler benchmarks because those can't be gamed. There is no dominant inner loop in a compiler that can be gamed to result in a massive speedup. Even my pet peeve with SPEC around autopar is not an issue in the gcc benchmark.

All the gaming of SPEC involved vendors reporting results using their own compiler. Maybe that still goes on but I haven't looked at results on SPEC's site for a while now. I'm more interested in stuff that doesn't qualify for posting under their run rules, like when Andrei was running SPEC in reviews here, or geekerwan's work. They are trying to be fair so they're using the same compiler everywhere (or using the "standard" compiler for each platform, i.e. Xcode/llvm on Mac, MSVC/C++ on Windows, gcc or llvm on Linux) and not using goofy non standard heap libraries.
 
sale-70-410-exam    | Exam-200-125-pdf    | we-sale-70-410-exam    | hot-sale-70-410-exam    | Latest-exam-700-603-Dumps    | Dumps-98-363-exams-date    | Certs-200-125-date    | Dumps-300-075-exams-date    | hot-sale-book-C8010-726-book    | Hot-Sale-200-310-Exam    | Exam-Description-200-310-dumps?    | hot-sale-book-200-125-book    | Latest-Updated-300-209-Exam    | Dumps-210-260-exams-date    | Download-200-125-Exam-PDF    | Exam-Description-300-101-dumps    | Certs-300-101-date    | Hot-Sale-300-075-Exam    | Latest-exam-200-125-Dumps    | Exam-Description-200-125-dumps    | Latest-Updated-300-075-Exam    | hot-sale-book-210-260-book    | Dumps-200-901-exams-date    | Certs-200-901-date    | Latest-exam-1Z0-062-Dumps    | Hot-Sale-1Z0-062-Exam    | Certs-CSSLP-date    | 100%-Pass-70-383-Exams    | Latest-JN0-360-real-exam-questions    | 100%-Pass-4A0-100-Real-Exam-Questions    | Dumps-300-135-exams-date    | Passed-200-105-Tech-Exams    | Latest-Updated-200-310-Exam    | Download-300-070-Exam-PDF    | Hot-Sale-JN0-360-Exam    | 100%-Pass-JN0-360-Exams    | 100%-Pass-JN0-360-Real-Exam-Questions    | Dumps-JN0-360-exams-date    | Exam-Description-1Z0-876-dumps    | Latest-exam-1Z0-876-Dumps    | Dumps-HPE0-Y53-exams-date    | 2017-Latest-HPE0-Y53-Exam    | 100%-Pass-HPE0-Y53-Real-Exam-Questions    | Pass-4A0-100-Exam    | Latest-4A0-100-Questions    | Dumps-98-365-exams-date    | 2017-Latest-98-365-Exam    | 100%-Pass-VCS-254-Exams    | 2017-Latest-VCS-273-Exam    | Dumps-200-355-exams-date    | 2017-Latest-300-320-Exam    | Pass-300-101-Exam    | 100%-Pass-300-115-Exams    |
http://www.portvapes.co.uk/    | http://www.portvapes.co.uk/    |