Discussion Apple Silicon SoC thread

Page 44 - Seeking answers? Join the AnandTech community: where nearly half-a-million members share solutions and discuss the latest tech.

Eug

Lifer
Mar 11, 2000
23,680
1,133
126
M1
5 nm
Unified memory architecture - LP-DDR4
16 billion transistors

8-core CPU

4 high-performance cores
192 KB instruction cache
128 KB data cache
Shared 12 MB L2 cache

4 high-efficiency cores
128 KB instruction cache
64 KB data cache
Shared 4 MB L2 cache
(Apple claims the 4 high-effiency cores alone perform like a dual-core Intel MacBook Air)

8-core iGPU (but there is a 7-core variant, likely with one inactive core)
128 execution units
Up to 24576 concurrent threads
2.6 Teraflops
82 Gigatexels/s
41 gigapixels/s

16-core neural engine
Secure Enclave
USB 4

Products:
$999 ($899 edu) 13" MacBook Air (fanless) - 18 hour video playback battery life
$699 Mac mini (with fan)
$1299 ($1199 edu) 13" MacBook Pro (with fan) - 20 hour video playback battery life

Memory options 8 GB and 16 GB. No 32 GB option (unless you go Intel).

It should be noted that the M1 chip in these three Macs is the same (aside from GPU core number). Basically, Apple is taking the same approach which these chips as they do the iPhones and iPads. Just one SKU (excluding the X variants), which is the same across all iDevices (aside from maybe slight clock speed differences occasionally).

EDIT:



M1 Pro 8-core CPU (6+2), 14-core GPU
M1 Pro 10-core CPU (8+2), 14-core GPU
M1 Pro 10-core CPU (8+2), 16-core GPU
M1 Max 10-core CPU (8+2), 24-core GPU
M1 Max 10-core CPU (8+2), 32-core GPU

M1 Pro and M1 Max discussion here:


M1 Ultra discussion here:


M2 discussion here:


Second Generation 5 nm
Unified memory architecture - LPDDR5, up to 24 GB and 100 GB/s
20 billion transistors

8-core CPU

4 high-performance cores
192 KB instruction cache
128 KB data cache
Shared 16 MB L2 cache

4 high-efficiency cores
128 KB instruction cache
64 KB data cache
Shared 4 MB L2 cache

10-core iGPU (but there is an 8-core variant)
3.6 Teraflops

16-core neural engine
Secure Enclave
USB 4

Hardware acceleration for 8K h.264, h.264, ProRes

M3 Family discussion here:


M4 Family discussion here:

 
Last edited:

amrnuke

Golden Member
Apr 24, 2019
1,181
1,772
136
It's the same thing. I didn't say anything different.

M1 CPU is the fastest, best performing, best overall, speediest laptop CPU out there. Period. Not controversial.

M1 chip (entire SoC) is the fastest, best performing, best overall, speediest laptop chip out there. Period. Not controversial.

Is it more clear to you now?

Edit: Added "period" and "not controversial" since you seem to get caught up when I either write it or not write it.
I mean... this is just not true. Also, I'm not the one who got caught up on the period. But I'd not expect you to realize that, since you seem to completely not address all the points I brought up.

Meh, it's not worth talking with someone who's entering this in bad faith, someone who's not open to actually discussing the topic. You're just not worth it.
 

IvanKaramazov

Member
Jun 29, 2020
56
102
66
I feel like a lot of the arguments going on here atm are pretty pointless. So here's a little totally unrelated analysis and some handwavey speculation. Big wall of text, my apologies.

I feel like much of the focus on the M1 has been on its ST performance, probably because of splashy Geekbench and SPEC numbers (and less splashy but still impressive showings in Cinebench etc.). I think the focus on raw ST performance is a poor way to analyze Apple Silicon's strengths and weaknesses, for a few reasons...

1) M1 has a real but marginal ST lead over TGL in best showings, and comes out close in others, 2) M1 matches Desktop Zen 3 in some places and lags it in others, but 3) M1 does all this with a process node advantage. In the real-world I expect Apple will maintain a node or half-node advantage over its rivals for the foreseeable future due to its arrangement with TSMC, and thus may maintain cutting edge ST performance at any point in time. But in a world where everyone was on the same node, as a thought experiment, I expect Intel's mobile chips would match or slightly outperform M1 in ST, and the Zen 3 mobile parts would likely match it as well. It is possible (though we don't know), that M1's ST performance in a Macbook Pro and in an iMac will be necessarily identical due to voltage, and lag Rocket Lake / Zen 3 desktop due to an inability to clock up like the competition.

Apple Silicon's real edge is in efficiency. I don't mean that in the "Arm is for low power uses and x86 is for performance" way that gets tossed around. But Apple's wide and slow approach is achieving their competitive ST performance at unbelievably low power draw, and when high-performance workflows are increasingly massively parallel, that means that Apple probably has a multi-threaded advantage over its competitors even when accounting for process node.

Let's use Cinebench as a test case, explicitly because it's a workflow where the M1 appears at a disadvantage relative to other benchmarks.

- The M1 pulls a sustained 3.8w in ST and 15w in MT for its scores of ~ 1520 / 7800. That MT load engages both the 4 performance cores and the 4 efficiency cores; I don't know what the breakdown is there, but I expect the p-cores are around 13w. There doesn't, as far as I can tell, appear to be any big power spikes during Cinebench, unlike in the x86 world.

- The 4750u pulls a sustained 28w MT, and scores around 8080. It also hit a high of 30w, so no huge spikes.

- The 4800u is frankly hard to find both in real-life and in reliable reporting, but it seems that it pulls a sustained 34w MT for its scores ranging from 9500 to 10150 MT. That slight sustained increase in power draw over the other Renoir-u chips makes sense. The 4800u appears to be aggressively binned (and thus so rare) but also slightly more aggressively clocked to achieve its impressive performance uplift.

- Tiger Lake is almost not worth comparing, and I can't find Cinebench specific numbers, but it appears to pull 28w sustained but well upwards of 50w in spikes under load, and has an r23 MT score of 6260.

We'll know soon enough, but based on these sorts of numbers an 8+4 variant of the M1 should be right on target for a sustained "TDP" of around 28w (accounting for CPU load only, of course), which is not coincidentally the TDP of the Intel chips Apple has traditionally put in the upper-end 13" MBP. Of course, Intel has been running their chips a good bit hotter for extended periods before throttling back to that, whereas I'd anticipate an M1X running consistently at 28w on something like Cinebench, based on the behavior of the M1.

I think Apple was smart to start with the M1, and in chasses where their strengths are balanced between battery life and performance. An M1X in the 13" model would probably have better battery life than an Intel variant because of the e-cores, but much less definitively so. What it would have instead would be a significant performance advantage.

A fairly conservative guess at performance scaling from 4+4 to 8+4 is around a 60% MT improvement for most workloads. Which is to say, I would guess that an M1X at 28w sustained versus a 4800u at 34w sustained should deliver around 12500 MT Cinebench r23 v 10000ish. I anticipate Cezanne will even that up a bit (but again, only for the most aggressively binned parts, and at a slightly higher power draw).

TLDR: As an architecture, I expect that on the same node Apple Silicon can't match the ST performance of Intel and AMD's best examples. But they have a significant efficiency advantage, such that Apple's high-end Macs will almost certainly be the fastest laptop chips for MT by a wide mile, and therein will lie the architectural advantage of Apple Silicon going forward.

Thank you for coming to my TED talk.
 

Mopetar

Diamond Member
Jan 31, 2011
7,931
6,226
136
I think there are a few things to consider when discussing this chip and comparing it to traditional x86 options available.

While it's certainly fair game to use dedicated hardware to speed up some workload (at the end of the day I don't care how you get the performance if it's faster) it does mean that any future chips with more cores aren't going to see a massive improvement in performance unless they include more of that dedicated hardware and problems can be parallelized to run more of it. Performance improvements are limited to the extent to which additional (or faster) cores can speed up the overall task, which may be a small percentage in some cases.

Apple will only target markets or categories that apple cares about. Even though their GPU tech is rather impressive, particularly given the power envelope for it, I doubt we'll see Apple release anything that is designed to compete against even the upper mid-range parts like GA-104 and Navi 22. Performance will be great for an ultrabook or small form-factor desktop machine, but it won't be scaled to the high-end. The install base may not be good enough for developers to be willing to optimize titles for Apple's GPUs either.

Ability to pick and choose may be more limited due to the way Apple designs the hardware. If you really just want a system with more RAM, you might also need to buy one with the faster and more expensive chip as well. To some degree Apple already does this with their hardware using x86 CPUs where you need to buy a more expensive CPU in order to pair it with the best GPU offerings. Apple tends to value simplicity through limitation of choices so there may be fewer, more arbitrary product categories in their future products.
 

Qwertilot

Golden Member
Nov 28, 2013
1,604
257
126
Also the excellent iGPU etc. A lot of Apple's designs were ending up rather hamstrung by weak integrated graphics performance.

Putting a very solid baseline underneath the entire range is likely to be very useful.

Like with phones - a top device should really be at least very good at nearly everything.
 

Eug

Lifer
Mar 11, 2000
23,680
1,133
126
I think there are a few things to consider when discussing this chip and comparing it to traditional x86 options available.

While it's certainly fair game to use dedicated hardware to speed up some workload (at the end of the day I don't care how you get the performance if it's faster) it does mean that any future chips with more cores aren't going to see a massive improvement in performance unless they include more of that dedicated hardware and problems can be parallelized to run more of it. Performance improvements are limited to the extent to which additional (or faster) cores can speed up the overall task, which may be a small percentage in some cases.

Apple will only target markets or categories that apple cares about. Even though their GPU tech is rather impressive, particularly given the power envelope for it, I doubt we'll see Apple release anything that is designed to compete against even the upper mid-range parts like GA-104 and Navi 22. Performance will be great for an ultrabook or small form-factor desktop machine, but it won't be scaled to the high-end. The install base may not be good enough for developers to be willing to optimize titles for Apple's GPUs either.
I was thinking about this again last night actually. I’m now convinced that the higher end machines will include not just an upgraded GPU, but also a more robust image signal processor, etc. for their multimedia creator customer base.
 

Doug S

Platinum Member
Feb 8, 2020
2,315
3,651
136
People here are getting off track from what matters. The mere fact that people are arguing over how M1 compares to the fastest x86 CPUs shows that Apple achieved their goal: no one buying an M1 Mac is going to feel like they are giving up any performance versus if that same model Mac had the latest x86 CPU instead. Rosetta2 is fast enough that no one is going to worry they're making big sacrifices in the performance of their old x86 applications that haven't been ported to ARM yet.

The only possible objections for Apple customers buying an M1 Mac (and since I'm talking about "Apple customers" the Apple hater arguments about "too expensive" etc. are irrelevant here) are those who need to run a lot of Windows software, or those who need one of the higher end models Apple has yet to release.

When rumors about this move were discussed in the past a lot of people said it would destroy Apple because they'd have to give up too much performance. Even when after years of rumors it finally became official in June there were still many doubters. Now no one doubts the ability of Apple's ARM designs in a Mac to go toe to toe with the best x86 out there, even if they believe Apple is behind they are forced to acknowledge they aren't that far behind.

Erasing doubt was Apple's goal, to make their customers feel comfortable that the migration to ARM will be a win and not require some sacrifice or compromise on their part to stick with macOS. I think pretty much everyone would agree now that Apple's ARM transition will be a success. The only question left in my mind is will they eventually be able to win over enough Windows customers to get the Mac marketshare into double digits, or will the performance (especially performance/watt) advantages of Apple's ARM Macs not be seen as enough to overcome the "but it isn't Windows" resistance to change.
 

Eug

Lifer
Mar 11, 2000
23,680
1,133
126
People here are getting off track from what matters. The mere fact that people are arguing over how M1 compares to the fastest x86 CPUs shows that Apple achieved their goal: no one buying an M1 Mac is going to feel like they are giving up any performance versus if that same model Mac had the latest x86 CPU instead. Rosetta2 is fast enough that no one is going to worry they're making big sacrifices in the performance of their old x86 applications that haven't been ported to ARM yet.

The only possible objections for Apple customers buying an M1 Mac (and since I'm talking about "Apple customers" the Apple hater arguments about "too expensive" etc. are irrelevant here) are those who need to run a lot of Windows software, or those who need one of the higher end models Apple has yet to release.

When rumors about this move were discussed in the past a lot of people said it would destroy Apple because they'd have to give up too much performance. Even when after years of rumors it finally became official in June there were still many doubters. Now no one doubts the ability of Apple's ARM designs in a Mac to go toe to toe with the best x86 out there, even if they believe Apple is behind they are forced to acknowledge they aren't that far behind.

Erasing doubt was Apple's goal, to make their customers feel comfortable that the migration to ARM will be a win and not require some sacrifice or compromise on their part to stick with macOS. I think pretty much everyone would agree now that Apple's ARM transition will be a success. The only question left in my mind is will they eventually be able to win over enough Windows customers to get the Mac marketshare into double digits, or will the performance (especially performance/watt) advantages of Apple's ARM Macs not be seen as enough to overcome the "but it isn't Windows" resistance to change.
This is an important point, and tangentially one where my logic was off. I didn't think Apple would suffer, but as mentioned I said that Apple could go with 2+4 parts and still get decent enough performance for the low end to maintain sales, at least with mainstream users. While true for native apps, 2+4 might have sucked for Rosetta 2 apps, and this would have left a bad taste in some people's mouths.

While we are marvelling at M1's native speeds, what sometimes impresses me even more is just how fast it is in Rosetta 2. It beats older Intel hardware under Rosetta 2, so people may feel comfortable to upgrade even if they use more heavy duty apps that won't go native for a while yet.

OTOH, on the iPad this was not an issue. They can get away with 2+4 iPads because everything on those is native already. In retrospect, I could imagine the hypothetical situation where people are running Intel applications under Rosetta on iPads, and that would suck. Even just surfing could suck in that situation.

A few posts above I mentioned that I think performance on an iPad 7 - A10 is decent, but it would likely absolutely suck under Rosetta. A14 with Rosetta would undoubtedly suck much less, but in many instances it wouldn't really be much of an improvement over existing lower end Macs, so some reviews might not be kind in that scenario.

Interesting.

As I suspected, M1 is not magic, and memory still really matters. There are claims out there that M1 just doesn't need 16 GB, but obviously that doesn't make much sense.* The other factor to note was that the swap was bigger on the 8 GB model. However, it just so happens that the SSD is very fast at >3 GB/s so people will notice the swap less.
I was watching a Blender review on an M1 machine with 8 GB RAM. Performance was very good even despite the fact it was running under Rosetta, but what caught my eye was the occasional beachball, although it would only last a few seconds.

Here is an example of a beachball. Note that the video has been sped up, so the beachball is longer than it seems in the video.


I'm not a Blender user so I don't know how prone Blender is to beachballing, but nonetheless I know that older Intel machines with slower SSDs tend to sit in the beachballing twilight zone quite a while with they do beachball, and the only solutions are either to close applications or else to use a machine with more memory. This seems to support my idea that part of the reason the 8 GB devices seem so fast is because they have 2X to 3X the SSD speed now at over 3 GB/s, as compared to older lower end Intel Macs.

Also, I note that in these reviews of pro applications when they pull up Activity Monitor, they almost always have a swap file, and sometimes those swaps are many GB too. Meanwhile, on older Intel machines with say 32 GB in reviews previously, IIRC most of them either have no swap files or swap files less than 1 GB in size.
 

Heartbreaker

Diamond Member
Apr 3, 2006
4,233
5,242
136
1) M1 has a real but marginal ST lead over TGL in best showings, and comes out close in others, 2) M1 matches Desktop Zen 3 in some places and lags it in others, but 3) M1 does all this with a process node advantage.

...delted wall of text.

TLDR: As an architecture, I expect that on the same node Apple Silicon can't match the ST performance of Intel and AMD's best examples. But they have a significant efficiency advantage, such that Apple's high-end Macs will almost certainly be the fastest laptop chips for MT by a wide mile, and therein will lie the architectural advantage of Apple Silicon going forward.

I think your giant wall of text missed a critical element that invalidates what you set up, and your final conclusion.

Sure, M1 scores a small single thread win with a "node advantage", But the only alternatives that came close, were running comparatively unconstrained power, more than double M1 power.

That process node advantage will mainly give a small power savings in the 20%-30% range at best.

When the competition also gets its 20-30% node improvement in power usage, they will still be behind, because they are behind by much more than the 20-30% node advantage on the element that node improvement impacts.
 

DrMrLordX

Lifer
Apr 27, 2000
21,706
10,980
136
TLDR: As an architecture, I expect that on the same node Apple Silicon can't match the ST performance of Intel and AMD's best examples.

We can probably confirm that once AMD comes out with Zen4, though AMD will have one more generation of development work under their hat so I'm not sure it's a fair comparison. If Apple had ever released a MacOS product with A13, we could probably confirm/deny your hypothesis.

Also observe what AMD was able to accomplish staying on the same node (7nm). Node advancements don't necessarily account for everything.

Erasing doubt was Apple's goal, to make their customers feel comfortable that the migration to ARM will be a win and not require some sacrifice or compromise on their part to stick with macOS.

For the select bits of software that people will probably use pretty often on their Macbooks, yeah, I think they've done pretty well. Edge case users may reach some level of frustration. I don't think we can really accurately guess how many will fall into which categories, but rest assured that Apple will continue to provide support for key developers to move their ARM-based MacOS platform forward. In the long run, developers who get little to no direct support will have to follow suit to continue making money.

The only question left in my mind is will they eventually be able to win over enough Windows customers to get the Mac marketshare into double digits

M1 will have to be followed up by something better than what's available. It isn't just Windows, its the lack of configurability and the prices. Woof. There's a reason many people just don't like Apple.
 

Doug S

Platinum Member
Feb 8, 2020
2,315
3,651
136
Edge case users may reach some level of frustration.

Name one "edge case" where you can demonstrate this "level of frustration" you are claiming. I have yet to see anything where the M1 Mac is slower the x86 Mac it is replacing other than something using hand optimized x86 SIMD code. Even when running under Rosetta2 few things are slower on the M1 than on the x86 Mac being replaced.

The hand optimized SIMD code case won't matter in six months, because once the developer ports and optimizes their application for the M1 it will easily beat the x86 SIMD version - perhaps by several hundred percent (as per Apple's claims) since they will not only have the M1's ARM cores but also have a GPU, NPU and IPU at their disposal.

Sure, you can point to benchmarks for recently released Windows laptops using top of the line CPUs not available in Mac laptops (and cost over half of what entire M1 Mac costs) that may beat it in some things, but like AMD performance that's not a valid comparison for a Mac user looking to replace their Mac with another Mac. Which is all that Apple is concerned about at this time.
 

amrnuke

Golden Member
Apr 24, 2019
1,181
1,772
136
Sure, M1 scores a small single thread win with a "node advantage", But the only alternatives that came close, were running comparatively unconstrained power, more than double M1 power.

That process node advantage will mainly give a small power savings in the 20%-30% range at best.

When the competition also gets its 20-30% node improvement in power usage, they will still be behind, because they are behind by much more than the 20-30% node advantage on the element that node improvement impacts.
It's not just a node advantage - the Zen2 core is also a 1.5 year old core. Zen2->3 was a 20-25% performance improvement (CB20 ST being 20-25% depending on specific Zen2 vs Zen3 comparison), and N7->N5 is a 30% efficiency improvement or 15% speed improvement at iso-power. That calculates to be 20% performance * 15% performance = 38% total performance gain at same TDP.

So we take a theoretical 5800U-XT, a 15W TDP Cezanne (Zen3 core) on 5nm. 4800U (15W) CB ST was 1199. +20% for Zen3, +15% for N7->N5, that's a 38% total improvement, which makes the CB20 ST score estimate 1654. M1 native was 1522. I mean, I think that's completely in the realm of possibility that a Zen3-powered 15W APU on 5nm could actually have faster ST performance than M1. (Heat be damned, anyway.)

The chief issue, of course, is that M1 is actually available. Cezanne has no release date, and Cezanne is on 7nm anyway, so it's all just pie-in-the-sky dreaming. Granted, it's do-able, as the Zen3 core exists and 5nm exists. But this whole conversation could go down a very long rabbit-hole of what-ifs that are pretty irrelevant right now because I can't buy a 5nm Zen3 laptop chip, and may never be able to.

As of now, we only have Renoir, TGL, and M1. And M1 clearly has better efficiency and better single/lightly threaded performance and competes well enough in MT stuff with the 4800U and 4900HS that we can reasonable conclude that an 8+4 M1X or whatever would actually be damn stellar.
 

LightningZ71

Golden Member
Mar 10, 2017
1,629
1,898
136
On the node advantage for M1 vs. the competition (N5 vs N7), realize that the node advantage also caters more to Apple's approach to chip design. For the M1, Apple has thrown a ton of circuit resources at their cores to enable them to extract as much IPC as possible from each of their fewer clocks. While the node advantage only offers 10-20% better power draw characteristics on two equivalent designs on the new and old node, the extra capacity to pack logic into their cores by making them wider, by having deeper buffers, by having larger internal low level caches (iCache, dCache), etc. It also helps them with secondary effects, by allowing them to maintain their use of denser libraries because of their lower clock targets. At lower clocks, N5 is allowing them to be very sense, enabling a 4+4 design on a very small piece of silicon.

To put that in to perspective: if M1 was implemented on N7, it would either be a much larger chip, physically, reducing the number of die per wafer greatly. This would adversely affect the per chip profit, and make the die plus ram package larger and more expensive. Alternately, they could have maintained the same die size, and per wafer chips, but it would have likely cost them two of the Firestorm cores and at least some of the L2 and SLC. Power draw in single core scenarios would be notably higher, and multi threaded benches would be much slower.

It is my opinion that the node advantage for Apple in this situation is a very big deal. I'm not criticising Apple here, as they paid the money and took the risk to be on that node, and it has paid off. Zen3 on N5, as it currently exists, wouldn't gain a whole lot. It would consume a few less watts. It could sustain higher MC boost clocks. It would be a smaller chip. That's about it. But Zen3 wasn't built with N5 in mind. It was built with the constraints of N7 in mind. Zen4 will be built for N5. It will have extra circuit budget for more CPU resources. It will be power optimized for it's targeted node.

But, it isn't here today. And that's the reality of the market.
 

Heartbreaker

Diamond Member
Apr 3, 2006
4,233
5,242
136
It's not just a node advantage - the Zen2 core is also a 1.5 year old core. Zen2->3 was a 20-25% performance improvement (CB20 ST being 20-25% depending on specific Zen2 vs Zen3 comparison), and N7->N5 is a 30% efficiency improvement or 15% speed improvement at iso-power. That calculates to be 20% performance * 15% performance = 38% total performance gain at same TDP.

Zen 3 is already here. It's matching ST performance of M1, using what? 3x-4x the power??

All zen three gains is power improvement which won't even put it in the ballpark. Naturally there will be design improvements as well, but Apple is probably not done improving their chips either.

People seem to be missing the scale of the power usage differences.

Anandtech is measuring the Whole iPad Mini at the Wall. The whole computer maxing out ST uses 15 Watts, including PSU losses.

AMD/Intel power numbers are package power being through sofware. Intel will use up to 50 watts in a single thread, package power, to match the ST performance of the Mini that uses 15 watts for the whole computer.
 
Reactions: Viknet

amrnuke

Golden Member
Apr 24, 2019
1,181
1,772
136
Zen 3 is already here. It's matching ST performance of M1, using what? 3x-4x the power??

All zen three gains is power improvement which won't even put it in the ballpark. Naturally there will be design improvements as well, but Apple is probably not done improving their chips either.

People seem to be missing the scale of the power usage differences.

Anandtech is measuring the Whole iPad Mini at the Wall. The whole computer maxing out ST uses 15 Watts, including PSU losses.

AMD/Intel power numbers are package power being through sofware. Intel will use up to 50 watts in a single thread, package power, to match the ST performance of the Mini that uses 15 watts for the whole computer.
Well, we don't know what the Zen3 power usage will be in Cezanne. It looked like Zen2 in Matisse was a big power sucker based on the per core power draw I believe you're referencing, right? But the Renoir chips running full-bore Furmark and Prime95 draw peak power draw (at the wall) of 37.2W - powering a screen as well (reference). The Mac mini pulled 31W in a compute MT load and the power of the monitor is not included. Did Andrei mention what his "compute MT workload" was actually measuring? I didn't see that. Granted, Furmark + Prime95 might not be as taxing as Andrei's compute MT workload. But it sure seems pretty taxing.

Zen3 doesn't only gain power improvement, we already know ST scores are 20-25% higher, IPC is 15-20% higher. To say it only gains power improvement seems wrong. Where did you get that information?

So, sure, single-thread workload uses 15W. But multithread workload uses 31W. And MT+GPU workload on Renoir draws 37.2W (yes, that's at the wall, using a Metrahit - not using the software-provided numbers).
 
Reactions: Tlh97 and Carfax83

thunng8

Member
Jan 8, 2013
153
61
101
Well, we don't know what the Zen3 power usage will be in Cezanne. It looked like Zen2 in Matisse was a big power sucker based on the per core power draw I believe you're referencing, right? But the Renoir chips running full-bore Furmark and Prime95 draw peak power draw (at the wall) of 37.2W - powering a screen as well (reference). The Mac mini pulled 31W in a compute MT load and the power of the monitor is not included. Did Andrei mention what his "compute MT workload" was actually measuring? I didn't see that. Granted, Furmark + Prime95 might not be as taxing as Andrei's compute MT workload. But it sure seems pretty taxing.

Zen3 doesn't only gain power improvement, we already know ST scores are 20-25% higher, IPC is 15-20% higher. To say it only gains power improvement seems wrong. Where did you get that information?

So, sure, single-thread workload uses 15W. But multithread workload uses 31W. And MT+GPU workload on Renoir draws 37.2W (yes, that's at the wall, using a Metrahit - not using the software-provided numbers).
Note that the 37.2W is using the 4750U, 4800U laptops can be much higher. eg. 57.6W here:

 

IvanKaramazov

Member
Jun 29, 2020
56
102
66
I think your giant wall of text missed a critical element that invalidates what you set up, and your final conclusion.

Sure, M1 scores a small single thread win with a "node advantage", But the only alternatives that came close, were running comparatively unconstrained power, more than double M1 power.

I actually agree with you here, my suggestion was that while the M1 might not match the others for raw ST performance (at hugely different power levels, obviously), the efficiency of the M1 is exactly what should set it up to scale so well to more cores and higher raw MT performance. In other words, my expectation is that if one ignores efficiency and ppw entirely, the Apple Silicon chips may not be the clear best in ST, but likely will be in MT once larger chips come out. I obviously could be wrong.
 

ultimatebob

Lifer
Jul 1, 2001
25,135
2,445
126
They asked Linus Torvalds if he was planning on buying an M1 Macbook Air, and he said no (as expected) because he cannot run Linux on it.

Something tells me that someone will get Linux working on this thing eventually. If you can run Linux on Nintendo Switch, getting it working on an ARM Macbook should be easy!
 

ultimatebob

Lifer
Jul 1, 2001
25,135
2,445
126
I'm not DrMrLordX, but I'm a Linux user. My level of frustration is not being able to replace MacOS with Linux even though I do like Apple's approach to hardware design most of the time.

If I got one of these M1 Macbook systems for job, it would royally piss me off. It can't run Docker (yet), nor can it run VMWare virtualization software. I also have a ton of oddball IT administration software that probably aren't compatible with Big Sur yet.

Of course, this laptop wasn't made for my workload. The 16" Macbook Pro was, though. Which is why I bought one

I found this site, which will (hopefully) tell me when M1 native versions of my applications are available:

 
Last edited:

Heartbreaker

Diamond Member
Apr 3, 2006
4,233
5,242
136
Something tells me that someone will get Linux working on this thing eventually. If you can run Linux on Nintendo Switch, getting it working on an ARM Macbook should be easy!

Switch runs on a standard NVidia Tegra chip, and NVidia actually provides "Linux for Tegra". It's not intended for switch but it would probably be a great starting place for all important drivers.

But for M1 Macs. Apple isn't providing Linux drivers, and M1 isn't based on HW that has Linux drivers, there isn't even much of stating place for drivers.

No doubt someday futher away, it will happen, but likely not in a very usable form.
 

amrnuke

Golden Member
Apr 24, 2019
1,181
1,772
136
Note that the 37.2W is using the 4750U, 4800U laptops can be much higher. eg. 57.6W here:

The example I gave is one of an entire laptop, screen included, under full load - including GPU - using less power (37.2W) than a 3700X under a 2-thread load (39.63W - reference). Which sufficiently makes my point that using the power consumption of a desktop chip to criticize the potential power consumption of its laptop companion is a fool's errand.

Non-sequitor, and perhaps straying off-topic from what I intended:

CB20 MT score of the HP EliteBook 4750U I referenced is 3264 which is actually better than Lenovo's Yoga Slim you linked that draws 57.6W, which only got 3213 points. (There is much to be said about manufacturer implementations.) [Begin speculation] Assuming that translates to CB23 MT, the 4750U would get roughly the same score as the 4800U, which is 9286. (Some 4750U implementations carry a TDP of 45W and can hit 9716, but I don't think it's fair to include that.)

If we assume that both the Mac Mini and the EliteBook in question both draw their full peak loads while running CB23 (again, speculation):

4750U CB23MT score of 9286 @ 37.2W = 249.6 CB23/W
M1 CB23MT score of 7833 @ 31W = 252.7 CB23/W
 
Last edited:
Reactions: Carfax83 and mikegg

thunng8

Member
Jan 8, 2013
153
61
101
The example I gave is one of an entire laptop, screen included, under full load - including GPU - using less power (37.2W) than a 3700X under a 2-thread load (39.63W - reference). Which sufficiently makes my point that using the power consumption of a desktop chip to criticize the potential power consumption of its laptop companion is a fool's errand.

Non-sequitor, and perhaps straying off-topic from what I intended:

CB20 MT score of the HP EliteBook 4750U I referenced is 3264 which is actually better than Lenovo's Yoga Slim you linked that draws 57.6W, which only got 3213 points. (There is much to be said about manufacturer implementations.) Assuming that translates to CB23 MT, the 4750U would get roughly the same score as the 4800U, which is 9286. (Some 4750U implementations carry a TDP of 45W and can hit 9716, but I don't think it's fair to include that.

If we assume that both the Mac Mini and the EliteBook in question both draw their full peak loads while running CB23 (which may be a bit unfair to the EliteBook since that peak power draw includes the GPU under Furmark, and powering the screen too):

4750U CB23MT score of 9286 @ 37.2W = 249.6 CB23/W
M1 CB23MT score of 7833 @ 31W = 252.7 CB23/W
M1 in a mini draws 15W for cinebench.


I can't find the video source since I watched so many videos, but the macbook air when thermally throttled (at the end of the 30min run) draws 7W for a score of ~5900.
 
Reactions: Heartbreaker

mikegg

Golden Member
Jan 30, 2010
1,809
423
136
The example I gave is one of an entire laptop, screen included, under full load - including GPU - using less power (37.2W) than a 3700X under a 2-thread load (39.63W - reference). Which sufficiently makes my point that using the power consumption of a desktop chip to criticize the potential power consumption of its laptop companion is a fool's errand.

Non-sequitor, and perhaps straying off-topic from what I intended:

CB20 MT score of the HP EliteBook 4750U I referenced is 3264 which is actually better than Lenovo's Yoga Slim you linked that draws 57.6W, which only got 3213 points. (There is much to be said about manufacturer implementations.) Assuming that translates to CB23 MT, the 4750U would get roughly the same score as the 4800U, which is 9286. (Some 4750U implementations carry a TDP of 45W and can hit 9716, but I don't think it's fair to include that.

If we assume that both the Mac Mini and the EliteBook in question both draw their full peak loads while running CB23 (which may be a bit unfair to the EliteBook since that peak power draw includes the GPU under Furmark, and powering the screen too):

4750U CB23MT score of 9286 @ 37.2W = 249.6 CB23/W
M1 CB23MT score of 7833 @ 31W = 252.7 CB23/W
This is so wrong on so many levels. I can't even...

A lot of AMD fanboys in denial here.



Use of the word fanboys is not allowed.


esquared
Anandtech Forum Director
 
Last edited by a moderator:

amrnuke

Golden Member
Apr 24, 2019
1,181
1,772
136
M1 in a mini draws 15W for cinebench.


I can't find the video source since I watched so many videos, but the macbook air when thermally throttled (at the end of the 30min run) draws 7W for a score of ~5900.
So now we're ok with reported power draw? In any case, that entire segment was a complete non-sequitor, as I mentioned, since we don't have actual power draws under CB for both chips, we aren't apparently using the same measurement system, and so on. Continuing down that path is silly, but I thought interesting enough to consider! I really would love to have more data to make a better comparison, but all we have right now is speculation when comparing with contemporary laptop chips.

Anyway, do you care to address my primary response?

"The example I gave is one of an entire laptop, screen included, under full load - including GPU - using less power (37.2W) than a 3700X under a 2-thread load (39.63W - reference). Which sufficiently makes my point that using the power consumption of a desktop chip to criticize the potential power consumption of its laptop companion is a fool's errand."
 
Last edited:
Reactions: Tlh97 and Carfax83
sale-70-410-exam    | Exam-200-125-pdf    | we-sale-70-410-exam    | hot-sale-70-410-exam    | Latest-exam-700-603-Dumps    | Dumps-98-363-exams-date    | Certs-200-125-date    | Dumps-300-075-exams-date    | hot-sale-book-C8010-726-book    | Hot-Sale-200-310-Exam    | Exam-Description-200-310-dumps?    | hot-sale-book-200-125-book    | Latest-Updated-300-209-Exam    | Dumps-210-260-exams-date    | Download-200-125-Exam-PDF    | Exam-Description-300-101-dumps    | Certs-300-101-date    | Hot-Sale-300-075-Exam    | Latest-exam-200-125-Dumps    | Exam-Description-200-125-dumps    | Latest-Updated-300-075-Exam    | hot-sale-book-210-260-book    | Dumps-200-901-exams-date    | Certs-200-901-date    | Latest-exam-1Z0-062-Dumps    | Hot-Sale-1Z0-062-Exam    | Certs-CSSLP-date    | 100%-Pass-70-383-Exams    | Latest-JN0-360-real-exam-questions    | 100%-Pass-4A0-100-Real-Exam-Questions    | Dumps-300-135-exams-date    | Passed-200-105-Tech-Exams    | Latest-Updated-200-310-Exam    | Download-300-070-Exam-PDF    | Hot-Sale-JN0-360-Exam    | 100%-Pass-JN0-360-Exams    | 100%-Pass-JN0-360-Real-Exam-Questions    | Dumps-JN0-360-exams-date    | Exam-Description-1Z0-876-dumps    | Latest-exam-1Z0-876-Dumps    | Dumps-HPE0-Y53-exams-date    | 2017-Latest-HPE0-Y53-Exam    | 100%-Pass-HPE0-Y53-Real-Exam-Questions    | Pass-4A0-100-Exam    | Latest-4A0-100-Questions    | Dumps-98-365-exams-date    | 2017-Latest-98-365-Exam    | 100%-Pass-VCS-254-Exams    | 2017-Latest-VCS-273-Exam    | Dumps-200-355-exams-date    | 2017-Latest-300-320-Exam    | Pass-300-101-Exam    | 100%-Pass-300-115-Exams    |
http://www.portvapes.co.uk/    | http://www.portvapes.co.uk/    |