Discussion Intel Meteor, Arrow, Lunar & Panther Lakes Discussion Threads

Page 782 - Seeking answers? Join the AnandTech community: where nearly half-a-million members share solutions and discuss the latest tech.

Tigerick

Senior member
Apr 1, 2022
734
689
106






As Hot Chips 34 starting this week, Intel will unveil technical information of upcoming Meteor Lake (MTL) and Arrow Lake (ARL), new generation platform after Raptor Lake. Both MTL and ARL represent new direction which Intel will move to multiple chiplets and combine as one SoC platform.

MTL also represents new compute tile that based on Intel 4 process which is based on EUV lithography, a first from Intel. Intel expects to ship MTL mobile SoC in 2023.

ARL will come after MTL so Intel should be shipping it in 2024, that is what Intel roadmap is telling us. ARL compute tile will be manufactured by Intel 20A process, a first from Intel to use GAA transistors called RibbonFET.



Comparison of upcoming Intel's U-series CPU: Core Ultra 100U, Lunar Lake and Panther Lake

ModelCode-NameDateTDPNodeTilesMain TileCPULP E-CoreLLCGPUXe-cores
Core Ultra 100UMeteor LakeQ4 202315 - 57 WIntel 4 + N5 + N64tCPU2P + 8E212 MBIntel Graphics4
?Lunar LakeQ4 202417 - 30 WN3B + N62CPU + GPU & IMC4P + 4E012 MBArc8
?Panther LakeQ1 2026 ??Intel 18A + N3E3CPU + MC4P + 8E4?Arc12



Comparison of die size of Each Tile of Meteor Lake, Arrow Lake, Lunar Lake and Panther Lake

Meteor LakeArrow Lake (N3B)Lunar LakePanther Lake
PlatformMobile H/U OnlyDesktop & Mobile H&HXMobile U OnlyMobile H
Process NodeIntel 4TSMC N3BTSMC N3BIntel 18A
DateQ4 2023Desktop-Q4-2024
H&HX-Q1-2025
Q4 2024Q1 2026 ?
Full Die6P + 8P8P + 16E4P + 4E4P + 8E
LLC24 MB36 MB ?12 MB?
tCPU66.48
tGPU44.45
SoC96.77
IOE44.45
Total252.15



Intel Core Ultra 100 - Meteor Lake



As mentioned by Tomshardware, TSMC will manufacture the I/O, SoC, and GPU tiles. That means Intel will manufacture only the CPU and Foveros tiles. (Notably, Intel calls the I/O tile an 'I/O Expander,' hence the IOE moniker.)



 

Attachments

  • PantherLake.png
    283.5 KB · Views: 24,023
  • LNL.png
    881.8 KB · Views: 25,511
Last edited:

511

Golden Member
Jul 12, 2024
1,993
1,753
106
Well if you have a leading edge fab business Iike Intel and have outsourced 30% of your wafers you will bleed money AMD gave up on their Fab biz and that was a good decision for them.
Not to mention if your prior nodes are very expensive and they are the majority of volume they just have problems lined up.
 
Last edited:

MS_AT

Senior member
Jul 15, 2024
616
1,265
96
I have been thinking that code compilation might be one of those tasks as you have pointed out. I have some non-trivial C++ projects as well, but honestly, even those compile pretty quickly .... even when I force a "Build All".
By build all, do you mean you do a clean rebuild or just build everything? As most of sane build systems are leaning heavily on the incremental builds ensuring that you don't rebuild anything that has not changed so in best case scenario only the cpp file that was modified It has then to be then linked once again but depending on the size of the project and linker you use might give different meaning to "pretty quickly"

As I said it's very project specific so it's hard to compare unless we put here much more details or refer to some sort of open source project that can be used as a benchmark.

In general the compilation scales pretty well with number of cores until it doesn't and I don't want to take the thread off-topic by discussing everything that can slow you down, what can be done to optimize the build time, why some of these tricks cannot be universally applied etc.
 

OneEng2

Senior member
Sep 19, 2022
539
766
106
The reason developers should get the great machines is they run in debug mode. Optimization takes a back seat. If it runs ok on an above average machine, it should run fine on a lesser machine in release build. Running a profiler is not super taxing but if you're doing it over and over those minutes or even few seconds add up.

I probably wouldn't notice a lesser machine for every day compilation. Even incremental builds on large projects wouldn't be horrible. I'd rather spend the extra few bucks for a good processor so when I evaluate a large project or do updates, it takes less time.

The simple truth is. You can dumb down a faster machine (limit threads and such) but you can't smart up a slower one. A developer machine should at least meet the best metrics a program is being designed for and probably exceed it for good measure.
While there is some wisdom to making developers use an average machine (to avoid the "it works find on my machine" syndrome ), the loss of productivity will always drive managers to purchase the best machines (laughably today that means a laptop) for developers to compile on.
By build all, do you mean you do a clean rebuild or just build everything? As most of sane build systems are leaning heavily on the incremental builds ensuring that you don't rebuild anything that has not changed so in best case scenario only the cpp file that was modified It has then to be then linked once again but depending on the size of the project and linker you use might give different meaning to "pretty quickly"

As I said it's very project specific so it's hard to compare unless we put here much more details or refer to some sort of open source project that can be used as a benchmark.

In general the compilation scales pretty well with number of cores until it doesn't and I don't want to take the thread off-topic by discussing everything that can slow you down, what can be done to optimize the build time, why some of these tricks cannot be universally applied etc.
Yea, I meant everything regardless of if it has been touched or not. Generally I only do this on the build machine for a final release step, and this is done using the command line interface and a build script vs the IDE.

I don't think that it is off topic though. Even this use case runs into scalability issues past a certain number of cores. My thought is that the IO becomes the bottleneck in and out of the disk system. Again, a workstation would likely be a better option than a high core count desktop.

The question I am wondering about is are there ENOUGH use cases where a 52 core desktop would be worth the silicon to the OEM and the price tag to the user where the use case would not drive the user to a workstation instead?
 

reb0rn

Senior member
Dec 31, 2009
280
90
101
Everything other than games I have run that use 16 threads can scale to more is it compiling, some calculations, crypto, encryption, encoding.... there is maybe some that need more optimization but so far in my limited use i seen none
 

Schmide

Diamond Member
Mar 7, 2002
5,688
921
126
While there is some wisdom to making developers use an average machine (to avoid the "it works find on my machine" syndrome ), the loss of productivity will always drive managers to purchase the best machines (laughably today that means a laptop) for developers to compile on.

Another thing that justifies above average machines in the developers hands.

Emulation, or moreover platform replication. There are projects were you will have to run your own server, database, or other asset often in an unoptimized state. You will design it, test it, break it, reload it over and over all on one machine.

Though now that I think of it. Developers just need two machines. (Recurse till all the machines are mine)
 
Reactions: OneEng2

Hitman928

Diamond Member
Apr 15, 2012
6,617
12,144
136
Everything other than games I have run that use 16 threads can scale to more is it compiling, some calculations, crypto, encryption, encoding.... there is maybe some that need more optimization but so far in my limited use i seen none

Many compilations won’t scale that high. Video encoding won’t. I won’t say all because I haven’t tested them all, but many encryption algorithms won’t scale like that either.
 

OneEng2

Senior member
Sep 19, 2022
539
766
106
Many compilations won’t scale that high. Video encoding won’t. I won’t say all because I haven’t tested them all, but many encryption algorithms won’t scale like that either.
Yea, I don't know the actual number of real world applications that do scale to higher than 16c/32t, but my gut feeling is that most of them that DO are likely good candidates for a workstation vs high core desktop.
 

reb0rn

Senior member
Dec 31, 2009
280
90
101
Why would I pay workstation price if I can get same or almost same for 40%

@Hitman928 Maybe some do not scale but my user case is not limited to one app, my 9 PC are heavy loaded and if the price is right I would rather have 5 PC with 64 core at decent price
more so most multithreaded app just need minor tweak to scale other could be limited by ram or nvme that not the same
 

OneEng2

Senior member
Sep 19, 2022
539
766
106
Why would I pay workstation price if I can get same or almost same for 40%
I am speculating that you can't get "almost the same" in most apps without having the extra bandwidth that the workstation multiple memory channels gives you.

Additionally, I am speculating that for the kinds of applications that DO scale, many will be the kinds of work where the people doing it will be very happy to pay for a real workstation for the added productivity.

We will see next year. If Intel launches a 52 core part in H1 2026, we will see if it sells.... and what price it sells at ...... and how well practical applications scale.
 

reb0rn

Senior member
Dec 31, 2009
280
90
101
It will sell as intel need to make it work so price will be very competitive, its other unknown is how well new process will be, I am mostly interested perf/watt in multithreaded use + avx10

We know we lost any hope that any new intel node just do not compute and have a lots issue, they still have dozen tech that are lead edge and they sure need some luck and to get their fab in order
 

coercitiv

Diamond Member
Jan 24, 2014
7,134
16,554
136
Meanwhile Intel shaved $100 from the price of Ultra 7:

This is what happens when you have extra cores but not the consistent ST performance uplift that users were expecting.
 

coercitiv

Diamond Member
Jan 24, 2014
7,134
16,554
136
At this moment, do we have a reliable / fresh source of information on whether the MC is on the compute tiles or the SoC tile? I remember a while ago there was talk about it moving to the compute tile, but in the light of this dual compute tile SKU I find it hard to believe it. Quad channel RAM and distributed memory controller sounds like a very complex and expensive solution for a niche consumer product. Makes very little sense to me, unless this is meant for HEDT / workstation and not consumer.
 

eek2121

Diamond Member
Aug 2, 2005
3,330
4,894
136
Didn't work for AMD. ST performance was more important to more people.
It absolutely did work for AMD. Ryzen trailed in single core performance for Zen, Zen+, and Zen 2, while leading in core counts. Zen only became a single core beast with Zen 3 and X3D.
The same will apply to this 52c NVL-S, when compared to an optimized design it will sacrifice gaming perf for productivity perf. Mem controller stays on SoC tile for a start. The tiles are identical 8+16, in order for a dual tile to perform well in gaming it would need to be exclusive P tile and exclusive E tile. This would also increase MT perf since resulting core count would be something like 12P+40E due to distribution on somewhat identically sized tiles. The obvious problem with asymmetrical tiles would be design cost (financial, manpower, time to market). Ironically AMD is in a better position to execute such a setup with a 12+24 chip but I really doubt they'll do it until Intel has something on the shelves that challenges their 3D cache setup.
AMD’s memory controller is also on the SoC (IO die). Intel is working on stacked cache. Just because you add more cores doesn’t mean single core performance has to suffer. AMD could drop a 64-96 core Threadripper part that hits 5.7ghz if they wanted. They don’t due to targeting the pro/workstation market. We do get 5.5ghz parts, however.
 
Reactions: lightmanek
sale-70-410-exam    | Exam-200-125-pdf    | we-sale-70-410-exam    | hot-sale-70-410-exam    | Latest-exam-700-603-Dumps    | Dumps-98-363-exams-date    | Certs-200-125-date    | Dumps-300-075-exams-date    | hot-sale-book-C8010-726-book    | Hot-Sale-200-310-Exam    | Exam-Description-200-310-dumps?    | hot-sale-book-200-125-book    | Latest-Updated-300-209-Exam    | Dumps-210-260-exams-date    | Download-200-125-Exam-PDF    | Exam-Description-300-101-dumps    | Certs-300-101-date    | Hot-Sale-300-075-Exam    | Latest-exam-200-125-Dumps    | Exam-Description-200-125-dumps    | Latest-Updated-300-075-Exam    | hot-sale-book-210-260-book    | Dumps-200-901-exams-date    | Certs-200-901-date    | Latest-exam-1Z0-062-Dumps    | Hot-Sale-1Z0-062-Exam    | Certs-CSSLP-date    | 100%-Pass-70-383-Exams    | Latest-JN0-360-real-exam-questions    | 100%-Pass-4A0-100-Real-Exam-Questions    | Dumps-300-135-exams-date    | Passed-200-105-Tech-Exams    | Latest-Updated-200-310-Exam    | Download-300-070-Exam-PDF    | Hot-Sale-JN0-360-Exam    | 100%-Pass-JN0-360-Exams    | 100%-Pass-JN0-360-Real-Exam-Questions    | Dumps-JN0-360-exams-date    | Exam-Description-1Z0-876-dumps    | Latest-exam-1Z0-876-Dumps    | Dumps-HPE0-Y53-exams-date    | 2017-Latest-HPE0-Y53-Exam    | 100%-Pass-HPE0-Y53-Real-Exam-Questions    | Pass-4A0-100-Exam    | Latest-4A0-100-Questions    | Dumps-98-365-exams-date    | 2017-Latest-98-365-Exam    | 100%-Pass-VCS-254-Exams    | 2017-Latest-VCS-273-Exam    | Dumps-200-355-exams-date    | 2017-Latest-300-320-Exam    | Pass-300-101-Exam    | 100%-Pass-300-115-Exams    |
http://www.portvapes.co.uk/    | http://www.portvapes.co.uk/    |