Or maybe ArsTechnica forums…I’m guessing AnandTech forums.
Or maybe ArsTechnica forums…I’m guessing AnandTech forums.
Or maybe ArsTechnica forums…
Yeah, I’ve only ever seen it as Ars over 20+ years…Oh yeah, I never think about that abbreviated as AT since they already have Ars.
AMD's approach to chiplet based CPUs has high (idle) power draw, that's not what Apple wants. How is Intel's?Apple can probably avoid many of Intels and AMD's mistakes
My last PC certainly won the vertical drop award.The M2 Ultra is honestly embarrassing for a $4K desktop computer. You can get vastly more performance in a PC for that price. Apple Silicon was made for laptops and phones. There’d be no shame for Apple if they threw in the white flag and admitted that on desktop, where performance per watt is irrelevant, Apple silicon can‘t compete.
As far as I can tell AMD's high idle power draw is a thing of the past in Zen 4/5. My impression is that the main culprit was the bad GloFlo node the IO die was forced to use on Zen 3 and previous chips. Now they use TSMC N6 dies for IO. I've got this nagging feeling that there was another potential culprit as well, but regardless, high power draw at idle seems to have been solved.AMD's approach to chiplet based CPUs has high (idle) power draw, that's not what Apple wants. How is Intel's?
I'm coming from the Other side, and places like AT have serious, extensive discussion on this and other Ua and Litho technologies than I'm seeing in Apple forums.
If I am understanding the queries here, folks are wondering if Apple moving to tiles is going to somehow improve performance?
Generally, no.
Tiles are primarily an economic benefit via increased yield of wafers while defect densities may remain the same per nm/sq between tiles and monolithic full die designs.
A tile that has a defect only affects that tile/die area, small amount of silicon to be trashed.
A SOC/Full fat design monolithic with a similar defect affects multiples of that tile/die area, large amount of silicon to be trashed or binned down if sufficient redundancy has been included.
With Tiles being so small, and high yields may also make it possible to reduce the levels of redundancy they've learned to add from Monolithic die experience, and end up saving even more space and being even less expensive.
So far, I've only run across a handful of folks referenced on AT to X/Twitter posters who actually get into the nitty-gritty details doing comparisons across Apple and other micro-architectures and foundry lithographies.
If Apple is moving to tiles, then it is likely primarily because their monolithic dies are now becoming large enough that adding additional redundancy to overcome defect densities is becoming to high to maintain their cost basis.
Going with tiles is not a get out of jail free card, the downside is that moving from monolithic means you need inter-tile fabrics that are a can of worms on their own and so far have generally led to lower performance than monolithic dies in latency and throughput.
Can be alleviated with high-speed fabrics, and Apple can probably avoid many of Intels and AMD's mistakes with careful R&D. However I might be wrong but I'm not aware of anyone generally beating monolithics inherent latency/throughput advantage with tile use.
I'm coming from the Other side, and places like AT have serious, extensive discussion on this and other Ua and Litho technologies than I'm seeing in Apple forums.
If I am understanding the queries here, folks are wondering if Apple moving to tiles is going to somehow improve performance?
Generally, no.
Tiles are primarily an economic benefit via increased yield of wafers while defect densities may remain the same per nm/sq between tiles and monolithic full die designs.
A tile that has a defect only affects that tile/die area, small amount of silicon to be trashed.
A SOC/Full fat design monolithic with a similar defect affects multiples of that tile/die area, large amount of silicon to be trashed or binned down if sufficient redundancy has been included.
With Tiles being so small, and high yields may also make it possible to reduce the levels of redundancy they've learned to add from Monolithic die experience, and end up saving even more space and being even less expensive.
Apple ALREADY use chiplets, in the sense that Ultra's are a chiplet design. If Apple switch to a more aggressive version of such a design (disaggregated GPU vs CPU) it will be for OPTIONALITY (ie the ability to create more SKU variants) than because it saves money.So far, I've only run across a handful of folks referenced on AT to X/Twitter posters who actually get into the nitty-gritty details doing comparisons across Apple and other micro-architectures and foundry lithographies.
If Apple is moving to tiles, then it is likely primarily because their monolithic dies are now becoming large enough that adding additional redundancy to overcome defect densities is becoming to high to maintain their cost basis.
Going with tiles is not a get out of jail free card, the downside is that moving from monolithic means you need inter-tile fabrics that are a can of worms on their own and so far have generally led to lower performance than monolithic dies in latency and throughput.
Can be alleviated with high-speed fabrics, and Apple can probably avoid many of Intels and AMD's mistakes with careful R&D. However I might be wrong but I'm not aware of anyone generally beating monolithics inherent latency/throughput advantage with tile use.
That's true, packaging has its own costs. You can save additional money though by doing things like separating out the IO Die and manufacturing it on a cheaper node (like say TSMC N6) and also say repurposing a die without having to completely redesign a larger monolithic design (getting a bit into your later point about packaging smaller die together being useful for flexibility). AMD has said they definitely save money using this approach - it also probably depends on the type of package tech, the size of the dies and what the hypothetical monolithic die might be, and the yields of all of the aforementioned.Not true.
Any decent modern SoC is designed for resiliency, so that almost all manufacturing flaws can be worked around. Sometimes this is obvious (eg have a way to program out a CPU or GPU core that is flawed, and sell the result as 8 core rather than 10 core, or whatever), sometimes it's more subtle like designing all SRAMs with one or two redundant columns, so that you can "program out" the use of a redundant column.
Meanwhile even apart from the additional power consumed by chiplets (it's always going to cost more to cross a chiplet boundary than to stay within chiplet) packaging at the level required is not cheap AND you can lose yield during the packaging, perhaps more so than you save from your smaller chiplets hitting fewer manufacturing defects.
Two data points:
- Intel makes vastly less profit on its chiplet based SoCs than on its monolithic SoCs. Some of this is paying TSMC, but some of it is just the fact that chiplets are NOT a way to save money. They have various uses, but saving money is not one of them.
search for Intel Lunar Lake. (Though the entire article is interesting)
- Why is HBM so expensive? The general consensus is that "fairly priced" HBM should be about 3x the price of DRAM. Why so high? Because you lose some wafer space to the TSVs and other chiplet paraphernalia, you lose a fair number of chips to *packaging yield*, and the cost of the packaging.
Apple ALREADY use chiplets, in the sense that Ultra's are a chiplet design. If Apple switch to a more aggressive version of such a design (disaggregated GPU vs CPU) it will be for OPTIONALITY (ie the ability to create more SKU variants) than because it saves money.
It's a finely balanced issue whether it makes sense. Depends on the cost of masks, the cost per sq-mm of N2, etc.
It's POSSIBLE that given Apple's "traditional" volumes for Max and Ultra, a monolithic design (even with extra silicon "wasted" on duplicate ISPs, Secure Enclaves, more E-cores than makes sense, etc) was optimal.
BUT if Apple recently (say two or three years ago) felt it made sense (maybe even just for internal use) to build up a large number of data warehouse specific machines, and to specialize these, the most pressing being the equivalents of GB300s for training, with some P-compute, maybe no E-compute, and as much GPU compute (maybe along with some ANE?) then maybe the equation changes?
Given the costs of GB300 (IF you can even buy them) and the volumes Apple may want to install, running the numbers may show that, at least for one or two design iterations a chiplet design is actually the cheapest fastest way to get to where Apple wants to be, in that it allows them to do some degree of internal experimentation with the balance of warehouse designs, even if none of the results are ever seen by the public?
Not sure why you're trying to disagree with me when everyone has been switching to Tiling because it does in fact reduce losses from chip defects.Not true.
Any decent modern SoC is designed for resiliency, so that almost all manufacturing flaws can be worked around. Sometimes this is obvious (eg have a way to program out a CPU or GPU core that is flawed, and sell the result as 8 core rather than 10 core, or whatever), sometimes it's more subtle like designing all SRAMs with one or two redundant columns, so that you can "program out" the use of a redundant column.
Sure, it could. Not in this universe, of course. 😅Studio M4 max could start from $1999
MBP 14" M4 Max 14C 36GB/1TB is 3199 USD. I guess the equivalent Studio setup would be at least 2199 - unless they will try to match M2 Max Studio 30C GPU 32GB/512GB at 1999 USD, but it seems unlikely.Why not?
So far, I don't think I've seen any cases where any of these comms technology have bested what is available when a design is Monolithic.
So as I said, Apple moving to more of a Tile approach in future is likely do to them reaching common limits where defect density is starting to effect the economics they want to see.
Apple bought PA-Semi years ago and ended up spanking ARM/World, and then lost a lot of 0.1% talent which made the M2/M3 lag quite apparent.
Its already happenedSure, it could. Not in this universe, of course. 😅
Yes, because M2 Max and M4 Max are the same thing.Current M2 Max Mac studio starting price
(Btw, in Europe the prices have always been 600-700 Euros higher than what it should have been based on USD/Euro equivalence. E.g. the above MBP and Mac Studio are priced at 3849 Euros and 2419 Euros, respectively!)
So the prices listed in Apple's website don't include the VAT? Interesting, I never payed attention to that, since I was not going to buy overseas something that is available locally anyways. Thanks for the info!That’s because EU prices include the VAT and likely additional tax (recycling etc.). If you subtract the VAT, you end up with 3180 euros base price in Germany base price of that MacBook in Germany or less than $100 USD difference.
Mbp M2 pro/max and Mbp M4 pro/max that are not the same thing kept the same starting priceYes, because M2 Max and M4 Max are the same thing.
So the prices listed in Apple's website don't include the VAT? Interesting, I never payed attention to that, since I was not going to buy overseas something that is available locally anyways. Thanks for the info!
Both.Listed prices in the US never include the tax. It's very confusing from the European perspectiveI suppose it makes some sense for online shopping, as tax differs by state, but it's the same in supermarkets and restaurants. No idea how it actually works with online purchases, whether the tax gets added to your shopping cart or whether you are supposed to pay it separately.
I remember seeing a couple of youtube videos from High Yield and Asianometry discussing "hybrid bonding" - is that in Apple's near future? Can we expect ''stacked" ultra pieces in an iPhone/iPad within five years?Why this confidence that they are moving to tiles? As I mentioned before, there are much more interesting packaging solutions they can pursue.
I think you have it wrong. I think reduction in feature size will be more linear. I see this as the future..Assuming a 19.5 month refresh cycle
- M1: Q4 2020 5nm
- M2: Q3 2022 5nm
- M3: Q1 2024 3nm (N3)
- M4: Q4 2025 2nm (N2)
- M5: Q2 2027 1.4nm (A14)
- M6: Q4 2028 1.4nm (A14)
- M7: Q3 2030 1nm (A10)
- M8: Q2 2032 0.7nm (A7)
- M9: Q4 2033 0.5nm (A5)
- M10: Q3 2035 0.3nm (A3)
I would put this in the category of optionality. The way AMD uses chiplets allows them to create a wide range of SKUs from a few base components. This makes economic sense if you don't sell enough in a certain market to justify a dedicated, optimized mask set.That's true, packaging has its own costs. You can save additional money though by doing things like separating out the IO Die and manufacturing it on a cheaper node (like say TSMC N6) and also say repurposing a die without having to completely redesign a larger monolithic design (getting a bit into your later point about packaging smaller die together being useful for flexibility). AMD has said they definitely save money using this approach - it also probably depends on the type of package tech, the size of the dies and what the hypothetical monolithic die might be, and the yields of all of the aforementioned.
Everyone has NOT switched to tiles!Not sure why you're trying to disagree with me when everyone has been switching to Tiling because it does in fact reduce losses from chip defects.
If your redundancy isn't enough to allow you to scavenge an 8 core to a 6 core, or other failure, your loss is a small x tile sized area vs potentially a full CPU or SOC depending on where the failure is.
Packing losses are there, known, and included in Cost Basis and still everyone and their grandmother is making Tiling the defacto Standard, are they also wrong?
I'll see your cite with my own: https://arxiv.org/html/2203.12268v4
So as I said, Apple moving to more of a Tile approach in future is likely do to them reaching common limits where defect density is starting to effect the economics they want to see.
And going with Tiles instead of Monlothic means they will have to do the work on fabric interconnects and interposers, etc, just like AMD and Intel have done.
So far, I don't think I've seen any cases where any of these comms technology have bested what is available when a design is Monolithic. Perhaps Apple, Intel or AMD (TSMC) have some uber-inter-tile link in the works, however I'm happy to be shown to be wrong and some cheap solution is on the way which will make this point moot.
That being said, I'll reiterate that moving to Tiles en masse as Apple might be doing, is not automatically going to increase performance as I thought I was reading. Apple will however be able to look at all the work previously done by others and avoid a lot of costly mistakes/time either using a newer proven technology or doing some heavy lifting with TSMC or other and adapting one in a uniquely Apple way.
Apple bought PA-Semi years ago and ended up spanking ARM/World, and then lost a lot of 0.1% talent which made the M2/M3 lag quite apparent. I've not heard or read anything indicating they've bought out someone with some new, world changing interconnects design though, so I would expect them to be using one of the better, probably more expensive technologies.