It should be..since m4 was the only one who didn’t from the m family. Now its time
Probably the studio display rumor can be true with support for tb5 will help ipads and Mba base customers
Yeah while I think it’s funny my 1 min calculations turned out pretty much correct, I am merely pointing out how easy it is to make ”educated assumptions” by just using putting one and one together. And since that is the case, I dont see the point of these ”leakers” status at all. They should be fired and do something worthwhile instead.WHy are you repeating and quoting yourself a bit of narcissism ?
Gurman is not predicting benchmarks but devices
We also can say that based on existing A19 pro and M5 that M5 pro metal will be around 150k...but thats math not predictions
But if you really want to be a true leaker or a Messiah tell us from now what M6 will be in both cpu and gpu departments ( scores/ nr of cores/ frequencies/ cash)
Yes, but we already have TB5 into the M4pro/MaxWe went M1 -> M3 with no TB4 support so Apple could easily drag out TB5 implementation on M5.
Yes, but we already have TB5 into the M4pro/Max
Maybe besides M5, TB5 could be a good addition for the ipad pro users along side with the "so called" Studio Display
and support for native DisplayPort 2.1 output over USB‑C
ST 4600 MT 24000 GPU 94000But if you really want to be a true leaker or a Messiah tell us from now what M6 will be in both cpu and gpu departments ( scores/ nr of cores/ frequencies/ cash)
Sure, but if it is SoIC, then all bets are off? Probably safe to assume the result will be better than what the monolithic SoC approach would have achieved, but I imagine it would be hard for even an expert to predict. I guess I'd study what AMD is doing and try to extrapolate from there?
That’s not what I meant to say. I mean, if they are using SoIC for Pro/Max/Ultra, then it must be to gain some kind of advantage. We can assume Apple’s priorities haven’t changed. So what does SoIC allow them to do better?SoIC is going to make the GPU inherently better how? For RDNA 4 AMD dropped the chiplet approach they dropped chiplets for their dGPU offerings.
For the APU ( "Halo") class the CPU cores were already in a chiplet in the first place. The "everything else" is basically monolithic once exclude covering the CPU cores.
Apple going down the path where the CPU cores are pushed off into an exclusive island doesn't really buy much for them. The Pro , Max are very highly skewed toward being the GPU ( GPU cores + Display Engines+ media coprocessors ) anyway in the first place.
THey could do tiles like Intel.SoIC is going to make the GPU inherently better how? For RDNA 4 AMD dropped the chiplet approach they dropped chiplets for their dGPU offerings.
For the APU ( "Halo") class the CPU cores were already in a chiplet in the first place. The "everything else" is basically monolithic once exclude covering the CPU cores.
Apple going down the path where the CPU cores are pushed off into an exclusive island doesn't really buy much for them. The Pro , Max are very highly skewed toward being the GPU ( GPU cores + Display Engines+ media coprocessors ) anyway in the first place.
That’s not what I meant to say. I mean, if they are using SoIC for Pro/Max/Ultra, then it must be to gain some kind of advantage. We can assume Apple’s priorities haven’t changed. So what does SoIC allow them to do better?
Yup, basically my points 2 and 3 are this. AMD has also done basically done all of these at some point.THey could do tiles like Intel.
These are not mutually exclusive:
1) Wafer flexibility: e.g. a) potentially stacking cache made on cheaper nodes onto logic made from more expensive nodes, possibly allowing for even more cache (especially SLC - think AMD 3D-cache) for cheaper; b) using cheaper wafers for I/O vs logic, wouldn't necessarily be 3D stacked but still using the SoIC platform
2) Die reuse: if part of the die is going to be the same across all the variants, it could make sense to separate it out and attach the other pieces to it
3) Die flexibility: this almost a 2a) but you could imagine that say Apple wanted to build a GPU-heavy SOC variant, you would attach a CPU die to two GPU dies. A variation of this idea with even more flexility would be personally the most exciting but I wouldn't necessarily get my hopes up that Apple's first iteration of chiplets/tiles for the Pro/Max will allow for this as Apple may be more conservative, testing it out first but building otherwise familiar Pro/Max chips (in fact what I've described here would be just a formalization of the current M4 Pro/Max design).
So number 1) could lead to performance improvements if it allows for extra cache (it could also allow for cheaper cache) and number 3) could lead to performance improvements in a sense because the user could better tailor exactly what kind of performance they want to spend their money on - CPU-focused or GPU-focused or both.
Yup, basically my points 2 and 3 are this. AMD has also done basically done all of these at some point.
That’s not what I meant to say. I mean, if they are using SoIC for Pro/Max/Ultra, then it must be to gain some kind of advantage. We can assume Apple’s priorities haven’t changed. So what does SoIC allow them to do better?
There are a couple of recent articles at SE on stacking. In first forays... "an easier first step" (used by Qualcomm) for "20% to 30% improved power efficiency" is mentioned. Something else Apple might be inclined towards...
https://semiengineering.com/first-forays-into-true-3d-ic-designs/
Apple's priorities so far has been
1. Perf/Watt.
2. Fewer/Narrower package solutions delivered over wider set of products. (economies of scale in volume and length of individual package design usage/deployment ). Also ease to product at volume matters.
Tightly coupled to that at the product line up level : don't make everything for everybody. Even the list of products being distributed to is not very large.
3. Relatively narrow range binning for segmentation for a specific package/die. 1-2 CPU cores binned out. Or 2-4 GPU cores binned out.
Product segmentation also restricts and confines I/O. ( so not uniform across in a product line up. )
Isn't B flipped with the most recent V-Cache design, eliminating the logic die hotspot concern? I don't see why Apple couldn't do the same.Apple's priorities so far has been
1. Perf/Watt.
2. Fewer/Narrower package solutions delivered over wider set of products. (economies of scale in volume and length of individual package design usage/deployment ). Also ease to product at volume matters.
Tightly coupled to that at the product line up level : don't make everything for everybody. Even the list of products being distributed to is not very large.
3. Relatively narrow range binning for segmentation for a specific package/die. 1-2 CPU cores binned out. Or 2-4 GPU cores binned out.
Product segmentation also restricts and confines I/O. ( so not uniform across in a product line up. )
So in turn...
Perf/Watt : Versus monolithic SOIC raises or lowers Perf/Watt. SOIC is lower power than previous die packaging interconnect, but have the previous ones been better than monolithic?
Could SOIC help delivery a next gen "UltraFusion 2" that had better Pref/Watt characteristics? Yes. But that wasn't monolithic in the first place.
Two ( c and d) out of the three examples TSMC list for SOIC on this diagram are for horizontal (more a interposer connect than stacking of denser logic) at least as much as vertical moves.
View attachment 2562591
TSMC-SoIC® - Taiwan Semiconductor Manufacturing Company Limited
3dfabric.tsmc.com
B is somewhat over representative because probably cannot cover 'logic' lower layer chip relatively hot spots with a cache chip (e.g., AMD "3D V-Cache" ).
Package reuse :
M1 in iPad Pro ( then iPad Air) , MBA , Mini , and iMac.
As go bigger die the product spread gets narrower.
die reuse M2 Max M2 Ultra.
Something like AMD's 3D V-cache augments a line up of 5 Ryzen products by another 3. Apple only has M1 , Pro, Max and sometimes Ultra. Apple has less base offerings before even adding the V-Cache options. Mix in AMD Epyc and even wider SpC product line up that is being covered. Is Apple out to sell every SoC perumation possible? Probably not.
A failed V-Cache bonding and if can fuse off the layered cache part , then still have a working product. If all (or vast majority ) of cache was in broken connection then not so much. The stacked die could be cheaper (old fab process and/or just much smaller) , but it isn't from defects either ( sometimes going to loose two or more good dies).
The other question would be what is the run rate TSMC can do with SOIC.
It also costs money to do.
"...
..."
- SoIC-P is based on 18-25μm pitch μbump stacking and is targeted for more cost-sensitive applications, like mobile, IoT, client, etc.
![]()
TSMC 2023 North America Technology Symposium Overview Part 3 - Semiwiki
TSMC's 3DFabric initiative was a big focus at the symposium, as it should be. I remember when TSMC first went public with CoWos the semiconductor ecosystem, including yours truly, let out a collective sigh wondering why TSMC is venturing into the comparatively low margin world of packaging. Now...semiwiki.com
If 'dirt cheap' then wouldn't need a more affordable option. if already have a managable size die ( sub 200mm^2) then is 5-8% bigger on N2 more expensive than the overhead costs of 'gluing' that extra 8% on afterwards?
Also back to tsmc SOIC page.
View attachment 2562592
SoIC isn't making InFO_PoP necesarily disappear in most applications. ( pretty good chance that POP is cheaper in many of those DRAM to SoC connection contexts. Although it could help thermals to take the DRAM off the top and put it to the size of a smaller 2-D footprint chip. ( doesn't help M-series much though because not stacking DRAM to screw up thermals in the first place. )
Relatively narrow range of binning :
Apple charges a hefty premium for most "built to order" options. At the core they don't like complicated and/or expensive inventory. ( i.e., not out to sell everything to everybody). N2 wafers costing more , 16A costing thousands more , 14A costing thousands more will push Apple into SoIC to control costs. But as long as Apple can manage to pass along increase base costs into the product, they'd probably stick with the simpler inventory 2D/monolithic approach. ( e.g, Apple isn't covering Ultra at every iteration. Highest end stuff churns slower in Mac market. )
Isn't B flipped with the most recent V-Cache design, eliminating the logic die hotspot concern?
They have different classes of products, with different tradeoffs. For example, very low baseline power consumption is essential to Ax and Mx series, but less so for Mx Max.
Looking at the wording from Apple patents describing stacked die solutions, a focus appears to be on increasing the effective transistor budget without increasing the package area. This is why they empathize 2.5D stacking instead of flat tiles. I think @crazy dave is spot on here — if I can move secondary blocks like SLC, memory caches, display controller, etc. to a secondary die (maybe using 5nm process), then I have more effective area for compute. This approach will certainly use more power than a monolithic die on a cutting-edge process, but it might be worth it for prosumer chips.
Is AMD's Infinity cache on Halo Strix like Apples SLC?It wasn't eliminated. AMD worked out some bugs of layering SRAM on SRAM that required the to down clock the whole chip at bit. They still are not covering the more substantive hot spots (e.g. cores or very long distant signal drivers ) on the chip with SRAM.
The MI300 series there is some cache on the bottom but not on the top and those packages are thermal problem children compared to what interiors to Apple products.
AMD has a deeper need for CPU V-cache because they pulled the CPU cores away from memory and have uniformly slower access to minimize the NUMA issue. Apple's designs have always been cache 'heavy'. Pouring more ketchup on top of already large dose of ketchup probably won't see the same kind of improvements that AMD gets.
AMD also roughly collects all the bottom level SRAM into one place on the lower die. (created one big landing target for the V-cache). Apple's basic set up doesn't really do that. SLC cache is being shared by several different types of cores and is used to 'front' the memory system (which is relatively large (compared to AMD) and distributed around the die. )
If take the fab density gains to make the cores/logic smaller then in a zero sum die size puzzle they can add more area budget to the cache ( grow them larger in the monolthic die). N3P is also suppose to incrementally improve SRAM (not quite to N3B levels , but better than the N5 stuff. )
Ah I see your confusion. I wasn't suggesting that the logic dies would be vertically stacked - SoIC allows for both vertical and 2.5D interposer connects (as Apple already uses for its fusion connector). That would be option c) in your figure above. I admit I didn't explicitly mention that for logic in parts 2/3, but I assumed that was obvious and I did say it when discussing moving things like I/O to a different die (which I said would not be necessarily by vertically stacked, that should have been a stronger negative). I will edit it that to make it more clear. Vertical stacking, if it is coming at all, is likely restricted to cache (and maybe capacitors as mentioned in the interviews provided by @treehuggerpro - also as stated in his links, 3D stacking of logic, is likely to remain exotic for quite some time).the Max's construints certainly bleed into the Mac Pro constraints. Apple's notion that these packages are perfectly optimized to for each product is a bit overblown.
Pushing the Max power consumption higher will backslide the mBP 14" ( and lessor extent the 16"). Perhaps will get some power savings elsewhere ( screen) to offset.
His proposed variant of mix-and-match CPU/GPU tiles are not going to vertically stack well. So that isn't spot on.
Additionally Apple wanting to keep the package size the same and "do more" inside the package is also aligned with not wanting more SoC skus/variants.
If Apple wanted to pack more into a Ultra (two Max die) set up then, yes what you outlined would be what they would be pulled into. But this is still largely just clinging to the very chunky, really non-chiplet approach to construction. Pretty good chance it would be herded into just the Max just like UltraFusion was.
Very top of the "prosumer" market means Apple would probably treat it as being more price elastic and being more expensive would just get passed along to endusers (e.g., Mac Pro base price going up 100% in 2019). That isn't exactly along with most of the normal product line strategy.
The Pro being stuff into the super shrunk Mini chassis means it isn't particular free to backslide on that front even though a "desktop". If the Max is causing the battery life on MBP to backslide then doing the same to the Pro doesn't really help that product subsegment.
P.S. the die size bloat creep on the Max has been real
M1 Max 432 mm^2 ( 2x this was creeping close to being a reticle limit bust )
M2 Max 510 mm^2.
M3 Max 600+ mm^2
M4 Max ?? ( even if came backwards a bit ... it is still quite large after the M3 "pork out")
When High NA EUV phases in eventually, this would largely be the same problem as what drove UltraFusion in first place ( 429 mm^2 reticle. down from 858mm^2 )
YesIs AMD's Infinity cache on Halo Strix like Apples SLC?
This rumor appears to be based on a single source, a blog post here:Supposing M5 use N3P like the A19, do you think it is likely the M5pro and up use N3X (I read somewhere qualcomm is using it for the high end X2 elite to reach 5GHz, so it should be ready for early next year) ?
Supposing M5 use N3P like the A19, do you think it is likely the M5pro and up use N3X (I read somewhere qualcomm is using it for the high end X2 elite to reach 5GHz, so it should be ready for early next year) ?
N3 vs N5 | N3E vs N5 | N3P vs N3E | N3X vs N3P | |
Power | -25% ~ -30% | -34% | -5% ~ -10% | -7%*** |
Performance | 10% - 15% | 18% | 5% | 5%, Fmax @1.2V** |
Density* | ? | 1.3x | 1.04x | 1.10x*** |
HVM | Q4 2022 | Q4 2023 | H2 2024 | H2 2025 |