Based on information Apple has released so far, I’m fairly confident that there won’t be any traditional dGPU system in any of the new Macs. Apples approach is more flexible and more efficient, so there is really no reason for them to use dGPUs.
The energy efficient wins are pretty broad, but more flexible? That is a reach. A discrete GPU chip can be used on an add-in-card or embedded onto the motherboard. That is two deployment profiles. iGPU don't have that flexibility since sharing the same memory controller with the CPU ( and other I/O function units in the package).
dGPUs can use DRAM , GDDR , or HBM RAM relatively easily. CPU's .... not so much (at least not a wide set of general purpose workloadds with mid-to-high workload concurrency ).
Apple is getting some 'wins' here but those are trade-offs ( those aren't "free" wins. They are paying something for them. One of which is flexibility).
Unified Memory doesn't require homogeneity (one type of memory) and single package ( SoC).
www.anandtech.com
or
"... AMD announced that future Epyc+Radeon generations will include shared memory/cache coherency between the GPU and CPU over the Infinity Fabric, ..."
Don't move the data.
www.tomshardware.com
is spreading control and access to memory over different packages "free". No; most design decisions come with trade-ffs. It has higher latencies , but it is more flexible. IBM and Nvidia have been doing something similar with their more proprietary memory coherence links between Power 8-9 and Nvidia. To lessor extent has 4-16 socket CPU set-ups from a variety of implementors earlier ( IBM Z (and previous), Power , Intel , etc. ) .
If thrown on constraints like "must fit in smaller area" , "most consume lowest power " , "most be lowest overall latency", etc. that is something different.
There are some slightly warped tasks on "flexibility" where being able to natively run iPhone apps and Mac apps on the same GPU with no semantic gaps in GPU code execution... Yeah that is more flexibility. However, bending "flexibility" to fit the largely self imposed design constraints that Apple has imposed on itself ( as opposed to primarily external customer driven) is a bit of hand waving.
Certainly not flexible in time. 3-4 years later, a faster GPU comes out and ..... stuck with the exact same GPU started out with. That is not particularly flexible in the slightest. Even in an embedded context. Let's say SoC package is over due for an update , but dGPU from other source has substantively iterated. If the SoC would couple to the dGPU then could iterate. So system delivery schedule is more flexible. ( Apple could render that moot buy taking position won't do any update unless all major subcomponents more ... but that is yet again self imposed , not market driven, constraint. )
Apple also probably has layered as security constraint aspect over this. So not only cache coherency is being simplified by running it through a single memory controller hierarchy , but also a single place to set the security policies ( which process can touch which pages. ). Yet again those that is more "flexibility" to implement the requirement on the table rather than boarder implementation options.
Looking at the performance projections, I‘d expect high-end Apple laptops to have GPU performance roughly comparable to a mobile RTX 3060.
Since Apple sells 70+ % laptops, that is very nice for that subset of the Mac product line up. Doesn't present much potential breadth of performance coverage for the desktop side though.
Apple probably will drive more future customers into laptops. Pretty good chance though that isn't a 'free lunch" at the top end of the desktop space. They probably won't loose much sleep, if any, over that outcome.