A GPU built into a SOC like Apple's A12X/A13 has to share memory with the CPU. A DGPU always has memory dedicated to it so it doesn't have to share with the CPU. Apple gets somehow around this using tile based rendering.
The only work around is to have HBM which can be put on the same package as the SOC. But HBM is much more expensive compared to GDDR6 even.
It all depends on what system RAM you use. For example:
- LPDDR5 is viable in the lower-tier, it gives you ~100GB/s which will outperform GDDR5
- HBM2 is viable in the upper tier, it offers bandwidth competitive with high-end GDDR6 setups (or even surpassing them), while offering low latency (suitable for CPU workflows) and low power consumption.
Yes, HBM is expensive, but then again so are higher-end Apple computers. And you are saving some circuitry by not having to manage multiple chips with their dedicated RAM. It's probably not too far-fetched to assume that dropping the dGPU, VRAM, DDR system RAM and the power circuitry to feed the dGPU will save enough money to offset the higher costs of HBM + interposer. And, it would make the logic board layout more compact. Besides, that would allow Apple to keep the current 16" MBP price while actually making the value proposition more attractive. One is more at ease spending $3000 on a laptop knowing that its CPU has access to 300+ GBps bandwidth memory.
Also, tile rendering is great and all, but you still need GPU RAM bandwidth for compute. HBM as system RAM will solve this, not to mention that it will allow completely new levels of system performance.