Really looking forward to the M4 announcement. I think the GPU is bound for a nice upgrade. The M3 set the stage with the new abilities (dynamic caching etc), now the arch and cores themselves are due for an upgrade.
M3 is a new architecture. Even Apple refers to it as a major "new GPU architecture". Not sure how they would do dynamic caching with absolutely zero collaboration between the cores and the cache ( as if the cache is completely detached from architecture anyway).
What Apple is a bit stale on is the memory subsystem ( still basic LPDDR5) , not the GPU arch. I would expect some tweak fixes and adjustments to higher raw RAM bandwidth access ( and updated on chip backhaul network. ). But it would be surprising to see a "let us toss that out and do a new yet another architecture" move on M4.
Different clocking and some change behind the GPU instruction decode and rebalanced cache to new bandwidths seems more likely. Very unlikely going to get a better general process node density (even if wait to sync up with N3P. ). If there is no node density increase then going to be kind of tough to add "more stuff" . The design aid tools may be better at dealing with N3-like contexts so could squeak out some incremental wins in some subsets of the logic design, but substantially large increases in transistor budget likely are not coming next iteration.
Pretty good chance this will be an "improve what you have" generation ( like M2 over M1 ).
Apple released the M1 Ultra before WWDC that year and had a substantive number of hiccups of developers being about to optimize for that GPU hardware. They are basically in the same boat now. Throwing out new hardware where the tools and the developer knowledge is lagging isn't going to help much try to trade "hype headline blows" with Nvidia (or AMD).
If M4 turns out to be just about AI then I could not be more disappointed.
Probably not. If uplift the raw RAM backhaul then all the different computational subcomponents should get more data. That should 'lift all boats'; not just a single anointed one. A bigger NPU would more easily share the chip network bandwidth with the GPU ( and other units. ).
And now more publicly obvious Apple has got some substantive security holes to fill. Would be a good idea to fix those in M4.
Might want to get leveled up on I/O to USB4,2 ( TBv5 on non plain SoCs. Still suspect the plain M4 won't meet TBv4 requirements ( or perhaps hand waves around them with the lid closed). )
If they are going to continue to charge $400/TB ...maybe read/write speeds competitive with PCI-e v4/v5 SSDs that will increasingly be standard in 2025-26 timeframe.
Apple is still slacking on AV1 encode.