You make it sound like "5nm" and "extremely large caches" are completely decoupled from one another. Pragmatically, the are not. Using denser fab process allows the cache size to grow larger without causes some losses in other areas. ( e.g., sacrifice core count to have larger caches with a given fixed transistor budget and die size).
True, but I suspect that’s the beginning if the story, not the end of it. Apple can afford making more expensive chips since they are the sole consumer. That’s a big advantage as well.
Other vendors are now aggressively increasing the caches etc. because they see that Apples model works and they feel threatened. It will be interesting to see how these things play out as Intel and AMD move to the more advanced nodes.
For example both AMD and Nvidia GPUs are rumors to substantively boost their L2 caches in the next gen GPUs coming at the end of the 2022.
Well, I sure hope so, GPU L2 has traditionally been a joke. I mean, small L2 worked due to how GPUs operated historically, but AMD has demonstrated that fast large cache can dramatically reduce reliance on power hungry RAM.
But the playing field isn't just about laptops. Apple and Intel mostly sell laptops , but that isn't the whole market. Especially where "single threaded top end performance" is a strong selling point. Apple also was talking 'smack' about how they conquered desktop performance. Can't move the goal posts and crawl back into solely laptop land without a retreat there.
That’s true. I wonder whether Apple has a response here going forward. I guess they could yield the fastest single core desktop place and instead focus on ultra compact form factors and multicore performance. Or maybe they have something else planned.
Putting a higher performance GPU on the same memory bus as a CPU core trying to hit "beat everybody" single threaded throughput is a dual edged sword. That is why Apple puts a bandwidth cap on the CPU cores. It is a graphical user interface operating systems so at some point the GPUs 'wins' the limited bandwidth contrast when both sides want "too much".
True. But then again, giving CPU cores more bandwidth would also mean increasing the bandwidth of the caches, which would probably result in higher power consumption. Apple is quite conservative here. Their L1D for example is "only" around 153GB/s (48B/cycle at 3.2Ghz). Other Cpus can go much faster. But Apple compensates with larger cache size, lower latencies and their cluster architecture with shared/virtual L2 caches. The point is, 200GB/s to a CPU core cluster is hardly a limitation if you look at the full picture.
In the general desktop market it isn't likely that most of the buyers are going to be keen to throw modularity out the window for efficiency. Some of the efficiency trade-offs here are markets that Apple is tossing aside. As long as Intel and AMD are shooting at broader market coverage, they will likely continue to make different trade offs.
Oh, no doubt about it.
That real strength doesn't come for 'free'. It is all integrated... but it is all fixed integrated. Again scoped down to just laptops a more reasonable trade-off than as scale up the desktop product space.
One way out might be focusing more on multi-chip technology. Apple clearly has the fundamental tech down, let's see what they can do with it.
Errr..... it isn't like Intel and AMD don't have fused on die solutions. "x86" in and of itself doesn't inhibit heterogeneous compute solutions. Neither AMD or Intel have "bet the whole farm" on it, but it isn't like they haven't worked on it. (and some of this has to do with operating support and security.... not CPU core design. )
I think all the big players see MCM as the future. There is a lot of work in this area.