SoCs are usually structured as a bunch of nodes (things like CPU clusters, GPU clusters, Apple Neural Engine, DRAM controller, all the peripherals, etc) which all attach to a common interconnect. The industry likes to call this interconnect a NoC, short for network-on-chip. NoCs do have the somewhat network-like property that any node can reach any other node simply by issuing a request to the NoC. The addressing used in the NoC is physical memory addresses, so the "network switches" do simple address decoding to route incoming requests to the appropriate port.Or am I wildly off base and missing something fundamental?
The NoC in a complex SoC is partitioned to avoid needing a single giant switch that serves all nodes in the system. Instead, there's always several switches, which connect both to each other and to leaf nodes. If there's a subsystem that needs less bandwidth, for example, you might connect its nodes to a secondary switch that connects to the main high bandwidth switch through a narrow, low-bandwidth port.
What travels across UltraFusion is almost certainly one or more switch-to-switch links to bridge one Max chip's NoC to the other. At minimum, I'd expect the biggest, highest bandwidth switch that's connected to the DRAM controllers gets linked to its counterpart in the other die. That might be all they did, in fact, as it's all that's necessary to make the system function.
The link is probably not 10K wide, it's more likely to be structured as a bunch of narrower links, allowing the system to forward many requests and responses simultaneously.