Hey guys,
I'm afraid it will be very hard (or, rather, impossible) to comment on such deep technical details due to a number of NDAs.
What I can say (and this is not specifically about Apple or the M1 Ultra but a general comment about our dev philosophy) is that we want Redshift to run as efficiently as possible on all the GPUs it supports today - or the ones it will support in the future. For this reason, we try to keep close communications with the hardware vendors. In some cases those communications result in discussions about possible optimizations that are fairly easy/straightforward - so we go ahead and do them immediately. In some other cases, the optimizations might require more work and are pushed further down the line. And, in yet other cases, the optimizations are deemed unrealistic given the existing architecture of Redshift.
The specific problem discussed here (which is "why doesn't the M1 Ultra render exactly twice as fast as the M1 Max") does not belong to the unrealistic category. I.e. there might indeed exist workarounds. But these won't be simple modifications either!
So we will need to run some experiments, see what kind of perf gains we can get and weight the dev effort against it . . .