What's interesting, in MKBHD's new video, he asked about the integration with OpenAI, and how much is done on device, etc.
He was told that almost everything is done on device (so this is clearly why RAM is a factor in things), from Apple built models, and that gives fast responses, etc (whereas the Android devices pushing most of it to the cloud will have slower responses and poor signal will make that even worse) - so to me, it's clear they're aiming for a best user experience with this..... they say that anything that is too complex to run on device, or outside of Apple's on device model's expertise it can try and use Apple's Private Cloud Compute.
If it thinks it's something ChatGPT can do better, it will ask the user if you want to use ChatGPT - and it will ask each individual time. It's not a 'say yes once and it will use ChatGPT whenever', it asks you every single time to confirm if you want to use ChatGPT - which is nice to know, if you never want to use ChatGPT, you just say no. Also, OpenAI is not allowed to ever store any of the requests and your IP address will be obscured, so OpenAI cannot connect multiple requests to profile you.