I went down a rabbit hole about local LLM. It seems that the new apple silicon is especially well suited to run these due to the unified memory (able to assign large amount of 'ram' to the GPU) and high(er than normal pc) bandwidth.
These days it's rather easy to use these LLM without touching any terminal work with apps like LM Studio (text) and Diffusion Bee (picture/images). Within the apps you can download general or specialized models that can aid you to create or respond to your requests. It has been rather fun to try it out. (It's like chatgpt but you can choose a model that is especially suited for a certain coding language or creating pictures as real life as possible or focus on real animals).
I am trying to swap my 16" to a 14" MPB. Initially trying to figure out whether to go for a Pro or Max with 30GB+ ram. Now knowing that these models can be quite useful running locally, I am suddenly more focused on the highest bandwidth and as much ram as I can/want to afford.
These days it's rather easy to use these LLM without touching any terminal work with apps like LM Studio (text) and Diffusion Bee (picture/images). Within the apps you can download general or specialized models that can aid you to create or respond to your requests. It has been rather fun to try it out. (It's like chatgpt but you can choose a model that is especially suited for a certain coding language or creating pictures as real life as possible or focus on real animals).
I am trying to swap my 16" to a 14" MPB. Initially trying to figure out whether to go for a Pro or Max with 30GB+ ram. Now knowing that these models can be quite useful running locally, I am suddenly more focused on the highest bandwidth and as much ram as I can/want to afford.