I agree with what you are saying, but I think we will have to wait two or three years for the very nice scenario you laid out for us.
Earlier this year I popped for a M2 Pro 32G. I am amazed at what even that can run, but I have to use Ollama to run individual models for general text processing and NLP, one for medical advice, one vision model, one large uncensored model to get an understanding of general utility of censored vs. uncensored models (I wish all models I used were uncensored and trained just on English. I have only used non-English language support for a handful of interactions).
I also want Apple to develop very strong support for cloud models for my Apple Watch. With the excitement and disappointment for the new AI alternative pin computers, etc., I say we already have something useful in the Apple Watch so just keep improving that. I like to go about my day phone-free and the Apple Watch makes that work out OK for me.
Earlier this year I popped for a M2 Pro 32G. I am amazed at what even that can run, but I have to use Ollama to run individual models for general text processing and NLP, one for medical advice, one vision model, one large uncensored model to get an understanding of general utility of censored vs. uncensored models (I wish all models I used were uncensored and trained just on English. I have only used non-English language support for a handful of interactions).
I also want Apple to develop very strong support for cloud models for my Apple Watch. With the excitement and disappointment for the new AI alternative pin computers, etc., I say we already have something useful in the Apple Watch so just keep improving that. I like to go about my day phone-free and the Apple Watch makes that work out OK for me.