That’s what they all say. But a lot of these so called AI features require power more than what a phone has. Offloading to a server is sometimes a must.
Quantised models can be surprisingly small. And if Apple aren’t targeting LLMs for local use, more specific/tailored models absolutely can run on device.
That said, given the precedent sent by Siri, their next progression of Siri into an LLM will absolutely require network connection and be executed server side.
Samsung’s version on One UI 6.1 lets you toggle between running the local models on the phone’s NPU versus connecting to their servers.
The local version is slightly slower and produces worse results, but can be used for privacy or without the internet. The remote version is what you’d expect.
The thing is, these AI features are just features already present in some way or another, just emphasizing content generation and slapping AI branding.
Sure if you’re running large models like gpt, smaller models tailored to specific use cases can absolutely run on phones. Whether or not they get there implementation down right is a different story though
That’s what they all say. But a lot of these so called AI features require power more than what a phone has. Offloading to a server is sometimes a must.
Quantised models can be surprisingly small. And if Apple aren’t targeting LLMs for local use, more specific/tailored models absolutely can run on device.
That said, given the precedent sent by Siri, their next progression of Siri into an LLM will absolutely require network connection and be executed server side.
Samsung’s version on One UI 6.1 lets you toggle between running the local models on the phone’s NPU versus connecting to their servers.
The local version is slightly slower and produces worse results, but can be used for privacy or without the internet. The remote version is what you’d expect.
The thing is, these AI features are just features already present in some way or another, just emphasizing content generation and slapping AI branding.
Sure if you’re running large models like gpt, smaller models tailored to specific use cases can absolutely run on phones. Whether or not they get there implementation down right is a different story though