In the coming years, we need phones and computers to have a "local AI" service that's running and has minimal expected LLM functionality, such as chat and other LLM features.
Then, route all LLM calls to this service. If the user wishes to pay for premium LLM capabilities, that's forwarded on to the internet - the LLM of your choosing - trust wise, price, etc.., otherwise it stays local.
Then, route all LLM calls to this service. If the user wishes to pay for premium LLM capabilities, that's forwarded on to the internet - the LLM of your choosing - trust wise, price, etc.., otherwise it stays local.