Google is enhancing Android apps to align with user expectations for artificial intelligence, similar to advancements in Windows 11. Developers received a preview of this initiative, which includes a new feature called AppFunctions. This feature allows Android apps to expose public interfaces for specific functionalities, enabling seamless interaction with AI agents and system-level services. AppFunctions are analogous to the Model Context Protocol (MCP) for cloud-based AI interconnectivity and will be accessible through Google's Jetpack library and platform APIs, ensuring local interactions on devices.
AppFunctions are currently in early development, with initial examples implemented in the upcoming Gemini version for the Samsung Galaxy S26 series and other Samsung devices running OneUI 8.5 and higher. Users will interact with Calendar, Notes, and Tasks using AppFunctions to streamline activities. Google is launching an early preview of AppFunctions through a beta feature in the Gemini app, available on the Galaxy S26 series and select Pixel 10 devices, allowing users to delegate tasks to AI agents by double-pressing the power button. The initial rollout will focus on apps in food delivery, grocery, and rideshare sectors in the US and Korea.
AppFunctions are expected to be integrated into Android 17, with a stable release anticipated around mid-year.