2 Sources
2 Sources
[1]
Gemini gets one step closer to controlling other apps on your Android phone
Google could refer to the functionality as "screen automation." AI agents are the hottest trend in tech, and the increasing momentum leaves no room for doubting that your mobile devices will soon be controlled by them. Back at I/O 2025, Google demonstrated how it plans to put Gemini in charge of controlling your phone for you. This was dubbed Project Astra and showed Gemini's capabilities at not just viewing the text and media on your phone, but also scrolling and tapping when needed, basically displaying a spectrum of agentic force in action to help you with simple tasks. While Project Astra is still in its development phase, we're seeing evidence detailing how it might work.
[2]
Google wants Gemini to start using your Android apps
Rajesh started following the latest happenings in the world of Android around the release of the Nexus One and Samsung Galaxy S. After flashing custom ROMs and kernels on his beloved Galaxy S, he started writing about Android for a living. He uses the latest Samsung or Pixel flagship as his daily driver. And yes, he carries an iPhone as a secondary device. Rajesh has been writing for Android Police since 2021, covering news, how-tos, and features. Based in India, he has previously written for Neowin, AndroidBeat, Times of India, iPhoneHacks, MySmartPrice, and MakeUseOf. When not working, you will find him mindlessly scrolling through X, playing with new AI models, or going on long road trips. You can reach out to him on Twitter or drop a mail at [email protected]. Google has been working on Gemini's agentic capabilities, enabling it to carry out actions on our behalf. It recently rolled out an AI-powered Auto Browse feature in Chrome for users on its AI Pro or Ultra tiers. The company is also working on bringing similar capabilities to Gemini on Android, letting it control apps directly on your phone. A new report now sheds more light on this feature. Gemini has already made many complex tasks faster and easier. But this is just scratching the surface of what's possible, as the AI will eventually evolve to take actions on our behalf. A teardown of the latest Google app beta (v17.4) by 9to5Google details how the "Get tasks done with Gemini" feature will work. Codenamed bonobo, it will initially be available as a Labs feature, with the ability to book rides and place orders through screen automation. In practice, this could let you ask Gemini to book an Uber to your office or order dinner from Uber Eats without ever opening the app yourself. Reportedly, Gemini's agentic capabilities will only work in "certain apps" initially. Given that app UIs change frequently, it makes sense for Google to initially limit Gemini's screen automation feature on Android to a small set of apps. Beyond food delivery and ride-hailing, early support could be limited to first-party Google apps. The feature will require at least Android 16 QPR3 to work, since Google has laid the groundwork for screen automation with this release. You will always remain in full control While you can ask Gemini to carry out tasks on your behalf, you'll remain in full control and can manually stop or take over at any time. The strings in the code provide more insight into the feature and the privacy implications. Google notes that, "When Gemini interacts with an app, screenshots are reviewed by trained reviewers and used to improve Google services if Keep Activity is on." AP Recommends: Subscribe and never miss what matters Tech insights about everything mobile directly from the Android Police team. Subscribe By subscribing, you agree to receive newsletter and marketing emails, and accept our Terms of Use and Privacy Policy. You can unsubscribe anytime. Google will also warn users not to enter payment information into Gemini Chats and to use screen automation during emergencies. While not mentioned, Gemini's agentic capabilities on Android could be limited to users on the Pro and Ultra tiers. The company will most likely announce this feature in beta alongside the release of Android 16 QPR3 in March.
Share
Share
Copy Link
Google is building screen automation capabilities that let Gemini directly control Android applications on your behalf. The feature, codenamed bonobo, will initially support booking Uber rides and ordering from Uber Eats, with users maintaining full control to stop or intervene at any time. The functionality requires Android 16 QPR3 and raises new privacy implications as screenshots may be reviewed by trained reviewers.
Google is advancing its vision of AI agents by developing screen automation features that enable Gemini to directly control Android applications on behalf of users
2
. The functionality, internally codenamed bonobo, represents a significant step toward AI taking actions on behalf of users rather than simply providing information or suggestions2
.A teardown of the latest Google app beta (v17.4) reveals that this capability will initially launch as a Labs feature with support for booking rides and placing food orders through screen automation
2
. In practical terms, users could ask Gemini to book an Uber to their office or order dinner from Uber Eats without ever opening the apps themselves2
.The development builds on Google's Project Astra demonstration at I/O 2025, which showcased Gemini's ability to view text and media on phones while also scrolling and tapping when needed
1
. This spectrum of agentic force in action helps users complete simple tasks by functioning as an AI agent to assist with tasks across the device1
.
Source: Android Authority
Google could refer to the functionality as "screen automation," though Project Astra remains in its development phase
1
. The feature will require at least Android 16 QPR3 to work, as Google has laid the groundwork for screen automation with this release2
.Gemini to start using your Android apps will initially work only in "certain apps" due to the frequent changes in app interfaces
2
. Beyond food delivery and ride-hailing services, early support could be limited to first-party Google apps2
.Users will remain in full control and can manually stop or take over at any time while Gemini performs tasks
2
. This safeguard addresses concerns about autonomous AI agents operating without oversight, ensuring humans maintain final authority over actions taken on their devices.Related Stories
The feature introduces notable privacy implications that users should understand before enabling screen automation
2
. When Gemini interacts with an app, screenshots are reviewed by trained reviewers and used to improve Google services if Keep Activity is enabled2
.Google will warn users not to enter payment information into Gemini Chats and advise against using screen automation during emergencies
2
. While not confirmed, Gemini's agentic capabilities on Android could be limited to users on the Pro and Ultra tiers, similar to the Auto Browse feature recently rolled out in Chrome2
. Google will most likely announce this feature in beta alongside the release of Android 16 QPR3 in March2
.Summarized by
Navi
[1]
[2]
22 Nov 2024β’Technology

02 May 2025β’Technology

20 Jan 2026β’Technology

1
Business and Economy

2
Technology

3
Technology
