Critiqs

Google pushes AI forward with Gemini 2.5 for smarter digital agents

Google pushes AI forward with Gemini 2.5 for smarter digital agents
  • Google’s Gemini 2.5 model helps developers build digital agents that mimic real user actions in apps and sites.
  • It can automate tasks like form filling and button clicks while ensuring actions are double checked for safety.
  • Early results show faster development, streamlined automation, and safer interface testing in real world use.

A new artificial intelligence model is stepping up to tackle digital tasks that once demanded hands-on attention.

Google’s latest release, the Gemini 2.5 Computer Use model, is now available through its API, giving developers a powerful tool for building agents that interact with software just as a user would.

Handling jobs that need more than a simple script, the model can fill out forms, click on buttons, scroll through content, and even sign in to web platforms. This boost goes beyond plain programming and bridges the gap between automated tasks and real-world browsers or apps. The company says it delivers faster and more reliable results on performance tests for web and mobile tasks compared to other options on the market.

This new tool runs on a loop: the agent receives a user request, a snapshot of the current app or webpage, and a record of previous actions. Developers can zero in on exactly which UI functions the system should use, or add custom ones if their situation calls for it. The model then reviews the environment and picks the right move, such as typing into a field or clicking a button, before the cycle repeats with a refreshed screenshot.

Safety and Early Results

Of course, letting artificial intelligence agents roam free in user interfaces comes with serious responsibility. To guard against misuse, Google designed the Gemini 2.5 Computer Use model with built-in safety elements. Every action it proposes runs through an external check before it actually takes place. Developers can insist the system double-checks with humans before doing higher risk work, such as transactions or system changes.

The company urges anyone using these tools to test thoroughly before sending their new agents into the wider world. “The only way to build agents that will benefit everyone is to be responsible from the start,” a Google spokesperson explained.

Testers have already been putting the model through its paces. Some Google teams use it for quality assurance work, harnessing the tool for automated interface checks that speed up development in ways that would be impossible with traditional testing.

Others, early access developers, see strong results in areas like workflow automation and digital assistants, with the system able to arrange appointments, categorize virtual sticky notes, and carry out rapid multi-step tasks that would tax a human’s patience.

Developers interested in exploring these capabilities can find the Computer Use tool in Google AI Studio and Vertex AI. The hope is that with careful implementation, this new class of digital agent will not just handle repetitive chores, but open up the next wave of smarter user experiences.

SHARE

Add a Comment

What’s Happening in AI?

Stay ahead with daily AI tools, updates, and insights that matter.

Listen to AIBuzzNow - Pick Your Platform

This looks better in the app

We use cookies to improve your experience on our site. If you continue to use this site we will assume that you are happy with it.

Log in / Register

Join the AI Community That’s Always One Step Ahead