Technology

Google Consolidates Its AI Agent Strategy, Folding Project Mariner Into Gemini 2

Google is moving computer-use capabilities from its experimental Project Mariner into Gemini 2, its flagship AI model, as part of a shift toward making AI agents practical tools. The company is also e

Martin HollowayPublished 37m ago4 min readBased on 3 sources
Reading level
Google Consolidates Its AI Agent Strategy, Folding Project Mariner Into Gemini 2

Google Consolidates Its AI Agent Strategy, Folding Project Mariner Into Gemini 2

Google has reorganized the team behind Project Mariner, its experimental AI system that can browse the web, click buttons, and complete tasks on its own. Rather than keeping it as a separate research project, the company is folding Mariner's technology into Gemini 2, Google's newest flagship AI model, as part of a broader push into what the industry calls "agentic AI"—systems that can work through multi-step tasks with little human involvement.

The consolidation signals that Google now views AI agents not as long-term research, but as near-term products. Some Mariner capabilities have already moved into other Google products, including the newly launched Gemini Agent.

Gemini 2 Is Google's Answer to Agentic AI

The reorganization arrives alongside Google's announcement of Gemini 2, a new version of its main AI model that has been specifically trained to plan and execute tasks across computers and the web. The model shows Google's most serious commitment yet to agentic AI—essentially, AI that can figure out a goal and work through the steps needed to reach it.

To show what Gemini 2 can do, Google introduced two specialized agents: one built for coding tasks and another for data science work. Both showcase the model's ability to understand complex instructions, break them into smaller steps, and interact with software tools to get the job done.

The technical foundation comes straight from Project Mariner's research. Mariner operated as a Chrome browser extension that could look at web pages, understand what it saw, click buttons, fill in forms, and move between sites to complete goals a user specified. That infrastructure—the visual understanding and the ability to interact with web elements—now powers Google's broader agent ambitions.

Chrome Gets More AI Features

At the same time, Google has added multiple new AI capabilities to its Chrome browser, including a button that opens the Gemini chatbot directly from the browser window. These features rolled out first to paying Google subscribers in May, and are now available to all desktop users in the United States who browse in English.

The logic here is straightforward: most people spend their workday inside a web browser. By building AI tools directly into Chrome, Google positions itself to offer these AI agents right where users are already working—in web applications, email, productivity tools, and information searches.

This approach follows a pattern Google has used many times before. Think of how Gmail brought search directly into email, or how Android deeply embedded Google services into phones. The company typically wins by adding new capabilities to the platforms people already use, rather than asking them to switch to something entirely new.

What the Shift Tells Us About Google's Confidence

The move of Mariner's capabilities into Gemini 2 suggests Google has made real progress in getting AI to understand and interact with websites reliably. Browser automation has always been tricky—it requires spotting small on-screen elements precisely, and websites change constantly. Google's ability to turn this research into actual products indicates the company has solved some genuinely hard problems in how AI understands what it sees on a screen and adapts when things change.

The shift from standalone research to built-in product features also shows Google believes people will actually use AI agents for everyday tasks—booking trips, managing email, researching products, handling paperwork. That is a meaningful bet. The company could have played it safer and kept these as long-term research questions. Instead, it is integrating the work into its main products.

For companies and organizations, this opens up possibilities. Imagine automating routine web-based work—the kind of task a person does the same way every time—just by describing it in plain English to an AI. That would replace older automation tools that required custom code or specialized software, and would let organizations automate work without hiring engineers to build custom integrations.

Why This Matters Now

The broader context here is worth considering. Google's move arrives while other major companies are pursuing similar capabilities. Anthropic, maker of Claude, recently added computer use features to its models. Microsoft has integrated AI agents into its Copilot ecosystem. The industry appears to be reaching a point where agentic AI is shifting from interesting research to practical tool.

The technical challenge that Google and its competitors still face is making these agents reliable and safe. An AI agent with access to your computer could delete files, make purchases you did not authorize, or leak sensitive information—even by accident. Google's approach of building agents into existing products, rather than releasing standalone agent software, suggests the company is thinking carefully about managing these risks.

Whether Google's consolidated strategy succeeds will depend on whether people actually want to use AI agents for their work, and whether the company can keep a good balance between giving the AI useful power and keeping humans in control. The technical work is now in place, but the harder part—getting people to trust and adopt this—is still ahead.