3 Sources
[1]
The modern data stack was built for humans asking questions. Google just rebuilt its for agents taking action.
Enterprise data stacks were built for humans running scheduled queries. As AI agents increasingly act autonomously on behalf of businesses around the clock, that architecture is breaking down -- and vendors are racing to rebuild it. Google's answer, announced at Cloud Next on Wednesday, is the Agentic Data Cloud. The architecture has three pillars: * Knowledge Catalog. Automates semantic metadata curation, inferring business logic from query logs without manual data steward intervention * Cross-cloud lakehouse. Lets BigQuery query Iceberg tables on AWS S3 via private network with no egress fees * Data Agent Kit. Drops MCP tools into VS Code, Claude Code and Gemini CLI so data engineers describe outcomes rather than write pipelines "The data architecture has to change now," Andi Gutmans, VP and GM of Data Cloud at Google Cloud, told VentureBeat. "We're moving from human scale to agent scale." From system of intelligence to system of action The core premise behind Agentic Data Cloud is that enterprises are moving from human‑scale to agent‑scale operations. Historically, data platforms have been optimized for reporting, dashboarding, and some forecasting -- what Google characterizes as "reactive intelligence." In that model, humans interpret data and decide what to do. Now, with AI agents increasingly expected to take actions directly on behalf of the business, Gutmans argued that data platforms must evolve into systems of action. "We need to make sure that all of enterprise data can be activated with AI, that includes both structured and unstructured data," Gutmans said. "We need to make sure that there's the right level of trust, which also means it's not just about getting access to the data, but really understanding the data." The Knowledge Catalog is Google's answer to that problem. It is an evolution of Dataplex, Google's existing data governance product, with a materially different architecture underneath. Where traditional data catalogs required data stewards to manually label tables, define business terms and build glossaries, the Knowledge Catalog automates that process using agents. The practical implication for data engineering teams is that the Knowledge Catalog scales to the full data estate, not just the curated subset that a small team of data stewards can maintain by hand. The catalog covers BigQuery, Spanner, AlloyDB and Cloud SQL natively, and federates with third-party catalogs including Collibra, Atlan and Datahub. Zero-copy federation extends semantic context from SaaS applications including SAP, Salesforce Data360, ServiceNow and Workday without requiring data movement. Google's lakehouse goes cross cloud Google has had a data lakehouse called BigLake since 2022. Initially it was limited to just Google data, but in recent years has had some limited federation capabilities enabling enterprises to query data found in other locations. Gutmans explained that the previous federation worked through query APIs, which limited the features and optimizations BigQuery could bring to bear on external data. The new approach is storage-based sharing via the open Apache Iceberg format. That means whether the data is in Amazon S3 or in Google Cloud , he argued it doesn't make a difference. "This truly means we can bring all the goodness and all the AI capabilities to those third-party data sets," he said. The practical result is that BigQuery can query Iceberg tables sitting on Amazon S3 via Google's Cross-Cloud Interconnect, a dedicated private networking layer, with no egress fees and price-performance Google says is comparable to native AWS warehouses. All BigQuery AI functions run against that cross-cloud data without modification. Bidirectional federation in preview extends to Databricks Unity Catalog on S3, Snowflake Polaris and the AWS Glue Data Catalog using the open Iceberg REST Catalog standard. From writing pipelines to describing outcomes The Knowledge Catalog and cross-cloud lakehouse solve the data access and context problems. The third pillar addresses what happens when a data engineer actually sits down to build something with all of it. The Data Agent Kit ships as a portable set of skills, MCP tools and IDE extensions that drop into VS Code, Claude Code, Gemini CLI and Codex. It does not introduce a new interface. The architectural shift it enables is a move from what Gutmans called a "prescriptive copilot experience" to intent-driven engineering. Rather than writing a Spark pipeline to move data from source A to destination B, a data engineer describes the outcome -- a cleaned dataset ready for model training, a transformation that enforces a governance rule -- and the agent selects whether to use BigQuery, the Lightning Engine for Apache Spark or Spanner to execute it, then generates production-ready code. "Customers are kind of sick of building their own pipelines," Gutmans said. "They're truly more in the review kind of mode, than they are in the writing the code mode." Where Google and its rivals diverge The premise that agents require semantic context, not just data access, is shared across the market. Databricks has Unity Catalog, which provides governance and a semantic layer across its lakehouse. Snowflake has Cortex, its AI and semantic layer offering. Microsoft Fabric includes a semantic model layer built for business intelligence and, increasingly, agent grounding. The dispute is not over whether semantics matter -- everyone agrees they do. The dispute is over who builds and maintains them. "Our goal is just to get all the semantics you can get," he explained, noting that Google will federate with third-party semantic models rather than require customers to start over. Google is also positioning openness as a differentiator, with bidirectional federation into Databricks Unity Catalog and Snowflake Polaris via the open Iceberg REST Catalog standard. What this means for enterprises Google's argument -- and one echoed across the data infrastructure market -- is that enterprises are behind on three fronts: Semantic context is becoming infrastructure. If your data catalog is still manually curated, it will not scale to agent workloads -- and Gutmans argues that gap will only widen as agent query volumes increase. Cross-cloud egress costs are a hidden tax on agentic AI. Storage-based federation via open Iceberg standards is emerging as the architectural answer across Google, Databricks and Snowflake. Enterprises locked into proprietary federation approaches should be stress-testing those costs at agent-scale query volumes. Gutmans argues the pipeline-writing era is ending. Data engineers who move toward outcome-based orchestration now will have a significant head start.
[2]
Google Cloud databases power the agentic enterprise - SiliconANGLE
Google Cloud databases are being rebuilt for the age of AI agents AI agents depend on the quality and accessibility of the data behind them, placing Google Cloud databases at the center of that equation. The agentic data cloud introduced at Google Cloud Next 2026 underscores Google's push to position its databases as the foundation of AI-driven enterprise architecture. Models are advancing, but without enterprise data, they lack real value, according to Sailesh Krishnamurthy (pictured), vice president of engineering for databases at Google Cloud. "The models are amazing. The models surprise us every day, they can do a lot of work, but they don't have all the context," Krishnamurthy said. "The context is in the data. The heart of the data is actually stored in these systems. You need to provide that context in order to answer the questions." Krishnamurthy spoke with theCUBE's John Furrier at Google Cloud Next, during an exclusive broadcast on theCUBE, SiliconANGLE Media's livestreaming studio. They discussed how Google Cloud databases are evolving from passive storage systems into intelligent context hubs that power agentic AI applications. (* Disclosure below.) The shift reshaping database architecture is not merely one of scale -- it is one of purpose. For 50 years, databases had one job: store data and return exact results on demand, Krishnamurthy noted. AI breaks that contract entirely. Today's applications need the best results, not just exact ones, demanding that graph traversal, vector embeddings, full-text search and relational operations coexist in a single system rather than forcing costly, unnecessary data movement. "When you have this opportunity to look at data as a graph, look at data with vector embeddings, do semantic search or full-text search -- all of a sudden, it's not about getting the exact results, but getting the best results and the best quality," Krishnamurthy said. "You don't need to unnecessarily move the data just to organize it in a different way. I think [that's] the big change for databases." Google also announced Spanner Omni, a downloadable edition of its globally distributed database that can run on-premises or across rival clouds, extending Google-scale infrastructure to wherever enterprise data lives. Agentic migration tooling powered by Gemini is also accelerating the path for organizations looking to move existing workloads, Krishnamurthy explained. Migration agents can now handle not just schema and data but the application layer -- including embedded SQL queries -- dramatically compressing timelines that once required months of manual effort. "Today, what you can do with agents is you can dramatically change how fast you can migrate your systems," he said. "It's not just the database. When you think about database migration, you have schemas and you have data, but you have the application with complexity -- the application has SQL queries embedded in it. Today, with the power of Gemini, we are excited that people are able to migrate their whole application stack so much faster." Here's for the complete video interview, part of SiliconANGLE's and theCUBE's coverage of Google Cloud Next.
[3]
Google delivers connective tissue for autonomous AI agents to access data without restrictions - SiliconANGLE
Google delivers connective tissue for autonomous AI agents to access data without restrictions Google Cloud is turning the traditional enterprise data platform on its head, unveiling the Agentic Data Cloud infrastructure platform that aims to act as a kind of central nerve center for the era of artificial intelligence agents. In a blog post, Andi Gutmans, Google's vice president and general manager of Data Cloud, explains that existing data infrastructures were designed to act as "static repositories," where information just sits until it's asked a question by a human. But in the era of AI, this kind of "human-scale" infrastructure is no longer fit for purpose. To that end, Google has designed the Agentic Data Cloud to work as a "system of action" that evolves data infrastructure into a dynamic reasoning engine that enables autonomous agents to get to work, rather than just think about the problems they're trying to solve. Announced at Google Cloud Next 2026 this week in Las Vegas, the Agentic Data Cloud will provide the connective tissue AI agents need to work across the enterprise without hindrance, and it's built on three main pillars: a universal context engine that aims to prevent agents from "hallucinating," a suite of agentic-first developer tools, and the cross-cloud lakehouse platform that unifies data from across any cloud environment. According to Gutmans, one of the biggest hurdles with deploying AI agents today is the so-called "context gap." If an agent doesn't understand a company's specific definition of what something like "gross margin" actually means, it's probably going to end up making expensive mistakes. To fix this, Google has evolved its Dataplex Universal Catalog into the Knowledge Catalog, which is a kind of map of business meaning that's meant to inform AI agents of the peculiarities of the organization they serve. The catalog scans all of a company's documents, including its accounts, PDFs, PowerPoint presentations and images, extracting entities and studying the relationships within them to build a navigable schema that agents can use. Also helping with this is BigQuery Measures and a new LookML Agent that will help to bake business logic into the entire Agentic Data Cloud stack. By aggregating all of these metrics into a single, governed data foundation, Google says that when an AI agent queries company data, it will use the same "source of truth" each time. This new context engine is already powering Google's new Deep Research Agent, enabling it to perform multistep reasoning across web assets and internal documents to create complex research reports that would take human analysts weeks. The lives of developers are being made easier, too. The company has announced a new Google Cloud Data Agents Kit that brings "agentic skills" directly into the tools developers already use, including platforms such as Claude Code and VS Code. With the Data Agent Kit, developer environments can autonomously orchestrate outcomes, including selecting frameworks such as Apache Spark or dbt, while generating production-ready code based on Google's best practices. Three new, highly specialized AI agents were also announced to make life easier for developers. They include a new Data Engineering agent for building and governing complex data transformations, a Data Science agent for automating AI model lifecycles across BigQuery and Spark, and a Database Observability agent that acts like a "guardian," tasked with diagnosing and repairing data infrastructure issues. Gutmans said Google has embraced the Model Context Protocol to ensure these agents play nicely with one another. "[It] provides a secure, universal interface that allows any agent to safely discover and use your data assets across our core engines, including: BigQuery, Spanner (Preview), AlloyDB, Cloud SQL (GA) and Looker MCP (Preview)," he said. "MCP for Google Cloud uses our security stack, governing agent interactions based on your existing IAM policies, VPC Service Controls, and data residency requirements." Finally, Google is trying to address the problem of AI agent "gravity." This refers to how agents lose their autonomy when they're slowed down by cross-cloud latency or prevented from accessing data trapped in other cloud platforms. Gutmans introduced the new "cross-cloud Lakehouse," which aims to provide a borderless data environment for AI agents. It integrates with Google's Cross-Cloud Interconnect service directly into the data plane, and employs the Apache Iceberg REST catalog to connect to the Amazon Web Services and Microsoft Azure clouds. What this means is that AI agents can treat data stored in Azure data lake or in an S3 bucket as if it were sitting locally in Google Cloud, without the usual headaches associated with data migration and egress fees. To aid data mobility further, Google also introduced bi-directional federation capabilities for Databricks Unity Catalog, Snowflake Polaris and AWS Glue to break down proprietary data siloes. It's also unchaining its Spanner Omni database, allowing it to run on-premises or in rival clouds.
Share
Copy Link
Google Cloud announced Agentic Data Cloud at Cloud Next 2026, transforming enterprise data stacks from human-scale reporting systems into agent-scale systems of action. The platform features Knowledge Catalog for automated semantic metadata curation, a cross-cloud lakehouse that queries data across AWS and Azure without egress fees, and Data Agent Kit that enables developers to describe outcomes rather than write code.
Google Cloud announced its Agentic Data Cloud at Cloud Next 2026 on Wednesday, marking a fundamental shift in how enterprise data stacks are architected
1
. The platform addresses a critical reality: traditional data infrastructure was built for humans running scheduled queries, but as AI agents increasingly act autonomously on behalf of businesses around the clock, that architecture is breaking down1
. "The data architecture has to change now," Andi Gutmans, VP and GM of Data Cloud at Google Cloud, told VentureBeat. "We're moving from human scale to agent scale."1
The Agentic Data Cloud represents Google's push to position its databases as the foundation of AI-driven enterprise architecture, transforming them from passive storage systems into intelligent context hubs that power agentic AI applications
2
. The platform is built on three core pillars designed to act as connective tissue for autonomous AI agents to access data without restrictions across any cloud environment3
.
Source: VentureBeat
One of the biggest hurdles with deploying autonomous AI agents today is the context gap—if an agent doesn't understand a company's specific definition of metrics like "gross margin," it will likely make expensive mistakes
3
. The Knowledge Catalog, an evolution of Google's Dataplex governance product, automates semantic metadata curation by inferring business logic from query logs without manual data steward intervention1
.This automated approach scales to the full data estate, not just the curated subset that a small team of data stewards can maintain by hand
1
. The catalog scans all of a company's documents—including accounts, PDFs, PowerPoint presentations and images—extracting entities and studying relationships to build a navigable schema that agents can use3
. It covers BigQuery, Spanner, AlloyDB and Cloud SQL natively, and federates with third-party catalogs including Collibra, Atlan and Datahub1
.The cross-cloud lakehouse addresses what Google calls AI agent "gravity"—how agents lose autonomy when slowed by cross-cloud latency or prevented from accessing data trapped in other cloud platforms
3
. Using the open Apache Iceberg format, BigQuery can now query Iceberg tables sitting on Amazon S3 via Google's Cross-Cloud Interconnect with no egress fees and price-performance comparable to native AWS warehouses1
.
Source: SiliconANGLE
Gutmans explained that previous data federation worked through query APIs, which limited the features and optimizations BigQuery could apply to external data. The new storage-based sharing approach means "whether the data is in Amazon S3 or in Google Cloud, it doesn't make a difference"
1
. Bidirectional data federation in preview extends to Databricks Unity Catalog on S3, Snowflake Polaris and the AWS Glue Data Catalog using the open Iceberg REST Catalog standard1
.Related Stories
The Data Agent Kit ships as a portable set of skills, MCP tools and IDE extensions that drop into VS Code, Claude Code, Gemini CLI and Codex
1
. Rather than writing a Spark pipeline to move data from source A to destination B, data engineers describe the outcome—a cleaned dataset ready for model training or a transformation that enforces a governance rule—and the agent selects whether to use BigQuery, the Lightning Engine for Apache Spark or Spanner to execute it, then generates production-ready code1
."Customers are kind of sick of building their own pipelines," Gutmans said. "They're truly more in the review kind of mode, than they are in the writing the code."
1
Three specialized AI agents were announced: a Data Engineering agent for building and governing complex data transformations, a Data Science agent for automating AI model lifecycles across BigQuery and Spark, and a Database Observability agent for diagnosing and repairing data infrastructure issues3
.For 50 years, databases had one job: store data and return exact results on demand, noted Sailesh Krishnamurthy, vice president of engineering for databases at Google Cloud
2
. AI agents break that contract entirely. Today's applications need the best results, not just exact ones, demanding that graph traversal, vector embeddings, full-text search and relational operations coexist in a single system2
.
Source: SiliconANGLE
"The models are amazing, but they don't have all the context," Krishnamurthy said. "The context is in the data. The heart of the data is actually stored in these systems of action."
2
Google also announced Spanner Omni, a downloadable edition of its globally distributed database that can run on-premises or across rival clouds2
. Agentic migration tooling powered by Gemini now handles not just schema and data but the application layer—including embedded SQL queries—dramatically compressing timelines that once required months of manual effort2
.The shift from human-scale to agent-scale operations demands infrastructure that enables semantic search, supports vector embeddings, and provides the governed context necessary for AI agents to take action rather than simply answer questions. By addressing data access, business context, and developer productivity simultaneously, Google Cloud is positioning the Agentic Data Cloud as the foundation for agent-centric action-oriented platforms that can operate autonomously across multi-cloud environments.
Summarized by
Navi
[1]
06 Aug 2025•Technology

30 Apr 2026•Technology

22 Apr 2026•Technology

1
Technology

2
Health

3
Policy and Regulation
