Curated by THEOUTPOST
On Thu, 10 Apr, 12:13 AM UTC
3 Sources
[1]
Google will let companies run its Gemini AI models in their own data centers
Thomas Kurian presents at the Google's I/O Developer's Conference in Mountain View, Calif, on May 10, 2023. Google's cloud unit said Wednesday that clients will be able to run its Gemini artificial intelligence models in their own data centers. Early access to Google Distributed Cloud will be available in the third quarter, Google said in a blog post. The service is targeted at clients that want to use Google's cloud technology while retaining control of their data. As part of the announcement, Google said Nvidia will bring Gemini models to the company's Blackwell graphics processing units, or GPUs. Companies can buy the chips through Google or other channels. OpenAI and Anthropic, two of Google's leading rivals in developing large language models, have shied away from providing access in physical data centers because it gives them less control over the quality and speed of the technology. Cohere has an option for customers to deploy models on their own infrastructure, but the AI startup says it's slower to set up than going through the company or using clouds. Google's gesture may be attractive to a new set of potential customers. Many companies, schools and governments still maintain their own data center hardware, although cloud services have become common in recent years. Even customers that adhere to the secret and top secret U.S. government classification levels will be able to use Gemini through an air-gapped version of Google Distributed Cloud that's disconnected from the internet. In 2023, cloud infrastructure spending totaled $140 billion, according to technology industry researcher Gartner. Google had 8% of the market in 2023, while Amazon controlled 39% and Microsoft held 23%, Gartner said. Last month, Google announced a $32 billion agreement to acquire cloud security startup Wiz. "Our commitment to multi-cloud, along with our investments in infrastructure and in AI, are some of the reasons we're seeing tremendous movement with customers," Google Cloud chief Thomas Kurian said on a conference call after the Wiz announcement. Google's Gemini models can process text, audio and video feeds, with support for more than 100 languages.
[2]
NVIDIA Brings Agentic AI Reasoning to Enterprises With Google Cloud
Google's Gemini models soon will be available on premises with Google Distributed Cloud running with NVIDIA Confidential Computing on NVIDIA Blackwell infrastructure. NVIDIA is collaborating with Google Cloud to bring agentic AI to enterprises seeking to locally harness the Google Gemini family of AI models using the NVIDIA Blackwell HGX and DGX platforms and NVIDIA Confidential Computing for data safety. With the NVIDIA Blackwell platform on Google Distributed Cloud, on-premises data centers can stay aligned with regulatory requirements and data sovereignty laws by locking down access to sensitive information, such as patient records, financial transactions and classified government information. NVIDIA Confidential Computing also secures sensitive code in the Gemini models from unauthorized access and data leaks. "By bringing our Gemini models on premises with NVIDIA Blackwell's breakthrough performance and confidential computing capabilities, we're enabling enterprises to unlock the full potential of agentic AI," said Sachin Gupta, vice president and general manager of infrastructure and solutions at Google Cloud. "This collaboration helps ensure customers can innovate securely without compromising on performance or operational ease." Confidential computing with NVIDIA Blackwell provides enterprises with the technical assurance that their user prompts to the Gemini models' application programming interface -- as well as the data they used for fine-tuning -- remain secure and cannot be viewed or modified. At the same time, model owners can protect against unauthorized access or tampering, providing dual-layer protection that enables enterprises to innovate with Gemini models while maintaining data privacy. This new offering arrives as agentic AI is transforming enterprise technology, offering more advanced problem-solving capabilities. Unlike AI models that perceive or generate based on learned knowledge, agentic AI systems can reason, adapt and make decisions in dynamic environments. For example, in enterprise IT support, while a knowledge-based AI model can retrieve and present troubleshooting guides, an agentic AI system can diagnose issues, execute fixes and escalate complex problems autonomously. Similarly, in finance, a traditional AI model could flag potentially fraudulent transactions based on patterns, but an agentic AI system could go even further by investigating anomalies and taking proactive measures such as blocking transactions before they occur or adjusting fraud detection rules in real time. While many can already use the models with multimodal reasoning -- integrating text, images, code and other data types to solve complex problems and build cloud-based agentic AI applications -- those with stringent security or data sovereignty requirements have yet been unable to do so. With this announcement, Google Cloud will be one of the first cloud service providers to offer confidential computing capabilities to secure agentic AI workloads across every environment -- whether cloud or hybrid. Powered by the NVIDIA HGX B200 platform with Blackwell GPUs and NVIDIA Confidential Computing, this solution will enable customers to safeguard AI models and data. This lets users achieve breakthrough performance and energy efficiency without compromising data security or model integrity. Scaling agentic AI in production requires robust observability and security to ensure reliable performance and compliance. Google Cloud today announced a new GKE Inference Gateway built to optimize the deployment of AI inference workloads with advanced routing and scalability. Integrating with NVIDIA Triton Inference Server and NVIDIA NeMo Guardrails, it offers intelligent load balancing that improves performance and reduces serving costs while enabling centralized model security and governance. Looking ahead, Google Cloud is working to enhance observability for agentic AI workloads by integrating NVIDIA Dynamo, an open-source library built to serve and scale reasoning AI models across AI factories. At Google Cloud Next, attend NVIDIA's special address, explore sessions, view demos and talk to NVIDIA experts.
[3]
NVIDIA and Google enable local deployment of Gemini AI models
Analyst uses a computer and dashboard for analysis of information on complex data. Google Cloud and NVIDIA have announced a significant collaboration that aims to reshape how enterprises deploy and scale artificial intelligence. The companies unveiled plans to bring agentic AI capabilities to highly regulated industries by allowing enterprises to run Google's Gemini language models locally on NVIDIA-powered infrastructure. The announcement, made at Google Cloud Next 2025, centers on enabling enterprises to securely deploy AI models on-premises using Google Distributed Cloud and NVIDIA's latest Blackwell architecture.
Share
Share
Copy Link
Google Cloud and NVIDIA collaborate to enable enterprises to run Gemini AI models locally using NVIDIA Blackwell GPUs, offering enhanced security and compliance for regulated industries.
In a significant move for enterprise AI adoption, Google Cloud has announced that it will allow companies to run its Gemini artificial intelligence models in their own data centers 1. This development, set to be available in early access during the third quarter, marks a strategic shift in how advanced AI models can be deployed and utilized by businesses with specific data control and security requirements.
Google's initiative is bolstered by a collaboration with NVIDIA, which will bring Gemini models to NVIDIA's cutting-edge Blackwell graphics processing units (GPUs) 2. This partnership aims to provide enterprises with the computational power necessary to run sophisticated AI models while maintaining control over their data and infrastructure.
The collaboration introduces the concept of "agentic AI" to enterprise environments. Unlike traditional AI models that simply perceive or generate based on learned knowledge, agentic AI systems can reason, adapt, and make decisions in dynamic environments 2. This capability opens up new possibilities for problem-solving and automation across various industries.
To address security concerns, the solution incorporates NVIDIA Confidential Computing, which provides a dual-layer protection mechanism. This allows enterprises to innovate with Gemini models while maintaining data privacy and protecting against unauthorized access or tampering of both user prompts and fine-tuning data 2.
This on-premises deployment option is particularly significant for highly regulated industries, government agencies, and organizations with strict data sovereignty requirements. It enables them to leverage advanced AI capabilities while complying with regulatory standards and maintaining control over sensitive information 13.
Google's move to offer on-premises deployment sets it apart from some competitors like OpenAI and Anthropic, which have been hesitant to provide such access due to concerns over quality control and performance 1. This strategy could potentially attract a new set of customers who prioritize data control and security.
The service will be available through Google Distributed Cloud, allowing for flexible deployment options including air-gapped versions for highly secure environments 1. NVIDIA's HGX B200 platform with Blackwell GPUs will power these on-premises installations, offering high performance and energy efficiency 2.
Google Cloud is also working on enhancing observability for agentic AI workloads by integrating NVIDIA Dynamo, an open-source library designed to serve and scale reasoning AI models 2. Additionally, Google has announced a new GKE Inference Gateway to optimize the deployment of AI inference workloads, integrating with NVIDIA Triton Inference Server and NeMo Guardrails for improved performance and governance 2.
As enterprises continue to explore the potential of AI, this collaboration between Google and NVIDIA represents a significant step towards making advanced AI models more accessible, secure, and compliant with various industry requirements.
Reference
[2]
[3]
Google has announced significant updates to its AI offerings, including the integration of Gemini 1.5 into enterprise contact centers and new AI-powered features for Google Workspace. These advancements aim to revolutionize customer engagement and boost productivity in the workplace.
9 Sources
9 Sources
Google introduces Gemini 2.5 Flash, a new AI model optimized for speed and efficiency, alongside updates to its AI ecosystem and agent technologies.
8 Sources
8 Sources
Google has announced the release of new Gemini models, showcasing advancements in AI technology. These models promise improved performance and capabilities across various applications.
2 Sources
2 Sources
Google introduces Gemini 2.0 Flash Thinking, an advanced AI model with enhanced reasoning capabilities, multimodal processing, and transparent decision-making, positioning it as a strong competitor in the AI landscape.
22 Sources
22 Sources
Google and NVIDIA announce a deepened collaboration to advance AI technology and its applications across various sectors, including cloud computing, robotics, drug discovery, and energy grid optimization.
2 Sources
2 Sources
The Outpost is a comprehensive collection of curated artificial intelligence software tools that cater to the needs of small business owners, bloggers, artists, musicians, entrepreneurs, marketers, writers, and researchers.
© 2025 TheOutpost.AI All rights reserved