Google and NVIDIA Partner to Bring Gemini AI Models On-Premises with Enhanced Security

3 Sources

Share

Google Cloud and NVIDIA collaborate to enable enterprises to run Gemini AI models locally using NVIDIA Blackwell GPUs, offering enhanced security and compliance for regulated industries.

News article

Google Brings Gemini AI Models to On-Premises Data Centers

In a significant move for enterprise AI adoption, Google Cloud has announced that it will allow companies to run its Gemini artificial intelligence models in their own data centers

1

. This development, set to be available in early access during the third quarter, marks a strategic shift in how advanced AI models can be deployed and utilized by businesses with specific data control and security requirements.

NVIDIA Collaboration Enhances AI Infrastructure

Google's initiative is bolstered by a collaboration with NVIDIA, which will bring Gemini models to NVIDIA's cutting-edge Blackwell graphics processing units (GPUs)

2

. This partnership aims to provide enterprises with the computational power necessary to run sophisticated AI models while maintaining control over their data and infrastructure.

Agentic AI and Enhanced Security Features

The collaboration introduces the concept of "agentic AI" to enterprise environments. Unlike traditional AI models that simply perceive or generate based on learned knowledge, agentic AI systems can reason, adapt, and make decisions in dynamic environments

2

. This capability opens up new possibilities for problem-solving and automation across various industries.

To address security concerns, the solution incorporates NVIDIA Confidential Computing, which provides a dual-layer protection mechanism. This allows enterprises to innovate with Gemini models while maintaining data privacy and protecting against unauthorized access or tampering of both user prompts and fine-tuning data

2

.

Implications for Regulated Industries

This on-premises deployment option is particularly significant for highly regulated industries, government agencies, and organizations with strict data sovereignty requirements. It enables them to leverage advanced AI capabilities while complying with regulatory standards and maintaining control over sensitive information

1

3

.

Market Position and Competition

Google's move to offer on-premises deployment sets it apart from some competitors like OpenAI and Anthropic, which have been hesitant to provide such access due to concerns over quality control and performance

1

. This strategy could potentially attract a new set of customers who prioritize data control and security.

Infrastructure and Deployment Options

The service will be available through Google Distributed Cloud, allowing for flexible deployment options including air-gapped versions for highly secure environments

1

. NVIDIA's HGX B200 platform with Blackwell GPUs will power these on-premises installations, offering high performance and energy efficiency

2

.

Future Developments and Enhancements

Google Cloud is also working on enhancing observability for agentic AI workloads by integrating NVIDIA Dynamo, an open-source library designed to serve and scale reasoning AI models

2

. Additionally, Google has announced a new GKE Inference Gateway to optimize the deployment of AI inference workloads, integrating with NVIDIA Triton Inference Server and NeMo Guardrails for improved performance and governance

2

.

As enterprises continue to explore the potential of AI, this collaboration between Google and NVIDIA represents a significant step towards making advanced AI models more accessible, secure, and compliant with various industry requirements.

TheOutpost.ai

Your Daily Dose of Curated AI News

Don’t drown in AI news. We cut through the noise - filtering, ranking and summarizing the most important AI news, breakthroughs and research daily. Spend less time searching for the latest in AI and get straight to action.

© 2025 Triveous Technologies Private Limited
Instagram logo
LinkedIn logo