AI chatbots are exposing real phone numbers and addresses, sparking privacy alarm

Reviewed byNidhi Govil

4 Sources

Share

AI chatbots like Google Gemini and ChatGPT are revealing people's real phone numbers, home addresses, and other personal details without consent. Privacy experts warn that personally identifiable information in training data is creating new risks for AI doxxing, with DeleteMe reporting a 400% increase in customer complaints about generative AI exposing personal information in just seven months.

AI Chatbots Giving Out Phone Numbers Creates Privacy Crisis

AI chatbots are revealing people's real phone numbers and home addresses, creating what privacy experts are calling a new form of AI doxxing. A Reddit user reported receiving constant unwanted calls from strangers "looking for a lawyer, a product designer, a locksmith" after Google's generative AI allegedly started distributing their personal number

1

. In March, an Israeli software developer found his WhatsApp flooded with messages after Google Gemini incorrectly listed his number as customer service for a payment app

1

. Meanwhile, a University of Washington PhD candidate discovered that Google Gemini readily provided her colleague's private cell phone number when prompted

1

.

Source: MIT Tech Review

Source: MIT Tech Review

The scope of this problem extends far beyond isolated incidents. DeleteMe, a data removal company, reports that customer queries about generative AI exposing personal information have surged by 400% in the last seven months, reaching several thousand cases

1

. Among these privacy concerns, 55% specifically reference ChatGPT, 20% mention Google Gemini, 15% involve Claude, and 10% relate to other AI tools

1

. Rob Shavell, DeleteMe's cofounder and CEO, says complaints typically fall into two categories: customers asking chatbots innocuous questions about themselves and receiving accurate home addresses, phone numbers, family members' names, or employer details, or discovering that AI chatbots reveal personal information about others, sometimes generating "plausible-but-wrong contact information"

1

.

How AI Chatbots Expose Personally Identifiable Information

Experts believe these privacy lapses stem from personally identifiable information being embedded in training data, though the exact mechanisms remain difficult to pinpoint

1

. A 2025 Cornell University study revealed that at least five leading AI companies—Anthropic, Google, Meta, Microsoft, and OpenAI—automatically use users' inputs to train their chatbots unless users opt out

2

. Even more concerning, Meta and OpenAI retain user data indefinitely, meaning these AI models train not just on publicly available information like old phone books, but also on information users shared with chatbots years ago

2

.

Source: CNET

Source: CNET

CNET staffers conducted tests to see how easily AI chatbots reveal personal information, with alarming results

2

. Grok proved the most willing to share data, pulling multiple present and past addresses within seconds, though it added a note stating "Home addresses are private; I recommend contacting him through professional channels"

2

. ChatGPT provided one CNET staffer with an old landline phone number and easily revealed a relative's cellphone number and address

2

. In a separate test, journalist Matt Novak discovered that ChatGPT accurately delivered a real phone number he hadn't used in years, apparently pulled from a 2016 FOIA request PDF

3

. Google Gemini and Claude showed more restraint, refusing to provide personal information in most tests

2

.

Fraudsters Exploit AI Chatbots for Scams

The data privacy risks extend beyond accidental exposure. Virgin Media O2 reported that scammers are deliberately planting fake customer-service numbers online for AI chatbots to regurgitate to users

4

. "AI tools are creating new opportunities for fraudsters to create realistic-looking fake numbers that appear through search results or chatbots, putting people at risk of calling a criminal rather than their trusted provider," warned Murray Mackenzie, the company's fraud prevention director

4

.

Source: New York Post

Source: New York Post

Researchers at AI security company Aurascape explain that attackers accomplish this through "seeding poisoned content" across the web. "Attackers are quietly rewriting the web that AI systems read," said lead security researcher Qi Deng. "When you ask an assistant how to call your airline, it does exactly what it was designed to do, but with a customer support and reservations number that leads straight to a scammer instead of the real company"

4

. This manipulation of publicly available information creates a dangerous environment where users trust AI-generated results that lead them directly to fraudsters.

Limited Options for Victims

For those affected by AI doxxing, recourse remains frustratingly limited. The Reddit user whose number was distributed reported that "standard support forms are a complete dead end" and hadn't received a single response while the harassment continued daily

4

. A Google spokesperson told MIT Technology Review that the company has safeguards to prevent personal information from appearing in AI features and reviews removal requests, but victims report difficulty getting help .

A ClearNym spokesperson emphasized that "Gemini's problem is not a defect. It's the result of unchecked years of data brokerage practices that meet generative AI," noting that years of harvested personal data now collide with AI systems trained on massive internet datasets

4

. University of Washington researcher Yael Eiger highlighted the shift in accessibility: while her information technically existed online before, it was buried deep enough that almost nobody would find it. "Having your information be accessible to one audience, and then Gemini making it accessible to anyone" creates an entirely different privacy landscape

4

. As AI systems become more sophisticated, experts warn that the intersection of data brokerage practices and generative AI will continue creating new vulnerabilities that individuals have little power to control.🟡 Atkinson),

Today's Top Stories

TheOutpost.ai

Don’t drown in AI news. We cut through the noise - filtering, ranking and summarizing the most important AI news, breakthroughs and research daily. Spend less time searching for the latest in AI and get straight to action.

Instagram logo
LinkedIn logo
Youtube logo
© 2026 TheOutpost.AI All rights reserved