AI Can Unmask Two-Thirds of Pseudonymous Accounts, Raising Alarm Over Online Privacy

Reviewed byNidhi Govil

2 Sources

Share

Research by Anthropic and ETH Zurich reveals that large language models can identify real-world identities behind pseudonymous accounts with alarming precision. The AI successfully unmasked two-thirds of users on platforms like Hacker News and Reddit by analyzing writing patterns and post content alone—a task that previously required hours of human investigation.

Large Language Models Enable Mass Deanonymization at Scale

A yet-to-be-peer-reviewed study conducted by researchers at ETH Zurich and AI company Anthropic has uncovered a troubling capability: large language models can perform AI deanonymization at scale, effectively ending the era of practical obscurity that has long protected pseudonymous accounts online. The research demonstrates that AI agents can "re-identify" users on popular platforms like Hacker News and Reddit based solely on their pseudonymous online profiles and conversations—work that would traditionally take hours for a dedicated human investigator

1

2

.

Source: Futurism

Source: Futurism

The results proved alarming: the AI agent successfully unmasked an astonishing two-thirds of users in their experiments. In some cases, the models achieved up to 68% recall with approximately 90% precision, meaning the AI correctly identified many accounts while maintaining relatively low error rates

2

. Conventional deanonymization methods in the same experiments achieved close to zero success, highlighting how dramatically AI has shifted the landscape.

How AI Agents Connect Real-World Identities to Anonymous Profiles

The research by Anthropic and ETH Zurich involved collecting datasets from public social media sites to test their deanonymization capabilities. Researchers linked Hacker News posts to LinkedIn profiles using references in user profiles, then anonymized the datasets by removing identifying information. They trained an LLM to link posts back to original authors by analyzing personal interests, demographic clues, writing style, and incidental details revealed in posts

1

2

.

"What we found is that these AI agents can do something that was previously very difficult: starting from free text (like an anonymized interview transcript) they can work their way to the full identity of a person," coauthor and ETH Zurich AI engineer Simon Lermen told Ars Technica. This represents a significant departure from previous approaches, which generally required structured data and two datasets with similar schemas that could be linked together

1

.

Even when fed extremely general data, like responses to an Anthropic questionnaire about how people use AI in their daily lives, the LLM could identify people approximately seven percent of the time. When analyzing comments from various movie communities on Reddit, the AI demonstrated an astonishing rate of precision in linking pseudonymous online profiles—the more users discussed movies, the easier it became for the system to deanonymize them

1

.

Threats to Online Privacy and the Cost of Surveillance

The implications for online privacy are considerable, as researchers estimate the cost of identifying an online account using their experimental pipeline could fall between $1 and $4 per profile. This means mass expose of anonymous internet accounts could be conducted relatively cheaply, dramatically lowering the barrier to large-scale surveillance

2

.

"The average online user has long operated under an implicit threat model where they have assumed pseudonymity provides adequate protection because targeted deanonymization would require extensive effort," the researchers wrote. "LLMs invalidate this assumption." The study warns that governments could link pseudonymous accounts to real identities for surveillance of dissidents, journalists, or activists. Corporations could connect seemingly anonymous forum posts to customer profiles for hyper-targeted advertising, while attackers could build sophisticated profiles of targets at scale to launch highly personalized social engineering scams

1

.

What This Means for Online Anonymity and Future Protections

The research highlights how AI democratizes deanonymization capabilities that were once limited to well-resourced investigators. "Across Hacker News, Reddit, LinkedIn, and anonymized interview transcripts, our method identifies users with high precision -- and scales to tens of thousands of candidates," Lermen wrote in a blog post accompanying the paper

1

.

While the authors acknowledge limitations—including small sample sets that require verified identity links and difficulty distinguishing what the LLM gathered from web search versus search engine embeddings—they maintain their findings paint an alarming picture. "Users, platforms, and policymakers must recognize that the privacy assumptions underlying much of today's internet no longer hold," the paper states

1

.

Looking ahead, researchers suggest that individuals may need to rethink how much personal information they reveal online, even in spaces that appear anonymous. Potential solutions could include improved privacy tools, stronger platform safeguards, or AI systems designed to anonymize sensitive data before it is shared publicly

2

. The advent of AI has ushered in a new era that calls for enhanced safety measures—or could potentially signal the death knell of online pseudonymity as we know it.

Today's Top Stories

TheOutpost.ai

Your Daily Dose of Curated AI News

Don’t drown in AI news. We cut through the noise - filtering, ranking and summarizing the most important AI news, breakthroughs and research daily. Spend less time searching for the latest in AI and get straight to action.

© 2026 Triveous Technologies Private Limited
Instagram logo
LinkedIn logo