Open-source AI models vulnerable to criminal misuse as hundreds operate without guardrails

4 Sources

Share

A 293-day study by SentinelOne and Censys exposes widespread criminal misuse of open-source AI models, with hundreds of deployments stripped of safety features. Researchers found 7.5% of system prompts could enable harmful activity, including phishing, disinformation campaigns, and fraud. The findings reveal an accountability gap in the open-source AI ecosystem.

News article

Open-Source AI Models Face Criminal Exploitation

Thousands of computers running open-source AI models are operating outside the security constraints of major artificial intelligence platforms, creating significant vulnerabilities that criminals can exploit, according to new research from cybersecurity companies SentinelOne and Censys. The study, conducted over 293 days and shared exclusively with Reuters, reveals that open-source AI models vulnerable to criminal misuse represent an overlooked threat in the AI security landscape.

Hackers can commandeer these systems to conduct spam operations, phishing content creation, or disinformation campaigns, evading the platform security protocols that govern mainstream AI services

2

. The research offers a troubling window into the scale of potentially illicit activities enabled by thousands of open-source LLM deployments, including hacking, hate speech and harassment, violent content, data theft, fraud, and in some cases child sexual abuse material

3

.

Hundreds of Models Deployed Without Guardrails

While thousands of open-source LLM variants exist—including significant deployments of Meta Llama and Google Gemma—researchers identified hundreds of instances where guardrails were explicitly removed. The study analyzed publicly accessible deployments through Ollama, a tool that allows people and organizations to run their own versions of various large-language models.

Researchers were able to examine system prompts, the instructions that dictate how models behave, in roughly a quarter of the LLMs they observed. Of those, they determined that 7.5% could potentially enable harmful activity

4

. Roughly 30% of the hosts observed are operating out of China, while about 20% are in the U.S., highlighting the global scope of this challenge.

An Iceberg of Potential Misuse Lacking Crucial Security Features

Juan Andres Guerrero-Saade, executive director for intelligence and security research at SentinelOne, characterized AI industry conversations about security controls as "ignoring this kind of surplus capacity that is clearly being utilized for all kinds of different stuff, some of it legitimate, some obviously criminal". He likened the situation to an iceberg of potential misuse that isn't being properly accounted for across the industry and open-source community.

This gap in oversight matters because it exposes a fundamental tension in the AI ecosystem between innovation and security. While open-source models democratize access to AI capabilities, they also create opportunities for criminal misuse when deployed without adequate safeguards.

Shared Responsibility Across the AI Ecosystem

Rachel Adams, CEO and founder of the Global Center on AI Governance, emphasized that once open models are released, shared responsibility for what happens next extends across the ecosystem, including the originating labs

2

. "Labs are not responsible for every downstream misuse (which are hard to anticipate), but they retain an important duty of care to anticipate foreseeable harms, document risks, and provide mitigation tooling and guidance, particularly given uneven global enforcement capacity," Adams noted.

Meta declined to respond to questions about developers' responsibilities for addressing downstream abuse but pointed to its Llama Protection tools and Meta Llama Responsible Use Guide. Microsoft AI Red Team Lead Ram Shankar Siva Kumar acknowledged that while open-source models "play an important role," the company remains "clear-eyed that open models, like all transformative technologies, can be misused by adversaries if released without appropriate safeguards." Microsoft performs pre-release evaluations to assess risks for internet-exposed, self-hosted scenarios where criminal misuse can be high

3

.

What This Means for AI Security

The findings raise urgent questions about how the industry balances open innovation with security. As deployers continue to run models without guardrails, the potential for illicit activities grows. The research suggests that cybersecurity measures need to extend beyond the major platforms to address the thousands of self-hosted deployments that currently operate in a regulatory gray zone. Observers should watch for potential policy responses, increased scrutiny of tools like Ollama, and whether major AI labs will implement stricter controls or monitoring for downstream use of their open-source models.

Today's Top Stories

TheOutpost.ai

Your Daily Dose of Curated AI News

Don’t drown in AI news. We cut through the noise - filtering, ranking and summarizing the most important AI news, breakthroughs and research daily. Spend less time searching for the latest in AI and get straight to action.

© 2026 Triveous Technologies Private Limited
Instagram logo
LinkedIn logo