OpenClaw's viral rise exposes users to prompt injection attacks and credential theft

Reviewed byNidhi Govil

25 Sources

Share

OpenClaw, the open-source AI assistant that exploded to 150,000 GitHub stars in months, has become a security nightmare. Researchers discovered 341 malicious skills on ClawHub stealing credentials, while exposed instances leak API keys and OAuth tokens. The platform's 770,000 AI agents on Moltbook face prompt injection attacks that could herald a new era of AI worms.

OpenClaw Emerges as Major Security Threat in AI Assistant Landscape

The rapid ascent of OpenClaw, an open-source AI assistant created by Austrian developer Peter Steinberger, has exposed critical vulnerabilities that security researchers are calling an "absolute nightmare" and a "security dumpster fire."

1

Since launching in November 2025, the project has accumulated over 150,000 GitHub stars, making it one of the fastest-growing AI open-source projects on the platform.

2

But this breakneck popularity has created an ecosystem ripe for exploitation, with researchers identifying hundreds of malicious skills and exposed instances leaking sensitive data across the internet.

Source: Inc.

Source: Inc.

Unlike cloud-based chatbots, OpenClaw runs locally on users' devices with deep system integration, connecting to messaging platforms like WhatsApp, Telegram, and Slack while performing autonomous tasks at regular intervals.

2

The AI assistant harnesses the power of Anthropic's Claude and OpenAI's ChatGPT models, but its organizing code runs on individual computers, granting it access to email, messages, and file systems. This architecture requires users to grant system-level controls and account permissions, creating what Cisco researchers describe as an extended attack surface where threat actors can craft malicious prompts that cause unintended behavior.

2

Malicious Skills and Supply-Chain Attacks Target OpenClaw Users

A security audit conducted by Koi Security revealed 341 malicious skills across multiple campaigns on ClawHub, the marketplace designed for OpenClaw users to find and install third-party extensions.

4

Of these, 335 skills use fake prerequisites to install Atomic Stealer (AMOS), a commodity macOS stealer available for $500-1000 per month that harvests data from infected hosts. The malicious skills masquerade as legitimate tools with professional-looking documentation, instructing users to download trojan files on Windows or execute obfuscated shell commands on macOS that fetch next-stage payloads from attacker-controlled infrastructure.

Source: 404 Media

Source: 404 Media

The problem stems from ClawHub being open by default, allowing anyone with a GitHub account at least one week old to upload skills. Security researcher Jamieson O'Reilly demonstrated how trivial it would be to backdoor a skill, publishing a minimal "ping" payload and artificially inflating its download count to become the most popular asset.

5

Within eight hours, 16 developers in seven countries had downloaded the artificially promoted skill, illustrating the ease of supply-chain attacks against the platform.

4

Exposed Instances Leak API Keys and Credentials Across the Internet

Pentester Jamieson O'Reilly discovered hundreds of OpenClaw Control admin interfaces exposed online due to reverse proxy misconfiguration.

5

Because OpenClaw auto-approves "local" connections, deployments behind reverse proxies often treat all internet traffic as trusted, allowing unauthenticated access to sensitive data. O'Reilly found instances with no authentication protection whatsoever, leaking Anthropic API keys, Telegram bot tokens, Slack OAuth credentials, signing secrets, and complete conversation histories.

2

Source: VentureBeat

Source: VentureBeat

OpenClaw has already been reported to have leaked plaintext API keys and credentials, which can be stolen by threat actors via prompt injection or unsecured endpoints.

2

The lack of sandboxing by default means the AI assistant has the same complete access to data as the user, with credentials stored in plaintext under ~/.clawdbot/.

5

Token Security claims that 22 percent of its enterprise customers have employees actively using OpenClaw, likely without IT approval, raising concerns about corporate data leakage via AI-mediated access.

5

Prompt Injection Attacks and the Rise of AI Worms

Researchers at Simula Research Laboratory identified 506 posts on Moltbook—the simulated social network where OpenClaw agents interact—containing hidden prompt injection attacks, representing 2.6 percent of sampled content.

1

Moltbook now hosts over 770,000 registered AI agents controlled by roughly 17,000 human accounts, creating the first large-scale network of semi-autonomous AI agents that can communicate through major communication apps.

1

Prompt injection attacks require an AI assistant to read and execute malicious instructions hidden in source web material or URLs, potentially causing the agent to leak sensitive data, send information to attacker-controlled servers, or execute tasks with the privileges it has been granted.

2

Security researchers have predicted the rise of self-replicating adversarial prompts among networks of AI agents—what might be called "prompt worms" or "prompt viruses"—that spread through networks of communicating AI agents similar to how traditional worms spread through computer networks.

1

Palo Alto Networks warned that OpenClaw represents what British programmer Simon Willison describes as a "lethal trifecta" that renders AI agents vulnerable by design due to their access to private data, exposure to untrusted content, and the ability to communicate externally.

4

With persistent memory, attacks are no longer just point-in-time exploits but become stateful, delayed-execution attacks where malicious payloads can be fragmented, written into long-term agent memory, and later assembled into executable instructions.

4

Unexpected Costs and Security Advisories Mount

Beyond security vulnerabilities, OpenClaw users are discovering unexpected financial burdens. Benjamin De Kraker, an AI specialist at The Naval Welding Institute, reported burning through $20 worth of Anthropic API tokens while he slept, simply by checking the time.

3

A "heartbeat" cron job set up to issue a reminder checked the time every 30 minutes, sending around 120,000 tokens of context to Claude Opus 4.5.2 model at approximately $0.75 per check, amounting to nearly $20 for 25 checks. The potential cost to run reminders over a month would be about $750.

3

In the past three days alone, the project has issued three high-impact security advisories: one one-click remote code execution vulnerability and two command injection vulnerabilities.

3

Creator Peter Steinberger has since rolled out a reporting feature allowing signed-in users to flag skills, with skills receiving more than three unique reports being auto-hidden by default.

4

However, deploying OpenClaw safely requires isolating the AI instance in a virtual machine and configuring firewall rules for internet access, rather than running it directly on the host OS with root access.

5

Today's Top Stories

TheOutpost.ai

Your Daily Dose of Curated AI News

Don’t drown in AI news. We cut through the noise - filtering, ranking and summarizing the most important AI news, breakthroughs and research daily. Spend less time searching for the latest in AI and get straight to action.

© 2026 Triveous Technologies Private Limited
Instagram logo
LinkedIn logo