Skip to main content
Daniel J Glover
Back to Blog

ClawdBot Security Risks Explained

9 min read

ClawdBot security risks have dominated cybersecurity headlines this week - and for good reason. The open-source personal AI assistant went viral almost overnight, racking up over 20,000 GitHub stars in a single day and spawning a wave of enthusiastic adopters buying Mac Minis purely to host their own instances. Within days, security researchers found hundreds of exposed instances leaking credentials, API keys, and private conversations to the open internet.

I run ClawdBot myself. I use it daily for email triage, research, and task automation. So this is not a hit piece from someone who has never touched the tool. It is an honest assessment from an IT leader who sees both the genuine productivity gains and the very real risks that most adopters are glossing over.

The broader lesson here extends well beyond one tool. ClawdBot - now rebranded to Moltbot following a trademark request from Anthropic - is the first viral example of a pattern that will define 2026: personal AI agents with deep system access, running on consumer hardware, managed by people who may not fully grasp the security implications.


What ClawdBot Actually Does

For those unfamiliar, ClawdBot is an open-source AI assistant that connects large language models to your personal infrastructure. You control it through messaging apps like Telegram, WhatsApp, or Signal. It can read and respond to emails, manage calendars, browse the web, execute shell commands, and interact with dozens of services through its extensible skills system.

The appeal is obvious. Instead of switching between fifteen apps and manually processing information, you delegate to an always-on agent that handles the grunt work. It is genuinely useful.

But "always-on agent with access to everything" is also a perfect description of a security nightmare. To function, ClawdBot needs credentials for your email, messaging platforms, calendar, and any other service you connect. It stores conversation history, user profiles, and operational state locally. It can execute commands on its host machine.

This is not a chatbot. It is infrastructure.


Five Security Risks That Matter

1. Exposed Instances and Misconfiguration

The most immediate ClawdBot security risk is basic misconfiguration. Jamieson O'Reilly, founder of red-teaming firm Dvuln, discovered hundreds of ClawdBot instances exposed to the public internet via Shodan scans. Of those he examined manually, eight had zero authentication - full administrative access available to anyone who found them.

The root cause was a classic localhost trust assumption. A combination of reverse proxy setups and default configurations caused remote connections to be treated as local, bypassing authentication entirely. While the developers have since patched the specific vulnerability O'Reilly reported, the damage illustrates a persistent pattern: consumer-friendly installation that masks enterprise-grade configuration requirements.

Eric Schwake, Director of Cybersecurity Strategy at Salt Security, put it bluntly: "A significant gap exists between the consumer enthusiasm for ClawdBot's one-click appeal and the technical expertise needed to operate a secure agentic gateway."

2. Plaintext Secrets on the Filesystem

Hudson Rock's research team published a detailed analysis of how ClawdBot stores sensitive data. The findings should concern every security professional.

ClawdBot maintains persistent state in directories like ~/.clawdbot/ and ~/clawd/. Unlike encrypted browser credential stores or OS keychains, these files - Markdown documents, JSON configuration files, conversation logs - are readable by any process running under the same user account.

This means API keys, gateway tokens, VPN credentials, and session data all sit in plaintext files. Hudson Rock documented that the gateway token alone, if compromised, could allow remote code execution on the host machine.

The comparison to the Change Healthcare breach of 2024 is apt. That $22 million ransomware payout traced back to a single compromised VPN credential found on an infected machine. If an AI agent had stored those same credentials in a plaintext memory file, the attack surface would have been even more accessible.

3. Cognitive Context Theft

This is the genuinely novel threat. Hudson Rock coined the term "Cognitive Context Theft" to describe what happens when infostealers target AI agent memory files.

Traditional credential theft gives attackers passwords and tokens. Stealing an AI agent's memory files gives them something far more valuable: a psychological dossier of the user. ClawdBot's MEMORY.md and daily log files contain what the user is working on, who they trust, their private concerns, their communication patterns, and their decision-making context.

This is social engineering gold. An attacker with access to someone's AI memory files could craft phishing attacks of unprecedented specificity. They would know exactly which colleagues the target trusts, which projects they are anxious about, and which communication style would bypass their suspicion.

Major malware-as-a-service families are already adapting. Hudson Rock reports that RedLine Stealer, Lumma, and Vidar are developing specific targeting for ClawdBot's directory structure and file naming conventions.

4. Supply Chain Poisoning via ClawdHub

O'Reilly's second disclosure was arguably more alarming than the exposed instances. He demonstrated a proof-of-concept supply chain attack against ClawdHub - the community skills library where users download extensions for their ClawdBot instances.

He uploaded a benign skill, artificially inflated its download count to over 4,000, and watched developers from seven countries install it. The skill executed code on their ClawdBot instances. Had it been malicious, it could have exfiltrated SSH keys, AWS credentials, and entire codebases.

ClawdHub's developer notes state that downloaded code is treated as trusted - there is no moderation or vetting process. This mirrors the slopsquatting risks we see in traditional package registries, but with an even more permissive trust model. At least npm and PyPI have basic malware scanning. ClawdHub, at the time of writing, has none.

5. Memory Poisoning and Persistent Backdoors

If an attacker gains write access to a ClawdBot instance - whether through an exposed admin panel, an infostealer with RAT capabilities, or a poisoned skill - they can do something uniquely dangerous: alter the AI's behaviour permanently.

By modifying core files like SOUL.md (which defines the agent's personality and rules) or injecting false information into MEMORY.md, an attacker can instruct the agent to trust malicious domains, exfiltrate future data, or behave in ways that serve the attacker's interests.

This effectively turns the user's personal AI assistant into a persistent insider threat - one that the user continues to trust because it looks and behaves normally on the surface. As I discussed in yesterday's analysis of AI agents as the new insider threat, this class of attack is particularly difficult to detect because the compromised entity has legitimate access to everything.


The Expertise Gap Problem

The fundamental tension with ClawdBot is that installation feels like downloading a consumer app, but safe operation requires the mindset of a systems administrator.

Prompt Security's analysis highlights this gap clearly: "The strongest outcomes tend to come from users who already understand these tradeoffs. As autonomy increases, operational discipline becomes more important, not less."

Most publicly shared setup guides focus on getting ClawdBot running. They demonstrate responsiveness and capability. They rarely cover pairing approvals, security audits, dependency management, sandbox configuration, or token scoping. Skipping these steps does not just leave an agent loosely configured - it places an always-on system with broad access into a hostile environment without meaningful guardrails.

This is the shadow AI governance crisis made concrete. When employees can spin up a personal AI agent on a work laptop - or worse, on a personal device that also connects to corporate resources - the security boundary dissolves entirely. The agent does not distinguish between personal and corporate data. It processes both with equal enthusiasm.


What IT Leaders Should Do Now

Having run ClawdBot for several months and watched this security discourse unfold, here is what I would recommend to any IT leader navigating this space.

Establish an AI Agent Policy

If your organisation does not have a specific policy covering personal AI agents, you need one yesterday. This is distinct from your general AI acceptable use policy. Personal AI agents have persistent access, execute autonomously, and store state locally. They require their own risk assessment framework.

At minimum, your policy should address which AI agents are permitted on corporate devices, what data they may access, and what network segmentation is required for machines running them.

Inventory and Detect

You cannot manage what you cannot see. Start scanning for ClawdBot and similar tool signatures on your network. Look for the characteristic directory structures (~/.clawdbot/, ~/clawd/), the gateway ports, and the messaging platform integrations.

Endpoint detection and response (EDR) tools should be configured to alert on new services listening on unusual ports and on the creation of configuration files matching known AI agent patterns.

Enforce Secrets Management

The plaintext secrets problem is solvable. If employees must use personal AI agents, mandate that credentials are stored in encrypted vaults - not Markdown files. Hardware security modules, OS keychains, and encrypted environment variables all provide better protection than plaintext on the filesystem.

Network Segmentation

Any machine running a personal AI agent should be treated as a higher-risk endpoint. Segment it appropriately. Do not allow it unfettered access to corporate resources. Apply the principle of least privilege aggressively.

This aligns with zero trust principles - verify explicitly, assume breach, and never trust based solely on network location.

Monitor for Supply Chain Attacks

If your developers are installing skills from ClawdHub or similar unvetted repositories, treat those installations with the same scrutiny you would apply to any third-party dependency. Code review, sandboxed testing, and signature verification should be standard practice.


A Balanced Perspective

It would be easy to read this analysis and conclude that personal AI agents should be banned outright. That would be a mistake.

ClawdBot and tools like it represent a genuine shift in personal productivity. The ability to delegate routine administrative work to an AI agent is transformative for knowledge workers. Banning these tools will not prevent adoption - it will drive it underground, which is worse from a security perspective.

The right approach is controlled enablement. Acknowledge the productivity gains. Quantify the risks. Implement appropriate controls. And recognise that this is the beginning of a much larger trend.

Palo Alto Networks' Chief Security Intelligence Officer called AI agents 2026's biggest insider threat. She is right. But insider threats are not new - we have managed them for decades. What is new is the speed, scale, and autonomy. The organisations that adapt their security posture now will be better positioned than those who wait for the first breach to force their hand.

ClawdBot's viral moment is a wake-up call. The question is not whether personal AI agents will become ubiquitous - they will. The question is whether your security strategy is ready for them.

Share this post

DG

Daniel J Glover

IT Leader with experience spanning IT management, compliance, development, automation, AI, and project management. I write about technology, leadership, and building better systems.

Let's Work Together

Need expert IT consulting? Let's discuss how I can help your organisation.

Get in Touch