USING AI

OpenClaw

The viral AI agent that went from Clawdbot to Moltbot to OpenClaw in five weeks—and exposed why autonomous agents running on your machine put API keys, credentials, and patient data at risk.

~40 min read Security Focus
Why This Matters Now

OpenClaw became one of the most popular open-source projects in the world in a matter of weeks. Within 48 hours of going viral, security researchers found installations leaking API keys in plaintext. Its creator made 6,600 commits in a single month—almost all AI-generated, almost none reviewed. Then he joined OpenAI. This is the story of what happens when autonomous AI agents get access to everything on your machine.

March 2026 Update

Since this module was first published, the OpenClaw story has accelerated dramatically. Malicious skills on ClawHub have jumped from roughly 341 to over 1,184—many now deploying commodity macOS malware that harvests keychain passwords, browser credentials, and SSH keys. Two new critical CVEs have been disclosed, including a pre-authentication remote code execution flaw affecting 42,000 publicly exposed servers. Meanwhile, OpenClaw has surpassed 250,000 GitHub stars, and the broader agentic AI threat landscape has produced real-world harm: a healthcare AI was jailbroken into tripling opioid doses, and a consumer AI browser was found vulnerable to zero-click calendar-invite exploits.

The sections below have been updated to reflect these developments. New material is marked with [March 2026] where applicable.

The OpenClaw Story

In January 2026, an open-source project called Clawdbot appeared on GitHub. Built by Peter Steinberger—founder of PSPDFKit, a well-known mobile software company—it was a personal AI agent that could write code, manage your calendar, control smart home devices, and interact across WhatsApp, Telegram, Slack, and a dozen other platforms. Within weeks it had over 157,000 GitHub stars—eventually surpassing 250,000, more than React—making it one of the most popular open-source projects in the world.

Then things got complicated. Anthropic sent a trademark complaint—the name was too close to "Claude." Steinberger renamed it to Moltbot, then two days later to OpenClaw. A companion product, Moltbook, followed the same path. The drama made headlines everywhere from CNBC to TechCrunch. Three names in five weeks. The internet couldn't look away.

But it was a quote from Steinberger that crystallized the real conversation. In a widely-read Pragmatic Engineer interview, he said:

"I ship code I don't read."

He had made over 6,600 commits in January 2026 alone—almost all generated by AI agents. The quote became a lightning rod. Supporters saw it as the logical conclusion of vibe coding: if AI can write code, why read it? Critics saw something more dangerous: autonomous software running on your machine with access to everything you have access to.

On February 15, 2026—just weeks later—Steinberger joined OpenAI. Sam Altman announced it personally. OpenClaw moved to an independent open-source foundation to continue without him.


What OpenClaw Actually Is

OpenClaw isn't a coding tool in the way that Replit or Claude Code are coding tools. It's a full personal automation platform. At over 430,000 lines of code, it does everything:

It's model-agnostic—it can run on Anthropic's Claude, OpenAI's models, or open-source alternatives like Kimi 2.5. This flexibility means OpenClaw inherits whatever safety guardrails (or lack thereof) the underlying model provides.

The ambition is staggering. So is the attack surface.

How OpenClaw Differs from Other Agents

Claude Code (covered in our Vibe Coding module) is a coding-specific agent with a built-in permission system and safety hooks. It asks before running commands and can be configured to block access to sensitive files.

OpenClaw is broader—it's not just a coding tool, it's a life automation platform with access to your messaging, calendar, email, and filesystem. More capability means more attack surface.


What Went Wrong

Here's what makes agents like OpenClaw fundamentally different from a chatbot like ChatGPT or Claude.ai: they run on your machine.

When you use ChatGPT in a browser, it can only see what you paste into it. When you run OpenClaw on your laptop, it can see everything your machine can see:

This isn't a theoretical concern. It's the design. OpenClaw needs access to your messaging apps, your calendar, your filesystem to do what it promises. The question is: what guardrails exist?

Within 48 hours of going viral, security researchers had their answer. OpenClaw installations were found storing API keys in plaintext configuration files accessible to anyone on the network. Third-party "skills"—community-contributed plugins—contained credential stealers and backdoors. Users were giving a 430,000-line program they'd never audited access to their most sensitive accounts.

[March 2026] ClawHavoc: The Supply Chain Gets Worse

In late February 2026, the malicious skills problem on ClawHub—OpenClaw's community plugin marketplace—escalated dramatically. Researchers tracking the "ClawHavoc" campaign found that the number of malicious skills had jumped from roughly 341 to over 1,184.

More concerning than the numbers is what the skills now do. According to Trend Micro, 335 skills now install Atomic Stealer (AMOS)—a commodity macOS info-stealer available for $500–$1,000 per month on criminal marketplaces. AMOS harvests keychain passwords, browser credentials, SSH keys, and API keys, then exfiltrates them silently.

The key shift is in the attack model. Earlier malicious skills relied on the user to run them. The newer skills manipulate the AI agent itself as a trusted intermediary—embedding instructions that cause the agent to install the malware on the attacker's behalf. The user sees the agent working normally. The agent doesn't know it's been compromised. The malware runs with the user's full permissions.

[March 2026] New Critical Vulnerabilities

Two new CVEs have added to OpenClaw's growing vulnerability catalog (now tracked at github.com/jgamblin/OpenClawCVEs):

42,000 Exposed Servers

Internet scanning found 42,000 OpenClaw instances publicly accessible on the open internet—up from roughly 1,000 in early February. Each of these is a machine running an AI agent with access to its owner's files, credentials, and connected services, now reachable by anyone. If you've installed OpenClaw, verify it is not exposed on a public IP address or port.

[March 2026] Rapid Security Hardening

To its credit, the OpenClaw project has responded with a burst of security-focused releases. Versions v2026.2.17 through v2026.2.25 shipped transport security improvements, session cleanup fixes, and WebSocket authentication lockdowns. The WebSocket vulnerability was patched within 24 hours of public disclosure. The onboarding experience now warns that OpenClaw is "personal-by-default" with a single trusted operator boundary—an acknowledgment that multi-user or public-facing deployments were never the intended use case.

These are positive steps. But they also underscore the fundamental tension: a project moving at 6,600 commits per month will always be patching vulnerabilities faster than most users can update.

And it wasn't just OpenClaw. Across the broader AI agent ecosystem:

Why This Hits Healthcare Harder

If you're a clinician building tools with AI—even personal prototypes—your machine may contain credentials that connect to healthcare systems. An API key for a clinical database. OAuth tokens for an EHR sandbox. Test credentials for a HIPAA-covered service.

A leaked API key on a personal project is embarrassing. A leaked API key that grants access to patient data is a HIPAA violation—with penalties that start at $100 per record and scale to millions.

The Shadow AI Problem

"Shadow AI" describes AI tools that employees use without organizational oversight—no security review, no access controls, no audit trail. A 2026 study in PMC found that 97% of organizations that experienced AI-related security incidents lacked proper AI access controls.

[March 2026] The costs are now quantifiable: data breaches involving shadow AI tools cost an average of $670,000 more than standard incidents. Palo Alto Networks has labeled AI agents "2026's biggest insider threat."

When a clinician installs a coding agent on their laptop and connects it to Slack, email, and a cloud database—even for a side project—they've created shadow AI with elevated privileges that traditional security tools can't detect.

Hallucinated Dependencies: A New Attack Vector

There's another risk specific to AI-generated code that connects directly to security. When coding agents recommend installing software packages, roughly one in five recommended packages doesn't actually exist. The agent hallucinated the name.

Normally, that just means an error message. But attackers have figured out something clever: they register malicious packages using the fake names that AI models commonly hallucinate. Security researchers call this "slopsquatting." When an agent recommends a hallucinated package and then installs it—the malicious version runs on your machine with whatever permissions the agent has.

This is especially dangerous with autonomous agents that install dependencies without asking. A coding agent that auto-installs packages is one hallucination away from executing malicious code.

[March 2026] The Broader Agentic AI Threat Landscape

OpenClaw's vulnerabilities are not isolated. They are symptoms of a broader security crisis as autonomous AI agents proliferate across every industry. Several developments in early 2026 illustrate how quickly this landscape is evolving.

State-Level Actors Are Using AI Agents Offensively

Chinese state-affiliated group GTG-1002 has been documented using AI agents to execute 80–90% of attack lifecycle steps autonomously—from reconnaissance to exploitation. This is not theoretical; it represents a fundamental shift in how adversaries operate. When attacks are automated by AI, they scale faster and cost less to execute than human-driven campaigns.

Shadow AI Costs Are Quantifiable

Data breaches involving unauthorized "shadow AI" tools now cost organizations an average of $670,000 more than standard security incidents. Palo Alto Networks has called AI agents "2026's biggest insider threat"—not because the agents are malicious, but because they operate with broad permissions and minimal oversight inside enterprise networks.

Healthcare AI Is Already Being Exploited

In March 2026, researchers demonstrated that Doctronic, a healthcare AI system used for clinical decision support, could be jailbroken via prompt injection—the same attack class that affects OpenClaw skills. Researchers were able to triple recommended OxyContin doses by manipulating the system's instructions. This is the direct clinical consequence of the security patterns we see in the agent ecosystem.

Consumer AI Browsers Are Vulnerable Too

Perplexity's Comet browser was found to contain zero-click exploits: a malicious calendar invite could silently exfiltrate files from the user's machine with no interaction required. This reinforces a pattern: any tool that combines AI capabilities with local system access creates an attack surface that traditional security models were not designed to handle.

For Clinicians: Why This Pattern Matters

Every example above shares the same root cause: an AI system with broad permissions was manipulated through its inputs. Whether it's an OpenClaw skill installing malware, a healthcare AI recommending dangerous doses, or a browser leaking files via a calendar invite—the attack works because the AI acts as a trusted intermediary between the attacker and your data. If you use any AI tool that can access patient data, clinical systems, or credentials, these are not abstract risks. They are the current threat model.


Safer Alternatives to OpenClaw

OpenClaw's security problems sparked a wave of alternatives designed to prove you don't need 430,000 lines of unauditable code to build a useful AI agent.

NanoClaw launched in late January 2026 as a direct response to OpenClaw's vulnerabilities. Its entire core logic is roughly 500 lines of TypeScript—a developer (or a second AI) can audit it in under ten minutes. The key innovation is containerization: the agent runs inside isolated Docker containers with no access to your host filesystem or environment variables. Your API keys, SSH keys, and credentials stay walled off by default.

Nanobot, from the University of Hong Kong, takes minimalism further: approximately 4,000 lines of code, 45MB of memory, boots in under a second. A senior developer can read the entire source code in an afternoon and verify every action it takes.

The Size Spectrum

OpenClaw: 430,000+ lines. NanoClaw: 500 lines. Nanobot: 4,000 lines. Smaller doesn't always mean better—but it does mean auditable. When an agent has access to your terminal, your files, and your credentials, knowing exactly what it can do matters.


Protecting Yourself

Whether you use OpenClaw, a lighter alternative, or any AI agent, the same principles apply. Use them the way you'd use any powerful tool in medicine: understanding the risks and applying appropriate safeguards.

1. Separate Your Secrets

The single most impactful thing you can do: don't store API keys where agents can read them.

2. Use Permission Systems

Choose agents that ask before acting. Claude Code's permission modes let you require approval before file writes or command execution. Its hook system can block specific actions entirely—for example, preventing any write to files matching *.env or *credentials*.

NanoClaw's containerization approach goes further: the agent physically cannot access files outside its designated directory. Even if it's compromised, the blast radius is contained.

3. Verify Before You Install

When a coding agent recommends a package or dependency, verify it exists and is legitimate before installing. Check the package registry directly. Look at download counts and publication dates. A package with 3 downloads published yesterday is not the same as one with 3 million downloads maintained for years.

4. Review Before You Push

"I ship code I don't read" makes for a great headline. It's a terrible security practice. Before any code leaves your machine:

5. Understand What You're Connecting

Before giving any agent access to a messaging platform, cloud service, or database, ask yourself: what's the worst thing that could happen if this agent were compromised? If the answer involves patient data, financial records, or privileged system access—reconsider, or add additional safeguards.

The Medical Analogy

Think of coding agents like procedural sedation. Incredibly useful. High benefit-to-risk ratio when done properly. But you wouldn't administer it without monitoring, without knowing the reversal agent, without having thought through what could go wrong. The same discipline applies here: use the tool, but maintain situational awareness.

Cross-References

For more on protecting patient data when using AI tools, see our PHI, HIPAA, and AI module.

For how training data and bias affect AI-generated code, see Bias, Ethics, and the Training Data Problem.


Resources

The OpenClaw Story

Steinberger's own words on joining OpenAI, why OpenClaw must remain open source, and his vision for the project.
The definitive mainstream account—origin, viral growth, naming drama, and security fears in one piece.
Gergely Orosz's in-depth interview with Steinberger—6,600 commits in one month, vibe coding at scale.
Independent investigative journalism on how a hobby project became a security nightmare.
The latest chapter—Steinberger joins OpenAI, OpenClaw moves to an open-source foundation.

Security Research

The key credential analysis—cleartext API keys under ~/.clawdbot, backup files retaining deleted secrets, 300K–400K users affected.
The original CVE disclosure—CVSS 8.8, one-click remote code execution via WebSocket token theft.
Major antivirus vendor's verdict: 512 vulnerabilities (8 critical), malicious skills on ClawHub (now 1,184+).
How adversaries embed instructions in emails and webpages that OpenClaw ingests, enabling reconnaissance and exfiltration.
Analysis of 3,000+ OpenClaw skills found hundreds that were malicious—reverse shells, semantic worms, cognitive rootkits.
A researcher uploaded a malicious skill, inflated it to 4,000+ downloads, and watched developers from 7 countries execute it.

March 2026 Updates

The definitive account of the ClawHavoc supply chain campaign—skill counts, attack techniques, and AMOS deployment.
Technical analysis of how 335 skills deploy AMOS to harvest keychain passwords, browser creds, SSH keys, and API keys.
CVSS 9.8 pre-auth remote code execution via voice-call extension, 42,000 instances affected.
Community-maintained list of all known OpenClaw CVEs with status and affected versions.
Prompt injection applied to clinical AI—researchers manipulated prescribing recommendations, demonstrating direct patient safety risk.
Calendar invites used to silently exfiltrate files from AI-powered browser—no user interaction required.

Safer Alternatives

How 500 lines of TypeScript with container isolation addresses OpenClaw's 52+ modules and shared-memory architecture.
The source code: ~500 lines, Anthropic's Agents SDK, Apple container isolation, auditable in an afternoon.
~4,000 lines of Python, MCP-based, web search, scheduled tasks, memory—minimalist alternative from HKU's Data Intelligence Lab.

Podcasts

The definitive 3-hour interview: origin story, self-modifying agents, naming drama, and whether AI replaces programmers. Full transcript.
Steinberger on his workflow, philosophy, and 6,600 commits in January 2026.
Enterprise security perspective on how locally-run AI agents create brand-new attack surfaces.
Cybersecurity podcast covering the vulnerabilities and the Moltbook chaos—150,000 AI agents forming digital religions within 72 hours.
SANS Institute—gold standard in cybersecurity training—noted exposed OpenClaw installs found via internet scanning (now 42,000+).

Healthcare & Shadow AI

How clinicians are using unauthorized AI tools for documentation relief, creating HIPAA compliance risks.
Hard data: 86% of IT executives reported shadow IT instances, 17% of healthcare workers using unauthorized AI tools.
Best practices from 115 healthcare organizations on securing AI tools in clinical environments.

Conclusion: The OpenClaw Lesson

OpenClaw isn't a cautionary tale about a bad product. It's a well-built platform created by an experienced developer that attracted 250,000 stars for good reasons. The problem isn't OpenClaw specifically—it's what OpenClaw represents: a new category of tool that blurs the line between "software you use" and "software that acts on your behalf with your credentials."

Steinberger shipping code he didn't read was a choice about his own risk tolerance. But when 250,000 people install that same agent—many of them less technical, some of them handling sensitive data—the risk calculus changes. The agent that's a productivity miracle for one person is a security incident waiting to happen for another.

The answer isn't avoidance. AI agents are genuinely useful, and they're not going away. The answer is choosing tools with appropriate guardrails for your situation. If you're a clinician experimenting with AI tools, the question isn't "should I use agents?" It's "does this agent have the security controls my situation demands?"

The Bottom Line

OpenClaw showed the world what AI agents can do. It also showed what can go wrong when powerful tools ship without adequate security for their most vulnerable users. Before you install any agent on your machine, ask: what can it access, who audited the code, and what happens if it's compromised? Trust, but verify.

A Note on This Module

This module was written with AI assistance and verified by human review. All linked sources were checked at time of publication. Last updated March 2026 with ClawHavoc supply chain data, new CVEs, and broader agentic AI threat context. The OpenClaw story is still unfolding—some details may have evolved since writing.

Learning Objectives

  • Describe OpenClaw's rise and why it exposed critical security concerns
  • Explain how AI agents differ from browser-based chatbots in terms of system access
  • Identify how agents can access API keys, credentials, and sensitive files on your machine
  • Recognize specific attack vectors including prompt injection, slopsquatting, and credential exfiltration
  • Describe how supply chain attacks through AI agent marketplaces (e.g., ClawHavoc) use the agent as a trusted intermediary to deploy malware
  • Compare OpenClaw to safer alternatives like NanoClaw and Nanobot
  • Connect OpenClaw's vulnerabilities to the broader agentic AI threat landscape, including healthcare AI jailbreaks and state-level AI-driven attacks
  • Apply practical guardrails to protect sensitive data when using any AI agent