AI Agents & Security

The Claw Giveth, The Claw Taketh Away

OpenClaw promised developers an open-source AI agent that could actually do things on your machine. Two weeks later, it's doing things nobody asked for. A deep dive into the fastest-growing—and most controversial—AI tool of 2026.

Listen
A dramatic robotic claw emerging from a terminal, split between golden promise and crimson danger
An AI puppet with its strings being cut by crimson scissors, representing prompt injection vulnerability
01

Your Agent Isn't Working For You Anymore

Here's the nightmare scenario nobody talks about at AI demo days: your autonomous coding agent visits a webpage to research a bug fix, and the page tells it to exfiltrate your SSH keys instead. Not hypothetical. CrowdStrike's latest threat report demonstrates exactly this attack against OpenClaw, the open-source AI agent that's been tearing through the developer community since late January.

The attack is brutally simple. An attacker hides instructions in a webpage—white text on white background, invisible to humans, irresistible to AI. When OpenClaw's headless browser visits the page, it reads the hidden text as a user instruction and executes it with whatever permissions you've granted. Which, if you're running OpenClaw as most people do, is everything.

"The agent believes the website content is a user instruction, and it happily complies." — CrowdStrike Threat Report, February 2026

The "human-in-the-loop" safeguard—OpenClaw's permission prompt—was supposed to prevent exactly this. CrowdStrike found it trivially bypassable. The AI can be socially engineered by the injected text to frame the malicious action as routine maintenance. "Updating SSH config for security" sounds reasonable when your agent asks, and most developers click "Allow" without reading the fine print. The fundamental problem isn't a bug to patch. It's an architectural flaw: when your agent can both browse the web and execute shell commands, every webpage becomes a potential attack vector.

Three AI tools represented as different vehicles on different terrains
02

The Safety-First vs. Capability-First Schism

The developer community is splitting into two camps, and the fault line runs right through OpenClaw. On one side: the "capability maximalists" who want an AI agent that can do anything on their machine, consequences be damned. On the other: the "safety pragmatists" who'd rather trade some flexibility for not getting pwned.

SuperPrompt's detailed comparison captures the dynamic perfectly. OpenClaw is the wild off-road buggy—fully open-source, limitless access, community-driven, and occasionally on fire. Claude Code (Anthropic's offering) is the corporate sedan—managed, sandboxed, reliable, but you don't get to peek under the hood. And then there's Nanobot, the minimalist bicycle that emerged in direct response to OpenClaw's bloat and security scandals.

Radar chart comparing OpenClaw, Claude Code, Nanobot, and Manus across capability, security, open source, ease of use, community, and stability dimensions
AI agent tools compared across six dimensions. OpenClaw leads in capability and open-source freedom but trails badly on security and stability. Data from SuperPrompt comparative analysis, February 2026.

The quote that stung most: "Use OpenClaw if you want a coworker. Use Nanobot if you want a script. Use Claude Code if you want to keep your job." It's reductive, but it captures something real. The market is sorting itself by risk tolerance, and OpenClaw has positioned itself as the high-reward, high-risk option. Whether that's a viable long-term positioning—or just a phase before the inevitable security reckoning—depends on how fast the team can close the gaps exposed this week.

A fortress of server racks with hundreds of glowing red doors left wide open
03

923 Open Backdoors Into Developer Laptops

The number 923 is going to follow OpenClaw for a very long time. That's how many "Clawdbot gateways"—instances of the pre-rebrand agent—security researchers found exposed to the public internet without authentication. Anyone who stumbled across one of these IPs could issue commands to the host machine as if they were the owner. Full shell access. No password required.

It gets worse. Roughly 20% of these exposed instances were traced to corporate networks. Not hobbyists running the tool on personal machines. Developers at companies—likely as "Shadow IT" deployments that never went through security review. One researcher described it as finding "nearly a thousand open backdoors into developer laptops across the globe."

Bar chart showing security incident severity escalation from January 20 to February 6, 2026
Two weeks from launch to critical: the escalating severity of OpenClaw security incidents. Sources: CrowdStrike, NotebookCheck, 1Password Security Blog.

This is what "secure by default" means as a missing principle, not a missing feature. OpenClaw's documentation itself acknowledges the risk with disarming candor: "Running an AI agent with shell access on your machine is... spicy." Spicy. Corporate security teams are going to love that word in their incident reports. The emergency patches in version 2026.1.30 addressed the gateway exposure, but the reputational damage is the kind that triggers CTO-level bans. Shadow IT giveth, corporate policy taketh away.

A rocket ship made of GitHub stars soaring upward with some stars cracking and falling off
04

150,000 Stars and a Crisis of Legitimacy

Let's talk about the number everyone keeps citing: 150,000 GitHub stars. Forbes crowned OpenClaw "the fastest growing open source AI project of 2026 so far." The Discord is buzzing. X is flooded with demo videos of OpenClaw autonomously debugging production code. The vibes are immaculate.

But vibes aren't validation. Some analysts are flagging that the star count may be inflated—possibly conflating total downloads and forks, possibly reflecting bot activity. GitHub star farming is a well-documented practice in the open-source world, and a project that went from zero to 150k in under three weeks deserves scrutiny, not just celebration.

Line chart showing OpenClaw's explosive GitHub star growth compared to Claude Code and Nanobot
OpenClaw's GitHub star trajectory versus Claude Code and Nanobot. The ClawHub malware incident (Jan 29) barely dented the growth curve—suggesting hype may be outrunning due diligence. Sources: GitHub Trends, Forbes.

What's undeniable is that there's massive pent-up demand for an open, local, capable AI agent. Developers want the power that cloud-based tools promise but without the vendor lock-in, the token costs, and the feeling that someone else controls their workflow. OpenClaw tapped that nerve perfectly. The question isn't whether the demand is real—it is. The question is whether OpenClaw can mature fast enough to deserve the trust that 150,000 starred repos imply. Right now, starring OpenClaw feels less like endorsement and more like rubbernecking.

A cross-section diagram of a mechanical claw revealing intricate inner workings of circuits and neural pathways
05

Under the Hood: What OpenClaw Actually Does

Strip away the hype and the security drama and you find a genuinely interesting piece of engineering. DigitalOcean's technical deep-dive lays out what OpenClaw actually is: a Node.js application that connects to any major LLM—Anthropic's Claude, OpenAI's models, or local models via Ollama—and gives it hands.

Those hands include persistent project memory (local vector stores that remember your codebase context between sessions), web browsing via headless browser, and—the big one—full terminal command execution. "It doesn't just chat; it does," the tutorial explains. "It can grep your logs, find the error, Google the fix, and apply the patch—if you let it."

The architecture is genuinely novel in how it bridges chatbots and automation. Most AI coding tools either give you suggestions (Copilot-style) or run in isolated sandboxes (Claude Code). OpenClaw sits in between: a persistent agent with real system access and memory that accumulates over time. That's what makes it powerful. That's also what makes it dangerous. The same architecture that lets it chain together "read logs → search web → apply fix" also enables "read .env → visit malicious page → exfiltrate tokens." Tool is real. The hype is also real. The danger is very real. All three coexist.

A phoenix emerging from flames with wings made of code commits and version numbers
06

From Moltbot to OpenClaw: Rebirth by Emergency Patch

Version 2026.1.30 was supposed to be a victory lap. The official rebrand from "Moltbot" to "OpenClaw" (resolving a naming dispute that had plagued the project), new free AI model support via Kim K2.5, native Voyage AI integration, a slick token usage dashboard. Creator Peter Steinberger—best known for his PDF framework work—had a roadmap and momentum.

Instead, it became an emergency triage release. The same update that shipped the rebrand also shipped critical security patches for the exposed gateways. The changelog reads like two different products stitched together: "New feature: enhanced shell completion. Also: fixed a vulnerability that allowed unauthenticated remote code execution." That whiplash tells you everything about where OpenClaw is in its lifecycle. It's simultaneously shipping v1 features and patching existential security flaws.

Steinberger's public posture has been refreshingly honest. The documentation doesn't pretend the risks don't exist—it admits running a shell-access AI agent is "spicy." But honesty isn't the same as safety. The question enterprise teams are asking isn't "does the developer know the risks?" It's "can we trust that the mitigations are sufficient?" Right now, the answer is: not yet. But credit where it's due—most open-source projects at this stage would be hiding behind PR statements, not shipping patches and publishing candid docs.

A Trojan horse made of neon app icons being wheeled through a marketplace gate, with malware tentacles visible inside
07

The Marketplace That Shipped Malware on Day One

If you needed a single story to illustrate why the AI agent era is going to be a security minefield, this is it. OpenClaw launched "ClawHub," a community marketplace for "AgentSkills"—extensions that give the agent new capabilities. Think npm packages, but each one is an instruction set for an AI with near-root access to your machine.

Security researcher Jason Meller from 1Password found the problem fast. A top-downloaded skill disguised as a "Twitter Integration" contained a staged malware delivery chain. Step one: install the skill. Step two: the agent runs the skill's instructions. Step three: a macOS infostealer begins exfiltrating developer credentials and SSH keys. All while the developer thinks they're just setting up a Twitter bot.

"This isn't just a bad npm package; it's an instruction set for an AI that has root-adjacent access to your entire life." — Jason Meller, 1Password

This is worse than a compromised npm package, and here's why: with a bad package, the malicious code has to be hidden in actual source code that, in theory, someone could audit. With an AI agent skill, the "code" is natural language instructions. The skill literally tells the AI what to do, and the AI does it. There's no binary to decompile, no obfuscated function to flag. The attack surface isn't code—it's language. That's a paradigm shift in supply chain security, and nobody has good tooling for it yet.

Tool or Hype? Yes.

OpenClaw is a genuinely capable AI agent built on a genuinely dangerous architecture, riding a genuinely inflated hype wave into a community that genuinely needs what it promises. It's not a scam—the code works, the vision is coherent, and the demand is real. But it shipped trust before earning it, and the security incidents of the past two weeks aren't growing pains. They're warnings. The developers who'll benefit most from OpenClaw are the ones who approach it with the same caution they'd give any tool that asks for the keys to their entire machine: test in isolation, audit every skill, and never—ever—let it browse the web unsupervised. The claw is real. Handle with care.