AI Assistant OpenClaw Hit by "OpenDoor" Flaw Allowing One-Click System Takeover.
A catastrophic one-click vulnerability allows attackers to gain full, persistent system control via the trusted local AI agent.
February 4, 2026

The spectacular rise of the open-source AI assistant OpenClaw, formerly known as Clawdbot, has been violently tempered by the revelation of a systemic security flaw so severe it has been dubbed the "OpenDoor problem." Security researchers have demonstrated that the popular autonomous agent can be entirely compromised through deceptively simple means, allowing attackers to install permanent backdoors and gain complete, persistent control over a user's computer. The sheer breadth of the agent’s capabilities, which include autonomous command execution and full file system access, has turned it from a revolutionary productivity tool into what many in the cybersecurity community now call an unparalleled "security nightmare." The vulnerabilities expose thousands of self-hosted instances globally to a catastrophic level of risk, threatening not just data theft but complete system takeover with a single click or the processing of a single malicious document.
The core of the crisis lies in the unprecedented level of access OpenClaw is granted to function. As a "local-first" AI assistant, OpenClaw is designed to bridge large language models with real execution environments, enabling it to go beyond simple conversation and actually take action on behalf of the user, such as executing terminal commands, managing files, orchestrating complex workflows, and integrating with messaging applications[1][2][3][4]. To perform these tasks, the agent maintains persistent state and often runs with high privileges, routinely handling and storing critical credentials like API keys, bot tokens, and OAuth secrets[5][4]. Cybersecurity experts have labeled this architecture a "lethal trifecta" of risks: access to private data, exposure to untrusted content, and external communication capabilities[2][3]. The agent’s persistent memory files, which can include a user’s anxieties, projects, and private credentials in plaintext, create a unique threat vector known as "Cognitive Context Theft," providing a psychological dossier for perfect social engineering attacks[5].
The OpenDoor problem manifests through several distinct, yet equally devastating, attack vectors, each exploiting the trust placed in the autonomous agent. The most critical vulnerability identified, tracked as CVE-2026-25253, is a high-severity flaw with a CVSS score of 8.8, enabling one-click remote code execution (RCE)[6]. This technical issue stems from a token exfiltration vulnerability in the OpenClaw Control UI, which the developers themselves described as leading to a "full gateway compromise"[1]. Specifically, the Control UI was found to trust an unvalidated `gatewayUrl` from the query string and would auto-connect on load, sending the stored, highly privileged gateway token in the WebSocket connect payload[6]. This design flaw meant that a target user merely had to click a crafted malicious link or visit a nefarious website for their token to be sent to an attacker-controlled server[6]. With this token, an adversary could connect to the victim's local gateway, modify configuration and tool policies, and invoke privileged actions, effectively achieving full system control[6]. The flaw was addressed by the agent’s maintainer, Peter Steinberger, in version 2026.1.29[1][6].
Beyond the technical RCE flaw, the second major vector involves supply chain attacks through the agent’s own ecosystem. OpenClaw’s functionality is extended through "skills," which are plugin-like packages often distributed via a public registry like ClawHub[7][4]. This open-source distribution model has been ruthlessly exploited in a campaign researchers have dubbed "ClawHavoc," which saw 341 malicious skills identified on the platform[8]. These backdoored skills, often disguised as legitimate tools, install an infostealer or, in the worst cases, a reverse shell, granting an attacker full, interactive access to the victim’s machine[8]. The insidious nature of this attack is that the payload is often hidden in operational code, triggering during normal market search or use, bypassing superficial security analysis[8]. This problem is compounded by the agent’s fundamental inability to reliably distinguish between benign user instructions and malicious instructions embedded in untrusted inputs, such as manipulated documents or emails[9][10]. An attacker can embed hidden commands within a seemingly innocent document, causing the AI agent, with its full system access, to ignore its original task and execute the attacker’s instructions instead[10].
The scale of this exposure is significant and spans the globe. Security researchers identified over 4,500 exposed OpenClaw/Clawdbot instances internationally, with the highest concentrations found in key economic centers like the United States, Germany, Singapore, and China[11]. For enterprise environments, the risks are particularly acute; one audit found that 22 percent of corporate customers had unauthorized OpenClaw use, with more than half of those granting the agent privileged access[7]. This indicates a widespread "shadow IT" risk, where employees unknowingly introduce a massive security liability into their networks[7]. The ability for attackers to exfiltrate critical credentials and even messaging platform session credentials, enabling surveillance across channels like WhatsApp, underscores the severity of the compromise[11].
The OpenClaw debacle serves as a harsh, immediate lesson for the burgeoning AI agent industry. The era of autonomous AI assistants, while promising immense productivity gains, collapses traditional security boundaries by granting an application access to an entire operating system[4]. The sheer ease with which a critical vulnerability could be weaponized—a one-click exploit that bypasses authentication and leads to full RCE—highlights a profound failure in securing a system that inherently requires root-level trust[6]. Experts are now pushing for a fundamental shift in security paradigms, arguing that strategies must move away from static policy enforcement toward real-time behavioral governance, where the agent’s actions are continuously monitored and constrained[1]. Until such secure architectures are the norm, the cautionary tale of OpenClaw’s OpenDoor problem suggests that for AI agents with unfettered system access, users must be extraordinarily diligent, as running one of these powerful tools un-sandboxed is functionally equivalent to inviting malware directly onto their machines.
Sources
[2]
[3]
[4]
[7]
[8]
[10]
[11]