OpenAI launches GPT-5.5-Cyber to arm vetted defenders with unrestricted tools for advanced vulnerability research
By swapping rigid safety filters for identity-based access, OpenAI empowers vetted defenders to automate complex and autonomous vulnerability research.
May 8, 2026

OpenAI has officially launched GPT-5.5-Cyber, a specialized variant of its flagship large language model designed specifically for the high-stakes environment of digital defense. This release marks a significant strategic pivot for the company, as it addresses a long-standing friction point for security professionals: the tendency of AI safety filters to block legitimate research requests that resemble malicious activity. By introducing a model that is significantly more permissive in its response behavior, OpenAI is providing vetted defenders with a tool capable of performing deep vulnerability research, malware analysis, and even the execution of exploits against authorized test environments.[1] This move is widely viewed as a direct response to the increasing sophistication of state-sponsored cyberattacks and the emergence of similar specialized tools from competitors like Anthropic.
The core innovation of GPT-5.5-Cyber lies in its integration with a new identity-based security architecture known as the Trusted Access for Cyber framework.[2] Under this system, OpenAI has moved away from the one-size-fits-all safety guardrails that characterize its public-facing models.[3][4] For the general user, requesting a functional exploit for a known software vulnerability typically triggers a refusal message based on safety policies. However, for a researcher authenticated through the new framework, GPT-5.5-Cyber will fulfill the request, allowing for the rapid generation of proof-of-concept code and the identification of attack paths that a defender must understand to secure a network. This specialized access is currently limited to a select group of researchers and organizations responsible for protecting critical infrastructure, ranging from energy grids and telecommunications to global financial systems.[5]
The model’s performance in simulated environments suggests a major leap in autonomous reasoning. Recent evaluations conducted by the United Kingdom’s AI Security Institute revealed that GPT-5.5-Cyber is one of only two models capable of solving complex, multi-stage enterprise attack simulations from start to finish.[6][3] In one specific test, known as The Last Ones, the model successfully navigated a mock corporate network consisting of four network segments and twenty machines, performing tasks that would typically require a human expert approximately twenty hours to complete.[3] While the model occasionally falters in long-horizon reasoning, its ability to chain together multiple exploitation steps represents a new frontier for AI-driven security operations. By allowing the model to actively execute code in controlled settings, OpenAI enables a more dynamic form of red teaming that can surface vulnerabilities at machine speed.
This release places OpenAI in a heated competition with Anthropic, which recently debuted its own restricted model, Claude Mythos Preview. Comparative data suggests the two models are locked in a virtual dead heat for supremacy in the cybersecurity domain. While Anthropic’s Mythos has been praised for its ability to find decades-old vulnerabilities in hardened codebases, GPT-5.5-Cyber has shown a slight edge in isolated expert-level security tasks, scoring over seventy percent on advanced capture-the-flag benchmarks. A key differentiator for OpenAI is its approach to availability and cost. While Anthropic has kept Mythos extremely restricted, reportedly limiting access to fewer than fifty organizations under heavy government scrutiny, OpenAI is scaling its program to thousands of individual defenders. Furthermore, industry analysts note that OpenAI’s model is significantly more cost-effective to run, potentially making it the more viable option for large-scale enterprise security centers.
Strategic partnerships are central to the rollout of GPT-5.5-Cyber, with industry giants such as Cisco, CrowdStrike, and Cloudflare among the first to integrate the model into their defensive workflows. These companies are using the model to accelerate the "security flywheel," a process of continuous discovery, detection, and remediation. For instance, Cisco is leveraging the model to automate the analysis of massive traffic datasets to find subtle indicators of compromise that traditional heuristics might miss. CrowdStrike is employing the model’s binary reverse-engineering capabilities to deconstruct new malware strains in minutes rather than days. By putting these tools in the hands of the world’s most advanced security firms, OpenAI aims to ensure that the defensive use of AI scales faster than the offensive capabilities of malicious actors.
The decision to release such a permissive model has not been without controversy, sparking an urgent debate within the White House and among national security officials. The primary concern is the dual-use nature of the technology: a model that can write an exploit to help a defender patch a system can, in the wrong hands, be used to launch a devastating attack. To mitigate this risk, OpenAI has implemented rigorous vetting procedures.[7][8][9][4] Access is not granted to individuals but to organizations that can prove they are defenders of critical infrastructure. Additionally, starting in June, all users of the cyber-permissive models will be required to utilize advanced account security measures, including phishing-resistant hardware authentication. The company has also been transparent with government agencies, demonstrating the model’s capabilities to the U.S. Department of Commerce’s AI Standards and Innovation Center and various congressional committees.[10]
Beyond its immediate utility in finding bugs, GPT-5.5-Cyber represents a broader trend toward the democratization of high-level security expertise. Traditionally, the ability to perform deep vulnerability research was restricted to a small number of elite "red team" specialists. By encoding this expertise into a large language model, OpenAI is potentially lowering the barrier to entry for defensive work, allowing mid-level security analysts to perform tasks that previously required a decade of experience. This shift is critical as the global talent gap in cybersecurity continues to widen, leaving many organizations vulnerable simply because they cannot hire enough qualified personnel. The model acts as a force multiplier, giving smaller security teams the capabilities of a much larger, more sophisticated operation.
Looking ahead, the emergence of GPT-5.5-Cyber signals a permanent change in how the AI industry views safety and alignment. The original paradigm of "refuse all harmful requests" is evolving into a more nuanced "identity-based permission" model. This allows AI companies to push the boundaries of what their models can do without violating the ethical commitments that prevent them from releasing dangerous tools to the public at large. As these models become more autonomous and more integrated into the core infrastructure of the internet, the focus will likely shift from preventing the AI from doing harm to ensuring that the human operating the AI is trustworthy.
In the final analysis, OpenAI’s release of GPT-5.5-Cyber is a calculated gamble that the benefits of accelerating defensive capabilities outweigh the risks of creating more potent offensive tools. If successful, the model could lead to a more resilient digital world where vulnerabilities are found and fixed before they can be exploited. However, it also marks the beginning of a new era in which AI models are no longer just assistants but active participants in the ongoing conflict for control of the world’s digital infrastructure. The success of this initiative will be measured not just by the benchmarks it clears, but by the stability of the critical systems it is now tasked with protecting.