UK AISI Evaluates GPT-5.5 Cyber Attack Capabilities
- •The UK's AISI confirmed that GPT-5.5 can successfully execute complex network intrusion sequences.
- •GPT-5.5 achieved a 20% success rate in a 32-step corporate network attack simulation.
- •No current AI model has demonstrated the capability to fully compromise industrial control systems.
The rapid advancement of frontier AI models has introduced both unprecedented productivity gains and significant security concerns. A critical new report from the UK’s AI Security Institute (AISI) underscores this dual reality by detailing the cyber-attack capabilities of OpenAI’s latest flagship model, GPT-5.5. Researchers used advanced cyber-range environments to quantify how these systems execute complex, multi-stage attacks that were once reserved for human experts.
The assessment focused on a 32-step simulation called "The Last Ones," which mimics the process of taking over a corporate network. The test evaluated skills in reconnaissance, credential theft, and navigating internal system architectures. The fact that GPT-5.5 and Anthropic’s Claude Mythos Preview completed this end-to-end simulation marks a watershed moment, suggesting that models have evolved into sophisticated agents capable of autonomous, long-horizon planning.
However, the AISI emphasized that these findings occurred in a controlled, research-oriented sandbox, far removed from real-time enterprise defenses. While the models excelled at logical execution, they have not yet shown the ability to evade active, industrial-grade security operations. Furthermore, the models failed to compromise industrial control systems, indicating that specialized knowledge in operational technology remains a significant barrier for current-generation LLMs.
The study also revealed that performance tends to scale with computational investment. As models grow in intelligence, their capacity for offensive operations increases as a secondary, emergent effect of better multi-step reasoning. Consequently, the focus for policymakers and corporate security teams is shifting from attempting to prohibit these tools to proactively adapting to their existence.
The implications of this report extend far beyond the laboratory, as these capabilities will eventually be harnessed by security professionals for defensive purposes. Red teaming, automated penetration testing, and rapid vulnerability mitigation are poised for a massive upgrade. The AISI, in conjunction with the National Cyber Security Centre, is urging organizations to rethink their security infrastructure to prepare for a world where AI-powered threats are a standard reality.
Ultimately, this assessment serves as a reality check for the global business community. While current models have limitations, the trajectory of their development is clear. For university students and future technologists, understanding the intersection of AI safety, autonomous agency, and cybersecurity will be one of the most critical skill sets of the coming decade.