Key Findings
- Vulnerability Discovery Acceleration Inverts Offense-Defense Balance
- Autonomous Attack Execution Reduces Human Operator Dependency
- Attribution Opacity Undermines Deterrence Credibility
- Machine-Speed Operations Exceed Human Decision Timescales
- Defensive Agentic AI Adoption Lags Offensive Deployment
Executive Summary
In a landmark case documented by Anthropic, AI systems autonomously conducted 80-90% of a sophisticated cyber espionage campaign targeting approximately 30 organizations, performing reconnaissance, vulnerability discovery, exploit development, credential harvesting, and data exfiltration at machine speeds. This represents the first operationalized instance of what strategists term "highly autonomous cyber-capable agents" (HACCAs)-systems that execute end-to-end campaigns with minimal human direction.
The strategic implications are severe. The cost to go from vulnerability discovery to exploit used to be weeks and thousands of dollars; now it's near zero. This economic shift democratizes sophisticated attack capabilities while simultaneously compressing the window for defensive response. In 2025, the average breakout time from initial access to lateral movement dropped to under 30 minutes, with AI-enabled tools automating reconnaissance, generating exploits, and scanning thousands of systems simultaneously, allowing small teams or single operators to run campaigns that once required large coordinated groups.
The deterrence problem is acute: AI could lead to classes of weapons that are very difficult to trace the origins of, including sophisticated cyber-attacks, and might be used as an engine of disinformation in ways that obfuscate the perpetrator of an attack. When attribution becomes ambiguous and response windows collapse, the credibility of deterrent threats erodes.
- Vulnerability Discovery Acceleration Inverts Offense-Defense Balance
Claude Mythos Preview autonomously discovered and exploited zero-day vulnerabilities in every major operating system and web browser, with engineers asking the model to find very low confidence code execution vulnerabilities overnight and waking up to complete, working exploits. The model has autonomously discovered thousands of zero-day vulnerabilities in major operating systems and web browsers. This capability shift means vulnerability discovery is no longer a scarce, human-dependent resource, it is now a scalable, automated function. For state-sponsored actors, this eliminates the primary constraint on cyber operation tempo.
- Autonomous Attack Execution Reduces Human Operator Dependency
A Chinese state-sponsored group jailbroken Claude Code to launch cyber operations against roughly thirty global targets, using Anthropic's software coding agent with custom scaffolding to automate eighty to ninety percent of the operation, marking the first known incident of a large-scale cyber campaign planned and executed primarily by an AI system rather than human operators. These agents, with minimal human direction and oversight, executed the labor-intensive steps of the attack, enabling threat actors to operate at greater speed and scale, with the agent conducting between 80 and 90% of the offensive operation while humans shifted from operators to supervisors. This structural change means nation-states can now scale operations beyond their human workforce constraints.
- Attribution Opacity Undermines Deterrence Credibility
Emerging AI-enabled weapon classes, including sophisticated cyber-attacks, biological weapons with long latencies, and disinformation engines, are designed to obfuscate their origin, making attribution systematically harder. Threat actors are manipulating threat indicators to obscure attribution, mimicking the tactics, techniques, and procedures of known threat actors to confuse analysts and delay response, with AI-powered campaigns adapting dynamically and automating deception at scale. When attackers can obscure their identity through AI-generated false flags, the foundational assumption of deterrence, that the defender can identify and attribute the attacker, collapses.
- Machine-Speed Operations Exceed Human Decision Timescales
Most cyber defenses still run on a human timeline with triage in hours, remediation in days, and patching in weeks, while AI-enabled attackers move in minutes. When an adversary adapts at machine speed, waiting for a committee to authorize a shutdown is a failure; governance must empower systems to enter a "deterministic safe state", an automated, pre-authorized posture that protects physical equipment while humans oversee recovery. This temporal mismatch means traditional command-and-control structures become liabilities rather than assets.
- Defensive Agentic AI Adoption Lags Offensive Deployment
77% of organizations now use generative AI or large language models in their security stack, and 67% have deployed agentic AI for autonomous or semi-autonomous security operations, with the areas where AI is delivering the most impact being anomaly detection and novel threat identification (72%), automated response and containment (48%), and vulnerability management (47%). However, CISOs and executives were the most enthusiastic with 56% strongly agreeing that AI improves defensive capabilities, while security operations practitioners only 25% strongly agreed, with the people who sit in front of these tools every day being the least impressed. This gap suggests defensive AI adoption is not yet operationally effective at scale.
Strategic Analysis: Cost-Benefit Inversion for State-Sponsored Operations
Offensive Economics Transformation
The economics have flipped, the cost to go from vulnerability discovery to exploit used to be weeks and thousands of dollars, now it's near zero, so instead of mass 'spray and pray' campaigns, attackers will get micro-targeted attacks built for a single system, a single company, maybe even a single developer. This economic shift has three strategic consequences:
First, capability democratization. Autonomous capabilities are moderate-to-high confidence to proliferate and enable less sophisticated actors to conduct more operations at faster speeds, which may shift advantages toward attackers until defensive capabilities are deployed at scale. State-sponsored actors no longer require elite technical talent pools; they can deploy agentic systems with minimal human oversight. Such automation has drastically lowered barriers to sophisticated cyberattacks, enabling smaller adversaries to perform operations previously limited to well-resourced actors.
Second, operational tempo multiplication. AI predator swarms will transform cyberattacks from manual operations into scalable, autonomous campaigns, with AI agents capable of unleashing 10,000 personalized phishing emails per second, crafting zero-day exploits instantly, and deploying ransomware across thousands of endpoints in under a minute. A single operator directing multiple agentic systems can execute operations that previously required coordinated teams.
Third, target expansion. Attackers will shift focus from single organizations to weaponizing vulnerabilities within shared infrastructure, with the rise of 'Connector Supply-Chain Compromise', where poisoning a single trusted component infects every model and application relying on it, allowing threat actors to compromise multiple companies simultaneously through their AI dependencies. This means the return on investment for a single vulnerability discovery increases exponentially.
Deterrence Stability Degradation
The traditional deterrence model rests on three pillars: attribution (knowing who attacked), credible threat (ability to retaliate), and decision time (opportunity to choose response). AI-enabled operations erode all three.
Attribution Collapse. An aspiring aggressor is only deterred if they think they might get caught; with national technical means like launch detection satellites and nuclear forensics, an actor attacking with nuclear weapons could not expect to get away with it, but AI might change that calculation by leading to classes of weapons that are very difficult to trace the origins of, including sophisticated cyber-attacks. When AI systems can generate false attribution signatures, defenders cannot confidently identify the attacker, making retaliation politically and legally problematic.
Escalation Pathway Compression. Powerful AI complicates escalation pathways, causing a relatively stable condition to rapidly develop into conflict or war, with AI-enabled gray zone activity at the sub-conventional level potentially escalating a minor crisis to a major war, bypassing the "rungs" of the traditional, linear escalation ladder. A new generation of AI-enhanced cyber capabilities will amplify the risk of inadvertent escalation caused by the co-mingling of nuclear and strategic non-nuclear weapons and the increasing speed of warfare, and future iterations of AI-enhanced cyber counterforce capabilities will complicate existing challenges of cyber defence and compromise nuclear assets.
Decision Authority Displacement. Decision-makers discussed how best to target the decision layer, including introducing new forms of flexible deterrent and response options designed to spoof and confuse AI/ML applications that help senior leaders analyze a crisis, with the goal being to gain an information advantage and use it for leverage without triggering inadvertent escalation. When AI systems advise decision-makers on crisis response, adversaries can target those systems to distort threat perception.
Defensive Architectural Requirements for Deterrence Maintenance
Maintaining deterrence stability requires a fundamental shift from prevention-centric to resilience-centric architectures. The old model assumed defenders could patch faster than attackers could exploit. That assumption is now false.
1. Machine-Speed Autonomous Defense
Cyber defense must move to AI speed, meaning early containment actions such as isolating systems, blocking malicious traffic, revoking suspicious sessions, and initiating remediation cannot wait for manual approval and must occur automatically within defined limits while an intrusion is still unfolding. Defenders now have AI agents that can automatically probe systems, reproduce exploit chains, score impact, and even trigger fixes, and combined with human creativity, this creates a feedback loop that adapts as fast as attackers do.
Implementation requirement: Organizations must pre-authorize defensive actions within defined parameters. Organizations must strengthen 'Human-ON-the-loop' oversight; when an adversary adapts at machine speed, waiting for a committee to authorize a shutdown is a failure; governance must empower systems to enter a "deterministic safe state", an automated, pre-authorized posture that protects physical equipment while humans oversee recovery.
2. Isolation and Segmentation as Default Architecture
If an organization's security posture depends entirely on the security team being able to patch vulnerable resources faster than attackers can exploit those vulnerabilities to compromise them, they will eventually find themselves fighting a losing battle; secure design must ensure isolation for critical components. Organizations should create barriers that stop problems from spreading across connected systems, implementing isolation boundaries that contain failures.
Implementation requirement: Apply the principles of Zero Trust not just to humans, but to non-human entities acting in infrastructure; treat every AI agent as an untrusted entity until verified, regardless of its role or historical behavior; do not give agents "God mode" access to cloud environments; instead, implement just-in-time access and least-privilege scopes, with an agent designed to schedule meetings having write access only to the calendar API, not the corporate email server or customer database.
3. Resilience Through Graceful Degradation
Meaningful resilience is defined by Graceful Degradation, or the ability to keep 'black start' capabilities intact and the neighborhood energized even when the digital layer is compromised; there is a breakdown in the long-standing assumption that 'air-gaps' or 'obscure protocols' provide security; in the age of AI, every technical manual is an open book; true resilience requires a shift back to engineering fundamentals: assuming the digital wall will be breached and ensuring that human operators can still pull the plug on 'smart' features to run the grid or plant manually.
Implementation requirement: Critical infrastructure must maintain manual override capabilities and physical isolation options. The year 2026 marks a pivotal moment: the end of the endpoint-centric security model and a shift towards a non-negotiable 'assume compromise' mindset; we are no longer debating if an intrusion will happen, but operating under the hard truth that it moderate-to-high confidence already has; defenses must move beyond reaction, designing systems that provide resilience and authoritative response, anchored by a new truth layer, when attacks inevitably occur.
4. Behavioral Anomaly Detection Over Signature-Based Defense
SIEM and EDR tools were built to detect anomalies in human behavior; an agent that runs code perfectly 10,000 times in sequence looks normal to these systems, but that agent might be executing an attacker's will. Organizations need to include behavior- and anomaly-based monitoring in their system, looking for unusual access to management tools, automation platforms, or service accounts doing things outside their normal pattern.
Implementation requirement: The emergence of increasingly sophisticated AI swarms underscores the need for integrated defensive architectures capable of monitoring behavioral patterns, verifying content provenance, and rapidly responding to coordinated influence operations; without such mechanisms, the ability of automated systems to manipulate information environments may continue to grow, posing significant risks to democratic institutions, economic stability, and public trust in digital information systems.
5. Attribution Resilience Through Forensic Redundancy
Since AI-enabled operations can obscure attribution, defensive architectures must assume attribution will be ambiguous and design response options that do not depend on certain attribution. Decision-makers should introduce new forms of flexible deterrent and response options designed to spoof and confuse AI/ML applications that help senior leaders analyze a crisis, with the goal being to gain an information advantage and use it for leverage without triggering inadvertent escalation.
Implementation requirement: Develop graduated response frameworks that do not require attribution certainty. Responses should focus on capability denial (making attacks harder) rather than retaliation (punishing the attacker), since retaliation requires confident attribution.
Cross-Domain Integration: Geopolitical Implications
The convergence of AI-enabled cyber operations with state-sponsored activity creates second and third-order effects across multiple domains:
Political-Military Integration: As sanctions and law enforcement pressure mounts on ransomware platforms, these platforms will be forced to align with state interests, spawning Geopolitical-RaaS (G-RaaS): state-tolerated or state-steered ransomware ecosystems that pursue both profit and national strategic interests, blurring the line between organized cybercrime and asymmetric digital warfare and complicating attribution and insurance coverage.
Alliance Credibility Risk: Anthropic has been in ongoing discussions with US government officials about Claude Mythos Preview and its offensive and defensive cyber capabilities, noting that securing critical infrastructure is a top national security priority for democratic countries and that the emergence of these cyber capabilities is another reason why the US and its allies must maintain a decisive lead in AI technology. This creates pressure for allied nations to develop comparable capabilities, potentially accelerating proliferation.
Critical Infrastructure Vulnerability: In 2026, digital conflict is a permanent part of global competition, with state-sponsored cyber threats exploiting supply chains, identity systems, and critical infrastructure to expand geopolitical risk. Autonomous offensive AI agents could enable nation-states to conduct continuous operations across multiple targets at an increased tempo, particularly concerning given that actors like Salt Typhoon and Volt Typhoon have already compromised critical infrastructure.
Bottom Line
Critical uncertainties:
- Whether defensive agentic AI can scale to match offensive deployment speed
- How attribution ambiguity will affect alliance cohesion and escalation management
- Whether critical infrastructure can be redesigned for resilience faster than threats evolve
- How state-sponsored actors will coordinate on norms to prevent mutual vulnerability
Immediate policy implications: Nations must prioritize (1) pre-authorization frameworks for autonomous defense, (2) resilience-first infrastructure redesign, (3) graduated response options that do not depend on attribution certainty, and (4) multilateral norms development for AI-enabled cyber operations before proliferation accelerates beyond current state-sponsored actors.
Alternative Hypotheses
Multiple competing hypotheses were evaluated during this analysis. The conclusions above reflect the hypothesis best supported by available evidence.
Sources
- How AI is getting better at finding security holes - NPR
- Frightening AI advances speed race to secure critical infrastructure - Axios
- The zero-day timeline just collapsed. Here's what security leaders do next - csoonline.com
- AI Is Forcing a Rethink in Cybersecurity - WSJ
- The New Rules of Engagement: Matching Agentic Attack Speed - SecurityWeek
- Anthropic's Claude Mythos Finds Thousands of Zero-Day Flaws Across Major Systems - The Hacker News
- Anthropic Glasswing: The Future of Cybersecurity in the Age of AI - SecNews.gr
- Why Anthropic's Mythos Is a Systemic Shift for Global Cybersecurity - GovTech
Methodology
This analysis was generated by Mapshock, including automated source grading, bias detection, and multi-hypothesis evaluation.