AI Enables First Zero-Day 2FA Bypass for Mass Exploitation

First Known
AI-Developed Zero-Day
2FA Bypass
Critical Vulnerability
Thousands
APT45 CVE Analysis Prompts
5,000+
Vulnerability Cases in Training

Google's Threat Intelligence Group (GTIG) has confirmed what security researchers have long anticipated: threat actors successfully used artificial intelligence to discover and weaponize a zero-day vulnerability for mass exploitation. The vulnerability—a two-factor authentication bypass in a popular open-source web administration tool—represents the first documented instance of AI-assisted vulnerability discovery and exploit development deployed in active attacks.

The exploit targets a semantic logic flaw stemming from hard-coded trust assumptions, precisely the type of vulnerability that large language models excel at identifying. Google worked with the affected vendor to patch the flaw but has not disclosed the tool's name. The incident confirms that AI has fundamentally altered the threat landscape, compressing timelines between vulnerability discovery and exploitation.

Identifying AI-Generated Exploit Code

GTIG assessed with high confidence that an AI model generated the Python exploit script based on distinctive code characteristics. The script contained excessive educational docstrings, a hallucinated CVSS score, and structured Pythonic formatting typical of LLM training data. Additional markers included detailed help menus and a clean ANSI color class implementation—all hallmarks of AI-generated code rather than human-written exploits.

The 2FA bypass requires valid user credentials for exploitation, indicating the attackers likely planned to use it as part of a broader credential compromise campaign. The vulnerability's nature—a high-level semantic logic flaw—demonstrates AI's capability to identify complex architectural weaknesses that traditional scanning tools might miss.

Critical Reality Check

Ryan Dewhurst, watchTowr's Head of Threat Intelligence: 'AI is already accelerating vulnerability discovery, reducing the effort needed to identify, validate, and weaponize flaws. Discovery, weaponization, and exploitation are faster. We've been watching timelines compress for years. There is no mercy from attackers, and defenders don't get to opt out.'

PromptSpy: Autonomous AI-Driven Android Malware

Beyond exploit development, threat actors are deploying AI for autonomous malware operations. PromptSpy, an Android malware family, abuses Google's Gemini API to analyze screen content in real-time and determine optimal attack actions. The malware uses an autonomous agent module to navigate the Android interface, monitor user activity, and adapt its behavior dynamically without human intervention.

PromptSpy's capabilities extend to sophisticated anti-analysis techniques. It captures biometric data to replay authentication gestures including PINs and patterns. An 'AppProtectionDetector' module identifies the on-screen coordinates of uninstall buttons and serves invisible overlays to intercept touch events, making the button appear unresponsive. The malware maintains operational resilience by allowing runtime rotation of C2 infrastructure, Gemini API keys, and VNC relay servers without payload redeployment.

State-Sponsored AI Adoption Patterns

Multiple advanced persistent threat groups have integrated AI tools into their operations, demonstrating widespread adoption across nation-state actors:

Specialized Training Data for Exploit Development

Threat actors are creating purpose-built AI training resources to enhance exploit capabilities. Researchers identified a GitHub repository called 'wooyun-legacy' designed as a Claude code skill plugin containing over 5,000 real-world vulnerability cases from the Chinese WooYun disclosure platform (2010-2016). By priming AI models with historical vulnerability data, attackers enable in-context learning that steers the model to approach code analysis like a seasoned expert, identifying logic flaws that base models might overlook.

A suspected China-aligned threat actor deployed specialized agentic tools including Hexstrike AI and Strix in attacks against Japanese technology targets, demonstrating the development of domain-specific AI offensive tools.

Defense Imperative

Organizations must assume AI-accelerated exploitation is already occurring at scale. Implement zero-trust architectures, enforce phishing-resistant authentication, and reduce time-to-patch windows. Traditional 30-60 day patch cycles are obsolete in an AI-enabled threat environment.

Implications for Defensive Operations

The emergence of AI-discovered zero-days fundamentally alters the security equation. The traditional window between vulnerability disclosure and widespread exploitation—historically measured in weeks—will continue compressing toward hours or minutes. Security teams must recalibrate their assumptions about threat actor capabilities and resource requirements.

Defenders face an asymmetric disadvantage: while attackers need only one successful vulnerability discovery, defenders must identify and patch all potential weaknesses. AI multiplies this asymmetry by enabling attackers to scale vulnerability research operations that previously required specialized expertise and significant time investment. Organizations relying on obscurity or the complexity of their systems for protection should reassess that posture immediately.

Recommended Actions

Security teams should implement the following measures to address AI-accelerated threats:

The first confirmed AI-discovered zero-day represents an inflection point in offensive security capabilities. Organizations that fail to adapt their defensive strategies to this new reality will find themselves increasingly vulnerable to accelerated, scaled exploitation campaigns. The question is no longer whether AI will be used for vulnerability discovery and exploit development—it's how quickly your defenses can evolve to match this new threat velocity.

Questions about your exposure?

RedEye Security provides assessments for organizations that need to understand their real risk.

Talk to us