AI Breakthrough Tracker

AI-Generated Exploits and Security Threats

AI-Generated Exploits and Security Threats

Key Questions

What did Google's Threat Intelligence Group (TIG) discover about AI in cyberattacks?

Google's TIG identified the first criminal use of AI-generated code for a zero-day Python exploit to bypass 2FA. This marks a significant milestone in adversarial AI applications.

How are hackers using AI to bypass 2FA?

Hackers leveraged AI-generated code to exploit a zero-day vulnerability in Python, enabling them to circumvent two-factor authentication (2FA). Google's report highlights AI's growing role as a tool in cybersecurity threats.

What broader security risks does this highlight discuss?

The highlight addresses adversarial milestones, supply chain attacks on ML packages like Mini Shai-Hulud, and dual-use risks from agentic AI and code-generation tools. Miles Brundage noted it as a major development and the first frontier AI audit.

Google TIG first criminal AI zero-day Python 2FA bypass; adversarial milestones; supply chain ML pkgs (Mini Shai-Hulud); dual-use risks from agentic/code-gen.

Sources (2)
Updated May 14, 2026