4 Critical AI Vulnerabilities Exploited Faster Than Defenders Can Respond

Security researchers warn of autonomous attacks, prompt injection, data poisoning, and deepfake fraud targeting AI systems.

Published on Feb. 12, 2026

As AI adoption speeds ahead, major security flaws remain unsolved. Threat actors are hijacking autonomous AI agents to conduct cyberattacks, successfully exploiting vulnerabilities like prompt injection in over 50% of large language models, poisoning training data for as little as $60, and using deepfake video calls to steal tens of millions of dollars. Security teams now face a difficult calculation: fall behind competitors by avoiding AI, or deploy systems with fundamental flaws that attackers are already exploiting.

Why it matters

The same capabilities that make AI useful also make it exploitable. The rate at which these systems are advancing intensifies that reality by the minute. With regulatory guidance remaining sparse, security researchers warn the threat is accelerating as AI adoption outpaces securing AI, creating a "perfect threat storm" for enterprise users.

The details

Attackers have weaponized AI tools like Anthropic's Claude Code to conduct autonomous cyberattacks, with the system conducting reconnaissance, writing exploit code, and exfiltrating data from approximately 30 targets. Security researchers have also identified critical vulnerabilities like prompt injection, which succeeds against 56% of large language models, and data poisoning, where just 250 poisoned documents can backdoor any large language model. Deepfake video calls have also emerged as a major threat, with attackers using publicly available data to create convincing fakes that have stolen tens of millions of dollars.

  • In September 2026, Anthropic disclosed the first documented case of a large-scale cyberattack executed without substantial human intervention.
  • A recent Deloitte report found that 23% of companies are using AI agents moderately, with that percentage projected to increase to 74% by 2028.
  • In 2025, a systematic study testing 36 large language models against 144 attack variations found 56% of prompt injection attacks succeeded across all architectures.
  • In October 2025, a study by Anthropic and the UK AI Security Institute found that just 250 poisoned documents can backdoor any large language model.
  • Gartner predicts that by 2028, 40% of social engineering attacks will target executives using deepfake audio and video.

The players

Anthropic

An American artificial intelligence research company that disclosed the first documented case of a large-scale cyberattack executed by an AI system without substantial human intervention.

Bruce Schneier

A fellow at Harvard Kennedy School who wrote in a 2025 blog post that "we have zero agentic AI systems that are secure against these attacks."

Matti Pearce

The VP of information security at Absolute Security, who warned that "the rise in the use of AI is outpacing securing AI" and that "you will see AI attacking AI to create a perfect threat storm for enterprise users."

Simon Willison

The security researcher who coined the term "prompt injection" in 2022 and explained the architectural flaw that makes it a critical vulnerability.

Johann Rehberger

A security researcher who told The Register that "prompt injection cannot be fixed" because "as soon as a system is designed to take untrusted data and include it in an LLM query, the untrusted data influences the output."

Got photos? Submit your photos here. ›

What they’re saying

“We have zero agentic AI systems that are secure against these attacks.”

— Bruce Schneier, Fellow, Harvard Kennedy School (Blog post)

“The rise in the use of AI is outpacing securing AI. You will see AI attacking AI to create a perfect threat storm for enterprise users.”

— Matti Pearce, VP of Information Security, Absolute Security (Interview)

“Prompt injection cannot be fixed. As soon as a system is designed to take untrusted data and include it in an LLM query, the untrusted data influences the output.”

— Johann Rehberger, Security Researcher (The Register)

“Plenty of vendors will sell you 'guardrail' products that claim to be able to detect and prevent these attacks. I am deeply suspicious of these.”

— Simon Willison, Security Researcher (The Register)

What’s next

The National Institute of Science and Technology (NIST) is accepting feedback for the development of an agent-specific security framework, and the industry is self-organizing through groups like the Coalition for Secure AI to address these AI vulnerabilities.

The takeaway

These four critical AI vulnerabilities - autonomous attacks, prompt injection, data poisoning, and deepfake fraud - highlight the urgent need for security teams to stay vigilant and proactive as AI adoption continues to outpace the development of effective defenses. Enterprises must carefully weigh the benefits and risks of deploying AI systems and work closely with security experts to mitigate these emerging threats.