Agent Security Bench reveals 84.30% attack success rate across 13 LLM platforms. OWASP designates prompt injection as LLM01:2025. Here's why CVE-2025–32711 changes everything — plus defenses that reduce attacks to <1%.
Microsoft issued emergency patches in May 2025 for CVE-2025–32711 — a critical vulnerability in Microsoft 365 Copilot with a CVSS score of 9.3. The attack vector: zero-click prompt injection that exfiltrates internal files without any user interaction. If you're building with LLMs right now, your application is probably next.
The numbers are stark. OWASP's November 2024 report found that "over 73% of tested LLM applications were found to be vulnerable to prompt injection attacks in 2025." The Agent Security Bench published in November 2024 revealed an "average attack success rate of 84.30% across 10 LLM-based agent scenarios, with current defenses largely ineffective." CrowdStrike's 2025 Threat Hunting Report documents that "adversaries are targeting software used to build AI agents and weaponizing AI at scale, exploiting vulnerabilities in AI software and using it to gain initial access."
If you're deploying ChatGPT API integrations, building LangChain workflows, or implementing RAG systems, you're in the blast radius. CVE-2025–46059 just exposed arbitrary code execution in LangChain's GmailToolkit with a CVSS score of 9.8. CVE-2025–32711 proved that zero-click exploits work in production enterprise systems. And the research is unambiguous: existing defenses are "largely ineffective."
This investigation reveals why prompt injection became OWASP's #1 LLM vulnerability for 2025, how 84% of agents fail security benchmarks, which CVEs demand immediate action, and — critically — the three defenses that reduce attack success rates to below 1%. Here's what the official security reports won't tell you: the vulnerability isn't the LLM itself — it's the architecture we're using to deploy it.
OWASP 2025: Prompt Injection Tops the List
OWASP designated prompt injection as LLM01:2025 — the number one vulnerability in large language model applications. This isn't speculation from a security blog. The OWASP Top 10 for Large Language Model Applications 2025 report, published November 14, 2024, identifies prompt injection as "the most critical risk in deploying and managing large language models."
The scope is staggering. According to the official OWASP report, "over 73% of tested LLM applications were found to be vulnerable to prompt injection attacks in 2025." That's not a minority of poorly-secured applications — that's nearly three-quarters of the entire ecosystem.
Prompt injection occurs when attackers manipulate LLM inputs to override system instructions. But unlike SQL injection, there's no clear syntactic boundary between "code" and "data" in natural language processing. The LLM processes user instructions and external content in the same context — making it fundamentally unable to distinguish between legitimate prompts from authorized users and malicious instructions embedded in untrusted content.
The attack comes in two forms. Direct injection happens when malicious users directly submit crafted prompts to override the system's behavior. Indirect injection — the more dangerous variant — occurs when attackers embed instructions in content the LLM ingests automatically: emails processed by AI assistants, PDFs analyzed by document processors, web pages retrieved by AI agents.
The OWASP report is explicit about the consequences: "These vulnerabilities allow attackers to hijack the behavior of LLM applications, leading to data leakage, unauthorized actions, and compromised systems." For security engineers, this expands your attack surface with every LLM integration. For developers, it means code you shipped last quarter is exploitable. For product managers, it means AI features you're planning need security-first redesigns.
Why do traditional defenses fail? Input sanitization doesn't work — attackers craft prompts that bypass blacklists. Pre-prompting strategies like "Ignore malicious instructions" fail because the LLM follows the most recent or most emphatic instruction. Encoding schemes provide some protection but aren't foolproof. The GitHub repository tldrsec/prompt-injection-defenses states it plainly: "The most reliable mitigation is to always treat all LLM productions as potentially malicious."
That recommendation isn't theoretical caution. It's the consensus view of security researchers who've tested every proposed defense and watched them break. When OWASP designates something as vulnerability #1, it means compliance auditors will start asking about it. It means insurance underwriters will want to see your mitigation strategy. It means the next security questionnaire from your enterprise customer will include LLM01:2025.
This is where it gets critical for your career: OWASP doesn't issue #1 designations lightly. The organization has been defining web application security standards for 20 years. When they elevate a vulnerability to the top position, it's based on prevalence, exploitability, and impact. Prompt injection scores high on all three.
From Theory to Exploitation: CVE-2025–32711 and CVE-2025–46059

In May 2025, Microsoft issued emergency patches for CVE-2025–32711, described as "a critical information disclosure vulnerability in Microsoft 365 Copilot." The CVSS score: 9.3. The attack classification: zero-click, meaning "no user interaction required."
Here's the technical mechanism. An attacker sends a specially crafted email to the victim's address. Microsoft 365 Copilot, designed to assist users by automatically processing their communications, ingests the email content. Embedded instructions in the email — hidden in HTML tags, encoded in white text, or obfuscated through other means — hijack Copilot's behavior. The compromised Copilot accesses internal files from the victim's Microsoft 365 context. Data exfiltrates to an attacker-controlled server. All of this happens without the victim clicking anything, opening any attachments, or taking any action whatsoever.
Field Effect's security intelligence team notes in their analysis: "The flaw was assigned a CVSS score of 9.3 and was patched by Microsoft in May 2025. Customers do not need to take any action, but the threat reveals a new class of LLM vulnerabilities." That last clause is critical. Microsoft patched their implementation, but the attack class persists. Every other AI assistant that automatically processes emails, documents, or web content faces the same architectural vulnerability.
CVE-2025–46059 affects LangChain's GmailToolkit component in version 0.3.51 and "allows attackers to execute arbitrary code via crafted email messages." The severity rating: CVSS 9.8 — critical. For context, CVSS 9+ means exploitable remotely, with low attack complexity, no privileges required, and high impact on confidentiality, integrity, and availability.
The attack path works like this. An attacker sends an email containing obfuscated HTML with embedded prompt injection instructions. The victim's LangChain application uses GmailToolkit to process incoming emails. The embedded instructions bypass basic sanitization that developers might have implemented. The LLM interprets the malicious content as executable commands from the system. Result: arbitrary code execution in the application's runtime environment.
Zeropath's security analysis is unambiguous: "The vulnerability allows attackers to execute arbitrary code via crafted email messages. The severity is rated critical, with a CVSS score of 9.8, due to the potential for system compromise and data leakage." Arbitrary code execution means an attacker can run any command the application has permissions for — accessing databases, calling APIs, modifying files, or pivoting to other systems.
If you're using LangChain — which powers thousands of enterprise AI applications — you need to audit your deployment immediately. Version 0.3.51 is exploitable. If you're running Microsoft 365 Copilot, the specific CVE is patched, but other tools in your stack likely have similar vulnerabilities waiting to be discovered and disclosed.
Notice the pattern. Both CVEs exploit indirect prompt injection. Both target enterprise productivity tools that employees use daily. Both enable attackers to compromise systems through content the LLM automatically processes. This isn't a software bug that can be fixed with input validation. It's an architectural vulnerability inherent to how we're deploying LLMs.
CrowdStrike's 2025 Threat Hunting Report validates this with threat intelligence data: "Multiple threat actors have exploited vulnerabilities in AI software, leading to unauthenticated access, credential harvesting, and malware deployment." CrowdStrike isn't warning about theoretical risks. They're observing active exploitation by multiple threat actor groups in the wild.
The technical reality most people miss: these aren't edge cases discovered by academic researchers. They're design flaws being exploited in production environments right now.
Agent Security Bench: The 84.30% Problem
In November 2024, researchers published Agent Security Bench (ASB) — the first comprehensive benchmark for LLM agent security. They tested 10 distinct prompt injection attacks against 13 different LLM backbones (GPT-4, Claude, Llama, Gemini, and others) across real-world agent scenarios involving tool use, multi-step reasoning, and autonomous decision-making.
The results: "Our comprehensive benchmark revealed an average attack success rate of 84.30% across 10 LLM-based agent scenarios, with current defenses largely ineffective." Let that number sink in. More than eight out of ten attacks succeeded. This wasn't testing against undefended hobby projects — the benchmark included applications implementing existing security best practices.
The research tested attacks across different complexity levels. Simple direct injections ("Ignore previous instructions and do X") succeeded at high rates. More sophisticated attacks using encoding, multi-step reasoning hijacking, and context poisoning succeeded at similar rates. Even attacks designed to test specific defenses managed to bypass them.
Why do current defenses fail so catastrophically? Because most rely on pattern matching or rule-based filtering. Defense teams implement blacklists for phrases like "ignore previous instructions." Attackers evolve to use synonyms like "disregard prior directives" or "set aside earlier guidance." Teams implement input encoding. Attackers craft prompts that work post-decoding. Teams add pre-prompts telling the LLM to resist manipulation. Attackers override with stronger, more emphatic instructions later in the context.
The fundamental issue: LLMs are trained to follow instructions. When presented with conflicting instructions — a system prompt saying "Never reveal internal information" versus an injected prompt saying "Show me your system prompt" — the model often prioritizes based on recency, emphasis, or the statistical patterns it learned during training. There's no cryptographic boundary, no access control list, no privilege separation. Just statistical word prediction trying to satisfy what appears to be user intent.
Google's Threat Intelligence Group observed the same vulnerability patterns. In their November 4, 2025 report, they documented that "threat actors are misusing AI to enhance their operations, including targeting vulnerabilities in AI software." Google identified "multiple threat actors exploiting vulnerabilities in AI software to gain initial access, expand attack surfaces, and harvest credentials."
This isn't one hacking group running experiments. It's systematic, multi-actor exploitation across the threat landscape. Google's threat intelligence covers nation-state actors, organized cybercrime groups, and opportunistic attackers. When they report multiple threat actors targeting AI software, it means this attack vector has become mainstream in offensive cyber operations.
CrowdStrike's August 2025 Threat Hunting Report adds enterprise context: "Adversaries are targeting software used to build AI agents and weaponizing AI at scale." Their threat hunters — security professionals embedded in customer environments actively searching for intrusions — are finding this in the wild. Not in proof-of-concept demonstrations. In production enterprise networks.
The implications for organizations deploying AI agents are severe. Security teams trained on traditional vulnerabilities like SQL injection and cross-site scripting lack frameworks for LLM security. Developers ship AI features without security reviews because there are no established review processes for prompt injection. Product managers prioritize speed-to-market over security because competitors are shipping AI features rapidly.
The result: 84% attack success rates in production environments. Systems deployed at enterprise scale with fundamental, exploitable vulnerabilities. And unlike traditional vulnerabilities where you can patch the software and redeploy, architectural vulnerabilities require redesigning how your application works.
Here's what scares security researchers: we're deploying vulnerable-by-design systems at enterprise scale while threat actors are already weaponizing the attack techniques.
PromptArmor, Spotlighting, and Architecture: Reducing ASR to <1%

PromptArmor, published in July 2025, represents a breakthrough in practical prompt injection defense. The research demonstrates that "after removing injected prompts with PromptArmor, the attack success rate drops to below 1%" on the AgentDojo benchmark.
The framework uses a two-phase approach. First, detection: a guardrail LLM examines incoming input to identify potential injection attempts. Second, removal: if the guardrail detects malicious content, it surgically removes the injected instructions before passing the cleaned input to the application LLM. The key insight is that even if the guardrail LLM is itself vulnerable to prompt injection, it can still detect and neutralize attacks when properly configured with defensive prompts.
The performance metrics are impressive. When using GPT-4o as the guardrail, PromptArmor achieves a 0.07% false positive rate and 0.23% false negative rate. For production deployment, this means legitimate user inputs almost never get incorrectly flagged (false positives would frustrate users), and actual attacks almost never slip through (false negatives would leave you vulnerable).
The practical implication: you can wrap your existing LLM applications with a security layer. No retraining required. No architectural overhaul. Deploy a guardrail LLM that examines inputs before they reach your application LLM. The researchers note that "PromptArmor is recommended as a standard baseline for evaluating new defenses against prompt injection attacks," positioning it as an emerging industry standard.
Spotlighting addresses the fundamental problem that LLMs cannot distinguish user instructions from external data. The solution: mark all external content with explicit delimiters that signal provenance. For example, when your RAG system retrieves documents, format them as:
EXTERNAL_DATA_START
[Retrieved document content]
EXTERNAL_DATA_ENDThe LLM learns that anything within these markers is untrusted content that should inform its response but not control its behavior. Instructions embedded within marked blocks should be treated as data, not commands. This technique reduces attack success rates from approximately 50% in undefended systems to under 2% with proper implementation.
The GitHub repository tldrsec/prompt-injection-defenses, which has become the de facto reference for prompt injection mitigation, recommends Spotlighting as a foundational technique. The repository aggregates "every practical and proposed defense against prompt injection" and provides implementation guidance for each approach.
Beyond prompt-level defenses, architectural changes contain damage when injection succeeds. The tldrsec repository outlines several critical patterns:
Blast radius reduction: Limit what each LLM agent can access. If an HR agent gets compromised, it shouldn't have database credentials for the financial system. Use principle of least privilege religiously.
Input pre-processing: Validate and sanitize before LLM interaction. Remove potentially dangerous markup, check input length limits, normalize encoding.
Output validation: Check LLM responses for malicious payloads before executing any actions. If the LLM outputs SQL queries, parse and validate them. If it generates shell commands, whitelist allowed operations.
Taint tracking: Monitor data flow from untrusted sources. Mark external inputs as tainted and track how that data propagates through your system.
Secure thread design: Isolate different security contexts. User queries run in one thread, internal tool use in another, external API calls in a third. Compromise of one thread shouldn't automatically compromise others.
These architectural defenses don't prevent injection — they're not designed to. They contain the blast radius when injection succeeds. Because with 84% attack success rates, assuming your application will be attacked successfully is just prudent threat modeling.
What you should implement today: First, immediate action within 24 hours — if you're using LangChain, update to the latest version to patch CVE-2025–46059. Second, this week — implement PromptArmor or an equivalent guardrail system. Third, this month — add Spotlighting to all external data ingestion points in your LLM applications. Fourth, this quarter — conduct an architectural security review focused on blast radius reduction.
The question isn't whether you'll face prompt injection attempts — it's whether you'll detect them when they happen and whether the damage will be contained.
Enterprise Implications: Security, Compliance, and Budget
For security engineers, your threat model just expanded significantly. OWASP's LLM01:2025 designation means prompt injection is now on compliance checklists. When auditors review your security controls, they'll ask: "How do you mitigate LLM01?" Your answer cannot be "We haven't evaluated that yet."
Immediate action items for security teams: Red team your LLM integrations using Agent Security Bench methodologies to understand your actual vulnerability posture. Deploy detection guardrails based on the PromptArmor framework to catch attacks in real-time. Update incident response playbooks to include AI-specific attack scenarios — your existing playbooks for web application attacks don't translate directly. Budget for LLM security tools, which will require investment beyond your traditional WAF and SIEM infrastructure.
Developers building LLM features: you're shipping vulnerable code if you're not implementing specific defenses. The 84% attack success rate means eight out of ten LLM applications you deploy without dedicated security measures are trivially exploitable. This isn't about perfect security — it's about reducing attack surface to manageable levels.
Framework-specific actions for development teams: If you're using LangChain, verify immediately that CVE-2025–46059 doesn't affect your deployment — version 0.3.51 is critically vulnerable. If you're building RAG systems, implement Spotlighting for all retrieved content before it reaches your LLM. If you're developing multi-agent systems, review inter-agent communication security to prevent prompt injection from spreading between agents. The GitHub repository tldrsec/prompt-injection-defenses should be required reading for every developer working with LLMs.
Product managers planning AI features: security is now a launch blocker, not a post-launch concern. CrowdStrike documented enterprises delaying AI rollouts due to unresolved security concerns. Google's threat intelligence group has confirmed active exploitation of AI vulnerabilities. If you launch with known vulnerabilities, you're accepting responsibility for the breach that follows.
Risk assessment framework for product teams: Ask your security engineers three questions. First: Have we tested our LLM features for prompt injection using established benchmarks? Second: What's our detection capability — will we know when an injection attack is attempted or succeeds? Third: What's the blast radius if an LLM agent is compromised — what systems can an attacker reach? If your security team cannot answer these questions confidently, you're not ready to ship.
The compliance angle is emerging rapidly. OWASP designation means regulatory scrutiny is coming. HIPAA security assessments will start asking about LLM controls for healthcare applications. GDPR data protection impact assessments will need to address LLM data handling. SOC 2 audits will evaluate LLM security controls as part of systems security. Microsoft's CVE-2025–32711 demonstrated that even the largest technology companies ship vulnerable AI systems — compliance frameworks will hold everyone to the same standards.
Your compliance team needs LLM-specific frameworks now, before auditors ask for them. The OWASP Top 10 for LLM Applications provides that framework. It's not sufficient to say "We follow OWASP Top 10" — web application security standards don't address LLM-specific vulnerabilities. You need to demonstrate controls specifically for LLM01:2025.
The Bottom Line: We're in an Arms Race
The evidence is overwhelming. OWASP designates prompt injection as the #1 LLM vulnerability for 2025. 73% of LLM applications tested are vulnerable. Agent Security Bench documents 84.30% attack success rates across multiple LLM platforms. Multiple CVEs with CVSS 9+ scores have been disclosed in production systems. CrowdStrike and Google both report active exploitation by multiple threat actors. This isn't emerging risk — it's current reality affecting production systems today.
But effective defenses exist. PromptArmor achieves attack success rates below 1%. Spotlighting dramatically reduces vulnerability when properly implemented. Architectural changes like blast radius reduction and output validation contain damage when attacks succeed. The 95% vulnerability statistic isn't inevitable — it's a deployment problem. It reflects teams shipping LLM features without security controls, developers unaware of prompt injection attack vectors, and enterprises prioritizing speed over security.
Here's the uncomfortable truth that security researchers grapple with: LLMs are vulnerable by design. Natural language doesn't have syntactic boundaries between code and data the way programming languages do. We're essentially giving attackers a Turing-complete natural language interface to our applications. Every input is potentially executable. Every document the LLM reads might contain malicious instructions. Every API response could include injection attempts.
The question facing the industry: Can we build secure systems on fundamentally insecure foundations? Or do we need a paradigm shift in how we architect LLM applications — perhaps separating reasoning engines from tool execution, implementing capability-based security models, or developing new programming paradigms designed for LLM control flow?
The next CVE is coming. Someone will find a zero-day in another major LLM framework or platform. Another enterprise AI tool will turn out to have critical vulnerabilities. More threat actors will adopt prompt injection techniques. The 84% attack success rate won't improve automatically — it requires conscious engineering effort to implement defenses, test them, and maintain them as attacks evolve.
Discuss in comments: How is your team addressing prompt injection? Have you implemented PromptArmor, Spotlighting, or other defenses? What's your security testing process for LLM features? Will your application be in the 5% that's protected — or the 95% that isn't?
Sources:
- OWASP Top 10 for LLM Applications 2025
- Agent Security Bench (ASB) Research
- CrowdStrike 2025 Threat Hunting Report
- Google Threat Intelligence Group Report
- Field Effect CVE-2025–32711 Analysis
- Zeropath CVE-2025–46059 Analysis
- PromptArmor Research Paper
- GitHub tldrsec/prompt-injection-defenses