← All Articles
SecurityAgentic AI

OWASP Top 10 for LLMs (2025): Executive Summary and Why It Is Essential Reading for Agent Development

26 February 2026

If you are building AI agents in production, the OWASP Top 10 for LLMs and GenAI is not optional reading. It is one of the clearest, community-driven security baselines for how modern LLM and agentic systems fail, and how teams should reduce those failures in real deployments.12

OWASP now positions this work under the broader GenAI Security Project, explicitly covering LLMs, agentic AI systems, and AI-driven applications. That scope makes it directly relevant for agent architects, platform teams, security leaders, and engineering executives making delivery and governance decisions.13

Executive Summary

The 2025 OWASP Top 10 identifies ten recurring risk classes that show up across the LLM and GenAI application lifecycle: development, deployment, and operations.2

The main executive takeaway is simple: most critical failures in agent systems are not model-IQ problems. They are systems-security and control-plane problems.

When translated to agent development, the OWASP list says that teams must secure:

  • inputs and instructions (prompt injection, system prompt leakage)
  • data and model lineage (supply chain, data/model poisoning)
  • tool execution boundaries (improper output handling, excessive agency)
  • retrieval and memory surfaces (vector and embedding weaknesses)
  • decision quality and operational resilience (misinformation, unbounded consumption)
  • confidentiality controls (sensitive information disclosure)

In short, safe agent development is less about one strong prompt and more about layered controls around planning, tool use, memory, identity, validation, and monitoring.

This is critical for anyone building AI agents to understand before shipping to production.

OWASP 2025 Top 10 Through an Agent Development Lens

OWASP riskWhat it means in practiceWhy agent teams should care
LLM01: Prompt InjectionMalicious or crafted input manipulates model behaviorAgents can be pushed into unsafe tool calls, policy bypass, or hidden instruction following
LLM02: Sensitive Information DisclosureConfidential data leaks through prompts, context, or outputsAgents often touch internal knowledge, APIs, and user data, increasing leak impact
LLM03: Supply ChainCompromised models, datasets, plugins, or dependenciesAgent stacks depend on many third-party components and connectors
LLM04: Data and Model PoisoningTraining/fine-tuning/embedding data is tampered withPoisoned knowledge can silently alter agent behavior and trustworthiness
LLM05: Improper Output HandlingLLM output is consumed without validation/sanitizationUnsafe outputs can become executable actions in downstream systems
LLM06: Excessive AgencyAgents get too much autonomy or privilegeOver-scoped permissions turn normal mistakes into high-impact incidents
LLM07: System Prompt LeakageHidden instructions are exposed and exploitedRevealed policy logic makes bypass attempts easier and more targeted
LLM08: Vector and Embedding WeaknessesWeaknesses in retrieval/indexing pipelinesRAG-enabled agents can be manipulated through poisoned or adversarial retrieval content
LLM09: MisinformationConfident but incorrect outputs influence decisionsAgentic automation can propagate bad decisions at machine speed
LLM10: Unbounded ConsumptionUncontrolled usage of tokens/compute/toolingAgent loops and broad tool access can create cost spikes and service instability

Risk names and descriptions above are based on the OWASP 2025 Top 10 content and linked risk pages.24

Why This Is Essential Reading for Agent Development

  1. It gives a shared language across engineering, security, and leadership for AI risk prioritization.
  2. It reframes agent failures as architecture and governance issues, not only model behavior issues.
  3. It aligns security planning to the full lifecycle (build, deploy, operate), which is where most agent incidents emerge.
  4. It is directly connected to OWASP's broader agentic security work, making it practical for teams moving from LLM apps to autonomous agents.15

What to Do Next (Executive Action)

  • Adopt the OWASP Top 10 as a baseline control framework for all new agent initiatives.
  • Map each agent workflow to the Top 10 and identify where controls are currently missing.
  • Prioritize guardrails around tool invocation, privilege boundaries, output validation, and retrieval integrity first.
  • Add continuous monitoring for prompt attacks, data leakage, cost anomalies, and unsafe autonomous actions.
  • Re-review architecture decisions whenever agent scope or autonomy increases.

Agent development is now a security-critical discipline. Teams that treat this OWASP guidance as foundational design input, rather than post-launch compliance, will ship faster with fewer high-severity failures.

Footnotes

  1. OWASP Foundation. "OWASP Top 10 for Large Language Model Applications." 2 3

  2. OWASP GenAI Security Project. "2025 Top 10 Risk & Mitigations for LLMs and Gen AI Apps." 2 3

  3. OWASP GenAI Security Project. "OWASP Top 10 for LLM is now the GenAI Security Project and promoted to OWASP Flagship status" (March 26, 2025).

  4. OWASP GenAI Security Project. "LLM01:2025 Prompt Injection."

  5. OWASP GenAI Security Project. "OWASP Agentic Security Initiative (ASI)."

About the Author

Peter Wood

Peter Wood

Healthcare technology leader specialising in data platforms, operational intelligence, and agent-driven automation. Peter has led large-scale digital transformation programmes with major hospital groups and global technology partners, translating advanced analytics and AI into measurable improvements in clinical operations, capacity, and patient flow.