← All Articles
SecuritySafety

The OWASP GenAI Security Project: The Community That Is Setting the Standard for Secure AI

26 February 2026

Artificial intelligence is moving from experimentation into critical infrastructure. Boards are asking whether their organisations are safe. Regulators are starting to catch up. And yet, for most enterprises, the question of how to systematically evaluate and govern AI security risk still lacks a clear, authoritative answer.

The OWASP GenAI Security Project is the closest thing the industry has to one.

It is not a vendor product, a proprietary framework, or a compliance checkbox. It is an open, community-driven initiative that has become the most widely cited reference for LLM and generative AI security risk, used by security teams, architects, regulators, and AI builders across the world.1

The OWASP GenAI Security Project went from a lab initiative to one of OWASP's only Flagship-status projects in under 18 months, driven by over 600 contributing experts from more than 18 countries.

For executives making decisions about AI adoption, governance, and risk tolerance, and for architects designing systems that depend on LLMs and agents, understanding this project is not optional. It is foundational.


From Web Security to AI Security: A Brief History of OWASP

The Open Worldwide Application Security Project was founded on September 9, 2001, by Mark Curphey. The OWASP Foundation was formally incorporated as a United States non-profit on April 21, 2004.2

Its founding mission was direct: make application security visible so that organisations and individuals could make informed decisions about real security risks. In a landscape where most security knowledge lived inside vendors, consultancies, and closed research groups, OWASP's open approach was radical.

The early tools included the OWASP Guide, a practical security handbook, and WebGoat, a deliberately vulnerable application used for training. But the resource that made OWASP a household name in security was the OWASP Top 10: a regularly updated, community-curated list of the most critical application security risks.

Over two decades, OWASP grew into the pre-eminent open-source foundation for application security, with more than 250 local chapters worldwide and thousands of contributors producing standards, tools, guides, and cheatsheets consumed globally.2

In February 2023, the board voted to rename the organisation from the "Open Web Application Security Project" to the "Open Worldwide Application Security Project," reflecting that the scope had long since extended well beyond the web.2


The Birth of the GenAI Security Project

As large language models moved from research curiosity to enterprise deployment in 2023, it quickly became apparent that existing security frameworks did not adequately address the risks they introduced. Prompt injection, training data poisoning, model extraction, and excessive agent autonomy were not part of the standard threat model.

Steve Wilson, then of Contrast Security, launched the OWASP Top 10 for Large Language Model Applications in 2023 as a community response to this gap. What began as a small group of security professionals addressing an urgent need rapidly expanded. By 2024, the project had attracted more than 600 contributing experts from over 18 countries and nearly 8,000 active community members.3

This is critical for anyone building AI systems to understand: the OWASP GenAI Security Project is not a list of theoretical problems. It is a living record of how real AI systems fail, and what to do about it.

In March 2025, OWASP took the significant step of formally elevating the initiative: the Top 10 for LLM Applications became the OWASP GenAI Security Project, and it was promoted to Flagship status by vote of the Global Board of Directors.1 Flagship status is reserved for OWASP's most mature, active, and strategically important projects. The GenAI Security Project earned it in under 18 months.

The project now covers LLMs, agentic AI systems, red teaming, incident response, supply chain, data governance, and the full emerging landscape of AI-powered applications.


Key Resources: What the Project Produces

The OWASP GenAI Security Project is a prolific producer of open, vendor-neutral guidance. The table below summarises the primary resources available as of early 2026, with direct links.

ResourceDescriptionLink
OWASP Top 10 for LLM Applications 2025The core risk taxonomy for LLM applications, covering the ten most critical vulnerability classes with mitigationsView resource
OWASP Top 10 for Agentic Applications 2026The first globally peer-reviewed framework for autonomous and agentic AI security risks, developed with over 100 industry contributorsView resource
GenAI Red Teaming GuidePractical methodology for adversarial evaluation of LLM and GenAI systems, covering model evaluation, implementation testing, infrastructure, and runtime behaviourView resource
Vendor Evaluation Criteria for AI Red Teaming Providers & Tooling v1.0Criteria for assessing red teaming vendors and automated testing tools, from simple chatbots to advanced multi-agent and MCP architecturesView resource
Agentic AI Threats and MitigationsComprehensive threat taxonomy for agentic systems with mitigation guidance, covering planning, tool use, memory, and multi-agent coordinationView resource
Agentic AI Secure Development GuidePractical guidance for designing, building, and deploying secure agentic applicationsView resource
GenAI Incident Response GuideGuidelines and best practices for responding to security incidents involving GenAI applicationsView resource
State of Agentic AI Security and GovernanceA comprehensive landscape view of today's autonomous AI security and governance challengesView resource
Solutions Reference Guide (Q2–Q3 2025)Vendor-agnostic reference for organisations securing LLM and Agentic AI applicationsView resource
AI Security Solutions LandscapeA mapped view of the AI security product ecosystem aligned to the DevOps-SecOps intersectionView resource
OWASP AIBOM GeneratorOpen-source tool for generating AI Bills of Materials (AI SBOMs) to improve AI supply chain transparencyView resource
A Practical Guide for Secure MCP Server DevelopmentSecurity guidance for teams building on the Model Context ProtocolView resource
FinBot Agentic Security CTFA Capture the Flag challenge for hands-on learning about agentic AI securityView resource

All resources are published under Creative Commons Attribution-ShareAlike v4.0 and available at genai.owasp.org/resources.4


What Is Red Teaming — and Why Does It Matter for AI?

Red teaming is a concept with roots in military and national security practice. Historically, a "red team" was an independent group tasked with adopting the perspective of an adversary: to challenge assumptions, probe defences, and surface vulnerabilities that internal teams, anchored by familiarity bias, could not see.5

In cybersecurity, red teaming evolved into structured adversarial simulation: systematically attacking systems before malicious actors do, using the same techniques and mindset as real threat actors.

"GenAI Red Teaming combines traditional adversarial testing with AI-specific methodologies, addressing risks like prompt injection, toxic outputs, model extraction, bias, knowledge risks, and hallucinations." — OWASP GenAI Red Teaming Guide[^owasp-redteam]

For generative AI systems, red teaming takes on new dimensions. Traditional penetration testing looks for memory corruption, authentication bypass, and injection flaws in deterministic code. AI systems are probabilistic, context-sensitive, and instruction-following by design. Their attack surface includes inputs, system prompts, training data, retrieval pipelines, tool integrations, and the chain of automated decisions made by autonomous agents.

The OWASP GenAI Red Teaming Guide defines five threat categories that structured AI red teaming must address:

Adversarial Attacks

Attempts to manipulate model behaviour through crafted inputs. The most prevalent form is prompt injection, where malicious content embedded in user input, retrieved documents, or tool responses overrides the model's intended instructions.

In agentic systems, successful prompt injection can redirect an agent's entire task plan, cause it to leak sensitive information, invoke unintended tools, or take actions with real-world consequences such as sending communications, modifying data, or escalating permissions.

Red teaming in this category tests whether the system resists direct injection, indirect injection through data sources, multi-turn manipulation, and role-playing exploits.

Alignment Risks

Alignment failures occur when a model's outputs deviate from organisational values, policies, or safety requirements, even without an explicit attack. This includes generating harmful, biased, or ethically problematic content, inconsistently applying policy across different user populations or inputs, and following technically valid but unintended instructions that violate the spirit of the deployment.

Red teaming in this category probes edge cases in safety tuning, assesses consistency under adversarial framing, and evaluates whether the model's implicit values match the organisation's explicit requirements.

For executives, alignment failures represent reputational, legal, and compliance risk that may not surface through standard functional testing.

Data Risks

AI systems can expose sensitive information through several mechanisms: training data memorisation, retrieval pipeline leakage, system prompt exposure, and excessive verbosity in responses.

In enterprise deployments, data risks extend to customer PII, intellectual property, confidential business logic encoded in system prompts, and the contents of internal knowledge bases used in retrieval-augmented generation.

Red teaming in this category tests for extraction of training data, retrieval of out-of-scope documents, and leakage of system configuration through adversarial prompting strategies.

Interaction Risks

Interaction risks arise from the way users, models, and systems engage in real deployments: jailbreaking, social engineering of AI personas, multi-turn manipulation that gradually shifts model behaviour, and exploitation of role-playing or hypothetical framing to bypass guardrails.

In agentic contexts, interaction risks also include chain-of-thought manipulation, where an attacker poisons the reasoning steps an agent takes before arriving at a final action.

Red teaming in this category requires creative adversarial prompting, persona manipulation, and evaluation of whether guardrails hold under sustained pressure across multi-turn conversations.

Knowledge Risks

Knowledge risks cover the accuracy, currency, and reliability of the information an AI system produces or retrieves. This includes hallucination, misinformation generation, confident assertion of incorrect facts, and the potential for AI outputs to propagate errors at machine speed in automated workflows.

For agentic systems making decisions autonomously, knowledge risks can translate directly into operational failures: incorrect data written to databases, erroneous decisions communicated to customers, or flawed analysis driving executive decisions.

Red teaming in this category evaluates factual reliability, RAG retrieval quality under adversarial document injection, and the system's capacity to express uncertainty rather than confabulate.

Red teaming is not a one-time audit. It is a continuous practice that must evolve as models, architectures, and threat landscapes change. The OWASP framework provides the methodology to make it systematic.

The OWASP Red Teaming Guide is structured around four evaluation domains: model evaluation, implementation testing, infrastructure assessment, and runtime behaviour analysis. Together, these cover the full stack from the model itself to the operational environment in which it runs.6


The Active Initiatives: Where the Work Happens

Beyond published documents, the OWASP GenAI Security Project runs several active initiatives that produce ongoing research, tooling, and community engagement.

Agentic Security Initiative (ASI)

The Agentic Security Initiative is the project's most active research programme, addressing the security implications of autonomous AI systems: agents that plan, use tools, access memory, delegate tasks, and operate without continuous human supervision.

The ASI has produced the Agentic AI Threats and Mitigations taxonomy, the Secure Agentic Development Guide, the State of Agentic AI Security and Governance report, the OWASP Top 10 for Agentic Applications, and the FinBot CTF challenge. It is directly relevant to any organisation building or deploying AI agents in production.

Red Teaming Initiative

This initiative develops and maintains standardised red teaming methodology for AI systems. It produces benchmarks, testing frameworks, and evaluation criteria, and it published both the GenAI Red Teaming Guide and the Vendor Evaluation Criteria for AI Red Teaming Providers.

The initiative is specifically designed to help organisations avoid the trap of superficial "jailbreak-only" testing and establish a rigorous, full-spectrum adversarial evaluation practice.

AI Supply Chain and AIBOM

The OWASP AIBOM Generator addresses a growing concern in enterprise AI governance: the opacity of AI system dependencies. Just as Software Bills of Materials (SBOMs) became a compliance and security standard for software supply chains, AI Bills of Materials (AIBOMs) aim to provide transparency into the models, datasets, fine-tuning pipelines, and plugins that constitute an AI system.

This initiative is particularly relevant for regulated industries where provenance, auditability, and third-party risk management are compliance requirements.


Why This Matters: An Executive and Architecture Perspective

The OWASP GenAI Security Project is not a technical curiosity. It is a governance instrument.

For executives, it provides:

  • A shared risk language that can be used consistently across engineering, security, legal, and risk management
  • A baseline against which to assess vendor claims, system designs, and third-party AI integrations
  • Alignment with emerging regulation, as regulators and standards bodies increasingly reference OWASP frameworks in AI governance guidance
  • A due diligence record — having demonstrably mapped your AI systems against an open, peer-reviewed framework is increasingly expected in board reporting and risk assessment

For architects, it provides:

  • Threat models that account for the full attack surface of LLM and agentic systems, including surfaces that traditional threat modelling misses
  • Mitigation patterns for prompt injection, excessive agency, supply chain risk, and retrieval pipeline weaknesses
  • Red teaming methodology to validate that security controls work under adversarial conditions
  • Agentic-specific guidance for systems with autonomous planning, tool use, and multi-agent coordination
The organisations that are ahead on AI security are not necessarily those with the largest security budgets. They are those that started with the right frameworks early and applied them consistently across architecture, development, and operations.

Where to Start

The most effective starting point depends on where your organisation sits today.

Your current stageRecommended entry point
Early AI adoption, limited security reviewOWASP Top 10 for LLM Applications 2025
Deploying or planning AI agentsOWASP Top 10 for Agentic Applications 2026 + Agentic AI Threats and Mitigations
Evaluating AI security postureGenAI Red Teaming Guide + Vendor Evaluation Criteria
Building an AI governance programmeSolutions Reference Guide + State of Agentic AI Security and Governance
Developing on MCP or multi-agent frameworksSecure MCP Development Guide + Agentic Security Development Guide

The full resource library is at genai.owasp.org/resources. All materials are free, open, and vendor-neutral.


Footnotes

  1. OWASP Foundation. "OWASP Promotes GenAI Security Project to Flagship Status." March 26, 2025. 2

  2. OWASP Foundation. "About the OWASP Foundation." See also: Wikipedia. "OWASP." 2 3

  3. OWASP Foundation. "OWASP Top 10 for Large Language Model Applications."

  4. OWASP GenAI Security Project. "Resources Archive."

  5. RAND Corporation. "Red Teaming: The Art of Challenging Assumptions."

  6. OWASP GenAI Security Project. "GenAI Red Teaming Guide." January 2025.

About the Author

Peter Wood

Peter Wood

Healthcare technology leader specialising in data platforms, operational intelligence, and agent-driven automation. Peter has led large-scale digital transformation programmes with major hospital groups and global technology partners, translating advanced analytics and AI into measurable improvements in clinical operations, capacity, and patient flow.