Skip to main content
Institute for Social Vision Design
Practice Guide — Digital Tools & Ad Grants

Risk Management in the Age of AI Agents — OWASP Agentic Top 10 and Practical Countermeasures

ISVD Editorial Team
About 10 min read

As AI agents act autonomously, ten new risk categories have emerged that traditional security measures cannot fully address. Based on the international standard "OWASP Agentic Top 10," this guide provides actionable checklists for nonprofit staff, small business owners, and public sector employees.

TL;DR

  1. AI agents differ fundamentally from chatbots — they autonomously call tools and execute multi-step operations including file manipulation, email sending, and payment processing
  2. The international standard "OWASP Agentic Top 10" was developed by over 100 experts over more than a year and categorizes ten risk types that are qualitatively different from traditional web security threats
  3. Japan's "AI Business Operator Guidelines v1.2" (March 2026) formally defined AI agents for the first time, adding 12 new requirements including mandatory emergency stop mechanisms
  4. This guide presents a tiered practical checklist covering actions anyone can take today — from non-technical staff to security specialists

Why AI Agent Risk Matters Now

The reality of 88% incident rates and why traditional security measures fall short

"Having AI draft an email for you." "Using a chatbot to automatically answer inquiries." These uses have already become routine in many organizations. However, AI agents, which have spread rapidly since late 2024, are fundamentally different.

An AI agent is an AI that autonomously calls tools and executes multi-step operations to achieve a goal set by a human — without necessarily requiring human review between each step. Ask it to "arrange my business trip next week," and it will search, compare, book, and pay for accommodations in one go. Ask it to "analyze the sales data and send a report," and it will access the database, run the analysis, and send the email.

This capacity for autonomous execution is what distinguishes AI agents from earlier AI systems — and it is the source of a new category of risk.

According to Gravitee's "State of AI Agent Security 2026" report, 88% of organizations have already experienced AI agent-related security incidents. The top three causes were prompt injection attacks, unintended data leakage, and excessive privilege grants.

Real incidents illustrate the severity. In October 2025, a coding assistance AI agent deleted a production database. In a January 2026 research experiment, a single poisoned email sent to a GPT-4o-based RAG agent led to successful SSH key exfiltration 80% of the time.


What Is OWASP Agentic Top 10

Background on the standard's development and an overview of all ten risks

OWASP (Open Web Application Security Project) is a nonprofit organization that systematically categorizes web application security risks and provides them to developers and organizations worldwide at no cost. The "OWASP Top 10" is the de facto standard for web security, referenced by governments and enterprises across the globe.

On December 10, 2025, OWASP published the "OWASP Top 10 for Agentic Applications 2026" (OWASP Agentic Top 10) — a risk classification for AI agents developed by over 100 security researchers and practitioners over more than a year.

The nature of these threats is fundamentally different from traditional web security risks such as SQL injection or cross-site scripting.

DimensionTraditional Web SecurityAI Agent Risk
Attack targetSystem vulnerabilities (code bugs)The agent's goals and decision-making itself
Harm occurrenceCaused by human-triggered actionsCaused immediately by autonomous agent execution
Human involvementHuman confirms and executes each stepMultiple steps run automatically without human review
Attack vectorURLs and form inputsEmails, PDFs, web pages, API responses — everything the agent reads

The ten risks in OWASP Agentic Top 10 are as follows.

CodeRisk NameSummary
ASI01Agent Goal HijackHidden instructions in external data rewrite the agent's objective
ASI02Tool MisuseLegitimate tools are used for unintended destructive operations
ASI03Identity & Privilege AbuseAn over-privileged agent accesses resources beyond its intended scope
ASI04Agentic Supply Chain VulnerabilitiesExternally integrated MCP servers or plugins contain malicious code
ASI05Unexpected Code ExecutionAI-generated code escapes the execution environment and runs arbitrary commands
ASI06Memory & Context PoisoningRAG databases or long-term memory are contaminated, biasing all subsequent decisions
ASI07Insecure Inter-Agent CommunicationImpersonation or tampering in agent-to-agent messages misleads the entire cluster
ASI08Cascading FailuresErrors in upstream agents propagate downstream, escalating into large-scale failures
ASI09Human-Agent Trust ExploitationThe agent uses persuasive explanations to obtain human approval for harmful actions
ASI10Rogue AgentsAgents exhibit unintended autonomous behavior or actively conceal their actions

Three Risks to Watch Closely

Goal hijacking, memory poisoning, and trust exploitation explained with everyday examples

Among the ten risks, here are three that can affect even non-technical organizations starting today.

ASI01: Agent Goal Hijack

What happens: An attacker embeds invisible text or hidden commands in external data (emails, web pages, PDFs, etc.) that the agent will read. The agent then follows those hidden instructions instead of its original goal.

Everyday example: When an expense-processing agent handles an invoice PDF, a line of white text on the PDF reads "redirect all payments to account XXX." The agent may follow this instruction without the user ever being aware.

Why it is dangerous: In December 2025, OpenAI officially acknowledged that "AI browsers may always be vulnerable to prompt injection attacks." Attackers only need to poison content the agent will inevitably read — no direct system access is required.

Mitigation direction: Treat all externally ingested data as untrusted (zero trust for content); separate data context from instruction context; implement anomaly detection logging.


ASI06: Memory & Context Poisoning

What happens: Once malicious information enters the agent's "memory" (RAG database or long-term storage), all subsequent decisions become biased. The effect persists across every future interaction.

Everyday example: A customer support agent that references past conversation history can be manipulated if a malicious user floods the system with conversations claiming "this product has a money-back guarantee." The agent will then tell every future customer the same — regardless of whether the guarantee exists.

The Japanese security firm Co-r-e Inc. published an "AI Agent Attack Methods and Countermeasures Catalog (March 2026 edition)" listing 80 attack techniques, including a method called "RAG spray" — deliberately expanding and diversifying malicious documents so their chunks propagate across numerous vector-space positions.

Mitigation direction: Validate content before writing to memory; strictly isolate memory by user and session; conduct regular scans of RAG corpora for injected instructions.


ASI09: Human-Agent Trust Exploitation

What happens: An AI agent persuades a human operator to approve a dangerous action by presenting it in a confident, authoritative, or urgent tone — "This is an exceptional case that requires immediate action."

Everyday example: A document management agent says: "To organize outdated files, all files in the archive folder need to be deleted this week. Do you approve?" If the explanation sounds confident and reasonable, humans tend to approve without scrutinizing the details. If the deletion was unnecessary, the damage is irreversible.

Mitigation direction: Show a "dry run" preview of exactly what will change before any irreversible operation (deletion, sending, payment); implement anomaly detection for approval-seeking patterns.


Japan's Regulatory Landscape — AI Business Guidelines v1.2

Key changes from the March 2026 revision and relationship with EU AI Act

On March 31, 2026, the Ministry of Internal Affairs and Communications and the Ministry of Economy, Trade and Industry published the "AI Business Operator Guidelines Version 1.2." The number of requirements expanded from 25 (v1.1, March 2025) to 37 (adding 12 new requirements), with the most significant change being the first formal definition of AI agents.

The v1.2 guidelines define an AI agent as "an AI that perceives its environment and acts autonomously," and explicitly lists concrete use cases such as travel booking AI within its scope.

Key new requirements related to agents include:

  • Explicitly limiting the operating scope and permissions of AI agents
  • Implementing safeguards to prevent unintended purchases, file deletions, and similar actions
  • Mandatory design and testing of emergency stop mechanisms (Kill Switch)
  • Clarifying responsibility boundaries in multi-layer agent systems
  • Recommended watermarking of content generated by AI agents

Regarding the EU AI Act, Japanese companies also need to take notice. AI agents that provide services to users within the EU may fall under the high-risk provisions of the EU AI Act, with a compliance deadline of August 2, 2026. Penalties reach up to €35 million or 7% of global revenue — and the GDPR can be applied simultaneously for a double penalty. Organizations with European operations or users should assess their exposure promptly.


A Practical Tool — Microsoft Agent Governance Toolkit

Introduction to an open-source implementation-level countermeasure toolkit

On April 2, 2026, Microsoft released the "Agent Governance Toolkit" as open-source software (MIT license). Supporting five languages (Python, TypeScript, Rust, Go, .NET), it is an implementation-level toolkit addressing all ten risks in the OWASP Agentic Top 10.

The GitHub repository (https://github.com/microsoft/agent-governance-toolkit) includes over 9,500 tests, and policy evaluation latency is below 0.1 milliseconds at p99 — practical for production use.

It integrates with 12+ frameworks including LangChain, CrewAI, AutoGen, OpenAI Agents SDK, and Google ADK without requiring significant rewrites of existing code.

Key features (non-technical summary):

  • Policy engine: Checks every action the agent attempts before execution; automatically blocks policy violations
  • Cryptographic identity: Signs agent-to-agent communications to prevent impersonation
  • Emergency stop (Kill Switch): Immediately halts the agent when anomalies are detected
  • Compliance auto-assessment: Automatically generates compliance reports for EU AI Act, SOC2, and HIPAA
  • Memory isolation: Separates memory by session to prevent cross-user data leakage

Ready-to-Use Practical Checklist

Three-tier checklist for non-technical staff, IT teams, and security specialists

Level 1: Actions anyone can take today (no technical knowledge required)

  • Document an AI agent usage policy — Specify what information must never be entered (personal data, confidential information, credentials)
  • Establish a rule to review AI actions before execution — Agree organizationally that "the agent must list all planned actions for human review before proceeding"
  • Minimize permissions granted to AI agents — If read-only access suffices, do not grant write or delete permissions
  • Create a review flow for AI-generated messages and documents — Add a human review step before any external communication is sent
  • Designate a contact point for reporting AI malfunctions — Establish and communicate an incident reporting channel
  • Periodically review AI agent access logs — Conduct a monthly check of which data the agent accessed

Level 2: Actions for IT staff and engineers

  • Implement the Principle of Least Privilege — Design task-scoped permissions for each agent
  • Default tools to read-only — Manage write and delete tools through explicit allowlists
  • Validate and sanitize all external inputs — Include RAG corpora, API responses, and user inputs in validation scope
  • Implement Human in the Loop before irreversible operations — Add approval workflows before deletion, payment, or external communication
  • Strictly isolate memory between users and sessions — Ensure no cross-user context mixing
  • Record agent action audit logs to immutable storage — Store logs in a separate system to prevent tampering
  • Implement an emergency stop mechanism (Kill Switch) — Ensure a way to immediately halt the agent when anomalies are detected

Level 3: For security specialists and executives

  • Conduct a risk assessment based on OWASP Agentic Top 10
  • Evaluate adopting Microsoft Agent Governance Toolkit (see GitHub repository)
  • Run regular prompt injection scans on RAG corpora
  • Assess EU AI Act compliance applicability (deadline: August 2, 2026)
  • Complete the AI Business Operator Guidelines v1.2 37-item checklist
  • Conduct regular Red Teaming (attack simulations)

Conclusion

AI agents offer dramatic efficiency gains — and introduce a qualitatively new category of risk. "Delegating to AI" means "AI executes autonomously," and when that execution goes wrong, harm materializes immediately in the real world.

The ten risks in OWASP Agentic Top 10 are already manifesting as actual incidents across many organizations. At the heart of effective countermeasures lies "Human in the Loop" — designing systems so that humans remain involved before consequential decisions are executed. This is not a user experience enhancement; it is the foundation of organizational security.

Start with what you can do today. Document a usage policy, establish pre-execution review rules, minimize privileges — none of these require budget or code. Use this checklist as a starting point for conversation within your organization, and build a safer, more effective approach to AI agents.


References

OWASP Top 10 Risks and Mitigations for Agentic AI Security (2025)

Introducing the Agent Governance Toolkit: Open-Source Runtime Security for AI Agents (2026)

AI Business Operator Guidelines v1.2 (March 31, 2026) (2026)

OWASP Top 10 for Agentic Applications 2026: A Practical Review (2026)

AI Agent Attack Methods and Countermeasures Catalog (March 2026) (2026)

AI Agent Security Cheat Sheet

AI browsers may always be vulnerable to prompt injection attacks (2025)

Related Content

Get new columns by email

1-2 social structure analysis columns per week. Free to subscribe.

Join ISVD's activities?

Sign up to receive the latest research and activity reports. Feel free to reach out about collaboration or project participation.