ai · 8 min read · Apr 28, 2026

Agentic AI Security Requires Layered Defense, Not Just Prompt Guards

A new framework maps AI agent vulnerabilities across seven architectural layers and four time horizons, revealing that 93% of research ignores the slowest, most dangerous threats.

Source: arxiv/cs.LG · Kexin Chu · open original ↗

Agentic AI systems need security models that account for persistent memory, tool use, and multi-agent coordination across extended time horizons.

  • Layered Attack Surface Model (LASM) maps seven distinct architectural components vulnerable to different threat classes.
  • Attack temporality spans four classes: instantaneous, session-persistent, cross-session cumulative, and sub-session non-bounded.
  • Most dangerous threats cluster at high layers (governance, multi-agent, ecosystem) with slow-burn temporality.
  • Only 7% of 120 reviewed threat-defense pairs address the high-layer, slow-burn zone.
  • Covert agent collusion, long-term memory poisoning, and supply-chain compromise represent emerging high-risk vectors.
  • Existing defenses focus on low-layer, fast attacks (prompt injection, jailbreaking) leaving systemic gaps.
  • Agentic security requires distributed systems thinking, not stateless LLM security models.

Astrobobo tool mapping

  • Knowledge Capture Document your agent's memory persistence model, tool invocation log, and inter-agent communication patterns. Record which of these are currently unmonitored.
  • Focus Brief Create a one-page threat matrix: rows = LASM layers, columns = temporality classes (T1–T4). Mark which cells your current defenses cover and which are gaps.
  • Daily Log Set a weekly audit task: review agent memory mutations, tool execution logs, and peer-agent messages for anomalies (e.g., unusual data access patterns, new tool calls).

Frequently asked

  • Agentic systems maintain persistent memory, invoke external tools, coordinate with other agents, and operate over extended time horizons. These capabilities introduce new threat vectors—memory poisoning, supply-chain compromise, and covert collusion—that stateless LLM security models do not address. Traditional defenses like prompt injection filters are insufficient.
Share X LinkedIn
cite
APA
Kexin Chu. (2026, April 28). Agentic AI Security Requires Layered Defense, Not Just Prompt Guards. Astrobobo Content Engine (rewrite of arxiv/cs.LG). https://astrobobo-content-engine.vercel.app/article/agentic-ai-security-requires-layered-defense-not-just-prompt-guards-b093fe
MLA
Kexin Chu. "Agentic AI Security Requires Layered Defense, Not Just Prompt Guards." Astrobobo Content Engine, 28 Apr 2026, https://astrobobo-content-engine.vercel.app/article/agentic-ai-security-requires-layered-defense-not-just-prompt-guards-b093fe. Based on "arxiv/cs.LG", https://arxiv.org/abs/2604.23338.
BibTeX
@misc{astrobobo_agentic-ai-security-requires-layered-defense-not-just-prompt-guards-b093fe_2026,
  author       = {Kexin Chu},
  title        = {Agentic AI Security Requires Layered Defense, Not Just Prompt Guards},
  year         = {2026},
  url          = {https://astrobobo-content-engine.vercel.app/article/agentic-ai-security-requires-layered-defense-not-just-prompt-guards-b093fe},
  note         = {Astrobobo rewrite of arxiv/cs.LG, https://arxiv.org/abs/2604.23338},
}

Related insights