Back to Topic Hubs
AI SecurityTopic HubMay 1, 2026Yellow — detail controls

Agentic AI Security: Prompt Injection, Tool Hijacking, and Voice Agents

Quick Answer

This topic gathers richards.ai work on what changes when LLMs gain tools, memory, and peer agents. It collects the threat models, defenses, glossary terms, checklists, executive briefs, and source papers behind multi-agent prompt injection, tool hijacking, memory poisoning, cross-agent infection, and voice-agent jailbreaking. The page offers reading paths for practitioners, engineering leads, researchers, and newcomers, anchored to the current canonical defense reference.

Agentic AI Security: Prompt Injection, Tool Hijacking, and Voice Agents

This topic gathers the richards.ai work on what changes once an LLM gains tools, memory, and peer agents. The unifying observation across the cluster is that prompt injection in agentic systems is a confused-deputy and authority-propagation problem, not a string-filtering problem. Hardening Multi-Agent Systems Against Prompt Injection is now the canonical defense reference for the cluster and supersedes the earlier exploitation paper, which is retained here for historical context on how the threat model emerged. Some reproduction detail is withheld in the linked artifacts pending vendor coordination.

What this topic covers

The cluster spans threat models, defenses, glossary terms, an operational checklist, an executive brief, source papers, an interactive jailbreak demo, and a runnable capability-control reference implementation. In scope: multi-agent prompt injection, cross-agent infection, memory poisoning, tool hijacking, retrieval-channel exfiltration, and the voice-modality variant of the same authority-confusion failure. Out of scope: single-agent jailbreaking that does not involve tools, memory, or inter-agent channels, and vendor-specific orchestration guidance.

How to read this page

Newcomers should start with What Is Multi-Agent Prompt Injection? and the glossary entries before opening the papers. Practitioners reviewing or building a system should jump to the multi-agent prompt injection defense checklist, which is re-anchored to the hardening paper. Engineering leads and stakeholders who will not work through the full paper should read the executive briefing. Researchers should go to the papers group; the hardening paper carries the current architecture, and the earlier exploitation paper documents the attack surface it answers.

Where this topic sits

This cluster sits inside the security pillar alongside other agentic and LLM-application topics; the topics index lists adjacent shelves, and the papers index holds the broader source-paper set this cluster draws from.

Papers

2 members

Learn

4 members

Glossary

4 members

Checklists

1 member

Briefs

1 member

Tools

1 member