Autonomous AI agents are evolving rapidly, moving beyond simple chatbots to systems that can reason, utilize tools, and act with minimal human oversight. This evolution, however, dramatically amplifies the threat of prompt injection. Databricks has outlined a practical approach to securing these powerful agents, drawing on Meta's "Agents Rule of Two" framework. This methodology, also echoed in research like Simon Willison's "Lethal Trifecta" and discussed in relation to AI Agent Observability: New Rules, identifies the critical conditions that make AI agents vulnerable.
The core principle is that an AI agent becomes susceptible to prompt injection when it simultaneously possesses three key characteristics: access to sensitive systems or private data, exposure to untrustworthy inputs, and the ability to change state or communicate externally. To mitigate these risks, Databricks advocates for ensuring an agent has no more than two of these attributes, effectively breaking the attack chain.