Security researchers have demonstrated a significant vulnerability where autonomous AI agents can be manipulated into leaking sensitive company data. The successful exfiltration was achieved not by hacking the AI model itself, but by using its intended functions, such as web browsing, against it through a technique known as indirect prompt injection.
The demonstration showed that an agent with access to both confidential internal documents and the internet can be turned into a source of data leakage. This occurs when the AI processes external information, such as an email, that contains a hidden malicious instruction.
How the Data Leak Was Executed
In the documented research, an AI agent was provided access to a private file containing a confidential piece of information, such as a unique secret key. The agent’s task involved monitoring and summarizing incoming data. The researchers then sent a carefully crafted message to the agent that appeared benign on the surface but contained a hidden command embedded within the text.
This hidden command instructed the AI agent to use its web browsing tool. The instruction specifically directed the agent to perform a web search and to include the contents of the private file as part of the search query. The agent, following its operational protocols, complied with the instruction.
From Private Document to Public Server Log
By executing the malicious command, the AI agent created a URL for a web search that contained the secret key. When the agent visited this URL, the confidential data was transmitted directly to an external web server controlled by the researchers. The sensitive information was then recorded in the server’s publicly accessible access logs.
This research provides a concrete example of how AI agents can be compelled to exfiltrate data they are entrusted with. The event highlights a tangible security risk where the agent’s interaction with unchecked external data sources can lead to the public exposure of private information.
Source: https://www.helpnetsecurity.com/2025/10/29/agentic-ai-security-indirect-prompt-injection/