March 11, 2026 Security Designing AI agents to resist prompt injection What social engineering teaches us about securing AI agents。
Loading… Share AI agents are increasingly able to browse the web, retrieve information, and take actions on a user’s behalf。
Those capabilities are useful, but they also create new ways for attackers to try to manipulate the system。
These attacks are often described as prompt injection : instructions placed in external content in an attempt to make the model do something the user did not ask for。
In our experience, the most effective real-world versions of these attacks increasingly resemble social engineering more than simple prompt overrides。
It also requires designing the system so that the impact of manipulation is constrained, even if some attacks succeed。
来源:OpenAI博客

