AI agent security and blast radius management
9/10 CriticalProduction incidents show AI agents leaking internal data, shipping ransomware through plugins, and executing destructive actions (deleting repos). Security shifted from prompt injection to actual agent capabilities and operational risk.
Sources
- AI | 2025 Stack Overflow Developer Survey
- State of the API 2025: API Strategy Is Becoming AI Strategy
- 5 Major Pain Points AI Agent Developers Can't Stop Ranting About ...
- 2025 Recap: The Year Software Development Changed Shape
- 2. Controlled Agency And...
- A practical business guide to the OpenAI API in 2025
Collection History
AI agents remain highly vulnerable to prompt injection and jailbreak attacks, with success rates exceeding 90%. Security researchers discovered the first zero-click attack on AI agents through Microsoft 365 Copilot, where 'attackers hijack the AI assistant just by sending an email... The AI reads the email, follows hidden instructions, steals data, then covers its tracks'. Microsoft took five months to fix this issue.
managing permissions and making sure the bot only pulls information from the right, up-to-date sources is a huge security and maintenance headache. The last thing you want is your IT bot accidentally sharing sensitive HR info.
Agents leaking internal data within minutes, Malicious plugins shipping ransomware, Supply-chain bugs in AI tooling, Agents deleting repos or months of work