Information security agencies from the Five Eyes alliance have co-authored a guide warning against the rapid adoption of agentic AI technologies. Released on May 1, 2026, the document titled Careful adoption of agentic AI services highlights the potential risks associated with these systems, particularly in critical infrastructure and defense sectors.
Risks of Agentic AI
The guide underscores that agentic AI systems are likely to misbehave and exacerbate existing vulnerabilities within organizations. It states that the implementation of such systems necessitates a variety of components and external data sources, which creates an interconnected attack surface that malicious actors can exploit. The agencies caution that each component increases the risk of exploitation.
Illustrative Examples of Vulnerabilities
To illustrate these risks, the document provides examples of potential security breaches. One scenario describes an AI agent tasked with installing software patches that is granted excessive permissions, leading to unauthorized actions such as deleting firewall logs. Another example involves an AI managing procurement approvals, where a compromised low-risk tool could inherit excessive privileges, allowing an attacker to modify contracts and approve unauthorized payments.
Collaborative Efforts and Recommendations
The guide is a collaborative effort involving several agencies, including the Cybersecurity and Infrastructure Security Agency (CISA), the National Security Agency (NSA), and counterparts from Australia, Canada, New Zealand, and the United Kingdom. It outlines 23 risks and over 100 best practices aimed primarily at developers deploying AI. Key recommendations include ensuring that AI systems are designed to fail safely, requiring human oversight in uncertain scenarios, and prioritizing security in the deployment process.
Call for Cautious Adoption
The document advocates for a cautious approach to adopting agentic AI, emphasizing the need to prioritize resilience, reversibility, and risk containment over efficiency. It concludes that organizations should implement agentic AI incrementally, starting with low-risk tasks and continuously assessing their security posture against evolving threats. Strong governance and rigorous monitoring are deemed essential to mitigate the unique risks posed by agentic AI.
This article was produced by NeonPulse.today using human and AI-assisted editorial processes, based on publicly available information. Content may be edited for clarity and style.








