AI agents are the next logical step in the automation revolution. But it should come as no surprise that handing more autonomy to a still-emerging class of software comes with risks.
One of the biggest is Shadow AI, where employees or even whole departments deploy AI agents without proper cyber oversight. If agents are set loose and permitted to operate without proper guardrails, they might leak data, access systems without permission, or breach compliance. That could expose the firm to breaches, penalties, or worse.
Traditional cyber tools simply aren’t designed for the perils of intelligent hyper-automation. We asked the experts what it will take to keep AI’s rapidly growing army of independent programs on a shorter leash.
Key Takeaways
- Security experts are sounding the alarm on the potential risks of agentic AI.
- The power of AI agents rests in their autonomy, and that freedom to act could also mean freedom to harm.
- As firms rush to embrace AI and encourage employees to build it into their daily workflows, there’s a concern that an unauthorized shadow AI infrastructure is quietly building up.
- To reassert control, firms need to have dedicated monitoring and authentication tools in place that can keep AI agents from tripping up security protections.
AI Gone Wild
AI agents are advanced machine learning applications that can independently make decisions and act on them.
Like all software advances, they promise to improve operational efficiency and boost productivity, but agents take it further.
Rather than take inputs and process them according to algorithmic instructions, they can act on their own, making decisions, working with corporate IT systems and sharing data dynamically, learning and adapting as they do.
The potential benefits are easy to see:
- You might create an agent to scour the web for the best travel deals for a given set of destinations, dates, and price ranges, then allow it to book your next holiday when all the parameters are satisfied.
- In corporate settings, AI agents can improve customer service. They handle more customer queries simultaneously and resolve them faster, without compromising quality. Agents are available 24/7 and can scale on demand as call volumes rise. Customer satisfaction improves as agents capture data and mine it for insights, turning every inbound query into grist for the analytics mill.
But the potential worries are just as visible. In a world where cyber attacks are already growing in frequency and sophistication every year, giving a new form of software new levels of autonomy creates a whole new category of risk.
One of the biggest is shadow AI, where employees or entire departments start using an agentic or generative AI solution without permission.
From writing emails with ChatGPT on the sly to using a low or no-code solution to build an agent-powered social media management tool, using AI in a corporate environment without proper oversight could be a cybersecurity nightmare.
Because agents have permission to act autonomously, they can sidestep some of the guardrails traditional security frameworks depend on. That could mean exposing sensitive information, executing operations that might be defined as high risk, or creating security vulnerabilities for the systems they interact with.
Secret Agents
The main concern with agents is visibility. AI’s surging popularity makes it hard to know exactly where and how they’re being used within the enterprise. Machine learning models seem to be embedded in every office productivity suite, collaboration platform, and development tool going, making them difficult to monitor.
It’s an echo of the cyber fears that arose when bring your own device (BYOD) first reared its head. Security teams suddenly had thousands of unauthorized endpoints inside the network perimeter and little to no control over how they were used or which systems they could connect with.
The Shadow AI Surge: Study Finds 50% of Workers Use Unapproved #AI Tools – https://t.co/PKdMt4Ho3Z #shadowAI #genaI #shadowIT #dataprivacy #security #cybersecurity pic.twitter.com/27VkIOgfh8
— Vallum Software (@vallumsoftware) May 7, 2025
Today, the push to embrace AI means employees feel empowered to switch on their own AI-powered assistants, coding tools, or analytics dashboards without telling IT. AI only works if it has loads of data to draw from, and the data it’s being fed belongs to the company.
But what data from which databases, authorized by whom, and for what purposes?
Data governance and security rules demand answers to those questions in advance.
Without them, the risk of unauthorized access to sensitive information or valuable intellectual property, thanks to unmonitored, agent-driven decisions, becomes very real.
If a trader in a financial services firm creates an agent to automate high-velocity trades, will they end up feeding confidential market data into a third-party machine learning model for processing? Agents are designed to make decisions within given parameters, but security may not be high on the average user’s list of actions.
And building an agent of your own is a snap. You don’t need coding skills, and thanks to the magic of natural language processing (NLP), you can train one with verbal commands alone.
Securing the Agentic AI Era
While cybercriminals don’t seem to be deploying AI agents to scale up their hacks yet, researchers have shown that agents can be configured to execute complex attacks.
Anthropic has shown that its Claude LLM (and other LLMs) can be used to replicate the mode of attack used in the 2020 Equifax data breach. Cyber experts warn that firms should expect to see similar agent-driven infections escape the laboratory and infect the real world – soon.
To prepare for that inevitability, Sourabh Satish, Co-Founder & CTO of Pangea, a firm that secures AI applications, told Techopedia that organizations adopting agent-based tools or architectures need to implement essential security controls in three key areas:
- Authentication & authorization: “Agent systems require robust user authentication tied directly to authorization, with granular controls at document and data chunk levels. Role-based, relationship-based, and attribute-based access control models provide the necessary flexibility for enterprise environments. Implementing just-in-time access provisioning further reduces the risk profile by limiting access duration to only what’s necessary for task completion.”
- Visibility & monitoring: “Security teams need complete visibility into agent operations, including model versions, authentication events, prompts, behaviors, data citations, and all interactions with external systems. Real-time alerting for anomalous patterns and comprehensive logging for forensic analysis are essential components of a robust monitoring system.”
- Content protection: “Real-time content filtering capabilities must be implemented to prevent sensitive data exposure, detect malicious content, and protect organizational information assets. Sophisticated DLP (Data Loss Prevention) mechanisms should be deployed to recognize and redact sensitive information before it leaves controlled environments.”
How Real Is the Threat?
The time to act is now. A January blog post by the National Institute of Standards and Technology (NIST) warns that many AI agents are vulnerable to hijacking, using a type of indirect prompt injection attack previously used to take control of generative AI models. The attacker places malicious instructions into the data that an AI agent is likely to ingest, triggering unintended and likely harmful actions.
A project launched last year by AI research firm Palisade Research is designed to measure how real the threat is. LLM Honeypot masquerades as a location for sensitive government and military data, setting up vulnerable servers to attract AI agents and log them when they try to break in.
So far, it has captured more than 12.5 million attempted interactions. Most of those were benign, but they have identified nine potential agents and confirmed two, which appear to originate from Hong Kong and Singapore.
The Bottom Line
So it begins. AI agents threaten to introduce cross-enterprise business risks that could disrupt operations, violate privacy, breach compliance, and harm company reputations.
Building agentic AI-specific guardrails into cybersecurity postures before the first high-profile attack happens is the sensible approach, experts say.
A proactive approach will help ensure that AI adoption isn’t derailed and investments in new projects deliver the benefits they should.
FAQs
What is shadow AI?
What are the risks of shadow AI?
How to detect shadow AI?
What are the consequences of using shadow AI?
How to avoid shadow AI?
References
- AI Agents Market Size, Share & Trends (Grandviewresearch)
- Let The Service Management Agentic AI Race Begin (Forrester)
- Global Cybersecurity Outlook 2025 (Reports.weforum)
- 32% of Employees Using AI Hide It From Their Employer (SecurityMagazine)
- How Expensive Is It to Have Your Own AI Agent, Really? (Knowledge.insead)
- Progress from our Frontier Red Team (Anthropic)
- Chinese Hackers Charged in Equifax Breach (FBI)
- Technical Blog: Strengthening AI Agent Hijacking Evaluations (NIST)
- Not what you’ve signed up for: Compromising Real-Worlde (Arxiv)
- LLM Agent Honeypot (Ai-honeypot.palisaderesearch)