AI & TechArtificial IntelligenceBusinessCybersecurityNewswire

AI’s Rise to Autonomy Left Security Behind

▼ Summary

– Enterprise AI has moved from pilots to production, creating a major security gap between what AI agents can do and what security teams can control.
– The primary security challenges are autonomous agents causing damage, widespread data exposure via “shadow AI,” and the rise of prompt injection attacks.
– Existing governance frameworks like NIST AI RMF are insufficient as they lack the specific technical controls needed for agentic AI deployments.
– Organizations are advised to integrate continuous adversarial testing into their engineering lifecycle, automating it where possible and tiering agents by risk level.
– Implementing concrete, technical controls like scoped credentials and runtime guardrails is essential, mirroring the shift that reduced risk in cloud security.

The rapid shift of artificial intelligence from experimental pilots to core business systems has created a dangerous security void. AI agents now autonomously handle sensitive data and critical transactions, yet most security teams lack the visibility and controls to manage these dynamic systems effectively. This gap has led to substantial financial losses and new categories of risk that traditional security frameworks are ill-equipped to address.

A recent industry briefing, developed with insights from Stanford’s Trustworthy AI Research Lab and over forty security leaders, outlines the primary security challenges organizations faced in 2025 and forecasts the most pressing risks for the coming year. The data reveals a troubling landscape. An EY survey indicates that 64% of large enterprises have lost over one million dollars due to AI failures, with one in five reporting a breach directly linked to unauthorized “shadow AI” use.

Security practitioners are currently grappling with three dominant problem categories. The first is the agent challenge. Modern AI has evolved from simple assistants into autonomous agents that perform multi-step tasks, call external tools, and make independent decisions. This autonomy introduces failure modes that don’t require a malicious attacker. An agent with excessive permissions can cause significant damage through its normal operations. Surveys show 80% of organizations have observed risky agent behaviors, such as unauthorized system access, while only 21% of executives claim full visibility into agent permissions and data access.

The second major issue is the visibility challenge. A staggering 63% of employees admitted to pasting sensitive company data, including source code and customer records, into personal AI chatbot accounts in 2025. The average enterprise is estimated to have roughly 1,200 unofficial AI applications in use, with 86% of organizations having no visibility into the associated data flows. Breaches involving shadow AI cost an average of $670,000 more than standard incidents, primarily due to delayed detection and difficulties in assessing the damage.

Third is the trust challenge. The threat of prompt injection moved from theoretical research to a common production incident in 2025, topping the OWASP LLM Top 10 list. This vulnerability stems from the fundamental difficulty large language models have in distinguishing instructions from data inputs. The risk is amplified as over half of companies now use retrieval-augmented generation or agentic pipelines, each creating new potential surfaces for these attacks.

Existing high-level governance frameworks, while useful for organizational structure, fail to provide the specific technical controls needed for agentic AI. They do not address necessities like tool call validation, prompt injection logging, or containment testing for multi-agent systems. Research indicates that model-level safety guardrails are often insufficient, with fine-tuning attacks bypassing leading models in a majority of test cases. Technically specific controls that add input validation and action-level guardrails are essential to close this gap, much like how enforceable technical controls like multi-factor authentication proved more effective than policy statements alone in conventional cybersecurity.

To combat these risks, the briefing emphasizes integrating continuous adversarial testing directly into agent development and operational workflows. The operating model should combine strong platform defaults, automation, and targeted expertise. Baseline security, such as sandboxed tool execution, scoped credentials, and runtime policy enforcement, should be built into the platforms themselves, not require custom engineering for each deployment. Automated testing suites should trigger automatically with any model update or agent reconfiguration, allowing human experts to focus on investigating meaningful changes rather than manually executing entire test playbooks.

A practical approach involves tiering agents by their risk level. Agents with access to sensitive data or production systems warrant continuous testing and stricter review gates, while lower-risk agents can rely on standardized controls and periodic checks. The overarching goal is to make security validation an inherent part of the engineering lifecycle. For organizations with limited resources, starting with automated testing tied to deployment pipelines and implementing runtime guardrails before any sensitive agent goes live is a critical first step.

The evolution of identity and cloud security offers a valuable lesson. Shifting from broad policy statements to enforceable technical controls, like least-privilege access and short-lived credentials, dramatically reduced lateral movement and contained the impact of incidents. Applying this same principle to AI is paramount. An agent with tightly scoped capabilities and time-bound credentials simply cannot access resources it was never granted, creating a concrete and observable security improvement.

(Source: HelpNet Security)

Topics

ai security 100% Agentic AI 95% shadow ai 90% visibility challenges 85% prompt injection 85% adversarial testing 80% security frameworks 75% ai governance 70% data exposure 70% technical controls 70%