Research shows enterprises increasing investment while scaling cautiously until reliability can be proven in production
Dynatrace , the leading AI-powered observability platform, released The Pulse of Agentic AI 2026, an inaugural global study focused on how observability and reliability determine the successful operationalization of agentic AI. The survey of 919 senior global leaders responsible for agentic AI implementation reveals that enterprises are not stalling because they doubt AI, but because they cannot yet govern, validate, or safely scale autonomous systems.
A structural shift: reliability as the gating factor
The research found that approximately ~50% of projects are in Proof-of-Concept (POC) or pilot stage. Adoption is still early but growing rapidly with 26% of organizations having 11 or more projects. As organizations move beyond experimentation and into scaled deployment, they are increasingly seeking platforms that are reliable, trustworthy, and proven.
This shift is reflected in both ambition and execution, with 74% expecting budgets to rise again next year. These findings point to a structural inflection point where reliability, resilience, governance, and real-time insight define enterprise readiness for agentic AI.
Key findings from the report:
Almost half (48%) of the senior global leaders surveyed anticipate budget increases of at least $2M, suggesting investments are still prudent.
AI agents are most commonly deployed within IT operations and DevOps (72%), followed by software engineering (56%) and customer support (51%).
Of those surveyed, business leaders say improving decision-making with real-time insights is top priority (51%) when deploying agentic AI, followed closely by improving system performance and reliability (50%) and improving internal efficiency to reduce operational costs (50%).
The greatest ROI expected for agentic AI projects is in ITOps/system monitoring (44%), cybersecurity (27%) and data processing & reporting (25%).
The top two main barriers to agentic AI production at this time are security, privacy or compliance concerns (52%) and technical challenges to managing and monitoring agents at scale (51%), followed by shortage of skilled staff or training (44%).
Trust and human oversight
Organizations signal that human guidance remains a purposeful part of agentic AI strategy, even as they build toward greater autonomy. The report shows leaders expect a 50/50 human–AI collaboration for IT and routine customer-support applications and a 60/40 human–AI collaboration for business applications, signaling that human judgment guides the system by setting goals, defining boundaries, and ensuring accountability.
Additional findings include:
While over half (64%) of organizations deploy a mix of autonomous and human-supervised agents, 69% of agentic AI–powered decisions are still verified by humans, and 87% of organizations are actively building or deploying agents that require human supervision.
Only 13% of organizations use fully autonomous agents, and just 23% rely exclusively on human‑supervised agents.
The top validation methods include data quality checks (50%), human review of agent outputs (47%), and monitoring for drift or anomalies (41%).
44% still use manual methods to review communication flows among AI agents, highlighting the need for more automated, governed oversight mechanisms. Autonomy.
“Organizations are not slowing adoption because they question the value of AI, but because scaling autonomous systems safely requires confidence that those systems will behave reliably and as intended in real-world conditions,” said Alois Reitbauer, Chief Technology Strategist at Dynatrace. “With most enterprises now spending millions of dollars annually and planning further budget increases, agentic AI is becoming a core part of digital operations. At the same time, the data shows a clear shift underway. While human oversight remains essential , organizations are increasingly preparing for more autonomous, AI-driven decision-making. The focus is now on building the trust and operational reliability needed to scale agentic AI responsibly.”
Marketing Technology News: MarTech Interview with Michael McNeal, VP of Product at SALESmanago
Observability enables trust and scale for Agentic AI
As organizations scale agentic AI beyond pilot projects, observability is the crucial intelligence layer that helps to build trust by providing visibility across every stage of the agentic AI lifecycle, from development and implementation through to operationalization. The report found that observability is already used across the entire lifecycle, with the highest adoption during implementation (69%), followed by operationalization (57%) and development (54%), underscoring its role as a foundational capability as agentic AI moves into production.
Marketing Technology News: Complexity as a Cost Center: The Hidden Financial Burden of Fragmented Martech Stacks
Additionally, the report found:
Nearly 70% of organizations surveyed already use observability during agentic AI implementation to gain real-time visibility into agent behavior, system performance, and decision-making in production environments.
50% use agentic AI for both internal and external use cases, 33% for internal purposes only, and 18% for external purposes only.
50% have agentic AI projects in production for limited use cases, 44% have projects in broad adoption across select departments, and 23% have projects in mature, enterprise-wide integration.
“Observability is a vital component of a successful agentic AI strategy,” continued Reitbauer. “The Dynatrace AI Center of Excellence (AI CoE) works with many of our largest customers, and as organizations push toward greater autonomy, they need real-time visibility into how AI agents behave, interact, and make decisions. Observability not only helps teams understand performance and outcomes, but it provides the transparency and confidence required to scale agentic AI responsibly and with appropriate oversight.”
Write in to psen@itechseries.com to learn more about our exclusive editorial packages and programs.