Enterprise AI Paradox: Why Investments Fail

Closing the Enterprise AI Gap: From Experimentation to Execution

The Enterprise AI Paradox

Why are AI investments failing to deliver results?

It is a question that has long plagued the industry, as organizations across sectors invest billions in AI infrastructure to pursue transformation. A recent MIT study’s finding that nearly 95% of corporate AI initiatives do not yield meaningful returns further attests to this. Where progress is visible, it has largely been limited to narrow use cases, such as employee productivity tools (e.g., Copilot), developer assistants (e.g., GitHub), and a small number of consumer-facing applications.

These efforts improve convenience, but rarely change how work is organized or executed. AI is often layered onto existing processes rather than reshaping how work flows across the enterprise. This explains why board-level expectations remain unmet despite rapid advances in AI.

At this stage, the challenge is no longer access to models or infrastructure,  but how AI is integrated into enterprise design, governance, and execution. While there is increasing interest in more autonomous approaches, most organizations have yet to translate these ideas into scalable, governed modes of operation.

This disconnect between technological promise and operational reality, between experimentation and execution, is what I refer to as the Big Enterprise AI Gap.

Why Enterprises Are Stuck

Today, it’s not about a lack of AI capability. Models have become more powerful, accessible, and affordable. What has not evolved at the same pace is the work itself.

Most enterprise operating models are still built around fragmented systems, siloed data, and workflows designed for human handoffs. Instead of integrated into operations, AI is utilized as a tool, with limited orchestration across systems and data.

This structural mismatch recalls a familiar pattern: pilots show promise, but scaling stalls. Use cases remain isolated; institutional knowledge does not compound; and concerns about risk, compliance, and reliability restrict broader deployment. Hallucinations driven by data drift or model degradation further erode confidence. As a result, productivity gains remain incremental rather than systemic.

Until operating models evolve to support more autonomous execution, with appropriate oversight and accountability, increased investment alone will not close the gap.

Against this backdrop, interest in agentic AI has accelerated. Autonomous agents capable of taking on end-to-end tasks previously performed by humans are increasingly discussed as a way to align expectations with execution. Concepts such as ‘recruiting’ AI agents, training them, or forming hybrid human-AI teams are gaining visibility. For most enterprises, however, these ideas remain more aspirational than operational.

Agentic AI and the Redesign of Enterprise Execution

Agentic AI shifts the focus from isolated tasks to end-to-end execution. Agent-based systems are designed to execute sequences of work across systems, coordinate actions with limited human intervention, and improve through repeated execution. In effect, they introduce a new operating fabric for the enterprise.

Rather than accelerating existing processes, agentic AI enables organizations to orchestrate workflows and adapt continuously. Context is retained across steps, decisions move closer to execution, and knowledge accumulates instead of resetting with each new use case.

For enterprises still operating with fragmented systems and manual coordination, agentic AI changes the paradigm. It creates the possibility of orchestration at scale. Work can be executed end-to-end, learning compounds across systems, and tasks, data, and decisions are coordinated in real time through a shared enterprise memory.

Just as cloud became the foundation for scalable infrastructure, agentic AI is emerging as a foundation for scalable enterprise execution, governing how tasks flow, decisions are made, and value is delivered.

It is not a technical novelty, but an operating layer that enterprises are beginning to design around.

Trust, Risk, and Why ‘AI Delivered Right’ Matters

As AI systems take on more autonomous roles, enterprise leaders are shifting their focus from generating output to ensuring smooth execution. Productivity gains matter only when outcomes are consistent, explainable, and governed.

Across industries, this shift is already exposing new risk surfaces. As AI moves closer to core workflows, organizations are confronting issues that go beyond model performance to trust, accountability, and control.

One of the most visible risks is hallucination, where outputs that appear plausible but are incorrect or fabricated. In traditional machine learning systems, such failures were typically linked to data drift, model degradation, misclassification, or poor data quality. These risks were familiar and largely predictable.

Generative and agent-based systems introduce a different dynamic. They can produce information with no grounding in the enterprise context, ranging from inventing facts and misaligning responses to misinterpreting intent and introducing non-existent entities. These behaviors typically stem from incomplete data, weak contextual anchoring, outdated information, or overly generalized models.

For enterprises, the impact goes beyond accuracy. Hallucinations introduce operational and governance risks that affect scale:

  • Loss of trust in AI-driven decisions among leaders and regulators
  • Unclear accountability when outcomes cannot be reliably traced
  • Limited deployment into core processes due to risk and compliance concerns

Without built-in mechanisms for validation, drift management, and policy enforcement, autonomy becomes a constraint rather than an advantage.

This is why ‘AI delivered right’ matters. Risk management and governance must be integral to execution from the outset, not added after deployment.

For CXOs responsible for enterprise AI adoption, these considerations now sit at the center of enterprise leadership.

Designing Trust into Agentic Systems

Meeting that responsibility requires trust to be engineered into agentic systems by design from the start. Experience shows that predictable, auditable autonomy depends on a set of practical design choices:

  • Agent Specificity: Break work into narrowly scoped agents with clear mandates. Break work into narrowly scoped agents with clear mandates. As complexity increases, so does the risk of hallucination and unintended behaviour. Tighter scope reduces that risk and improves reliability.
  • Domain-Specific Skillsets: Equip agents with the tools and data relevant to their role, including enterprise data sources, search capabilities, vector stores, and APIs, so that execution remains grounded in the business context rather than relying on generalized inference.
  • Validated Deployment: Test agent workflows end-to-end in controlled environments before release. Predictability and alignment with business intent must be established before autonomy is expanded.
  • Independent Checks: Maintain a maker–checker pattern, with separate verification mechanisms to detect data and model drift, enforce policy and compliance, and validate outputs continuously.
  • Design for Change: Business rules, products, and regulatory constraints evolve constantly. Modular agent design allows workflows to be updated without rebuilding systems, ensuring governance keeps pace with execution.

Taken together, these design choices point to a familiar lesson from the cloud era: scale depends not just on capability, but on operating models that embed governance and economics by design. Experience from platform work at Tech Mahindra, including Orion, reinforces this view. Reliable agentic execution is achieved when autonomy, validation, and oversight are engineered to function together as a system.

The Human + Agent Operating Model

Designing trust into agentic systems reshapes how work is organized and enables a distribution of responsibility between humans and machines.

In this context, agentic AI should not be viewed as replacing humans, but rather as redefining how work is shared. In this operating model:

  • Humans focus on judgment as strategists, supervisors, and exception handlers
  • Agents handle execution that’s repetitive, rules-based, and data-intensive
  • Teams evolve into hybrid human-agent squads
  • Productivity comes not from more hard work, but from orchestration

As a result, benefits are both operational and human. There are fewer handoffs, greater transparency in accountability, and consistency in execution. Over time, this model unlocks new levels of throughput, quality, and speed without increasing complexity.

Aligning Enterprise AI with Economics

Even when agentic AI works, many initiatives stall because economic models don’t align with execution. In traditional models, organizations pay for usage or effort, even though value is created only when tasks are completed successfully.

Agent-based execution calls for a different approach. Consumption models should reflect outcomes rather than activities. Approaches such as service tokens tie cost to executed work, keeping consumption transparent and economics predictable as workloads adjust.

When AI is consumed as a utility and paid for execution rather than experimentation, organizations are better positioned to transition from pilots to sustained operations with improved returns.

In practice, commercial models shape whether AI moves from pilots to production. To scale reliably, consider:

  • Outcome-linked Consumption: Tie fees to executed work or business outcomes, rather than infrastructure use alone
  • Elasticity with Predictability: Enable workloads to scale while keeping costs and value transparent to the business
  • Clear Metrics and SLAs: Define what success looks like (accuracy, throughput, compliance) and make it measurable.

A Vision for the Enterprise of the Future

We stand at the threshold of a new era of enterprise. Productivity tools and AI copilots were the tip of the iceberg. Agentic AI, when delivered responsibly, has the potential to reshape how enterprises operate, how decisions are made, and how value is created. Real progress will come from redesigning operating models so intelligence is embedded directly into execution.

Enterprises that treat intelligence as a core part of how work is governed and executed will lead. This shift calls for a shared understanding among leaders, technologists, regulators, and the workforce.

That imperative is reflected in the theme of this year’s World Economic Forum Annual Meeting, A Spirit of Dialogue. Closing the big enterprise AI gap is ultimately a design challenge, not a technological one. With the right architecture, governance, and operating discipline, it becomes an opportunity to build enterprises that are more resilient, more adaptive, and better aligned with the realities of scale.

Endnotes

  1. Challapally, A. Pease, C. Raskar, R, Chari, P. (2025, July). The GenAI Divide: State of AI in business 2025 report (Version 0.1). MIT
About the Author
Sham Arora
Chief Technology Officer, Tech Mahindra

Sham is Chief Technology Officer (CTO) at Tech Mahindra, driving enterprise-wide innovation and engineering-led solutions. With deep expertise in digital transformation, cloud adoption, automation, and platform modernization, Sham is passionate about creating future-ready ecosystems that deliver measurable business outcomes.Read More

Sham is Chief Technology Officer (CTO) at Tech Mahindra, driving enterprise-wide innovation and engineering-led solutions. With deep expertise in digital transformation, cloud adoption, automation, and platform modernization, Sham is passionate about creating future-ready ecosystems that deliver measurable business outcomes. As CTO, Sham shapes Tech Mahindra’s technology vision and strategy, aligning innovation with business value creation and accelerating the company’s journey toward a digital-first, hyperconnected world using AI as a central lever for transformation. His focus on building resilient, scalable, and sustainable technology architectures ensures that global enterprises can modernize and thrive in an era of rapid change.

Read Less