AI, when done right, is not just transformational, it redefines what’s possible. Quantum leaps in employee productivity. Structural margin expansion. Entirely new AI-discovered revenue streams. A defensible competitive moat that compounds over time. All of this is in the realm of possibilities.
And yet, between 80 and 95 percent of companies report limited return on investment from AI to date, and only 13 percent have reached high AI maturity, according to “Humans at the Helm of AI,” the study of more than 500 Global 2000 executives we published jointly with HFS Research earlier this year. The findings drew coverage in Forbes, CIO Dive, Yahoo Finance, Diginomica and across the IT Brief network, and the conversation they started is still unfolding.
Why the gap? Because enterprises do not operate in the clean, greenfield conditions where AI demos shine. They operate in brownfield reality, and that reality is the single biggest bottleneck to scaling AI.
The brownfield problem is bigger than most leaders admit
Every enterprise running today is an occupied building under renovation. Decades of systems, data, processes, contracts, and people, all in motion, all interdependent. You cannot shut it down to rewire it. You cannot bolt AI onto it and expect the building to hold. And yet that is exactly what most AI strategies assume.
Knowledge is trapped in siloed systems. Technical debt blocks scale. Data is fragmented across platforms that were never meant to talk to each other. Operating models were not built for AI at the speed AI now moves. Teams continue to re-ask, re-analyze and re-build, compounding governance issues and risk with every cycle.
Meanwhile, AI is spreading faster than it is being governed. Employees are adopting tools that IT cannot see and security cannot control. Sensitive data is leaving the enterprise through everyday workflows. Attack surfaces are expanding without corresponding increases in defense. Compliance obligations are accumulating without audit trails to support them.
The headlines from the last ninety days make the point more bluntly than any study could. At one of the world’s largest cloud providers, an internal AI coding agent, operating with production access, decided the most efficient way to fix a bug was to delete and rebuild the live environment. It executed at machine speed, faster than any human could have intervened. The company’s own explanation was telling: not an AI failure, a permissions failure. Which is precisely the point. The governance model assumed a human at the keyboard. The AI did not wait for one.
At a global consumer technology company, the opposite failure mode played out. Engineers were encouraged, through internal leaderboards, to adopt AI coding tools aggressively. They did. Adoption ran so far ahead of forecasting that the full year AI budget was exhausted well before year-end. The CTO’s public admission was candid. Back to the drawing board. The productivity was real. The operating layer to govern it was not.
Two incidents, same root cause. AI moving at machine speed through enterprises whose operating models, permissions, review cycles, budget controls, accountability structures, were designed for humans at human speed. In my conversations with CEOs and boards, one theme keeps popping up. If you do not have an AI strategy, you have AI chaos. And the chaos is not theoretical anymore. It is already in the news, and it is already in the numbers.
Accountability has to be embedded at every layer, with a human at the helm
The second reality is just as uncomfortable. AI cannot scale inside an enterprise unless accountability is embedded at every layer, with a human at the helm at every level of the organization.
We already know how to do this:
- Engineering teams have accountability for what ships.
- Security teams have accountability for what is exposed.
- Operations teams have accountability for what runs.
- Finance has accountability for what is spent.
None of these functions scaled by removing human ownership. They scaled by making ownership explicit, legible, and enforceable in the system itself.
AI needs the same treatment, and most enterprises have not done the work. They have not defined what AI is authorized to decide versus what a human must decide. They have not made it clear to employees whether they are engaging with AI or deferring to it. They have not registered their models, traced their data lineage, or designed human checkpoints into workflows before incidents force them to. Accountability, when it shows up at all, shows up as a slide in a governance deck, not as a constraint wired into the operating layer.
This is how you get a workforce crisis on top of a technology bet. Employees stop questioning AI outputs because no one told them they were allowed to. Leaders lose the ability to answer the question that matters most to a regulator, a customer, or a board. Who decided this, and on what basis?
AI without engineering discipline is not acceleration, it is risk at speed
So the gap between AI’s potential and AI’s return is not a technology problem. It is an engineering execution problem, and it is an accountability problem. Both have to be solved at the same time.
For AI to scale inside a real enterprise:
- Security must be designed in from day one, not bolted on after.
- Enterprise grade integration demands deliberate architecture, not shortcuts.
- Quality must be measurable, with observability and closed loop feedback built into every deployment from the start.
- Accountability must be executable, enforced in code, not asserted in policy decks.
Without this, AI does not accelerate innovation. It accelerates risk. Faster chaos is still chaos.
The answer is not more tools, it is an operating system
The answer to this challenge is not more tools. It is a unifying system.
Just as an operating system once standardized how humans interact with computers, abstracting complexity, managing resources, enabling everything built on top, enterprises today need an AI operating system to standardize how their employees, processes, and systems interact with AI. An operating system:
- Abstracts AI and data complexity.
- Standardizes human and AI interaction.
- Manages models, governance, and orchestration across the enterprise.
Without the operating system, AI investments remain disconnected, ungoverned, and unable to compound into lasting advantage.
This is the conviction behind ALTi AIOS™, an AI operating system to enable every business to become an AI business. The operating system makes governance executable rather than aspirational. Policies enforced in code. Models registered, evaluated, and observable. Data lineage traceable. Human checkpoints designed into the workflow, not retrofitted after an incident.
On the human side, leaders still have to do three things most are postponing:
- Define what AI is and is not authorized to decide, at the level of specific workflows, not slide deck principles.
- Make accountability legible, so employees know whether they are engaging with AI or deferring to it.
- Accept that governance is a design constraint on the system, not a checkpoint at the end of it.
This is the point Phil Fersht, CEO and Chief Analyst of HFS Research, has been making with characteristic clarity. When we launched the study together, Phil put it this way:
Enterprises are scaling AI faster than accountability, and that gap is now a workforce crisis. When leaders don’t define what AI decides and what humans own, employees stop questioning it. That’s not augmentation, it’s abdication. Fix it now, or you’re not building an intelligent organization. You’re scaling unmanaged risk.
Neither side works alone. AI without braver leaders produces faster chaos. Braver leaders without an AI operating system produce opinions that cannot be enforced. Both together, done with engineering rigor, is what separates the enterprises that will compound AI advantage over the next three years from the ones still running pilots.
In the coming days, we will share more details on what we are building and how everyone can join the conversation. Exciting times.
Join the conversation
The “Humans at the Helm of AI” research started this conversation. On April 21, we are continuing it live. I will be joining Phil Fersht, Dana Daher of HFS Research, and Paul Daugherty, technology CEO and CTO, board director and author, for a candid executive level discussion on what is holding enterprises back from scaling AI into real business outcomes. We will get into the AI velocity gap, what “human in the loop” means when it is working, why accountability breaks down across partner ecosystems, and what the enterprises putting humans genuinely at the helm are doing differently.