top of page
Search

The 95 Percent: What the AI Failure Stat Actually Means

  • Writer: Ryan King
    Ryan King
  • 3 days ago
  • 7 min read

In August 2025, a research group at MIT published a number that has become the most quoted figure in enterprise software. Roughly ninety-five percent of generative AI pilots were failing to deliver measurable P&L impact. The work was based on 150 leader interviews, a 350-person employee survey, and a review of 300 public deployments. The number landed, traveled, and is now used by every consulting firm in the country, including the firms that sold the failed pilots.

Most leaders read the number and conclude that the technology is not ready, or that their organization is not ready, or that the pilots needed more time. That framing is incomplete. The data does not support that. The mechanism is different.

The 5 percent that worked do not share a model, a vendor, or an architecture. They share a structural posture toward scope and accountability. That posture is teachable. It is not what most enterprise AI programs are doing, and it explains the gap better than any technical variable.


The number is doing different work than people think

Read carefully, the MIT NANDA finding is not a verdict on the technology. It is a verdict on procurement. The successful 5 percent were disproportionately external partnerships rather than internal builds. They started with a single workflow, not a platform. They wrote down what success would look like in dollars before the contract was signed.

Compare that to the prevailing pattern. McKinsey's State of AI Trust in 2026 reports that organizations putting twenty-five million dollars or more into responsible AI initiatives report higher maturity scores. The framing is that responsible AI is a spending category. It can be. It can also be a workflow contract with one number on it.

Gartner's CIO Agenda 2026 runs the same arithmetic from another angle. Ninety-four percent of CIOs expect major plan changes inside twenty-four months. Only forty-eight percent of digital initiatives meet or exceed business targets. The fail rate is not novel and is not unique to AI. It is the long-running base rate for IT projects that begin without a measurable outcome attached.

That is the structural read. The 95 percent is not a story about model quality. It is a story about how mid-market and enterprise leaders contract for software that is supposed to do work.


What the successful 5 percent actually do

Three things, repeatedly. None of them are exotic.

  1. First, they pick a workflow that has a number on it. Not a function, not a department, not a "use case." A workflow. Accounts receivable past forty-five days. Renewal-quote turnaround. Inbound support resolution time. The success criterion is the change in that number, measured against a baseline taken before the work begins.

  2. Second, they cap the pilot at a length where outcomes can be observed. Six to twelve weeks. Long enough to install, short enough that no one can hide. Gartner's research on agentic AI governance notes that organizations with formal governance platforms are roughly 3.4 times more likely to achieve high effectiveness. The platforms are not the cause. They are an artifact of the leaders who already wrote down what they were measuring.

  3. Third, they negotiate the engagement with the outcome embedded in the price. If the workflow does not move, the bill changes. This is the part that exposes the rest of the program. Vendors who would not sign that contract told the buyer, in advance, that they did not believe their own pitch.

Together those three moves separate the 5 percent from the 95 percent. They are not technology decisions. They are procurement and governance decisions. The technology is downstream of the contract.


Why most leaders get this wrong

Three patterns recur:

  1. The first is that AI is treated as a horizontal capability rather than a workflow intervention. The board hears "we are deploying AI across the company" and approves a budget. There is no single workflow on the hook for results. Twelve months later, no single workflow has changed enough to defend the spend. McKinsey's State of Organizations 2026 reports that eighty-eight percent of organizations experiment with AI but eighty-one percent report no meaningful bottom-line impact. That is not an experimentation problem. It is a scoping problem.

  2. The second is that internal builds win the political argument inside IT and lose the economic one outside it. The MIT data is consistent on this. Buying from a specialist vendor and forming a partnership succeeds at roughly twice the rate of internal builds. The internal build feels safer because it preserves control. It is more expensive, slower, and less accountable. The mid-market in particular cannot afford that posture. A $90M industrial distributor does not have the engineering bench to compete with a focused vendor on a niche workflow. It can buy clarity faster than it can build it.

  3. The third is that pilots are scoped to prove the technology rather than to change the workflow. A pilot designed to demonstrate that an LLM can read invoices is not the same as a pilot designed to reduce DSO by ten days. The first ends with a slide deck. The second ends with cash. Most programs run the first kind, then ask why the board is unimpressed.

CIO Magazine in 2026 reported on a related dynamic: agentic AI systems do not fail catastrophically. They drift. Behavior changes incrementally as models update, prompts evolve, and tool integrations are added. A workflow that worked in week six can be quietly worse in week thirty. The drift only matters if someone is still measuring. Most pilots stop measuring after the press release.


The actual mechanism

The mechanism that separates signal from noise in enterprise AI is not the model. It is a small number of structural decisions made before the contract.

Pick one workflow. Define the baseline. Cap the engagement. Tie price to the outcome. Keep measuring after the headline.

Each of those is boring. None of them require a Fortune 50-caliber data science team. All of them are within reach of a $25M to $250M company. That is the quiet upside in the failure data. The mid-market does not need to compete with hyperscalers on infrastructure to win on AI. It needs to be more disciplined than the average enterprise on scope and accountability, which is a much easier bar to clear.

The Anthropic governance moment in early May 2026 illustrates the same point at a different altitude. Yale's Chief Executive Leadership Institute, working with Sonnenfeld and colleagues, used the release of Anthropic's most powerful model as a forcing function for a governance framework across banking, healthcare, retail, and supply chain. The framework is useful. It is also telling that it took a model release to get boards to write down what they were measuring. The companies that already had a governance posture did not need the model release. They had the posture because they had the contracts. The contracts had the numbers. The numbers were the governance.

The mid-market is where this mechanism runs cleanest. A $90M industrial distributor with one ERP, one CRM, and roughly fifteen senior leaders can pick a single workflow, name a baseline, and produce a measurement cadence inside a quarter. A Fortune 100 enterprise running twelve business units cannot. The friction at scale is not technical. It is political. Every workflow has three claimants and four dotted-line owners. Naming a single baseline becomes a negotiation. The 95 percent failure rate is partly an artifact of running the playbook at the size where the playbook is hardest to run. That is also why the mid-market wins faster on AI when it bothers to be disciplined. The size is the advantage.


What changes the picture

A mid-market CIO who is running a half-dozen AI pilots in May 2026 has a small number of moves available before the next budget cycle.

Audit each pilot against three questions:

  1. Which workflow does this change?

  2. What is the dollar baseline?

  3. What does the contract say happens if the dollar number does not move?

Pilots that cannot answer all three should be paused, not extended. Pilots that can answer all three should be funded harder, not diversified.

Treat AI governance as a measurement cadence, not a platform purchase. The billion-dollar AI governance market Gartner is forecasting will sell tools to companies that needed a measurement habit. The habit is the governance. The tool can be helpful only after the habit exists.

Resist the "agentic" framing as a budget category. Agentic systems are workflow tools. They drift. They require a person who owns the outcome and watches the number. The platforms do not change that. The procurement contract does.

The 95 percent is not a story about technology that failed. It is a story about contracts that should not have been signed, scopes that should not have been approved, and pilots that were graded on activity instead of outcomes. Read that way, the number is useful. It points directly at the move. The move is procurement discipline, not platform spend.

That is the entire game in mid-market AI right now.

A short note on how this lands inside an organization: Procurement teams are not, by training or incentive, the right owners of an outcomes contract on AI. Procurement optimizes for unit price. Outcomes contracts optimize for value capture, which often costs more on a sticker basis. The CIO who delegates this to a procurement-led RFP is buying the wrong thing for the wrong reason. The contracts that produce the 5 percent are negotiated by an operating leader who can name the workflow, the baseline, and the dollar number, then handed to procurement for terms. Reverse that sequence and the result is the 95 percent.


Sources


 
 
bottom of page