Why Most AI Pilots Fail (And What Creative Teams Can Do Instead)
The failure statistics are everywhere and the fearmongering is rife. What they actually mean for small and medium creative businesses is an entirely different conversation, and one we're keen to keep driving.
You have probably seen the numbers by now. MIT says 95% of generative AI pilots deliver no measurable financial return. RAND says 80% of AI projects fail. Gartner predicts over 40% of agentic AI projects will be cancelled by 2027. These figures are being cited in every AI strategy deck and conference keynote this year. They are useful. They are also, on closer inspection, less precise than they appear.
That matters, because what you do with these numbers depends on whether you understand what they are actually measuring.
What the failure statistics really say
The MIT figure comes from the NANDA initiative's "GenAI Divide" report, published in July 2025. It reviewed over 300 public AI deployments, interviewed 52 organisations, and surveyed 153 senior leaders at four industry conferences. The headline, 95% failure, has been widely repeated. But the methodology has drawn credible criticism: the sample skews towards executives already engaged with AI, the definition of "failure" conflates different outcomes, and the 95% figure appears in the executive summary without clear derivation from the data in the body of the report.
The RAND figure is even more commonly misattributed. The 2024 report, based on 65 interviews with data scientists and engineers, does not independently measure an 80% failure rate. It cites that number as a pre-existing estimate and then investigates the causes through qualitative research. The study is valuable for understanding why projects fail, but it is a qualitative investigation, not a statistical measurement of how many.
The Gartner prediction is exactly that: a prediction. Based on a January 2025 poll of 3,412 webinar attendees, it forecasts cancellation driven by escalating costs, unclear value, and poor risk controls. It is the most credible available estimate for agentic AI project cancellation risk, and as of March 2026, nothing in the data contradicts it.
What these numbers share: they are all directionally consistent with each other and with independent findings from McKinsey, BCG, Deloitte, and S&P Global. Only around 5-6% of organisations are achieving significant financial impact from AI. S&P Global found that 42% of companies scrapped most of their AI initiatives in 2025, up from 17% the year before. The average organisation abandoned 46% of AI proofs of concept before they reached production.
The precise percentages vary, but the direction is consistent. Most enterprise AI investment is not converting to business results.
The perception gap that explains everything
The most useful finding in the recent research has nothing to do with failure rates. BCG and Columbia Business School surveyed 1,399 employees in August 2025 and found that 76% of executives believe their workforce is enthusiastic about AI adoption. Only 31% of employees actually are. That is a two-and-a-half times overestimate. And it goes a long way towards explaining why so many rollouts stall.
If leadership genuinely believes the team is on board, they plan accordingly: enterprise platform, company-wide rollout, training sessions, adoption targets. The investment makes sense based on what they think is true. The problem is that what they think is true is wrong.
Gallup's independent research tells the same story from the other side. Employees whose managers actively support AI use are 8.8 times more likely to say AI helps them do their best work. But only about a quarter of frontline employees report getting that level of support. The appetite exists in pockets, not uniformly, and leadership is consistently overestimating how far those pockets extend.
Why "start small" is necessary but not sufficient
The evidence does support starting with focused, problem-specific automation rather than enterprise-wide rollouts. McKinsey's AI "high performers" are 3.6 times more likely to commit to transformative AI change, but they reach that commitment after demonstrating value at the use-case level. RAND's five root causes of failure (misunderstanding the problem, poor data quality, chasing new technology, inadequate infrastructure, and applying AI to problems too complex for it) are all more manageable at a small scale.
But the research also shows that small pilots alone do not produce transformation.
McKinsey's high performers commit more than 20% of their digital budgets to AI. They are five times more likely to make large bets. They do not stay small; they start small and then invest seriously in redesigning workflows around what works. The organisations stuck in the "experimentation trap" are the ones running scattered pilots without a plan for what comes next.
This distinction matters for creative businesses. The temptation is to read the failure statistics as an argument for caution. But inertia is just as dangerous as a bad rollout. The answer is a specific sequence: solve a real problem for someone who has it, build evidence and appetite from that, then redesign the workflow around the change.
What this looks like in creative businesses
The creative sector data is thinner than the enterprise research, but it exists and it tells a consistent story. BIMA's 2025 benchmarking study of 60+ agencies found that 80% actively use AI, but only around 5% are genuinely innovating, building new services, developing proprietary approaches, and transforming how they work. The rest are stuck in what BIMA calls the experimentation trap: scattered tool use without structural change.
The IPA Agency Census 2025 found that 88% of agencies say AI is having a considerable impact on how they work. The workforce consequences are already visible: 8% of agencies reduced headcount due to AI in 2024, and 24% expect to within the next 12 months. Creative and non-media agencies saw employment fall 14.3% between 2024 and 2025.
And here is the readiness gap - the AI Digital GenAI Media Benchmark found that only 11% of agency teams have a formal AI roadmap. Average confidence in AI conversations with clients was 5.6 out of 10.
These numbers describe a sector that is adopting tools quickly, losing people as a result, and largely doing so without a strategy. That is the worst combination: change without direction.
The UK State of AI Adoption in Agencies report found direct evidence for the value of even light structure: agencies with a defined AI role or strategy show markedly faster adoption, with 79% expecting significant usage growth compared with 58% of agencies with no structured approach.
You do not need a transformation team, but you do need someone asking the right questions about where AI fits, who it helps, and what changes as a result.
The agent washing problem
One more finding worth flagging, particularly for anyone being sold an enterprise agentic AI platform. Gartner found systematic "agent washing" across the market: existing chatbots, RPA tools, and basic AI assistants being rebranded as "agentic AI" without delivering genuine autonomous capabilities. Of the thousands of vendors marketing agentic products, Gartner estimates only about 130 offer anything that genuinely qualifies.
For a creative team evaluating tools, this has a practical implication. The label "agentic" on a product does not mean it does what the sales deck suggests. Ask for a specific demonstration on your workflow, with your data, before committing budget.
What the successful rollouts have in common
The counter-evidence matters too. Enterprise-wide AI rollouts do succeed in some organisations. The pattern in every successful case is the same three things.
First, problem specificity. They are solving a named problem with measurable outcomes, not buying "AI" as a category.
Second, genuine change management. Communication about AI is not the same as support for using it. The organisations where adoption sticks are the ones investing in manager capability, feedback loops, and time for people to learn.
Third, and this is the one that connects everything: they close the perception gap. They find out what their teams actually think, want, and need before designing the rollout. They do not assume enthusiasm exists; they build it.
For creative businesses under 50 people, the practical translation is simpler than it sounds. Talk to the people doing the work. Find out which tasks frustrate them, where time disappears, what they would automate if they could. Start there, with them. Let the evidence build. Then make the bigger decisions from a position of knowledge rather than anxiety.
The sequence that matters
The failure statistics are real, directionally if not precisely. What they describe is a pattern: organisations deploying AI at people rather than with them, assuming appetite where none exists, and investing at scale before earning the right to.
The creative businesses that avoid this pattern will be the ones that got the sequence right: evidence before investment, appetite before scale, people before platforms.
If your team is trying to work out where AI fits, we built SIGNAL to find the honest starting point, not the most impressive one.
A note on how this was made: Research drew on MIT's GenAI Divide report (2025), RAND Corporation's root causes study (2024), Gartner's agentic AI predictions (June 2025), BCG and Columbia Business School's employee adoption survey (2025), Gallup's manager support research (2025), McKinsey's State of AI (2025), S&P Global, BIMA's agency benchmarking study, the IPA Agency Census 2025, and the AI Digital GenAI Media Benchmark. Perplexity Deep Research was used for source validation and cross-referencing. Claude was used for research synthesis and drafting. The header image was generated using Google Gemini. The framing, the audience, the editorial judgement, and every significant decision along the way were mine.

