Why Gartner Predicts 60% of AI Projects Will Fail and the Fix Is Boring (But Necessary)

Why AI Projects Are Really Failing
Traditional data management was built for reporting. Data needed to be accurate enough to populate a dashboard or run a quarterly report. Close enough worked. A few duplicate customer records were annoying but manageable. Inconsistent product names across systems? Someone would fix it manually.
AI doesn't work that way.
AI learns from your data. It scales whatever is in there, including the errors, the inconsistencies, the gaps. A model trained on bad data doesn't produce bad outputs occasionally. It produces bad outputs systematically, at scale, and with confidence. That's far more dangerous than a bad report.
A recent Gartner survey found that 63% of organizations either don't have or aren't sure if they have the right data management practices for AI. That means most organizations currently investing in AI are building on a foundation they themselves aren't confident in.
The Problems That Keep Showing Up
When we work with enterprises across pharma, financial services, manufacturing, and energy, the same issues appear regardless of industry. They're not exotic. They're the basics.
Duplicate and conflicting records. When the same customer exists in three systems with three different addresses and three different account statuses, an AI system can't tell which one is right. It uses all of them, and the outputs don't reflect reality.
No single source of truth. Without master data management, there's no authoritative version of core business entities like customers, products, or suppliers. AI models end up reconciling contradictions that should have been resolved at the data layer.
No data quality measurement. You can't fix what you can't see. Most organizations don't have DQ scorecards or automated monitoring. They find out data is bad when an AI output is obviously wrong, which is the most expensive moment to discover it.
Ungoverned metadata. AI systems need context. What does this field mean? Where did the data come from? Who owns it? How reliable is it? Without a governed data catalog, that context doesn't exist.
No data lineage. Regulators and internal teams increasingly want to understand how an AI arrived at a decision. Without lineage, that question has no answer.
None of this is new. These problems have existed in enterprise data environments for a long time. What's new is that AI makes them consequential in a completely different way.
Why the Fix Keeps Getting Skipped
If the problem is understood, why aren't more organizations solving it?
Because the fix doesn't make for a good slide deck.
"We're investing in AI-powered predictive analytics" is a board-level conversation. "We're cleaning up our customer master data and implementing DQ scorecards" is not. But the second one is what makes the first one possible.
There's also a timeline problem. Launching an AI tool takes weeks. Building a properly governed data foundation takes months. In an environment where executives want AI wins quickly, the foundational work gets deferred. Right up until an AI project fails and the post-mortem points to data quality as the root cause.
The numbers back this up. McKinsey reports that nearly two-thirds of firms have failed to scale their AI projects. In a separate survey, 75% of data leaders said they don't fully trust their data for decision-making. You can't build AI you trust on data you don't trust.
What AI-Ready Data Actually Looks Like
Gartner's own recommendation is to build iteratively. Extend and improve existing data management practices to support AI use cases rather than rebuilding everything at once. That's sound advice, and it matches how we approach governance programs at Incept.
In practice, AI-ready data means a few specific things:
Automated data quality rules that are measured and visible, not just documented in a policy somewhere. DQ scorecards that show quality by domain, system, and business rule.
A governed data catalog where users can find data assets, understand what they mean, and trust where they came from.
Master data under control. One authoritative version of customers, products, and suppliers across all systems.
Active metadata that gives AI systems and their users the context behind every data asset.
Stewardship workflows so data issues get flagged, routed, and resolved rather than left to accumulate.
This is not theoretical. These are the components we have deployed for clients like Sanofi, Regeneron, Opella, and Dominion Energy on the Informatica IDMC platform. They are the reason those organizations' AI programs are working when comparable programs elsewhere aren't.
The Window Is Closing
Organizations investing now in data quality, governance, and MDM are the ones whose AI programs will be in the 40% that succeed. The ones deferring this work are building toward the 60% that Gartner predicts will be abandoned.
The good news is you don't need to rebuild everything at once. The right approach is targeted. Identify the data domains your priority AI use cases depend on, measure their current quality, and start closing the gaps.
That's exactly what we help organizations do through our 5-Day Discovery. It's a focused assessment that maps your current data governance maturity, identifies where AI-readiness gaps exist, and delivers a concrete, prioritized plan.
It's not glamorous. But it's the work that makes everything else possible.
Ready to understand where your data stands?
Book Your 5-Day Discovery with Incept
Incept Data Solutions is a Platinum Informatica Partner and the first in North America to deploy Informatica CDGC and CDQ. With 195 certifications and 100+ enterprise engagements, we make data governance actually work.
.png)

.webp)



.png)

