
A director of product analytics at a logistics company told me this story a few weeks ago. Her team spent four months building a demand forecasting model. The algorithm was solid. The architecture was clean. In testing, it predicted regional shipping volume within 8% accuracy. Everyone was excited.
Then they connected it to production data. Accuracy dropped to 40%. Not because the model was wrong, but because three different regional offices were entering shipment data into three different systems, using three different naming conventions, with no shared definition of what counted as a "completed order." The same customer showed up as three different records. Volume numbers were double-counted in one region and undercounted in another.
Four months of model work, undone in a day by data that had been broken for years. Nobody had noticed because humans were compensating for the inconsistencies with tribal knowledge and spreadsheets. The AI could not do that.
The data problem is bigger than most leaders realize
In a major enterprise survey, the top three challenges with AI were implementation issues, integrating AI into roles and functions, and data issues. All three of those are downstream consequences of the same upstream problem: the data environment was not built for AI, and nobody budgeted the time or effort to fix it first.
Here is the uncomfortable truth. Your data has been "good enough" for years because humans fill in the gaps. People know that "Acme Corp," "ACME Corporation," and "acme co." are the same customer. They know that the Q3 revenue spreadsheet from marketing and the one from finance use different date boundaries. They know which database to trust and which one to ignore. That institutional knowledge lives in people's heads, not in the data.
AI cannot do any of that. It takes your data at face value. And when the data is incomplete, inconsistent, or siloed, the AI does not compensate. It confidently produces wrong answers. That is worse than having no AI at all, because now leadership is making decisions based on outputs that look authoritative but are built on a cracked foundation. (I wrote about how to separate genuine signals from misleading noise in Signal vs. Noise in Product Metrics.)
The three data problems that kill AI projects
1. Inconsistency across sources. The same entity (customer, product, transaction) is defined differently in different systems. When you try to merge those sources for AI training, the model sees noise where it needs signal. This is the most common and most expensive data problem. If you cannot answer "What is a customer?" with one definition that every system agrees on, your AI will struggle with anything customer-related.
2. Data trapped in silos. Legacy systems create data silos that make it nearly impossible for AI to access the full picture. Experts in the field recommend incremental integration strategies, focusing on interoperability and modular upgrades rather than massive rip-and-replace projects. Moving data to the cloud often helps, because it removes the physical barriers between systems. But cloud migration without data governance just moves your mess to a more expensive address.
3. No quality feedback loop. Data degrades over time. Customer records go stale. Categorizations drift. New edge cases appear that nobody mapped. If you are not continuously profiling and monitoring your data, what was clean last quarter is dirty this quarter. The organizations that succeed with AI treat data quality as an ongoing practice, not a one-time cleanup project. (This connects to the broader shift from one-time delivery to continuous learning. For more on building that muscle, see From Delivery to Learning.)
A practical playbook for getting your data AI-ready
1. Start with one entity, not the whole lake.
Pick the single most important entity for your first AI use case. Customer, product, order, whatever it is. Map every system that touches it. Document how each system defines it, what fields it captures, and where the definitions diverge. Do not try to harmonize everything at once. Get one entity right, prove the value, then expand. This is the same "smallest experiment" thinking we use in product work, applied to data.
2. Assign a data owner, not just a data team.
Somebody needs to be accountable for the quality of each critical data domain. Not the IT department generically. A named person who defines what "good enough" looks like, sets the quality thresholds, and owns the monitoring. The most effective organizations I've seen pair a data steward with the product owner for each AI initiative, so the person responsible for the outcome is connected to the person responsible for the input.
3. Profile your data before you model it.
Before you hand data to a data scientist, run basic profiling: null values, duplicates, outlier counts, cardinality checks, format consistency. This takes hours, not weeks. It will surface 80% of the problems that would otherwise blow up your model months later. Make this a standard step in every AI project kickoff, not an afterthought when something breaks.
4. Build a feedback loop into the pipeline.
Data quality is not a project. It is a practice. Set up automated monitoring on your key data pipelines. Track completeness, freshness, and consistency over time. When something drifts, surface it early. A weekly 15-minute data health check (completeness percentage, new duplicates, schema violations) will save you months of debugging downstream.
5. Use AI to fix your data, not just consume it.
This is the part most organizations miss. The companies that are furthest ahead with AI are using AI itself to solve data problems: probabilistic matching to merge records across databases, automated data catalogs to help people find what they need, and anomaly detection to flag quality issues before they reach a model. You do not have to solve every data problem manually. Use the technology to improve its own foundation.
Common traps
The "clean it all first" trap. Trying to fix every data quality issue across the entire organization before starting any AI work is a recipe for never starting. Focus on the data that feeds your highest-priority use case. Get that right. Learn from it. Expand from there.
The "cloud will fix it" trap. Migrating to the cloud removes physical barriers to data access, which helps. But moving dirty data to the cloud just gives you dirty data in a more expensive location. Cloud migration and data governance are separate efforts, and you need both.
The "that's IT's job" trap. If your data quality effort lives entirely inside IT, disconnected from the business teams who generate and consume the data, you will clean things that do not matter and miss things that do. The people closest to the work know where the data lies. Bring them in.
Try this next week
Pick your most important AI use case (current or planned). Identify the one entity it depends on most. Then answer these three questions:
- How many systems touch this entity? List them. If it is more than two, you probably have a consistency problem.
- Is there one shared definition? Ask two people from different teams to define it. If their answers differ, that divergence is already in your data.
- When was the last time someone profiled this data? If the answer is "never" or "I don't know," start there before you start modeling.
The goal is not perfect data. The goal is data that is understood, owned, and good enough for the specific decision you are trying to improve. That is a much more achievable target, and it is the foundation everything else is built on.
This is part of a weekly series on AI for business leaders. If your team is working through data readiness for AI, check out our AI for Product Owners micro-credential, explore upcoming public courses, or reach out at big-agile.com/contact.
Your team is using AI to generate code, content, and recommendations. Can you trace where any of it came from?