Buyer's Playbook

Your data is why your AI project is behind

By Ashit Vora9 min read

What Matters

  • -Data readiness is the top barrier to AI success for 43% of enterprises. Most discover this mid-project, after budget is committed and timelines are set.
  • -Data readiness has four dimensions: accessible, clean, structured, and sufficient. A project can fail on any one of them.
  • -You don't need perfect data to start. You need data that is good enough for v1. The bar is lower than most teams think - but you need to audit it to know where you stand.
  • -Data problems are fixable. The mistake is hiding them inside the AI development timeline instead of making them a visible, separate phase with their own scope and budget.
  • -Starting the AI build and fixing data in parallel is possible - but only with a clear handoff plan. Without it, you end up rebuilding twice.

Your AI project is behind. The team is telling you it's the data. And they're probably right.

Gartner found that 43% of enterprises name data quality as their top barrier to AI success. That makes it the most common single cause of AI project failure - more common than the wrong model choice, wrong vendor, or insufficient budget. And the uncomfortable part: most organizations discover this after the project starts, not before.

You've committed budget. You've set a timeline. And now you're four weeks in and the data team is saying it'll take three months to clean the training set.

This article won't make your data problems disappear. But it'll help you understand what you actually have, what the AI needs, and how to make progress without building on a foundation that will crack.

What data readiness actually means

"Data readiness" sounds like a technical checkbox. It isn't. It's a business question: can the AI do the job your organization is asking it to do, with the data your organization actually has?

There are four dimensions to check. A project can fail on any one of them.

1. Accessible. Your data exists somewhere. Can the AI get to it? Data that lives in PDFs is not accessible without an extraction pipeline. Data in legacy systems with no API is not accessible without an integration layer. Data spread across 12 spreadsheets in 12 different formats is accessible only in theory.

The accessibility gap is often the most surprising one. Teams assume "we have the data" means "the AI can use the data." These are different things.

2. Clean. Clean data is consistently formatted, free of critical gaps, and free of systematic errors. The customer table has an "industry" field - but 40% of rows have it blank, 30% use one taxonomy, and 30% use a different one from before a 2021 rebrand. That field isn't clean. Feeding it to an AI classification model trains the model to be 40% uncertain before it starts.

Clean doesn't mean perfect. It means consistent enough for the model to learn from it.

3. Structured. Structure is about shape. A language model needs text. A classification model needs labeled examples. A recommendation engine needs interaction history - what users did, in what order, with what outcomes.

The data might be clean and accessible, but in the wrong shape for the AI's job. A common version of this: a company has years of support tickets in a well-maintained database, but they don't have labels - nobody ever tagged whether a ticket was resolved by automation or by a human. The data is clean. It's accessible. But the AI can't learn "what resolution looks like" because that information was never recorded.

4. Sufficient. Volume and variety matter. A document retrieval system can work with a few hundred well-written documents. A fine-tuned classification model typically needs 500-2,000 labeled examples per class minimum. A fraud detection model needs enough historical fraud cases to learn the pattern - and fraud is rare enough that this often means years of history.

The sufficiency bar changes with the approach. Retrieval-augmented systems (RAG) need less data than fine-tuned models. Smaller scopes need less than broad ones. But there's always a minimum below which the AI doesn't have enough to work with.

Data Readiness: What Good Looks Like vs. What Fails

Accessible
Ready
Structured database with API access, or documents in an indexed repository
Not Ready
Data trapped in PDFs, emails, legacy systems with no export, or 12 different spreadsheets
Clean
Ready
Key fields > 90% populated, consistent formatting, no systematic errors in critical columns
Not Ready
Fields 40%+ empty, multiple taxonomies mixed, inconsistent formatting across records
Structured
Ready
Labeled examples available, interaction history recorded, entity relationships mapped
Not Ready
Correct data exists but never labeled, or logged in free text with no consistent schema
Sufficient
Ready
500+ labeled examples per class, or 200+ documents per topic, or 12+ months of interaction data
Not Ready
Fewer than 100 examples per class, sparse interaction history, or data only covers recent edge cases

These are directional benchmarks. The specific bar depends on your AI approach and use case.

The audit you should run before you build

Most teams skip this step. They assume the data is ready because someone in a meeting said "we have all that data." The audit takes one week and saves months.

For every data source your AI needs, answer these questions:

Where does it live? Name the specific system, database, or file store. Not "our CRM" - which table, which database, which API endpoint.

Who controls access? Is there an API key, a DBA who needs to approve queries, a legal review for PII data? Access issues that take a week to resolve become blockers on day one if you don't surface them early.

How complete are the critical fields? Run a simple query: for the fields your AI needs, what percentage of rows are non-null? If a critical field is below 80%, that's a problem to fix, not to work around.

Is the labeling there? If your AI needs to classify, does the data have labels? If it needs to rank, is there a signal for "good" vs. "bad" outcomes? If it needs to retrieve, is the content indexed and searchable?

How far back does it go? For use cases that depend on history (recommendations, fraud detection, demand forecasting), how many months or years of records exist? Is the data from the current workflow or from a process that's since changed?

"We run this audit in week one of every engagement - before architecture, before model selection, before anything. We have never had a project where the audit found no gaps. But we have had projects where the audit found gaps big enough that the original scope needed to change. Finding that in week one costs nothing. Finding it in week eight costs everything." - Ashit Vora, Captain at 1Raft

The audit output isn't a pass/fail grade. It's a map of what you have, what the AI needs, and what gap you need to close between them.

What good enough actually looks like

Perfection is not the standard. Good enough for v1 is the standard.

For a retrieval-augmented system pulling from internal documentation: you need the documents accessible and indexed, with a consistent enough structure that search works. You don't need every document perfectly formatted. 80% quality across 200 documents beats 100% quality across 20.

For a classification model: you need labeled examples. Typically 500-2,000 per class for a fine-tuned model; fewer if you're using a foundation model with few-shot prompting. The labels don't need to be perfect - interannotator agreement above 85% is workable for most use cases.

For an AI agent pulling from transactional data: you need the key entities (customers, products, orders, events) in a queryable form with the critical fields populated. You don't need every edge case covered. Edge cases can be added to the eval set and addressed in v2.

The pattern across all of these: define what the AI needs to do in v1, work backward to the minimum data requirements, check against what you have. Closing a specific gap is manageable. Fixing "all data problems" is not.

How to start building while data is still being fixed

The question we get most often: "Can we start building before the data is ready?"

Yes - but only if the two workstreams are separated clearly.

The AI build and the data pipeline are different work. Different skills, different timelines, different definitions of done. You can run them in parallel if you treat them as separate projects with a defined handoff point.

What parallel looks like in practice: the data team builds the pipeline, the extraction logic, and the labeling workflow. The AI team builds the infrastructure, the model architecture, and the interface. They define the data schema together on day one - what format the data needs to be in, what fields are required, what the output label looks like. Then they work in parallel until the handoff: the data team delivers clean, structured, sufficient data in the agreed schema, and the AI team plugs it in.

What parallel does NOT look like: the data team works on "data cleanup" with no defined output, while the AI team builds against placeholder data, and everyone plans to "sync up when the data is ready." This approach produces two systems that don't connect, a surprise integration week, and a second build.

The handoff plan is the thing that makes parallel workstreams work. Without it, you're just delaying the collision.

The cost of getting this wrong

The RAND Corporation's analysis of AI project failures found that 80% of AI projects fail to reach production. Data problems are the single most cited cause. The projects don't fail because the team chose the wrong model or the wrong vendor. They fail because someone assumed the data was ready without checking.

The cost isn't just the project delay. It's the rebuild. Teams that discover data gaps mid-project have three options: wait for the data to be fixed (delays the AI build by months), proceed with bad data (builds a system that doesn't work), or rebuild the AI once clean data arrives (pays for the work twice). All three are expensive.

The audit costs a week. The rebuild costs months.

If your project is already behind because of data, it's not too late to run the audit now, separate the workstreams, define the handoff, and get both moving in a direction that doesn't require rebuilding. If you haven't started yet, run the audit before you commit to a timeline.

The data is fixable. The question is whether you find that out before you've built on top of it.

Our AI consulting team runs data readiness assessments as the first phase of every engagement. If you want to know where your gaps are before you commit budget to a build, that's exactly what a first conversation covers.

Share this article