Key Takeaways
- 95% of AI projects fail — not because of the technology, but because of the data underneath it.
- Data quality as an AI obstacle more than doubled in a single year — from 19% of organizations citing it in 2024 to 44% in 2025.
- AI doesn't fix bad data. It amplifies bad data — at machine speed, across every workflow it touches.
- The right sequence is: data quality first, governance second, integration third, AI fourth.
- Three platforms — Quickbase, Mendix, and RapidMiner — each play a distinct role in getting your data AI-ready. None of them is the same tool doing the same job.
We recently covered a harsh reality when it comes to the growth of AI in companies: 95% of AI projects fail to deliver on their promises. The issue? Bad data quality.
Gartner predicted — correctly, as it turned out — that 30% of generative AI projects would be abandoned after proof of concept by the end of 2025. The most common reason is data quality.
And in a single year — 2024 to 2025 — the share of organizations citing data quality as their primary AI obstacle more than doubled, from 19% to 44%.
That last number matters. It means the organizations that launched AI initiatives in 2024 came back in 2025 with a different answer about what went wrong.
They bought the tools. They ran the pilots. And then they hit the same wall.
Bad data was already expensive before AI entered the picture. Research estimates that poor data quality costs organizations $12.9 million annually on average. MIT Sloan puts it at 15 to 25% of revenue.
AI does not fix that. AI amplifies it.
Good data plus AI equals multiplied value. Bad data plus AI equals multiplied errors — at machine speed, across every workflow the model touches.
What Bad Data Does to Your Operational Environment
The IBM Watson for Oncology project shows just how risky bad data can be.
MD Anderson Cancer Center invested $62 million in an AI project that ultimately failed — because the model was trained on hypothetical patient cases and outdated records, producing treatment recommendations doctors found unsafe and unusable.
NIST has since made this a foundational point in its AI risk guidance: data provenance, explainability, and governance are as vital for AI trust as the models themselves.
Most operational environments don't have a single catastrophic data failure. They have six smaller ones running simultaneously.
- Accuracy — Does the data reflect what is actually happening? Or is it what someone entered last Tuesday based on their best recollection?
- Completeness — Are required fields populated, or are there blanks that create blind spots for any model trying to draw conclusions?
- Consistency — Does the same entity — a project, an asset, a customer — carry the same identifier across ERP, CRM, and project management tools? Or does it have three different names in three different systems?
- Timeliness — Is the data current when AI needs to act on it, or are decisions being made on records that are two weeks stale?
- Validity — Is data in the right format and aligned with business rules, or are critical fields filled with free-text notes that no model can interpret?
- Uniqueness — Are there duplicates skewing analysis, inflating counts, or distorting predictions?
In construction, only 26% of contractors rate their data quality as high — from a Dodge Construction Network and CMiC study. Data accuracy tops the list of AI adoption obstacles for 57% of construction firms.
The reason is structural: project data lives across Procore, Revit, SharePoint, email, field apps, and spreadsheets. Each is a silo with its own format, access controls, and naming conventions.
Manufacturing has a specific failure mode that most AI tools never account for: sensor drift.
Sensors gradually drift from calibrated baselines, and predictive maintenance models learn from systematically wrong data. Poor sensor data alone inflates unplanned downtime by 15–20% — a direct EBITDA hit.
The Sequencing That Most Organizations Get Wrong
The order matters: data quality first, governance second, integration third, AI fourth.
Every organization that skips to AI without the foundational work ends up in the 95% failure group. Buying an AI tool does not fix your data. It exposes every flaw in your data — consistently, at scale, in front of the stakeholders who funded the initiative.
Four questions to answer before any AI initiative launches:
- Do your core systems agree? Do ERP, CRM, project management, and field tools share consistent identifiers for the same customer, asset, or project?
- Is data entering your systems clean? Are there validation rules and required field types enforcing quality at point of entry, or is your data model wide open to free-text errors?
- Is your operational data structured? What percentage lives in PDFs, emails, and spreadsheets rather than queryable records?
- Do you have governance and auditability? Who owns each data domain? Are there audit trails? Can you trace a data change back to its source?
If you can't answer all four with confidence, an integration readiness assessment is a smarter first step than a new AI purchase. VeilSun's App Checkup is designed to answer exactly those questions.
An Example in Three Platforms
We're not anti-AI tools. We use them every day. But we've watched organizations buy platforms and skip the foundational setup that makes those platforms actually work.
There are three platforms we use most often to help operations-heavy organizations get their data foundation right before AI enters the picture. They're not interchangeable — they each do something fundamentally different, and they often work together.
Quickbase
Get the Data Out of Spreadsheets and Into Structure
Think of Quickbase as the gap-filler between your enterprise systems. Most organizations run an ERP, a CRM, a project management tool — and then fill every gap between them with spreadsheets. Those spreadsheets become the single biggest AI liability in the environment.
Quickbase replaces that middle layer with structured, governed applications:
- Custom data rules enforce validation at the table level, preventing invalid data from being saved regardless of how it enters the system
- Strongly typed fields eliminate free-text entry where structured data is required
- Required fields and conditional logic enforce completeness at the point of entry
For governance, Quickbase provides granular role-based permissions, field-level controls, and audit trails that capture who changed what, when, and from what prior value. Review workflows route sensitive data changes through approval processes before they're committed.
Quickbase is particularly effective for internal-facing applications, smaller user bases, and fast deployment — when you need to get operational data out of spreadsheets quickly and into a format AI can actually use.
When VeilSun builds or optimizes a Quickbase application, embedding these controls is foundational — not optional. It's what makes every downstream AI use case possible.
Mendix
Govern Data Quality Across Complex, Multi-System Environments
For organizations where AI must span multiple systems — SAP, Salesforce, production scheduling tools, field data platforms — Mendix provides the data governance layer that makes cross-system AI coherent.
Mendix enforces data integrity through validation rules, event handlers, and access rules at the data model layer — before data is committed to any downstream system.
OData wrapping lets Mendix normalize non-OData databases and APIs into a unified, governed data model. This directly solves the consistency problem when ERP and CRM use different entity identifiers for the same customer or project.
For regulated industries, Mendix supports HIPAA, PCI DSS, DORA, and GxP through configurable access controls, audit logging, electronic signatures, and full traceability.
Where Quickbase excels at fast internal applications, Mendix handles the broader, more complex use cases: customer-facing applications, multi-tenant environments, tens of thousands of users, and deeply integrated enterprise landscapes.
When VeilSun builds Mendix applications around ERP and CRM, Mendix Connect and the platform's validation architecture become the normalization layer that cleans multi-source data before AI ever touches it.
RapidMiner (Altair)
Lay an AI Fabric Across Everything You Already Have
Quickbase and Mendix get your data structured and governed inside applications.
RapidMiner asks a different question: what about the data that's still sitting in all the other systems — the ERP you're not replacing, the spreadsheets still in use, the legacy tools that will be there for years?
RapidMiner (now part of Altair) is an AI and data science platform that creates a data fabric across all of your systems — without forcing you to migrate, replace, or consolidate them. It doesn't care where the data lives.
It connects siloed sources into a unified analytical environment that you can run reporting, machine learning, and AI applications on top of.
In practice, this means:
- Cross-system reporting and dashboards that draw from Quickbase, Mendix, Procore, your ERP, and your spreadsheets simultaneously — without a data warehouse migration project.
- Machine learning models trained on your operational data for predictive maintenance, quality forecasting, and resource planning.
- AI assistants and chatbots that can answer questions using data scattered across systems — because RapidMiner has built the contextual fabric underneath them.
- Governance and lineage controls so you can trust what the AI tells you and trace it back to the source.
It doesn't care where the data is — even if it's still in spreadsheets — it builds a unified layer on top of it so you can report across everything and power AI that understands your entire operational environment.
This is the layer most AI implementations are missing. Quickbase and Mendix build the applications that generate and structure your data. RapidMiner is the intelligence layer that spans all of it.
Three Systems Solving Different Data Needs
These platforms aren't competing for the same job. They operate at different layers.
Here's how to think about them:
|
Platform |
Primary Role |
What It Solves |
|
Quickbase |
Low-code applications that get data out of spreadsheets and into governed, structured records |
Eliminates spreadsheet silos; enforces quality at data entry; ideal for internal ops teams and rapid deployment |
|
Mendix |
Enterprise low-code platform for complex, multi-system, high-user-count environments |
Normalizes data across ERP, CRM, and field systems; enforces cross-system governance; handles customer-facing and regulated use cases |
|
RapidMiner (Altair) |
AI/data fabric layer that spans all systems — including those you're not replacing |
Enables cross-system analytics, ML models, and AI assistants without forcing migration; adds governance and lineage across the full environment |
Data Foundation First. AI Second.
Every VeilSun AI engagement starts with the data, not the model.
Before a single AI workflow is designed, we look at how data enters the system, how it flows between systems, and whether it is structured, governed, and consistent enough to be trusted.
The organizations that end up in the 5% — the AI initiatives that actually deliver measurable ROI — are the ones that did the boring work first:
- They cleaned the data model
- They enforced validation at entry
- They resolved the naming inconsistencies between ERP and CRM
- They built audit trails before they needed them
None of that is glamorous. But all of it is what makes AI run effectively.
VeilSun builds the data foundations that make AI work — on Quickbase, Mendix, and now RapidMiner — for operations-heavy industries where data quality failures are measured in dollars and downtime.
If you're not sure whether your data is ready, start there. See what that assessment looks like at our AI Development practice, or run the diagnostic first with an App Checkup.
No pressure, no pitch. If your AI initiative is stalled and you suspect the data is the reason, let's talk.
Schedule A Strategy Call Now
FAQs
Why do most AI projects fail?
AI project failure stems primarily from poor data quality — incomplete, inconsistent, or ungoverned data — rather than algorithmic limitations, as evidenced by MIT and Gartner research.
What is data quality and why does it matter for AI?
Data quality hinges on six dimensions: accuracy, completeness, consistency, timeliness, validity, and uniqueness. Since AI models rely on this data, flaws in any dimension cause them to amplify errors instead of generating value — across all workflows, and at machine speed.
What is the difference between Quickbase, Mendix, and RapidMiner for AI readiness?
Each platform operates at a different layer. Quickbase is best for getting operational data out of spreadsheets quickly, with strong governance for internal applications.
Mendix governs data quality across complex, multi-system enterprise environments and handles regulated, high-user-count use cases.
RapidMiner creates an AI/data fabric across all systems — including ones you're not replacing — enabling cross-system analytics, machine learning, and AI assistants without forcing a migration.
How does Quickbase help with data quality?
Quickbase ensures data quality at entry with custom rules, typed and required fields, conditional logic, and role-based permissions, preventing bad data from entering.
Audit trails and review workflows provide governance for regulated environments. VeilSun embeds these controls into every Quickbase application it builds. Learn more about VeilSun's Quickbase services.
How does Mendix support data governance across enterprise systems?
Mendix ensures data integrity via data model validation, normalizes backend systems with OData wrapping, and offers enterprise data cataloging, ownership, and sensitivity tagging through Mendix Connect.
It supports HIPAA, PCI DSS, DORA, and GxP compliance with audit logging, electronic signatures, and full traceability. Learn more about VeilSun's Mendix services.
What is RapidMiner and how does it help with AI readiness?
RapidMiner (now part of Altair) is a data science and AI platform that creates a data fabric across all of your systems — ERP, CRM, low-code apps, spreadsheets, and legacy tools — without requiring you to migrate or replace them.
It enables cross-system reporting, machine learning models, and AI assistants that draw from your entire operational environment. For organizations with complex, distributed data landscapes, RapidMiner is often the layer that makes AI operationally coherent.
What should I do before starting an AI initiative?
Before any AI initiative, ensure your data is accurate, complete, consistent, timely, valid, and de-duplicated. Confirm validation rules and governance controls are in place at entry. Verify core systems share consistent entity identifiers. If unsure, VeilSun's App Checkup is the recommended first step.
