VistaVu Blog

Why 95% of AI Success is Actually a Data Problem

Written by Chris Brooks | Mar 5, 2026 8:43:14 PM

It has always been about the data. We often overlook the crucial groundwork behind the results. Large language models function because they process vast amounts of text from various sources to learn communication and reasoning. Essentially, AI models are an evolution of data, mirroring its inherent patterns and meanings.

Generic AI tools are powerful generalists, but to solve business-specific problems, your internal data foundation must be strong and well-structured.

In SAP-centered organizations, this becomes even more important because the most valuable AI outcomes depend on business-context rich data tied to real transactions, not just generic content. When core finance and operational processes run through SAP, the quality, governance, and connectivity of that data become a direct advantage for AI initiatives moving from experimentation to execution.

 

Why Aren’t People Talking About Data?

Most AI failures occur because organizations discover too late that their data isn’t structured, governed, or accessible to support AI initiatives. If the foundation is weak, your AI outcomes will be weak.

The reality is, organizations consistently underinvest in data readiness. A recent study by MIT (The GenAI Divide) shows that weak data governance and fragmented data flows are among the top reasons AI projects fail.

The issue is not only data quality in isolation. It is also whether data is unified and reusable across business domains. When finance, operations, supply chain, and customer data remain fragmented, organizations end up with fragmented decisions, and AI projects often stall after the initial pilot phase.

From our partnership with  SimpleFi, we understood that organizations should treat enterprise data as a strategic asset—building a governed foundation that connects finance, operations, and planning data to support analytics and AI initiatives.

 

Why Is Data So Important?

Even simple Generic AI deployments like Copilot succeed or fail based on data quality, permissions, and governance. A few years ago, VistaVu implemented CoPilot internally, and the project team was faced with a hard decision, do we pause the AI project for 6-12 months to improve our data practices, or is it good enough for us to move forward as is and improve as we go. We were ok, but only because we implemented better data practices in 2020.

For SaaS AI rollouts, like SAP Joule, data is still important, however since the system is defined and the SaaS AI understands this, and has the access, implementing these solutions are much simpler and more of a feature rollout of an existing tool than a system rollout.

For custom modeling, the dependency on data has a direct correlation with success. Even the best engineers and architectures cannot overcome inaccurate or incomplete datasets. Most companies are not doing custom modeling today, but they should be thinking about their future and what data now can be operationalized in a few years.

At VistaVu we are cataloging conversations because they contain high-value knowledge. Capturing this now builds the data foundation needed for future AI success.

 

What Can We Do About It?

Your data does not have to be perfect. Perfection is impossible. It only needs to be good enough to create useful outcomes. The fastest way to begin is by using Generic AI while being deliberate about what data it can access.

  • As an example, Microsoft Copilot’s Work Mode can access the same content you can. If SharePoint permissions are poorly managed, AI will surface information users should not have. For Copilot, you need to ensure access first, then scale your AI usage.
  • Another option is to use embedded AI in your existing tools, such as SAP Joule or Atlassian Rovo. Their scope is narrower, their risk profile is smaller, and they help teams get comfortable with AI in practical ways.

The most complete approach is a full data-governance program that covers ownership, quality standards, lifecycle management, and process changes. These programs can be large and time consuming, but they are not required for early AI wins. They simply unlock more sophisticated results over time and at larger scale.

Over time, the goal is to treat data as business infrastructure, not just something stored in systems. That means building reusable data assets, establishing a single source of truth where it matters, and creating governed data products that can support analytics, planning, automation, and AI as connected capabilities rather than separate projects.

 

How Do I Know What Data Is Valuable and What Isn’t?

Sometimes you do not know. Ten years ago, few people would have predicted that Wikipedia would become a core part of training AI systems capable of reasoning and conversation. The lesson is to look ahead and explore possible future scenarios.

One practical way to evaluate future value is to ask which datasets could help connect strategy to execution. In many organizations, finance and operational data are strong starting points because they can support forecasting, scenario modeling, working capital visibility, and more intelligent cross-functional decision-making as AI maturity grows.

In our own case as a professional services company, we examined whether corporate device network logs could be used to help predict time entries. Each device is tied to a user, and URL activity often maps to customer systems. Four hours of access to a customer environment likely indicates four hours of billable work. It is useful, but not reliable without human validation. We chose to prioritize other datasets, but the exercise highlights the thinking required to recognize potential value of data.

 

SAP’s Role in Data and AI

When talking about data readiness, it’s essential to recognize that SAP sits at the heart of global commerce.

As SAP notes in its corporate overview, about 77% of the world’s transaction revenue touches an SAP system, and SAP customers generate roughly 87% of total global commerce. That scale matters.

From Nike to Coca-Cola to Amazon, global enterprises rely on SAP to run their core processes. That means SAP systems don’t just store data — they capture the structured, operational backbone of the world’s businesses.

If anyone is positioned to leverage enterprise data at scale, it’s SAP.

 

SAP Business Data Cloud (BDC): Turning Data into an Asset

SAP is not just sitting on data. It is actively building the foundation to manage and activate it.

This is also where SAP-native data strategy matters. SAP data carries business context, governance, and semantic meaning from core processes. With a modern foundation such as SAP Business Data Cloud, organizations can integrate, model, and govern SAP and non-SAP data without losing business meaning, which improves the reliability of analytics and AI outcomes.

SAP Business Data Cloud (BDC) is a fully managed, cloud-native platform that:

  • Unifies SAP data across business processes
  • Connects seamlessly with third-party data
  • Preserves business context and semantics
  • Eliminates costly data duplication

BDC brings together:

  • SAP Datasphere
  • SAP Analytics Cloud
  • SAP Business Warehouse
  • Native Databricks integration

Instead of copying data into disconnected lakes, BDC enables federated access and zero-copy architecture — allowing data to live where it belongs while still being usable in real time. That’s not just infrastructure. That’s AI readiness.

 

AI Is Only as Good as the Data Beneath It

BDC is designed as an AI-ready foundation. With semantic models, governed data products, and embedded AI capabilities like Joule, organizations can:

  • Train AI on trusted business data
  • Preserve financial and operational context
  • Automate insights and planning
  • Reduce risk from fragmented datasets

This is where the narrative shifts. The future isn’t just AI. The future is structured, governed, contextual enterprise data.

And SAP, because of its footprint and its data platform investments — is uniquely positioned to help organizations build that future.

 

Conclusion

The strongest AI programs usually do not begin with the most advanced model. They begin with trusted access, clear governance, and a deliberate plan for which data assets will be reusable across use cases. That is what turns early AI experiments into long-term enterprise capability.

Data exists independently of AI, but AI cannot function without data. For organizations to unlock the true potential of AI, they must intentionally build toward the data they will need in the future, treating data as a strategic asset from the outset. The choices made today about what data to prioritize and how to manage it will directly shape and enable future AI initiatives, driving meaningful change across the business and ultimately impacting the bottom line.

Ready to Build an AI-Ready Data Foundation?

AI success starts with trusted enterprise data. VistaVu helps organizations modernize their SAP landscape, improve governance, and unlock the full potential of analytics and AI.

Prefer email? Reach out and we’ll connect you with the right VistaVu expert.