TL;DR: In the rush to deploy conversational and voice AI agents, many businesses overlook a critical truth: an AI agent is only as trustworthy as the data behind it.
We’ve all seen what happens when an enthusiastic chatbot goes off-script – confidently quoting the wrong price to a customer or citing a policy that doesn’t exist. These aren’t just harmless quirks; they’re “confident mistakes at scale,” and they can erode customer trust or even violate regulations.
The irony is that such failures often have little to do with the AI model’s sophistication and everything to do with the data (or lack thereof) feeding it. As one AI leader put it, “if the data underneath isn’t trustworthy, [AI systems] will make confident mistakes at scale”. In other words, even the smartest AI can turn into a liability if it’s drawing from a broken data foundation.
This post explores why robust data infrastructure and pipelines are the unsung heroes of successful AI agents in real business workflows. Whether it’s delivering information to customers, retrieving client data, or generating quotes and pricing, clean, connected, and well-governed data is what makes the difference between an AI that dazzles and one that disappoints.
We’ll look at industry evidence – from Gartner’s striking finding that 85% of AI projects fail due to data issues, to Salesforce’s recent $8 billion bet on data integration (its acquisition of Informatica) – all pointing to a new reality: investing in data plumbing matters more than obsessing over the latest large language model. We’ll also dive into a practical example from our work at Real Wave, where we built a shopping agent using our “context engineering” framework. This shows how engineered data pipelines can turn an AI chatbot into a reliable, revenue-driving agent.
The goal is to cut through the hype and deliver a forward-looking, business-savvy perspective: if you want AI that actually works in the enterprise, you must treat your data infrastructure with as much importance as the AI itself. It’s a message that resonates across sectors – from finance to retail to healthcare – because no matter your industry, bad data is a recipe for bad AI. Let’s unpack why the future of AI isn’t just about bigger models or clever prompts, but about building a solid data foundation that makes those models truly useful.
The Hidden Risk: AI That’s Confidently Wrong
The promise of conversational AI in business is huge. We imagine virtual agents handling customer inquiries, generating quotes on the fly, or retrieving client info with a friendly voice interface – all at scale. But with that scale comes a hidden risk: an AI agent can be confidently wrong, and it will be wrong at scale. Unlike a human employee who might show uncertainty or ask for help when unsure, a misinformed AI will happily deliver answers with unwavering confidence. This is how misinformation or errors can proliferate rapidly once an AI system is deployed widely.
Why do these confident mistakes happen? Often, it’s because the AI lacks the right data or context to give a correct answer. The AI might have a fantastically complex language model brain, but if it’s not grounded in accurate, up-to-date information from your business systems, it’s essentially guessing – and sometimes guessing wrong.
Michael Ham, a director at Los Alamos National Lab, captured this risk succinctly when discussing advanced “agentic AI.” “We’re using systems that can make a decision ... but if the data underneath isn’t trustworthy, they’ll make confident mistakes at scale,” he warned.
In fact, Ham emphasized that even as AI algorithms improve, their reliability “still depends on the integrity of human-curated data”, noting “we can’t expect to have effective AI if our data isn’t well-managed”.
For business decision-makers, the takeaway is clear: garbage in, garbage out, but with the twist that AI will present the “garbage” with a persuasive sheen of authority. Picture a voice assistant for an insurance firm confidently quoting an outdated policy premium, or an AI sales agent in retail cheerfully recommending a product that’s out of stock. These aren’t just minor glitches; they can damage customer relationships and brand credibility. The root cause in such cases is usually not a flawed model – it’s flawed data, or fragmented data, or unseen data silos that leave the AI flying blind. Before blaming the AI for messing up, ask: was the AI even given the proper, clean data and context to begin with?
Why 85% of AI Initiatives Fail: It’s the Data, Not the Model
This data problem isn’t just anecdotal – it shows up consistently in industry research as the top reason AI projects fail. Gartner and other analysts have observed shockingly high failure rates for AI initiatives, and they point to data quality and data silo issues as primary culprits. One recent study found that only 12% of organizations feel their data is adequately prepared for AI – and a staggering 85% of AI projects ultimately fail due to poor data quality, weak governance, or misalignment between business and IT.
In other words, the vast majority of AI efforts aren’t faltering because the algorithms can’t do the job, but because the data foundation is broken.
Think about that: if 85% of AI projects fail, it means companies are pouring time and money into AI solutions that never deliver real value. Why? Often because “bad data, not bad algorithms, kills AI projects.” As one data strategist put it, 85% of AI failures are strategic, not technical – the real problem is fragmented, biased data.
Many enterprises have data scattered across siloed systems, in incompatible formats, with inconsistent definitions and little oversight. Feeding that into even the most advanced AI is like putting low-octane fuel in a race car – it’s not going to perform, and it might even break down.
Gartner analysts explicitly urge leaders to shift focus from model-centric to data-centric strategies. “AI doesn't fail in the model. It fails in the data,” one report summarizes. Between 70% and 85% of AI failures stem from poor data foundations – things like incomplete or dirty datasets, lack of real-time data integration, and fragmented, ungoverned data that produces untrustworthy outputs. The message is loud and clear: if your AI initiative is floundering, chances are your data pipeline is the real issue.
It’s worth noting this isn’t about blaming IT alone; it’s a business leadership issue. Data governance and alignment between IT and business teams are frequently mentioned as keys to success (or failure). When projects are “misaligned between business and IT,” as the study above noted, it often means the AI was developed in a vacuum without connecting to the real operational data or understanding the business context. The result? An AI that might work in the lab but fails in the field. As the saying goes, “Your data strategy is your AI strategy” – ignoring that truth is why so many well-intentioned AI projects never make it past pilot stage.
Clean, Connected Data: The Bedrock of Reliable AI Workflows
For AI agents to truly add value in day-to-day business workflows, they must be grounded in clean, connected, and timely data. Consider a few scenarios across sectors:
- Lead Qualification & CRM: A conversational AI qualifying leads needs instant access to your CRM to know if a prospect is already a customer or has an open deal. If the data pipeline is broken, the AI might ask redundant questions ("What's your email?") to a VIP client who just emailed you yesterday. A connected pipeline ensures the AI recognizes the contact, pulls their history, and skips straight to value, rather than annoying them with basic data entry.
- Service Quoting & Pricing: An AI agent generating quotes for a service business (like HVAC or cleaning) must be connected to your real-time pricing logic. If it relies on a static training document from last year, it will quote 2024 prices in 2025, leading to lost margin or awkward "sorry, that was wrong" follow-ups. A robust pipeline calls your pricing engine in real-time, accounting for current rates, surcharges, and active promotions.
- Booking & Scheduling: An appointment-setting bot is useless if it can't see your actual calendar. It needs a live data link to your scheduling software to offer true availability. Without this, you risk the "double-booking disaster" where the AI books a meeting at 2 PM because it didn't know you manually blocked that time for a dentist appointment five minutes ago. Real-time data integration prevents these calendar collisions.
- Order Taking & Fulfillment: For an AI taking orders, it needs to know more than just the product list; it needs inventory and fulfillment status. If a customer asks "Can I get this by Friday?", the AI needs to check stock levels and shipping cutoff times instantly. A well-engineered pipeline connects the chat interface to your inventory system, so the AI never promises a delivery date you can't keep.
Across all these examples, the pattern is the same: without unified and trusted data, an AI agent is flying blind. No amount of “clever” natural language generation will save an agent that lacks the facts. This is why forward-looking organizations are doubling down on data infrastructure now. Gartner explicitly predicts that through 2026, 60% of AI projects will be abandoned if they don’t have “AI-ready” data supporting them. In other words, companies are realizing that unless they fix data integration and quality up front, their AI investments won’t pay off.
It means breaking down data silos and creating pipelines that consolidate information from various sources into a coherent, governed whole. It means instituting data quality checks, so your AI isn’t learning from or acting on erroneous information. It also means tracking data lineage and context – so the AI knows, for example, that this price came from that approved pricing database as of today, or that this customer record is the most recent across all departments.
Essentially, it’s treating enterprise data with the same rigor you’d treat financial reporting: no one would tolerate a CFO making decisions on 6-month-old partial ledgers; similarly, your AI agent shouldn’t be making decisions on stale or siloed data.
Salesforce’s $8 Billion Bet on Data Infrastructure Over Hype
If you want evidence that the industry is recognizing the primacy of data infrastructure, look no further than Salesforce’s recent moves. In May 2025, Salesforce – a company at the heart of the AI-for-business wave – announced it will acquire Informatica, a leading data management and integration firm, for a whopping $8 billion.
Salesforce isn’t buying Informatica for its machine learning algorithms; it’s buying it to bolster the data plumbing beneath its AI features. As Salesforce’s CEO Marc Benioff put it, “Together, Salesforce and Informatica will create the most complete, agent-ready data platform in the industry,” aiming to “enable autonomous agents to deliver smarter, safer, and more scalable outcomes for every company”.
Think about that message: agent-ready data platform. Salesforce explicitly talks about a “unified architecture for agentic AI” – integrating Informatica’s data catalog, integration, governance, and metadata management tools into the Salesforce ecosystem. The goal isn’t just to own another product; it’s to ensure that the next generation of AI agents (Salesforce even has an initiative called Agentforce) are running on a trusted, unified data foundation. In Salesforce’s words, this combined data platform will help enterprises “deploy and scale smarter, safer AI agents” across functions like shopping and customer service. In plain terms, Salesforce is validating what many in the industry have been saying: better data infrastructure leads to better AI outcomes.
This shift is notable because for the last few years much of the AI buzz was around model-centric investments – bigger models, more parameters, fancy prompt engineering tricks. But now we have one of the world’s top software companies spending billions not on a new AI model, but on data integration technology. It signals a recognition that reliable AI at scale requires heavy lifting in the data layer. As a business leader or CTO, it’s a cue to re-examine your own investments: are you disproportionately focused on AI algorithms while neglecting the data pipelines that feed them? The Salesforce-Informatica deal is a loud reminder that robust data plumbing is now seen as a competitive advantage. It’s like the gold rush realization that selling the shovels (or in this case, the data pipelines) is as valuable as finding the gold. The companies that win in AI will likely be those that invested in making their data clean, connected, and accessible to their AI systems, not just those who experimented with the flashiest chatbot.
Case Study: Clean Data Turns a Chatbot into a $3,000 Sales Closer
To see the power of robust data pipelines in action, consider a practical example. At Real Wave, we helped one of our clients – the third-largest cleaning services company in Canada – transform a basic chatbot into a revenue-generating Shopping Agent. The results were striking: this AI agent now handles over 470 different service variations with zero hallucinations, and converts what used to be simple $300 cleaning leads into $3,000+ deals through 24/7 automated quoting. In other words, by giving the bot the right data and context, it went from just answering questions to actually closing sales.
How did they achieve this reliability and performance? The secret was in engineering a structured data and workflow pipeline behind the AI – what we call “context engineering.” Instead of relying on one big, mysterious prompt or hoping a generic model would somehow know everything, we treated the AI like a well-coached team member in a defined process. The Shopping Agent was built with a clear playbook and a suite of data retrieval tools that ensured it only spoke with accurate, up-to-date information. Specifically, our framework included:
- A clear script and instructions: We didn’t just tell the AI “do your best.” We gave it a rundown of its role and rules – akin to a one-page brief for a news anchor. This script outlined exactly what the AI’s job was (e.g., convert leads to booked services, with certain upsell options) and what it should or shouldn’t do (never guess pricing, never go off approved info, etc.). This immediately set guardrails that prevented the agent from wandering off into creative but incorrect territory.
- Specialized data-retrieval tools integrated into the AI’s workflow: The Shopping Agent wasn’t left to guess prices or service details from memory. It had a “tool team” of functions it could invoke to get facts on demand. For example, if a customer asked for a quote, the AI would call a GetServicePricing tool to fetch the exact price for that service combination from the pricing database. It had a CalculateDiscount tool to apply any relevant promotions, a CheckAvailability tool to see open slots in the schedule, a BuildCart tool to compile the order, and even a CreateOrder function to finalize the booking. Each tool was essentially a direct pipeline into a backend system – be it pricing, scheduling, or CRM – so the AI’s responses were grounded in the same data a human employee would use. By engineering these pipeline connectors, we ensured the agent never had to “make up” an answer about, say, pricing or availability; it always pulled the truth from the source in real time.
- Just-in-time content curation: Rather than dumping the entirety of the company’s service catalog or policy manual into the AI’s prompt (which could overload it and lead to confusion), the system was designed to feed the AI only the relevant snippets of information at the right moment. This is like how a good assistant producer will hand a news anchor exactly the right document or quote needed for that segment, not a 500-page encyclopedia. In the Shopping Agent’s case, if a customer asked about a specific service variation, the AI would retrieve just that service’s details and present an answer based on it. This focused context meant the AI’s responses stayed accurate and on-topic, and it didn’t hallucinate details from unrelated data.
- Continuous summarization and context management: The agent maintained a rolling summary of the conversation and key facts (like customer preferences gathered so far), allowing it to handle long back-and-forth dialogues without forgetting earlier details or contradicting itself. This is a bit like a customer service rep taking notes during a call – the AI had an internal memory of important points, which was kept up-to-date as the conversation progressed. By structuring this memory, the agent could handle complex, multi-step customer interactions reliably.
The end effect of this engineered pipeline was an AI agent that behaved less like a stochastic parrot and more like a well-trained salesperson with all the right data at their fingertips. It quoted accurate prices, recommended additional services appropriately, and knew when to hand off to a human (e.g., for edge cases or final confirmation) with a complete context transcript. Notably, we saw the system achieve zero hallucinations across those 470+ service scenarios. That’s a powerful benchmark – it means the agent never invented a fake service or a wrong price, a feat only possible because every response was backed by a deterministic retrieval from a trusted data source. When you engineer the data pipeline correctly, the AI agent earns the right to be trusted by users.
Context Engineering: Why Your AI Agent Needs a Production Team
What we did with the Shopping Agent exemplifies a larger concept we champion, called “context engineering.” The idea is to treat an AI agent not as a whimsical oracle, but as a production system that you actively orchestrate and support with data.
I often use the metaphor of a news anchor: your AI is the talent (like Anderson Cooper), but it needs a production team behind it—producers to fetch facts, researchers to check data, and a script to set boundaries. Without this support, even the best talent can make confident mistakes.
For a complete breakdown of this framework—including the "Anchor, Producer, Script" model—read our full guide on Context Engineering: Why Your AI Needs a Production Team.
By feeding the AI only what it needs when it needs it (a "curated feed"), you reduce the chance of error and keep the AI efficient. This is a disciplined approach to providing context, analogous to how a chef prepares just the right ingredients for a dish rather than handing the cook the whole pantry.
The context engineering approach illustrates how structured and governed data flows enable AI agents to act intelligently even in complex, fragmented enterprise environments. Rather than hoping an AI can magically learn everything about your business (and never forget or hallucinate), you engineer the context around it so that it’s always supported by the right data at the right time. This turns the AI from a probabilistic guesser into a deterministic doer for many tasks. It’s a shift from the mindset of “let’s just prompt ChatGPT and see what happens” to “let’s design an end-to-end system where AI, data sources, and business logic all work in concert.”
For non-technical executives, the key point is this: successful AI agents are a whole-team effort. The model (the “AI brain”) is just one player. Without the data engineers, context providers, and rule-setting that surround it, that brain can quickly go off the rails. But with a robust supporting cast, the AI can reliably augment your workforce – answering customer questions with factual accuracy, retrieving client data on command, generating quotes that match what your billing system will charge, and so on, all without skipping a beat. It becomes a true autonomous agent you can trust, precisely because you’ve built a safety net of data and context under it.
The Bottom Line: Data is Your Competitive Edge in the AI Era
No matter what industry you’re in, the pattern is the same: those who harness their data effectively will unlock the real business value of AI, and those who don’t will be left with pilot projects and “AI theater” that never scales. We are entering an era where data infrastructure is the differentiator. It’s not that model innovation is over – far from it – but the low-hanging fruit of applying existing AI models to business workflows will only bear fruit if the underlying data is reliable. As one Gartner analysis bluntly stated, “Stop asking ‘What’s the best AI model?’ Start asking ‘Why is our data destroying every AI initiative?’” It’s a provocation for executives to reframe their thinking: your AI strategy is fundamentally a data strategy.
The evidence is all around us. Most AI projects fail not in the lab, but in production, when they hit the messy reality of enterprise data. On the flip side, the AI projects that succeed tend to be those where companies invested early in data readiness – integrating sources, cleaning data, setting up robust pipelines, and continuously monitoring data quality. These investments may not be as headline-grabbing as a new AI model announcement, but they pay off in AI that actually works day in and day out. They prevent those nightmare scenarios of an AI confidently misleading thousands of customers or making biased decisions because it was trained on skewed, siloed data.
Forward-looking businesses are already pivoting to this data-centric approach. Salesforce’s acquisition of Informatica is one high-profile validation, but even smaller companies are taking note. We see companies hiring Chief Data Officers, standing up data governance councils, and rebuilding data architectures with AI in mind. The mindset is shifting from “let’s experiment with GPT-variant X” to “let’s get our data house in order so we can deploy AI safely at scale.” The latter might not sound as exciting at first, but it’s what separates AI pilot purgatory from real AI-powered transformation.
In conclusion, making conversational and autonomous AI agents reliable in real business workflows comes down to treating data as a first-class citizen. Ensure your agent has access to all the relevant information (and nothing extraneous) it needs. Make that information consistent and governed, so the AI isn’t thrown off by conflicting sources or outdated entries. Keep the data pipelines flowing in real time, so your AI’s knowledge stays current. And build the surrounding context – instructions, tools, fail-safes – so that the AI’s “brain” operates in an environment of truth and clarity.
By doing so, you turn AI from a risky bet into a strategic asset. You avoid the “confident mistake” trap and instead get AI agents that you, your team, and your customers can trust. The reward isn’t just avoiding failure – it’s proactive advantage. Imagine being able to deploy an army of virtual agents that can handle customer requests, internal queries, or analytical tasks with precision and consistency. That’s a game-changer for efficiency and service quality. But it only happens when you build on a rock-solid data foundation. In the end, the companies that win with AI will not necessarily be those with the fanciest algorithms, but those with the cleanest, most connected data. After all, in business as in life, you don’t want to be making decisions (or having AI make decisions) on faulty information. As we move into this next chapter of AI adoption, it’s the behind-the-scenes data work – the unsexy stuff, the pipes and plumbing – that will determine who thrives and who trips up. The message to leaders is clear: invest in your data infrastructure now, and your AI agents will not only be impressive – they’ll be reliable, and reliability at scale is the real revolution.
Back to Real Wave Blog
