Welcome to storyz

Our blog, storyz, is where we write about operations, strategy and technology that leverage data for impactful marketing results. Join us as we navigate the digital landscape and unlock business opportunities.

dirty data impacts ai chatbots

Operations

Written by: Joseph Chapman

Published on: December 1, 2025

Internal AI Chatbots have a Dirty Data Problem

Organizations seeking to deploy internal AI chatbots to support business process automation face a rather sobering reality. The impressive capabilities of large language models can’t compensate for decades of accumulated data problems. While leaders at organizations marvel at controlled demos of AI chatbots answering employee questions in natural language, IT and data teams are quietly panicking about what happens when these systems start training on actual corporate data.

The promise is compelling. An AI chatbot could save thousands of employee hours instantly answering HR policy questions, summarizing customer histories or explaining complex procedures. Beneath the surface, though, a troubling question lurks. What happens when you point AI at legacy systems filled with duplicate records, improperly formatted fields, contradictory data, missing information or information that hasn’t been validated since 2015?

These issues and more cause what constitutes dirty data. For more than half a decade, I worked for Merkle (Dentsu), a prominent database marketing agency. As a program manager in the database technology group at Merkle, I led teams that built and maintained large marketing data warehouses for insurance, pharmaceutical and telecommunications companies. I have seen firsthand how dirty data interrupts warehouse updates, produces inaccurate intelligence reports or generates flawed campaign target lists.

The Amplification Effect

Traditional business tools have always struggled when confronting poor data quality, but in predictable ways. ETL processes, choking on dirty data, might come to a halt with an error message, prompting you to fix the problem. A report with bad data might show incorrect numbers, but the errors are contained within that specific investigation. AI chatbots, however, don’t merely load, process and report information. They synthesize, combine, and present it confidently using natural-sounding prose.

Consider a healthcare organization with patient records scattered across multiple systems. One database shows a patient’s primary physician as Dr. Smith. Another lists Dr. Johnson. A third contains an outdated entry for Dr. Williams, who retired three years ago. When an employee asks the AI chatbot “Who is the primary care physician for patient X?”, the system doesn’t simply throw an error. It might synthesize these conflicting records into a plausible-sounding but entirely wrong answer, delivered with the authority that makes AI responses so persuasive. This amplification effect means data quality issues that were merely annoying in traditional systems become genuinely dangerous when filtered through AI. The AI chatbot doesn’t just return false or incomplete data, it tries to work it into a convincing story.

The Compliance Nightmare

For organizations in regulated industries, dirty data intersects catastrophically with AI deployment and business process automation. Healthcare organizations must contend with HIPAA. Financial services face SEC regulations. Government contractors navigate complex security clearance requirements.

Training data for internal AI chatbots may include emails, documents, customer records, proprietary research and internal communications, precisely the places where sensitive information tends to hide. An employee might have inadvertently included Social Security numbers in a spreadsheet. Another might have copied patient details into a Slack channel for a quick question. Someone else embedded proprietary client information in presentation notes.

When this information becomes part of the training material, the AI doesn’t understand context or sensitivity. It simply learns patterns. Ask it a tangentially related question, and it might helpfully surface that sensitive information to any employee with access to the chatbot, potentially violating regulations that require strict access controls.

The problem compounds because dirty data could include information that should have been deleted. GDPR’s right to be forgotten, retention policies, and legal holds all create situations where specific data must be purged. But if that data has already been incorporated into an AI model’s training, deletion becomes more complex. Some organizations are discovering they need to retrain entire models because they can’t verify whether deleted information influenced the model’s responses.

The Trust Collapse

There is also a human problem. Employees quickly develop a sense for whether a system is reliable. Once they catch an AI chatbot providing incorrect information a few times, trust deteriorates. This creates a vicious cycle. Employees stop using the chatbot, so the organization can’t collect feedback to improve it. Without usage data, it’s harder to identify which responses are problematic. The expensive AI initiative becomes shelfware, and employees return to the informal networks and manual processes the AI chatbot was supposed to replace. According to a survey by S&P Global Market Intelligence⧉, the share of companies scrapping most of their AI initiatives “surged from 17% to 42%” between 2024 and 2025. To be fair, exactly how much dirty data issues contributed to the abandonment was not discussed. Based on research performed for this article, we speculate it had a significant role.

Compounding the problem, the bar for AI accuracy is higher than that for human-provided information. If an employee asks a colleague a question and gets a wrong answer, he mentally categorizes it as “John didn’t know.” It doesn’t mean the employee stops asking colleagues questions. But when an AI chatbot provides incorrect information, it quickly undermines confidence in the entire system. The technology is evaluated differently, fairly or not.

Practical Paths Forward

Organizations successfully deploying internal AI chatbots as a part of a larger business process automation initiative are adopting several strategies to manage data quality challenges. Many are pivoting from fine-tuning models on their entire data ecosystem to using retrieval-augmented generation (RAG) . Instead of training the AI on everything, they maintain curated knowledge bases that the AI queries in real-time. This allows for easier updates and gives organizations more control over what information the AI chatbot can access. Others are implementing strict data governance before deployment. This means auditing databases, standardizing formats, deduplicating records, and establishing clear ownership for different data domains. It’s not glamorous work, but it’s the foundation that helps makes AI more accurate and reliable.

Progressive rollouts also help manage risk. Instead of giving all employees access to a chatbot trained on all company data, organizations start with narrow use cases and clean datasets. An AI assistant that only answers questions about PTO policies based on the latest official HR handbook is much easier to validate than one trained on piles of company documents.

Keeping people in the loop provides a safety net. Critical responses get reviewed by subject matter experts in testing scenarios before the AI goes live. Users can flag incorrect information, creating a feedback loop that identifies data quality problems. Some organizations even use AI chatbots to identify potentially problematic responses for human review before they reach end users.

Conclusion

The dirty data problem forces organizations to confront an uncomfortable truth. Many have tolerated poor data quality for years because the consequences were manageable. Reports might have been slightly off or business processes might have required varying degrees of manual reconciliation. AI, however, changes the equation. It doesn’t tolerate dirty data, it amplifies it. This means organizations can no longer kick the data quality can down the road while simultaneously pursuing AI initiatives. The two are fundamentally incompatible.

This creates a choice. Organizations can invest in the tedious, time-consuming work of data cleanup and governance. Or they can watch their AI initiatives deliver unreliable results, erode employee trust, create compliance nightmares or potentially be cancelled. The most successful deployments recognize that AI isn’t primarily a technology challenge, it’s also a data challenge. The sophistication of the model matters far less than the quality of the data you feed it. In the age of AI, garbage in, garbage out isn’t just a warning. It’s a guarantee.

Contact us or learn about agentryz™ if you need help with your internal AI chatbot initiative.

Related reading