Data Governance for AI Must Be Executable
Why AI models stall between proof of concept and production, and what technology leaders can do about it
BLOG
Rick Hamilton, Jaswant Singh, and Naresh Nayar
3/4/20266 min read


The Problem Isn’t the Model
A customer-facing AI agent confidently answers a benefits question. The answer is wrong because it retrieved a superseded policy document from an unversioned, access-uncontrolled corpus. The business now has three concurrent problems: customer harm, regulatory exposure, and an internal investigation that cannot reproduce the retrieval context that produced the response. No one can say which version of which document the model saw, because provenance was never captured. The investigation drags on for weeks; the root cause, ungoverned data, remains in place for the next incident. In AI-enabled enterprises, this is not an edge case. It is the predictable outcome of deploying AI on a data substrate built for reporting, not autonomous action. Models are only as trustworthy as the data they ingest; in most organizations, that data is neither traceable enough to explain nor governed well enough to defend.
In 2024, one of the authors was serving as CTO of a healthcare research organization when the team deployed its first RAG solution. We anticipated the usual technical challenges: problematic chunking strategies, improper ranking algorithms, nonfunctional requirements like system performance. The biggest problem, however, was none of these. It was out-of-date and contradictory data sources, resulting in the system misrepresenting current scientific thinking and organizational policy. The technical architecture worked, but the data substrate beneath it was ungoverned. Now amplify this lesson across an enterprise deploying autonomous agents that depend on data sources spanning dozens of systems and domains, whose immediate responses and downstream decisions may never be reviewed by a human. The problems we expected in 2024 were largely architectural. The problem that actually mattered was upstream, as ungoverned, conflicting data quietly degraded output quality as the AI became more relied upon and essential to our business.
In this context, data governance and modern data management together form the trust architecture for AI-driven analytics and automation. Governance defines decision rights, standards, and accountability for data meaning, quality, provenance, and access. These policies and standards do not enforce themselves; they must be translated into technical controls that operate within the data ecosystem. Modern data management enforces those standards as controls across pipelines, catalogs, and APIs, and ideally, it produces an evidence trail that makes outputs explainable and defensible. When governance and management are disconnected, AI scales faster than trust, and small data defects become systemic failures.
Governance structures and operational risk frameworks define who is accountable and what must be monitored. This paper addresses an important question previously raised: whether the data infrastructure beneath those frameworks is engineered to make accountability and monitoring possible? For related concepts, see AI Governance is Broken: Here’s How to Fix It and AI Risks Don’t Wait for Committees.
The central problem for most enterprises is not a lack of data, but rather is the lack of executable governance: policies that are enforced automatically at the point of data movement and model use, rather than documented in frameworks that nobody operationalizes. Until governance is implemented as enforceable controls within systems, AI will continue to scale faster than trust. In working with organizations across healthcare, financial services, and insurance, we have found three structural failures that recur with striking consistency, and they persist not because leaders are unaware of data governance, but because their governance programs are designed to produce documentation rather than controls.
Three Things Most Organizations Have Wrong
Each of these recurring problems is well-understood in isolation. What makes them dangerous in combination is that they compound. Inconsistent semantic definitions feed ungoverned pipelines, which feed ungoverned AI artifacts, and the resulting errors become progressively harder to trace back to their origin. Three in particular define the gap between governance programs that exist on paper and governance that operates in production.
→ Governance is not a committee. It is a build artifact.
If a policy is not enforced automatically at the point of data movement or model inference, it is documentation. Documentation does not prevent incidents. The question is not whether the organization has a governance program. It is whether that program produces controls or produces PDFs.
→ Semantic drift is the silent failure mode of enterprise AI.
The most damaging data quality failures are not nulls and duplicates, but rather, they are semantic. “Active customer” means something different in the CRM than in the data warehouse. “Net revenue” changes definition when accounting policy changes, with no version event recorded. The model trained on the old definition produces outputs that are internally consistent and operationally wrong. Teams debug model performance when they should be debugging meaning.
THE SEMANTIC DRIFT FAILURE IN PRACTICE
A risk model in production begins producing anomalous scores. The model team investigates: the algorithm is unchanged, the pipeline is running cleanly, there are no obvious data quality issues. After a few weeks of investigation, a data engineer discovers that the feature store’s definition of “active member” was updated three months earlier to reflect a new product line. The model was trained on the old definition. The model is not broken. The semantic contract between the feature store and the model was broken, silently, with no lineage event and no downstream notification. Weeks of engineering time spent diagnosing a data governance failure.
Preventing semantic drift requires both shared definitions and structural controls. Master Data Management (MDM) creates authoritative representations of key business entities across various domains. Data contracts establish the mutually agreed-upon expectations between data producers and consumers, addressing parameters such as schema, semantic definitions, quality standards, and refresh intervals. By integrating these controls into data pipelines and APIs, governance standards become actionable mechanisms that help minimize the risk of unnoticed drift affecting models.
→ RAG and agents have expanded the governance perimeter beyond tables and reports.
Production AI systems increasingly depend on retrieval corpora, prompt libraries, vector stores, and agent interaction logs—assets that exist outside the model but directly shape its outputs. These artifacts influence what the system sees, retrieves, and generates. In many enterprises, they live in ad hoc repositories with no version control, no access management, and no audit trail. Within this ungoverned perimeter, the next significant incident is incubating.
The exposure does not end at model behavior. APIs are the distribution boundary for governed data products—the point at which AI and analytics outputs cross domain or organizational boundaries. When data controls are not enforced at this boundary, sensitive data can leak, entitlements can drift, and quality inconsistencies can propagate at scale. APIs are not a parallel governance discipline; they are the enforcement surface for data governance in motion.
These gaps take on new significance once AI systems move beyond pilots. Three forces are converging to make these gaps untenable.
Why This Matters Now: Regulation, Competitive Velocity, and Production Scale
Executable data governance is no longer theoretical. Regulatory expectations are becoming more concrete, deployment cycles are shortening, and boards are asking how AI investments translate into operational results. These pressures are converging on the same requirement: governance that functions in practice, not just on paper.
→ The Window is Closing: Regulation is Arriving in Phases.
The EU AI Act applies in stages through 2026 and 2027, depending on system category and risk classification. While the Act addresses a broad set of system-level obligations—including risk management, documentation, and oversight—many of these requirements depend directly on data governance capabilities: documented training data provenance, monitoring for data and distributional drift, transparency of data sources, and the ability to demonstrate that controls are operating as designed. Organizations that cannot produce this evidence at audit will face deployment restrictions, not just fines.
For most global enterprises, EU jurisdiction is not optional. If AI systems touch EU data subjects, or if partners or customers operate under EU law, the obligations apply. The time to build the evidence infrastructure is before the audit, not during one. Organizations that invest early in lineage, version control, and control verification will move from reactive compliance to operational readiness.
→ Governance as Accelerant: Organizations with Executable Governance ship faster.
The competitive consequence of the governance gap is not that organizations build worse models. It is that they operate them more slowly, defend them less confidently, and spend more engineering time on incidents that governed pipelines would have prevented. Organizations that have built the trust layer (lineage tracking, quality controls embedded in pipelines, provenance for AI artifacts) report deployment cycles that are substantially faster, because sign-off is a verification step rather than an investigation. Leadership sees this in shorter and more predictable approval cycles.
The underlying driver is governance infrastructure. Mature governance and data management transform risk into strategic advantage across four dimensions: speed to trusted AI through governed pipelines with lineage and provenance that accelerate model deployment; risk containment through runtime enforcement that detects anomalies before they reach production; monetization enablement through API governance frameworks that allow safe data product exposure and new revenue streams; and operational scalability through AI-augmented governance—automated classification, quality monitoring, and policy enforcement—that grows with data and model volume without proportional headcount increases.
→ The POC-to-production gap is now a board-level question.
Enterprise AI investment has shifted from “should we invest” to “where is the return.” Organizations running AI as a portfolio of pilots are facing pressure to demonstrate that the capability can be operationalized at scale. When the honest answer is “our models work in the lab but cannot ship because legal cannot sign off on the data,” that answer has a short shelf life with a board or an investment committee. Governance is what makes operationalization possible. Without it, the answer to “why is this still in testing” eventually stops being acceptable.
These pressures ultimately reduce to a specific operational requirement: the ability to reproduce and explain system behavior.
Together, we explore this important topic more thoroughly in the full Substack article, including the Executable Governance Model, and a five-step playbook on how to begin your transformation. Read the full article here.


