Skip to main content

Why AI Agents Go Rogue: The Data Constitution Problem

Why AI agents go rogue: data chaos. The data constitution problem breaks agentic governance. How to fix it before agents scale and governance collapses.

Jake MorrisonFeb 2, 20265 min read

Your AI agents aren't malfunctioning—they're doing exactly what you told them to do. The problem is what you told them.

As enterprises move from experimenting with generative AI to deploying autonomous agents, a new problem is emerging: agents making catastrophic decisions because the data they're trained on is garbage.

A loan approval agent denies loans to applicants in certain zip codes because historical data reflects discriminatory lending patterns. A customer support agent routes complaints to the wrong department because it was trained on mislabeled ticket data. A hiring agent screens out qualified candidates because the training set was biased toward a specific demographic.

These aren't bugs. They're symptoms of a missing layer: formal data governance.

The Core Problem: Data Is Untrusted, But We Treat It Like It Isn't

Here's what enterprises typically do:

  1. Collect data from production systems (CRM, HR systems, databases, logs)
  2. Feed it to an AI model
  3. Deploy the model as an autonomous agent
  4. Hope nothing goes wrong
  5. When something does, scramble to fix it

Here's what they should do:

  1. Assume all data is untrusted until proven otherwise
  2. Define a "data constitution"—a formal specification of what valid data looks like
  3. Implement a quality layer that validates and cleans data before it reaches the model
  4. Enforce this layer for every data source, every field, every record
  5. Monitor for violations and alert when bad data tries to pass through
  6. Only then train and deploy agents

What Is a Data Constitution?

A data constitution is a formal, machine-readable specification that defines:

  • Schema: What fields exist? What are their types? Are they required?
  • Business logic: What values are valid? What are the relationships between fields? If field A is "customer status = churned," what values must field B take?
  • Integrity constraints: Primary keys, foreign keys, uniqueness requirements
  • Semantic rules: If this is a date field, is it in the past? In the future? Within a specific range?
  • Fairness rules: Which fields might introduce bias? How do we detect discrimination in data patterns?

Example: A loan approval dataset might have a data constitution that says:

  • Loan amount must be between $1,000 and $500,000
  • Interest rate must be between 1% and 20%
  • Approval decision must be one of: {approved, denied, pending}
  • If approval is "denied," there must be a reason code from this specific set
  • Approval rates across zip codes must not deviate by more than 10 percentage points (fairness rule)
  • All applicants must have a credit score between 300 and 850

If data violates the constitution, it gets rejected or flagged for manual review.

Why This Matters: The Cost of Bad Data

Bad data in a batch report: Someone notices the report looks wrong, investigates, finds the bad data, fixes the report, maybe sends a corrected version. Cost: A few hours of analysis work.

Bad data in a deployed AI agent: The agent makes decisions for days, weeks, or months before anyone notices. Each bad decision affects a real person: a rejected loan, a missed customer service issue, a job interview that goes nowhere. If the agent is making discriminatory decisions, you have regulatory liability, PR liability, and legal liability.

The hidden cost: You need to audit every decision the agent made while using bad data. You might need to overturn decisions, send apologies, offer compensation. The operational cost is often 10x the initial data cleaning.

Building a Multi-Tenant Quality Layer

The architecture:

  • Data sources: CRM, HR system, database, logs, third-party APIs
  • Ingestion pipeline: Extract data from each source
  • Quality layer (the key): Validate against data constitution. Apply transformations. Detect anomalies. Flag violations.
  • Clean data warehouse: Only validated data flows into this layer
  • AI training/inference: Models train on clean data. Agents only see clean data.

What the quality layer actually does:

  1. Schema validation: Does each record have the required fields? Are the types correct?
  2. Value validation: Are the values within expected ranges? Are they from the allowed set?
  3. Business logic validation: Do the values satisfy business constraints?
  4. Outlier detection: Using statistical methods, find records that are unusual (might be corrupt, might be a new phenomenon)
  5. Bias detection: Using fairness metrics, find patterns that suggest discrimination or systematic issues
  6. Reconciliation: Cross-reference fields across sources to find inconsistencies

Invalid records get routed to:

  • A manual review queue (high-severity issues)
  • An automated correction engine (common, fixable issues)
  • A quarantine zone (unknown issues, human investigation required)

The Governance Philosophy: Assume Data Untrusted Until Proven Valid

This is the opposite of how most organizations work. Typically:

  • Data is assumed clean unless evidence suggests otherwise
  • Bad data is discovered during analysis or after a problem occurs
  • Data governance is reactive (we'll fix it when we find it)

For agentic AI, this is backwards. You need:

  • Data is assumed dirty. It must prove it's clean.
  • Bad data is caught at ingestion, before it can influence decisions
  • Data governance is proactive (we prevent bad data from ever reaching the model)

This requires investment upfront: defining the constitution, building validation rules, implementing the quality layer. But it prevents catastrophic failures down the line.

Real-World Example: Hiring Agent Gone Wrong

Scenario: A company deploys an AI agent to screen job applications. The model was trained on 10 years of historical hiring data. The agent rejects 40% of applications from one demographic group and 10% from another.

Root cause: The historical data reflected past discriminatory hiring practices. The agent learned the pattern.

With a data constitution: The quality layer would have detected that approval rates by demographic varied significantly. A fairness rule would have flagged this before training. The data would have been investigated and corrected (or historical biases would have been documented and the model would have been designed to avoid them).

Cost without constitution: Lawsuits, regulatory fines, reputational damage, mandatory re-interviews for all rejected candidates from protected classes.

Cost with constitution: A few days of investigation during quality layer setup, understanding the historical bias, consciously deciding how to address it.

How to Start: Three Steps

Step 1: Define Your Data Constitution

For each data source that will feed AI agents:

  • What fields must exist?
  • What are valid values for each field?
  • What relationships must hold between fields?
  • What fairness constraints apply?

Document this in a machine-readable format (JSON Schema, YAML, or your metadata tool of choice).

Step 2: Implement Validation

Build or adopt a data validation framework (Great Expectations, dbt, or custom):

  • Validate every incoming record
  • Log violations
  • Route invalid records to a handling process
  • Monitor validation metrics (% passing, % of each violation type)

Step 3: Continuous Monitoring

Even after data passes validation, monitor it:

  • Are fairness metrics still satisfied?
  • Are there new anomalies emerging?
  • Has the data distribution shifted?

Use statistical methods to detect drift and alert when something changes.

The Hard Truth

Agentic AI without data governance is like giving a driver's license to someone who's never taken a driving test. They might be fine—or they might crash and hurt people.

The enterprises that will thrive with agentic AI are the ones that invest in data governance first. They'll move slower initially, but they'll move with confidence. They'll know their agents aren't making catastrophic decisions based on garbage data.

Everyone else is gambling. Don't gamble with autonomous agents. Define your data constitution, build your quality layer, and assume all data is untrusted until it proves otherwise.

Your users—and your regulators—will thank you.

JM

Jake Morrison

Staff Writer

Writes weekly recaps and storylines across multiple beats. He brings a sharp eye for detail and a knack for finding the story behind the story.

You might also like