Menu

  1. Mar 1, 2026

    The Audit Trail Problem: Why Investment Committees Don't Trust AI

Investment committees exist to say no. Their function is to challenge assumptions, probe weaknesses, and reject deals that do not meet the bar. They are professionally skeptical, and that skepticism has kept firms out of bad investments for decades.

This skepticism now extends to AI. When an analyst presents a deal with "AI-assisted underwriting," the IC's first question is not about the property. It is about the data. Where did that rent number come from? How do we know the lease extraction is accurate? What happens if the AI made a mistake?

These are reasonable questions. They reflect a fundamental truth about investment decisions: the quality of a decision depends entirely on the quality of the information that informs it. An IC that cannot verify the information cannot trust the decision.

Most AI implementations in CRE fail this test. They produce outputs without explaining their origins. They generate summaries without citing sources. They create an efficiency gain while introducing an audit gap that sophisticated investors refuse to accept.

What Investment Committees Actually Need

To understand the audit trail problem, you must understand what ICs require to approve a deal. The requirements go far beyond a recommendation.

IC Requirement

What It Means

Why It Matters

Data provenance

Every number traceable to a source document

Enables verification; catches errors

Assumption transparency

Every assumption explicitly stated with rationale

Enables challenge; reveals optimism

Variance documentation

Every conflict between sources identified and resolved

Prevents garbage-in-garbage-out

Sensitivity visibility

Impact of changing key assumptions quantified

Reveals risk concentration

Comparable support

Market assumptions backed by relevant transactions

Validates pricing and exit assumptions

Risk identification

Material risks explicitly catalogued

Enables informed acceptance

An IC memo that lacks any of these elements will face questions. An IC memo where these elements cannot be verified will face rejection.

The standard is not "this looks right." The standard is "we can prove this is right, and if it turns out to be wrong, we can identify exactly where the error originated."

Why Current AI Fails the Test

Most AI implementations in CRE were designed for speed, not auditability. They optimize for producing outputs quickly, not for proving those outputs are correct.

The Black Box Problem

When an AI system extracts a rent figure from a lease and populates it into a model, the IC needs to verify that extraction. In most implementations, this verification is impossible or impractical.

Verification Question

What IC Needs

What Most AI Provides

What document did this come from?

Specific document name and page

Often nothing; sometimes document name only

Where in the document?

Page number, section, or visual highlight

Rarely available

What was the exact source text?

The original language that was interpreted

Usually absent

How confident is the extraction?

Probability or confidence score

Rarely exposed to users

Were there other possible interpretations?

Alternative readings that were rejected

Never shown

Without answers to these questions, the IC cannot verify the extraction. They must either trust the AI blindly (which they will not do) or re-read the source documents themselves (which eliminates the efficiency gain).

The Hallucination Risk

AI systems can generate plausible-sounding information that is factually incorrect. In a general context, this is an annoyance. In an investment context, it is a potential fiduciary breach.

ICs are acutely aware of this risk. When they see AI-generated content, they wonder: is this real, or did the system fabricate it? Without a mechanism to verify every claim against source documents, they cannot distinguish accurate extraction from confident hallucination.

Hallucination Type

Example

IC Concern

Invented data

AI states lease expires in 2028 when document shows 2027

Direct error in underwriting

Misattributed data

AI assigns Tenant A's rent to Tenant B

Tenant-level analysis corrupted

Inferred data

AI assumes renewal option exists when lease is silent

Overstated optionality

Aggregation error

AI sums rent incorrectly due to unit confusion

NOI calculation wrong

Context confusion

AI pulls data from superseded document version

Stale information in model

A single hallucination that reaches an IC memo and is later discovered destroys confidence in the entire system. ICs will revert to manual processes rather than risk undetectable errors.

The Reconciliation Gap

When documents conflict (as they routinely do), someone must determine which source is correct. In a manual process, the analyst makes this judgment and documents their reasoning. In most AI implementations, the system either picks a value silently or surfaces the conflict without context.

Reconciliation Failure

What Happens

IC Impact

Silent resolution

AI picks one value without noting alternatives existed

IC unaware of uncertainty

Unresolved conflict

AI surfaces conflict but provides no resolution

IC must investigate themselves

Missing context

AI notes conflict but not which source is authoritative

IC cannot evaluate resolution

No documentation

Resolution made but rationale not captured

Future questions unanswerable

ICs need to know: what conflicts existed, how they were resolved, and why that resolution is correct. Most AI systems provide none of this.

What a Proper Audit Trail Looks Like

An AI system that ICs can trust produces outputs with complete provenance. Every data point connects to its source. Every conflict is documented with its resolution. Every extraction includes a confidence assessment.

Source Citation

Every extracted value must link to its origin with sufficient specificity to enable verification in seconds, not minutes.

Citation Element

Purpose

Example

Document name

Identifies source file

"Rent Roll - January 2025.pdf"

Page number

Locates within document

"Page 3"

Location on page

Pinpoints exact position

"Row 15, Column D" or coordinates

Source text

Shows original language

"Base Rent: $25.00 PSF NNN"

Extraction timestamp

Establishes when processed

"2025-01-15 14:32:07 UTC"

Document version

Distinguishes from superseded versions

"Version received 2025-01-10"

With this citation structure, an IC member can click on any number in a memo and immediately see exactly where it came from. Verification takes seconds. Trust becomes possible.

Confidence Scoring

Not all extractions are equally reliable. A clearly formatted rent roll yields high-confidence extractions. A scanned, handwritten amendment yields low-confidence extractions. The system must distinguish these and communicate the distinction.

Confidence Level

Meaning

IC Implication

High (>95%)

Extraction almost certainly correct; standard format, clear text

Can rely on value; spot-check sufficient

Medium (80-95%)

Extraction likely correct but some ambiguity exists

Should verify if material to decision

Low (<80%)

Extraction uncertain; non-standard format or unclear source

Must verify before relying on value

Manual override

Human corrected AI extraction

Extraction verified by human review

Confidence scores enable risk-based verification. ICs can focus their limited time on low-confidence extractions that are material to the deal, rather than re-verifying everything or trusting everything.

Variance Register

Every conflict between documents must be captured in a structured register that documents what conflicted, how it was resolved, and who made the resolution decision.

Register Field

Content

Purpose

Field name

What data point is in conflict

Identifies the issue

Source A value

First document's value with citation

Shows one side of conflict

Source B value

Second document's value with citation

Shows other side of conflict

Variance magnitude

Size of discrepancy (absolute and percentage)

Indicates materiality

Resolution

Which value was used and why

Documents the decision

Resolver

Who made the resolution decision (AI or human)

Establishes accountability

Resolution date

When resolution was made

Creates timeline

The variance register serves two purposes. First, it ensures conflicts are addressed rather than ignored. Second, it creates a record that can be reviewed if questions arise later.

Assumption Tracing

Beyond extracted data, ICs need to trace assumptions. When the model shows 3% rent growth, where did that assumption come from? Is it a market benchmark, a historical average, or an analyst judgment?

Assumption Source

What It Means

How to Document

Extracted from documents

Assumption reflects contractual terms

Cite source document

Historical analysis

Assumption reflects past performance

Cite calculation methodology and data

Market benchmark

Assumption reflects market data

Cite data source and date

Comparable transactions

Assumption reflects similar deals

Cite specific comparables

Analyst judgment

Assumption reflects professional opinion

Note as judgment; require rationale

Tracing assumptions prevents the common failure where a favorable assumption enters a model without scrutiny. If every assumption must be sourced, unsupportable optimism becomes visible.

Building Systems ICs Will Trust

The audit trail is not a feature added to an AI system. It is an architectural requirement that must be present from the foundation.

Extraction Architecture

The extraction process must capture provenance at the moment of extraction, not reconstruct it later.

Architectural Requirement

Implementation

Source binding

Every extracted value stored with document reference, page, and location

Original text preservation

Source text captured alongside interpreted value

Confidence calculation

Extraction confidence computed and stored with value

Version tracking

Document version recorded; extractions linked to specific version

Immutable logging

Extraction events logged permanently; cannot be silently modified

When extraction is architected this way, the audit trail exists automatically. It does not require additional effort to produce because provenance is inherent in the data structure.

Reconciliation Architecture

Conflict detection and resolution must be systematic, not ad hoc.

Architectural Requirement

Implementation

Automatic comparison

System compares same fields across documents without human initiation

Materiality assessment

System evaluates whether variance is significant based on magnitude and field importance

Resolution workflow

Conflicts routed to appropriate resolver (automated rules vs. human judgment)

Resolution capture

Every resolution documented with rationale

Propagation tracking

When resolution changes a value, all downstream uses updated and logged

Systematic reconciliation ensures nothing falls through the cracks. Every material conflict surfaces. Every resolution is documented.

Reporting Architecture

The IC memo or underwriting package must present information with embedded verification capability.

Architectural Requirement

Implementation

Interactive citations

Every data point clickable to view source

Confidence visualization

Confidence levels visible in presentation

Variance summary

Conflicts and resolutions summarized in memo

Assumption registry

All assumptions listed with sources

Change tracking

What changed since last version and why

When reporting is architected this way, IC members can drill into any claim without leaving the memo. Verification is frictionless rather than effortful.

The Organizational Shift

Technology alone does not solve the audit trail problem. Organizations must also change how they think about AI-assisted work.

Quality Over Speed

Many firms implemented AI to go faster. ICs do not care about faster. They care about more reliable. An AI system that produces unverifiable outputs quickly is worse than a manual process that produces verifiable outputs slowly, because the IC will reject the AI outputs and demand manual verification anyway.

The value proposition must shift from "AI makes us faster" to "AI makes us more thorough while also being faster." The audit trail is how thoroughness is demonstrated.

Human-AI Collaboration

The audit trail clarifies the respective roles of humans and AI. AI extracts and cites. Humans verify and judge. AI detects conflicts. Humans resolve them. AI populates models. Humans set assumptions.

Task

AI Role

Human Role

Document processing

Extract with provenance

Verify low-confidence extractions

Conflict detection

Identify and quantify

Determine resolution

Model population

Populate from extractions

Validate and adjust

Assumption setting

Provide benchmarks and history

Make judgment calls

IC memo generation

Assemble data with citations

Write narrative; frame recommendation

This division makes accountability clear. When something is wrong, the audit trail shows whether it was an extraction error (AI) or a judgment error (human).

Continuous Improvement

The audit trail enables systematic improvement because errors can be traced to their source. When an IC catches a mistake, the organization can determine: was it a document classification error, an extraction error, a reconciliation error, or a judgment error? Each type has different fixes.

Error Type

How Identified

Improvement Path

Classification

Document mistyped; wrong schema applied

Improve classifier training

Extraction

Value incorrect given document content

Improve extraction model for document type

Reconciliation

Wrong value selected from conflicting sources

Refine resolution rules

Judgment

Human made poor assumption or decision

Training or process change

Without an audit trail, errors are discovered but not diagnosed. The same mistakes repeat. With an audit trail, every error becomes a learning opportunity.

Conclusion

Investment committees do not trust AI because AI has not given them reason to trust it. Systems that produce outputs without provenance, reconcile conflicts without documentation, and generate content without verifiable sources fail the basic requirements of investment decision-making. The solution is architectural: AI systems must be built from the ground up to capture provenance, document conflicts, and enable verification. When every number traces to a source, every conflict is logged with its resolution, and every assumption is explicitly sourced, ICs can verify rather than trust blindly. The audit trail transforms AI from a black box that committees reject into an infrastructure that makes their verification work easier and more thorough. That is how AI earns trust: not by asking for it, but by making verification trivially easy.

Request a Free Trial

See how Eagle Eye brings clarity, accuracy, and trust to deal documents.

Request a Free Trial

See how Eagle Eye brings clarity, accuracy, and trust to deal documents.