Blog/When the Output Looks Right but Isn't
·
AIFinanceGovernance

When the Output Looks Right but Isn't

Prabin Pandey · Fox School of Business, Temple University

Brief Context

Working through a series of applied AI assignments in a graduate finance program, the same tension kept surfacing. AI tools produce outputs that are fluent, structured, and internally consistent. In most fields, that's useful. In financial analysis, it introduces a specific kind of risk. Convincing is not the same as correct.

Central Question

The question isn't whether AI is useful in finance. It clearly is. The harder question is: at what point does a well-formed AI output become a liability rather than an asset?

Reasoning

The first observation came from understanding what AI models actually do. They predict the next word based on patterns in training data. They don't verify claims. So when a model states that a fund "is expected to underperform the benchmark," that sentence reads like an analytical conclusion. But it may have no supporting evidence. The model generated it because it fits the pattern — not because it's true.

This matters more in finance than in most fields. A DCF model built with AI-generated discount rates or growth assumptions can produce a complete, internally consistent output — and still be wrong. The numbers will hang together. The error won't be visible until something depends on it.

The second observation was about trust. Newer AI models don't just produce coherent responses — they mask incorrect information more convincingly than older models did. The output doesn't just sound right. It sounds authoritative. That's the actual problem.

Thinking through DCF workflows made this concrete. Some stages — discounting, formula links, equity value calculations — can be tested mechanically. A formula either references the right cell or it doesn't. The error is structural and identifiable.

But other stages — defining scope, designing scenarios, selecting benchmarks — involve judgment. No formula determines whether a growth assumption is appropriate for a specific business in a specific market. Only understanding the business does.

The implication is that AI works better as a structural tool than as an analytical one. It can build the scaffolding. It can write the code that computes the output. Humans still need to decide what goes inside the model and whether the output means what it appears to mean.

Key Observations

Narrative validation matters more than tonal analysis. In MD&A work, the most important check wasn't whether management sounded confident or cautious. It was whether the narrative was consistent with the audited financial statements. Management writes the MD&A. Auditors verify the numbers. Those are different processes with different incentives. When they diverge, the audited numbers take priority — always.

Overconfidence in simulation outputs is easy to miss. One annotated AI output claimed that tracking error "exhibits notable consistency across simulated paths." That sentence sounds like an observation. It isn't. Consistency across simulated paths says nothing about what happens in actual markets.

Temperature settings shape how reliable AI outputs are. At low temperature, the model produces focused, repeatable responses suited to tasks that require accuracy. At high temperature, the model explores more but varies more. For financial analysis where outputs will be used as inputs, low temperature is generally the safer default.

Practical Meaning

The governance frameworks that major financial institutions have built around AI aren't excessive caution. They're a practical response to a clear problem: AI cannot be accountable. Humans are. Requiring model registration, independent validation, and human sign-off before AI outputs are used in client-facing decisions puts accountability exactly where it needs to be.

For an individual analyst, the practical implication is simpler. Treat AI outputs as a first draft requiring review — not a conclusion requiring formatting. The review isn't about distrust. It's about recognizing where the model's confidence comes from and whether that confidence is earned.

Limits or Uncertainty

These observations came from academic assignments, not live production environments. Real financial workflows involve data quality issues, time pressure, team dynamics, and institutional constraints that change how AI tools are used in practice. What looks like a clear validation step in a structured exercise may be harder to apply consistently when a deadline is close and the output looks correct.

There's also an open question about where exactly AI-assisted analysis ends and AI-generated analysis begins. Whether that balance holds as AI tools become more capable and more embedded in standard workflows is genuinely unclear.

Closing Reflection

The most durable insight from working through these materials wasn't about what AI can do. It was about what makes AI specifically risky: not that it produces wrong outputs, but that it produces wrong outputs that look right. That's a different problem than accuracy. It's a credibility problem. Better models won't solve it. Better practices might.