Most coverage of artificial intelligence in finance focuses on what these tools can do. Less attention is paid to how they perform under scrutiny, particularly in financial modeling, where small errors can carry real consequences.
After testing Anthropic’s Claude in real-world modeling scenarios, one conclusion stands out: Claude produces outputs that look credible at first glance but contain structural flaws that only an experienced professional would catch.
That gap between appearance and reliability is where risk begins.
Where AI Performs Well
Claude handled several foundational elements of financial modeling competently. It was able to:
-
Build basic revenue models
-
Generate standard financial statements
-
Apply consistent formatting, labels and units
The outputs appeared polished and professional. In some cases, they resembled models produced by junior analysts. That is what makes them risky.
The models looked right. The structure appeared logical. Formatting signaled credibility. For a time-constrained professional, those cues can create trust before a full audit is completed.
Related:Good Vibes Only: How Financial Advisors Can Build Custom Tools With AI
The Errors That Hide in Plain Sight
A closer review revealed issues that would likely go unnoticed without technical expertise:
-
Broken linkages between financial statements
-
Hardcoded values instead of centralized assumptions
-
Non-dynamic formulas and inconsistent logic across periods
-
Balance sheets that did not balance
-
Timing mismatches between beginning- and end-of-period values
-
Circular reference issues in areas like revolving credit
These are not edge cases. They point to a broader issue. The model may function, but it is not built on a reliable or auditable foundation.
Where Best Practices Break Down
Beyond individual errors, the models often failed to follow core financial modeling principles:
-
Assumptions were not clearly separated from calculations
-
Error checks were largely absent
-
KPIs lacked depth and industry-specific nuance
-
Formula design was inconsistent or inefficient
These gaps affect more than presentation. They determine whether a model can be trusted, adapted and audited under pressure.
The Real Risk Is Overconfidence
The key distinction is not between AI and human-built models. It is between models that are understood and those that are not. When a professional builds a model, every assumption and linkage is intentional. Even limitations are typically known. With AI-generated models, that understanding is outsourced.
This creates a different kind of risk:
-
The logic behind the model may not be fully clear
-
The structure may not align with internal standards
-
The review process may be less rigorous because the output appears complete
Related:Citi Brings Google-Powered AI Avatar to Wealth Clients
In practice, credibility is inferred from how the model looks, not how it was built.
Reviewing Is Not the Same as Building
There is also a practical workflow issue. Reviewing an AI-generated model is not equivalent to building one.
When reviewing:
-
You are interpreting logic you did not design
-
Errors can be harder to trace
-
Inconsistent structure increases audit time
In some cases, it is faster to build a clean model from scratch than to fix a flawed AI-generated one.
What This Means in Practice
Financial models support decisions involving significant capital. Even small issues can cascade:
-
Misstated cash flows can distort debt capacity
-
Timing errors can affect liquidity assumptions
-
Weak KPIs can lead to incomplete analysis
There is also a question of accountability. Regardless of how a model is created, responsibility for its output remains with the professional using it.
Where AI Fits Today
AI tools can still be useful in financial modeling. They can help:
-
Speed up repetitive components
-
Generate starting points for analysis
But they are not a substitute for professional judgment. Nor are they ready to operate without close oversight. For now, their role is best defined as assistive, not authoritative.
Related:The WealthStack Podcast: AI, Capacity and the Future of Advice with Mark Swan
A More Practical View of AI in Finance
The conversation around AI in finance does not need more optimism or skepticism. It needs more precision. AI can produce outputs that are visually convincing and directionally correct. In financial modeling, that is not enough.
The real risk is not that AI makes mistakes. It is those mistakes that are easy to miss, especially when the output looks finished. For financial professionals, the takeaway is simple: treat AI-generated models as drafts, not decision-ready tools.