AI Risk

AI Model Validation: Testing Techniques That Actually Work for ML and LLM Models

Table of Contents

Most AI model validation done in financial services is theater. Check the box, produce a report, file it away. Then the model goes to production and something quietly breaks — a bias slips through, a model drifts off course, or an examiner pulls the validation document and asks why the out-of-sample testing methodology doesn’t match what’s actually running.

This guide cuts through the theater. Here’s what actually works for validating ML and LLM models under the regulatory frameworks that matter: SR 11-7, FFIEC examination procedures, CFPB adverse action guidance, and NIST’s AI Risk Management Framework.

Why Standard Validation Frameworks Fall Apart for AI

SR 11-7 — the 2011 Federal Reserve/OCC supervisory guidance that still anchors model risk management — was written before large language models existed. It covers traditional statistical models with reasonable assumptions: linearity, stationarity, known distributions.

AI models break those assumptions routinely. A gradient-boosted credit scoring model has millions of parameters and non-linear interactions regulators can’t easily peer inside. An LLM used for compliance document review produces outputs no one can fully predict in advance. Neither fits neatly into the “validate once, monitor annually” cadence that worked for logistic regression.

The GAO noted in its May 2025 report on AI in financial services (GAO-25-107197) that existing supervisory guidance — including SR 11-7 — can apply to AI models, but institutions bear the burden of figuring out how. That’s the gap this post fills.

The Validation Stack: What You Actually Need

A complete AI model validation has four layers. Skipping any one of them is where practitioners get into trouble.

Layer 1: Conceptual Soundness

Before touching data, validate the model’s logic.

For ML models, this means reviewing feature engineering choices, checking whether the problem framing (classification vs. regression, threshold selection) matches the business use case, and assessing whether the training objective aligns with regulatory requirements like fairness across protected classes.

For LLMs, conceptual soundness means documenting the intended use case precisely — not “compliance document review” but “extraction of regulatory reporting fields from merchant dispute notices.” Ambiguity in the stated use case is the root cause of most LLM validation failures. If you can’t write a one-paragraph description of what the model should do and who it affects, the model isn’t ready for validation.

SR 11-7 calls this “model theory” — the reasoning chain from inputs to outputs. That’s your starting point regardless of model type.

Layer 2: Technical Validation

This is where most practitioners spend all their time — and still miss critical components.

For ML Models

1. Back-testing and hold-out validation Split data chronologically, not randomly. In financial services, temporal leakage is a common flaw — using future data to train a model that will be deployed on past data. Hold out the most recent 20-30% of your data as a genuine out-of-time test set. If your model performs significantly worse on that hold-out than on training data, you have a problem.

2. Cross-validation with purpose k-fold cross-validation is standard, but for imbalanced credit datasets, stratified k-fold is non-negotiable. Standard k-fold can give you a false sense of performance if minority class examples cluster in certain folds.

3. Population Stability Index (PSI) SR 11-7 and FFIEC examiners expect this. PSI measures whether the distribution of model inputs in production has shifted significantly from training data. Industry standard: PSI > 0.25 on any score band signals material distribution change and triggers recalibration review.

4. Discriminatory accuracy testing For credit models, this means testing across protected classes under ECOA and the Fair Housing Act. CFPB’s August 2024 comment made clear: there’s no “fancy technology” exemption from fair lending laws. Courts have held that the decision to use an algorithmic model itself can constitute a policy that produces discriminatory outcomes under disparate impact theory.

Test for both disparate treatment (explicit use of protected characteristics) and disparate impact (neutral policy that disproportionately harms protected groups). The CFPB expects institutions to consider less discriminatory alternatives and document why they weren’t chosen.

5. Variable contribution analysis Know why your model makes the decisions it makes. At minimum, use SHAP (SHapley Additive exPlanations) values or permutation importance to identify top drivers of model output. For regulatory examiners, a model you can’t explain is a model that exposes you to adverse action claims under ECOA.

For LLM Models

LLM validation is fundamentally different because there’s no ground truth to compare against in the same way. A language model generating compliance summaries or flagging suspicious transactions doesn’t have a single “correct” output — it has a range of acceptable outputs, and that range is defined by judgment, not metrics.

1. Reference-free evaluation with LLM-as-a-judge The emerging standard here is using a second, often more capable, LLM to evaluate the first model’s outputs against predefined criteria — accuracy, compliance, bias, tone, adherence to business rules. This is documented in practice at firms like Galileo AI and others working in financial services AI governance. The judge model scores outputs on these dimensions; you set thresholds for acceptable scores before deployment.

2. RAG pipeline testing If your LLM uses retrieval-augmented generation — pulling context from internal documents — test the retrieval layer independently. Does the retriever pull relevant documents? Does the context window contain what the model needs to answer correctly? A model can fail not because it can’t reason but because it’s answering the wrong question from wrong context.

3. Adversarial prompt testing LLMs are vulnerable to prompt injection, jailbreaks, and context manipulation. For financial services, this isn’t theoretical — a manipulated prompt could cause a model to output incorrect regulatory guidance or approve a transaction that should be flagged. Test with:

  • Benign variations of the same query (does the model give consistent answers?)
  • Edge-case inputs (missing fields, ambiguous phrasing, foreign language)
  • Injection attempts (user trying to override system instructions)

4. Hallucination detection For LLMs used in any compliance-adjacent function, you need automated hallucination detection. Compare outputs against authoritative source documents where available. In production, ground responses in retrieved context — this is the core value of RAG — and log instances where the model generates content not present in the retrieved documents.

5. Output schema validation LLMs should produce structured outputs where the downstream system requires them. Validate that output fields match expected schemas, data types, and value ranges. A model that generates a transaction flag with a risk_score field returning a string instead of a float will break your monitoring pipeline silently.

Layer 3: Independent Review

SR 11-7 requires model validation to be independent of model development. This isn’t a bureaucratic nicety — it’s the only structural protection against confirmation bias in the development team.

For community banks, the OCC’s September 2025 bulletin (Bulletin 2025-26) clarified that validation frequency and depth should be commensurate with risk exposure and model complexity. A fraud-detection model used in real-time decisions needs more frequent and rigorous validation than a marketing score with limited customer impact.

Independence means:

  • Validators don’t report to the same manager as developers
  • Validators weren’t involved in model construction
  • Validation findings go to a governance body with authority to escalate or reject

For LLMs, independence is harder to operationalize. The business unit that deployed the LLM often doesn’t have the AI expertise to validate it. Consider whether a specialist third-party or internal AI governance team is needed for high-stakes LLM deployments — and document why or why not.

Layer 4: Ongoing Monitoring

SR 11-7 mandates continuous monitoring post-deployment. Most institutions do this poorly for AI models.

For ML models:

  • Run PSI on a monthly cadence at minimum
  • Track default rates and approval rates by demographic segment — any statistically significant shift is a trigger
  • Re-score the hold-out validation sample periodically to detect performance degradation

For LLMs:

  • Monitor output distribution (length, sentiment, topic drift)
  • Track escalation rates — how often does a human need to override the model’s output?
  • Run periodic regression suites against known-answer test cases (“gold datasets”)

LLMs especially can degrade silently. A model that passed validation in January may drift as the underlying distribution of real-world queries shifts, or as the model provider updates weights. Build regression testing into your operational cadence, not just your initial deployment checklist.

The Regulatory Knot You Can’t Untie

Here’s what regulators keep circling back to, and what your validation program needs to address:

Explainability. CFPB has made explicit: lenders must provide accurate, specific reasons when denying credit — regardless of whether the model is a black-box. The complexity of your model doesn’t reduce your obligation. SR 11-7 reinforces this — documentation must be sufficient for a knowledgeable third party to understand model limitations.

Fairness testing documentation. CFPB’s guidance requires regular testing for disparate treatment and disparate impact, and consideration of less discriminatory alternatives. Your validation report should show this testing was done, what it found, and what decisions were made as a result.

Adverse action compliance. If your model drives decisions triggering ECOA adverse action notice requirements, the reasons you provide must reflect the model’s actual decision factors. If your model uses 200 features but you can only articulate 5 reasons for denial, you have an explainability gap that creates legal exposure.

Model inventory and change management. SR 11-7 requires a complete model inventory with documentation for each model. Any material change — retraining, feature modification, threshold adjustment — triggers a validation update. Don’t let your model inventory become a snapshot from two years ago.

What “Actually Works” Boils Down To

The validation programs that survive examiner scrutiny share common traits:

  1. Documentation exists before deployment, not after. Regulators can tell when you’ve written the validation report as a compliance exercise rather than as a genuine assessment. Pre-deployment documentation — conceptual model description, testing plan, acceptance criteria — demonstrates that validation is baked into your development process.

  2. Testing matches the use case risk level. A model approving $50M commercial loans needs more rigorous validation than an internal chatbot. Document your risk-tiering framework and apply validation intensity accordingly.

  3. Findings lead to actions. A validation report that identifies a bias gap and then recommends “monitor for now” with no remediation timeline will be flagged. Regulators expect you to fix problems you find, or document why you accepted the residual risk.

  4. LLMs get the same governance rigor as ML models. The novelty of LLMs is not an excuse for lower standards. The FFIEC IT Examination Handbook doesn’t carve out exceptions for newer technology. If your institution uses LLMs in any customer-facing or compliance-adjacent function, that use case needs to appear in your model inventory with validation documentation.

Closing Thought

The financial services industry has had model risk management guidance since 2011. What’s changed in the past few years isn’t the framework — it’s the models. LLMs and deep learning systems behave in ways SR 11-7’s authors couldn’t have imagined. The regulators know this. They’re not expecting you to have all the answers, but they are expecting you to have a defensible process for managing model risk that covers what you’re actually deploying.

The institutions that get into trouble aren’t the ones trying new things. They’re the ones deploying AI without the validation infrastructure to catch problems before they become enforcement actions.

Build the infrastructure. Test it honestly. Document what you find.

Frequently Asked Questions

Does SR 11-7 apply to AI and LLM models?

Yes. The GAO confirmed in its May 2025 report (GAO-25-107197) that SR 11-7 applies to AI models, though institutions must determine how to adapt the guidance for non-traditional model types like LLMs and deep learning systems.

How often should AI models be validated?

Validation frequency should match model risk. The OCC’s 2025 Bulletin 2025-26 clarified that high-risk models (fraud detection, credit decisioning) need more frequent validation than low-impact models. Most institutions validate high-risk AI models quarterly and lower-risk models annually, with continuous monitoring in between.

What’s different about validating an LLM versus a traditional ML model?

LLMs lack a single “correct” output, making traditional accuracy metrics insufficient. LLM validation requires reference-free evaluation (LLM-as-a-judge), adversarial prompt testing, hallucination detection, and RAG pipeline testing — none of which apply to traditional statistical models.

Rebecca Leung

Rebecca Leung

Rebecca Leung has 8+ years of risk and compliance experience across first and second line roles at commercial banks, asset managers, and fintechs. Former management consultant advising financial institutions on risk strategy. Founder of RiskTemplates.

Immaterial Findings ✉️

Weekly newsletter

Sharp risk & compliance insights practitioners actually read. Enforcement actions, regulatory shifts, and practical frameworks — no fluff, no filler.

Join practitioners from banks, fintechs, and asset managers. Delivered weekly.