AI Risk

AI Explainability: What Financial Regulators Expect and How to Deliver It

March 29, 2026 Rebecca Leung
Table of Contents

TL;DR:

  • SR 11-7’s “effective challenge” principle requires that humans can understand and question AI model outputs — but the guidance never defined what “sufficient explainability” means for neural networks or LLMs.
  • CFPB Circular 2022-03 makes it explicit: you can’t use complex AI models if you can’t provide specific, accurate adverse action reasons to consumers.
  • The EU AI Act’s Article 13 transparency obligations for high-risk systems take effect August 2026 — and they apply to anyone deploying AI in credit, employment, or insurance decisions affecting EU residents.

Your model works. It passed backtesting. It outperforms the legacy scorecard by 15%. And your regulator just asked you to explain how it makes decisions.

If you’re running traditional logistic regression, that’s a 10-minute conversation. If you’re running gradient-boosted trees, XGBoost, or — increasingly — LLM-powered decision support, that conversation just became the single biggest risk in your model risk management program.

Here’s the problem: every major financial regulator now expects AI explainability, but none of them agree on exactly what that means. SR 11-7 says “effective challenge.” The OCC says “explainability is a key risk.” The CFPB says you must provide “specific and accurate” adverse action reasons. The EU AI Act says “appropriate type and degree of transparency.” And NIST’s AI RMF lists explainability as a core trustworthiness characteristic alongside safety and fairness.

This guide breaks down what each regulator actually expects, which explainability techniques work for which model types, and how to document explanations that survive an exam.

The Regulatory Landscape: Who Requires What

Not every regulator uses the same language, but they’re all converging on the same expectation: if a model influences a decision, someone needs to be able to explain why.

SR 11-7 and OCC Bulletin 2011-12: “Effective Challenge”

The Federal Reserve’s SR 11-7 (issued April 2011) established the foundational principle: model risk management requires “effective challenge” — defined as “critical analysis by objective, informed parties that can identify model limitations and produce appropriate changes.”

That was written for spreadsheet-era models. As GARP noted in its February 2026 analysis, SR 11-7 “emphasizes transparency sufficient to enable effective challenge, yet provides limited guidance on what constitutes ‘sufficient explainability’ for complex models.” The framework’s foundational principles — sound governance, independent validation, effective challenge — remain conceptually robust, but the nature of the models those principles must govern has changed fundamentally.

The practical implication: your validation team needs to understand how the model reaches its outputs well enough to challenge assumptions, identify limitations, and recommend changes. For a linear model, that’s straightforward. For a 175-billion-parameter LLM? That’s an unsolved problem — and examiners know it.

OCC’s Four Key AI Risks

In May 2022, the OCC’s Deputy Comptroller for Operational Risk Kevin Greenfield testified before Congress outlining four key risks associated with AI. Explainability was number one on the list.

The OCC’s position: “a failure to understand an AI process or outcome could result in the bank acting in a way that harms its customers or fails to comply with consumer protection requirements.” And critically: “a lack of explainability means that a bank may be unable to apply model risk management practices to an AI process or technology, which could impair safety and soundness.”

Translation: if you can’t explain it, you can’t validate it. If you can’t validate it, you can’t govern it. And if you can’t govern it, examiners will write it up.

CFPB Circular 2022-03: The Adverse Action Test

CFPB Circular 2022-03 drew the hardest line of any US regulator. The Bureau stated that “ECOA and Regulation B do not permit creditors to use complex algorithms when doing so means they cannot provide the specific and accurate reasons for adverse actions.”

This isn’t a suggestion. Under Regulation B, when a consumer is denied credit, the lender must provide the specific principal reasons for that denial. The CFPB’s 2023 Innovation Spotlight on adverse action notices with AI/ML models acknowledged that the example methodologies for determining principal reasons (issued in 1982) may not directly apply to modern ML models — but made clear the obligation itself hasn’t changed.

The bottom line: if your AI credit model can’t generate specific, accurate reason codes that a consumer would understand, you have a compliance problem regardless of how well the model performs.

EU AI Act Article 13: Transparency for High-Risk Systems

The EU AI Act’s Article 13 requires that high-risk AI systems “shall be designed and developed in such a manner as to ensure that their operation is sufficiently transparent to enable deployers to interpret a system’s output and use it appropriately.”

For financial services, high-risk categories include AI used in creditworthiness assessment, credit scoring, insurance pricing, and employment decisions. The rules for high-risk AI take effect in August 2026, with some categories extending to August 2027.

Providers of high-risk systems must include instructions of use covering the system’s capabilities, limitations, potential risks, and how to interpret outputs. If you’re a US-based firm deploying AI that touches EU residents — through cross-border lending, insurance, or employment — these obligations apply to you.

NIST AI RMF: Explainability as a Trustworthiness Characteristic

NIST’s AI Risk Management Framework (AI RMF 1.0) identifies explainability and interpretability as core characteristics of trustworthy AI, noting that risk “can be a result of the opaque nature of AI systems (limited explainability or interpretability), lack of transparency or documentation.”

NIST distinguishes between interpretability (the system’s outputs can be understood by humans in the context of their functionality) and explainability (the mechanism by which the system produces outputs can be described). A model can be interpretable without being fully explainable — you might understand what it does without understanding exactly how.

The MEASURE function of the framework specifically calls for evaluating explainability as part of ongoing model assessment — not just at initial deployment.

The Explainability Spectrum: From Glass Box to Black Box

Not every model requires the same level of explanation. The key is matching explainability approach to model complexity, decision criticality, and regulatory context.

Model TypeInherent ExplainabilityTypical Use CasesRegulatory Comfort Level
Linear/Logistic RegressionHigh — coefficients directly show feature impactCredit scoring, risk ratingHigh
Decision TreesHigh — decision path is visibleFraud rules, underwriting criteriaHigh
Random Forest / Gradient BoostingMedium — ensemble obscures individual pathsCredit risk, AML scoringMedium (with post-hoc XAI)
Deep Neural NetworksLow — millions of parameters, no clear decision pathImage recognition, NLP tasksLow (requires significant XAI)
LLMs (GPT, Claude, etc.)Very Low — billions of parameters, emergent behaviorDecision support, document analysisVery Low (active regulatory concern)

Ante-Hoc vs. Post-Hoc Explainability

The CFA Institute’s August 2025 report “Explainable AI in Finance” draws an important distinction between two approaches:

Ante-hoc methods build interpretability into the model from the start. These include inherently transparent models (linear regression, decision trees, rule-based systems) and constrained architectures that sacrifice some predictive power for interpretability. If you’re in a domain where regulators expect clear decision paths — like fair lending — ante-hoc methods are your safest bet.

Post-hoc methods are applied after a model has been trained, attempting to explain decisions the model has already made. These include SHAP, LIME, attention visualization, and counterfactual explanations. They’re essential for complex models but come with a critical caveat: they’re approximations of the model’s actual reasoning, not the reasoning itself.

Practical XAI Techniques: What Works and When

SHAP (SHapley Additive exPlanations)

What it does: Assigns each input feature a contribution value (positive or negative) to a specific prediction, based on cooperative game theory. SHAP values are theoretically grounded and satisfy key properties: consistency, local accuracy, and missingness.

Best for: Tree-based models (XGBoost, LightGBM, Random Forest), credit scoring, any model where you need feature-level attribution for individual decisions.

Regulatory fit: Strong. SHAP can generate the feature-level explanations needed for adverse action reason codes under Regulation B. A SHAP analysis might show: “This applicant was denied because debt-to-income ratio contributed -0.32 to the score, recent credit inquiries contributed -0.18, and credit history length contributed -0.15.”

Limitations: Computationally expensive for large datasets. Kernel SHAP (model-agnostic version) can be slow. And SHAP values explain a model’s computation, not necessarily the causal reason for a decision — a distinction examiners increasingly appreciate.

LIME (Local Interpretable Model-agnostic Explanations)

What it does: Creates a simple, interpretable model (typically linear) that approximates the complex model’s behavior in the neighborhood of a specific prediction. Essentially: “around this data point, the model behaves like this simpler model.”

Best for: Quick, intuitive explanations of individual predictions across any model type. Useful for consumer-facing explanations and initial diagnostic work.

Regulatory fit: Moderate. LIME is easier to communicate to non-technical stakeholders but less theoretically rigorous than SHAP. The local approximation may not capture global model behavior, which can be a concern for validation teams.

Limitations: Results can be unstable — running LIME twice on the same prediction may give different explanations depending on the perturbation sampling. This instability is a problem when you need consistent, defensible explanations for examiners.

Feature Importance (Global Methods)

What it does: Ranks features by their overall contribution to model predictions across the entire dataset. Methods include permutation importance, gain-based importance (for tree models), and integrated gradients (for neural networks).

Best for: Model documentation, validation reports, and answering “what drives this model overall?” questions. Essential for model cards and exam-ready documentation.

Regulatory fit: Strong for documentation purposes, but insufficient alone. Examiners want both global explanations (“what does this model generally rely on?”) and local explanations (“why did this specific applicant get denied?”).

Counterfactual Explanations

What it does: Answers “what would need to change for this prediction to be different?” For example: “If this applicant’s debt-to-income ratio were 36% instead of 45%, the model would have approved the application.”

Best for: Consumer-facing explanations, adverse action notices, and fair lending analysis. Counterfactuals naturally map to the “principal reasons” format that Regulation B requires.

Regulatory fit: High for consumer-facing use. The CFPB’s framing of adverse action notices essentially asks for counterfactual reasoning: what factors, if different, would have changed the outcome?

Attention Visualization (For Transformer Models)

What it does: Shows which parts of the input a transformer-based model “attended to” when generating output. For text-based models, this highlights which words or phrases most influenced the output.

Best for: NLP tasks, document analysis models, and LLM-powered decision support where you need to show which parts of an input document drove the output.

Regulatory fit: Low to moderate. Attention weights don’t necessarily represent causal reasoning — a model may attend to a word without that word being the “reason” for the output. Useful as supplementary evidence but shouldn’t be the sole explainability mechanism.

Explainability by Model Type: A Practical Framework

Traditional ML Models (Logistic Regression, Decision Trees)

These are inherently explainable. Your documentation should include:

  • Model coefficients or decision rules
  • Feature importance rankings
  • Sample decision paths for representative cases
  • Marginal effects of key variables

Validation approach: Direct review of model structure. Validators can trace any prediction back to specific inputs and weights.

Ensemble Models (Random Forest, XGBoost, Gradient Boosting)

These require post-hoc explainability layered on top:

  • SHAP values for individual prediction attribution
  • Global feature importance (permutation or gain-based)
  • Partial dependence plots showing feature-output relationships
  • Interaction effects between top features

Validation approach: Statistical testing of SHAP value consistency, comparison of post-hoc explanations against known model behavior, and adversarial testing of edge cases where explanations may break down.

Deep Learning Models (Neural Networks, CNNs)

These require significant XAI investment:

  • Integrated gradients or DeepLIFT for feature attribution
  • Layer-wise relevance propagation for understanding internal representations
  • Concept-based explanations (TCAV) for higher-level understanding
  • Surrogate models that approximate the neural network’s behavior with an interpretable model

Validation approach: Compare multiple XAI methods against each other. If SHAP, LIME, and integrated gradients disagree on which features matter, that’s a red flag that none of them are capturing the true model behavior.

LLMs and Generative AI

This is the frontier — and the hardest to solve:

  • Chain-of-thought prompting to surface reasoning steps
  • Retrieval-augmented generation (RAG) to ground outputs in cited sources
  • Output logging and human review of decision-influencing outputs
  • Confidence scoring to flag low-certainty outputs for human review
  • Structured output formats that force the model to provide reasoning alongside decisions

Validation approach: Focus on output validation rather than mechanism explanation. Test for consistency (does the model give the same answer to the same question?), groundedness (are cited sources real and relevant?), and boundary behavior (where does the model fail?). For LLMs, you’re often validating the system (prompts + guardrails + human review) rather than the model itself.

Documenting Explainability for Examiners

Good explainability isn’t just about running SHAP — it’s about documenting a coherent explanation story that examiners can follow. Here’s what your model documentation should include:

1. Explainability Approach Rationale

Document why you chose your explainability methods for this specific model. Include:

  • Model type and complexity level
  • Decision criticality (what happens if the model is wrong?)
  • Regulatory requirements that apply (Reg B, EU AI Act, etc.)
  • Tradeoffs considered (e.g., “SHAP chosen over LIME for consistency of explanations in adverse action notices”)

2. Global Explainability Report

  • Top 10 features by importance with magnitudes
  • Partial dependence plots for top 5 features
  • Feature interaction analysis
  • Population stability analysis showing explanation consistency over time

3. Local Explainability Samples

  • SHAP waterfall plots for representative approved/denied cases
  • Counterfactual explanations for denied cases
  • Edge cases where the model’s decision is close to the threshold
  • Cases where the XAI explanation seems counterintuitive (and why)

4. Limitations and Known Gaps

Be upfront about what your explainability approach can’t do:

  • Where post-hoc explanations may diverge from actual model reasoning
  • Model types or input patterns where explanations are less reliable
  • Plans for addressing gaps (e.g., “migrating to inherently interpretable model for fair lending decisions by Q3”)

5. Model Cards

The concept of model cards — standardized documentation of a model’s intended use, performance, limitations, and ethical considerations — is rapidly becoming a regulatory expectation. Include explainability information directly in the model card: what XAI methods are used, what they show, and what they don’t.

Building an Explainability Program: 30/60/90-Day Roadmap

Days 1-30: Inventory and Assessment

Owner: Model Risk Management lead or CRO

  • Inventory all AI/ML models currently in production or development
  • Classify each model by complexity tier (inherently interpretable → black box)
  • Map each model to applicable regulatory explainability requirements
  • Identify highest-priority gaps: models used in consumer decisions (credit, insurance, employment) without adequate explainability
  • Deliverable: Explainability gap assessment report with prioritized remediation plan

Days 31-60: Implementation

Owner: Data science team with MRM oversight

  • Deploy SHAP or equivalent post-hoc XAI for all Tier 2+ models (ensemble and above)
  • Build automated explanation pipelines for consumer-facing models (adverse action reason code generation)
  • Create model card templates with explainability sections
  • Update model documentation standards to include explainability requirements
  • Deliverable: XAI tooling deployed for top 3 highest-risk models; updated documentation standards

Days 61-90: Validation and Governance

Owner: Independent validation team

  • Validate XAI outputs against known model behavior (do the explanations make sense?)
  • Test explanation consistency over time (are SHAP values stable across validation periods?)
  • Establish ongoing monitoring for explanation drift (explanations shifting even when model performance is stable)
  • Update model governance policies to require explainability review at each stage gate
  • Deliverable: Validated XAI framework; updated governance policy; exam-ready documentation for first model

So What?

Explainability isn’t a nice-to-have feature you bolt on after deployment. It’s a regulatory requirement that shapes which models you can use, how you deploy them, and how much oversight they need.

The firms getting this right aren’t just running SHAP on everything and calling it done. They’re making explainability a design constraint — choosing model architectures that balance performance with interpretability, building explanation pipelines into their MLOps infrastructure, and documenting the explainability story before the examiner asks for it.

The firms getting it wrong are deploying black-box models into consumer-facing decisions, hoping post-hoc explanations will be “good enough,” and discovering during an exam that their XAI outputs can’t actually generate the specific adverse action reasons that Regulation B requires.

With the EU AI Act’s high-risk obligations taking effect in August 2026, the window for treating explainability as an afterthought is closing fast.

Need a structured framework for assessing AI risks — including explainability gaps? The AI Risk Assessment Template & Guide includes a complete risk assessment methodology with explainability evaluation criteria built in.

FAQ

What’s the difference between explainability and interpretability in AI?

Interpretability means a human can understand the system’s outputs in context — you grasp what it does. Explainability means the mechanism by which the system produces outputs can be described — you understand how it works. NIST’s AI RMF distinguishes these explicitly. A model can be interpretable (you understand the output) without being fully explainable (you can’t trace the internal mechanism). In practice, regulators want both: understand what the model does and be able to describe why it reached a specific decision.

Which XAI method should I use for adverse action notices?

SHAP values are the strongest choice for generating adverse action reason codes. They provide mathematically consistent feature-level attribution that maps directly to the “principal reasons” framework under Regulation B. Counterfactual explanations are a strong complement — they answer “what would need to change” in a format consumers can understand. LIME is viable but its instability across runs makes it harder to defend in an exam. Whatever method you choose, validate that it produces consistent, specific reasons — not vague categories like “creditworthiness factors.”

Does the EU AI Act require full model transparency for all AI systems?

No. The EU AI Act’s Article 13 transparency requirements apply to high-risk AI systems — those used in credit scoring, employment decisions, insurance pricing, law enforcement, and other specified categories. The requirements focus on enabling deployers to “interpret a system’s output and use it appropriately” through instructions of use, documentation of capabilities and limitations, and information about known risks. The rules for most high-risk systems take effect in August 2026. Systems not classified as high-risk face lighter or no transparency obligations, though separate rules apply to general-purpose AI models.

Rebecca Leung

Rebecca Leung

Rebecca Leung has 8+ years of risk and compliance experience across first and second line roles at commercial banks, asset managers, and fintechs. Former management consultant advising financial institutions on risk strategy. Founder of RiskTemplates.

Immaterial Findings ✉️

Weekly newsletter

Sharp risk & compliance insights practitioners actually read. Enforcement actions, regulatory shifts, and practical frameworks — no fluff, no filler.

Join practitioners from banks, fintechs, and asset managers. Delivered weekly.