Regulatory Compliance

AI Compliance Framework: From Policy to Audit-Ready Documentation

Table of Contents

TL;DR:

  • Having an AI policy isn’t compliance — having proof you follow it is. Regulators want model inventories, risk assessments, validation evidence, and monitoring records.
  • OCC Bulletin 2011-12, EU AI Act Article 11, Colorado SB 205, and SEC FY2026 exam priorities all demand specific AI documentation — and they’re checking.
  • This guide walks through the seven documentation layers that make up an audit-ready AI compliance framework, with specific deliverables and timelines.

Your AI governance policy is a Word doc gathering dust in SharePoint. Your model inventory is a spreadsheet that hasn’t been updated since Q2. Your validation records consist of an email thread where someone said “looks good.”

That’s not a compliance framework. That’s a liability.

Here’s the uncomfortable reality: only 25% of organizations have fully implemented AI governance programs, according to aggregated enterprise data analyzed in early 2025. Meanwhile, the SEC’s FY2026 Examination Priorities explicitly flag AI technologies as a cross-cutting risk area. The OCC’s Bulletin 2025-26 just clarified model risk management expectations — even for community banks. Colorado’s SB 205 took effect February 1, 2026, requiring algorithmic impact assessments for high-risk AI. And the EU AI Act’s high-risk system documentation requirements kick in August 2026.

The gap between “we have a policy” and “we can prove compliance” is where enforcement actions live. This guide closes that gap.

What an AI Compliance Framework Actually Is

An AI compliance framework is not your AI governance framework. Governance defines who makes decisions and how. Compliance is the evidence layer — the documentation, testing records, and audit trails that prove you’re actually doing what your governance framework says you should.

Think of it this way:

LayerWhat It AnswersKey Outputs
GovernanceWho decides? How do we operate?Policies, committee charters, RACI matrices
ComplianceCan we prove it? Will it survive an exam?Model inventories, risk assessments, validation reports, monitoring dashboards

If your governance framework is the blueprint, your compliance framework is the inspection report. Regulators don’t take your word for it — they want the receipts.

The Regulatory Documentation Landscape in 2026

Multiple regulators are converging on AI documentation requirements simultaneously. Here’s what’s active right now:

U.S. Federal Banking Regulators

OCC Bulletin 2011-12 and Federal Reserve SR 11-7, issued jointly in April 2011, remain the backbone of model risk management expectations. They require:

  • A comprehensive model inventory covering all models in use
  • Model development documentation — assumptions, data, methodology, limitations
  • Validation by qualified independent parties
  • Ongoing monitoring of model performance and stability
  • Outcomes analysis comparing model outputs to actual results

The OCC’s August 2025 Bulletin 2025-26 specifically clarified that community banks should tailor model risk management to their size and complexity — but the core documentation expectations remain.

SEC Division of Examinations

The SEC’s FY2026 Examination Priorities, released November 2025, explicitly list AI as a cross-cutting risk area alongside cybersecurity and customer information safeguards. The Division will examine:

  • Whether firms make accurate representations about AI capabilities (cracking down on “AI washing”)
  • How advisers integrate AI into portfolio management, trading, marketing, and compliance
  • Whether compliance programs adequately address conflicts of interest from AI-driven recommendations
  • Firms’ training and security controls for AI-related risks

EU AI Act (Article 11 + Annex IV)

For firms with EU operations, Article 11 of the EU AI Act requires technical documentation for high-risk AI systems that demonstrates compliance. Annex IV specifies what that documentation must contain — including system architecture, training data descriptions, design specifications, validation metrics, and human oversight measures. These requirements apply to high-risk AI systems starting August 2026.

Colorado SB 205

Colorado’s AI Act (SB 24-205) became enforceable February 1, 2026. It requires deployers of high-risk AI to conduct and document algorithmic impact assessments — covering the purpose, intended use cases, known risks of algorithmic discrimination, and mitigation steps taken.

GAO Findings

The GAO’s May 2025 report, Artificial Intelligence: Use and Oversight in Financial Services (GAO-25-107197), found that while banking regulators are using existing guidance to oversee AI, the NCUA lacks both detailed model risk management guidance and the authority to examine credit unions’ technology service providers. The report signals that regulatory expectations around AI documentation are only going to increase.

The Seven Documentation Layers of an Audit-Ready AI Compliance Framework

Here’s what “audit-ready” actually looks like — seven documentation layers, each with specific deliverables that satisfy the regulatory landscape above.

Layer 1: AI Model Inventory

What it is: A living register of every AI and algorithmic model your organization uses, develops, or procures.

What examiners look for: Completeness. The most common exam finding isn’t that your model inventory is wrong — it’s that it’s incomplete. Examiners will cross-reference your inventory against vendor contracts, IT systems, and business unit interviews. If they find models you didn’t know about, that’s a problem.

Required fields per model:

FieldDescriptionWhy It Matters
Model IDUnique identifierTraceability across all documentation
Model NameDescriptive nameQuick identification
Business OwnerNamed individual (not a team)Accountability
Risk TierHigh / Medium / LowDetermines validation frequency and depth
Model TypeML, rule-based, statistical, LLM, etc.Drives validation approach
Data InputsSource systems and data typesData quality and privacy assessment
Use CaseSpecific business decision supportedScope and materiality
Vendor / In-HouseWho built itThird-party risk considerations
Deployment DateWhen put into productionAge triggers review requirements
Last Validation DateMost recent independent validationCompliance with SR 11-7 validation cycles
Next Review DateScheduled reviewForward-looking compliance

Who owns it: At most mid-size banks, the Model Risk Management (MRM) team under the CRO. At fintechs, this typically falls to the Head of Compliance or VP of Engineering. The key: one person owns the inventory, every business unit contributes to it.

How to build it from scratch: Survey every business unit, including procurement and vendor management. Ask: “Do you use any system that makes predictions, scores, classifies, or automates decisions?” You’ll discover 2-3x more models than anyone expected. The shadow AI problem is real — the GAO report found that financial institutions are using AI across trading, lending, customer service, fraud detection, and compliance, often without centralized visibility.

Layer 2: Risk Classification and Tiering

What it is: A systematic method for assessing each model’s risk level and assigning proportionate oversight.

Deliverables:

  • Risk tiering methodology document (criteria, thresholds, scoring)
  • Individual risk assessment for each model in the inventory
  • Tiering decisions with documented rationale

Tiering approach:

Risk TierCriteriaValidation FrequencyDocumentation Depth
Tier 1 (High)Directly impacts consumer credit decisions, fair lending, capital adequacy, or regulatory reportingAnnual full validation + continuous monitoringComprehensive: full development docs, validation report, outcomes analysis
Tier 2 (Medium)Supports operational decisions, fraud detection, or internal risk managementFull validation every 18 monthsStandard: development summary, validation findings, monitoring metrics
Tier 3 (Low)Informational only, no direct decision impact, easily reversibleValidation every 2-3 yearsAbbreviated: scope memo, key assumptions, periodic check

Under Colorado SB 205, any AI system making “consequential decisions” in employment, education, financial services, healthcare, housing, insurance, or legal services qualifies as high-risk. If you deploy in Colorado, map your tiering to their definition.

Layer 3: Model Development and Design Documentation

What it is: The technical record of how each model was built, what decisions were made during development, and why.

What SR 11-7 specifically requires:

  • Documentation of the model’s purpose and design
  • Assumptions and limitations (explicitly stated, not buried)
  • Data sources, preparation steps, and any exclusions
  • Selection of methodology and variables
  • Testing during development (in-sample and out-of-sample performance)

For EU AI Act compliance (Annex IV), add:

  • System architecture description
  • Training data provenance and labeling methodology
  • Design specifications including what the system optimizes for
  • Human oversight measures required by Article 14
  • Pre-determined change management procedures

Pro tip: If you’re using vendor or third-party models, you still need this documentation. OCC examiners have consistently required banks to understand the models they use, even when developed by vendors. The Consumer Bankers Association, in October 2025 comments to OSTP, flagged the practical difficulty of obtaining model-specific information from vendors — but the regulatory expectation remains. Document what you know, document what the vendor won’t disclose, and document the compensating controls you’ve put in place.

Layer 4: Validation and Testing Evidence

What it is: Independent proof that each model works as intended, doesn’t discriminate, and performs within acceptable parameters.

Core validation artifacts:

  1. Validation scope memo — what’s being tested and why
  2. Conceptual soundness assessment — is the approach appropriate for the use case?
  3. Data quality analysis — input data completeness, accuracy, representativeness
  4. Performance testing — accuracy, precision, recall, AUC, or relevant business metrics
  5. Sensitivity and stability analysis — how outputs change under different conditions
  6. Bias and fair lending testing — disparate impact analysis across protected classes
  7. Benchmarking — comparison against alternative approaches or challenger models
  8. Validation report — findings, issues identified, recommended remediation

The independence requirement: SR 11-7 requires that validation be performed by parties not involved in model development. For Tier 1 models, this means a separate MRM team or external validators. For smaller institutions, the OCC’s 2025 Bulletin clarified that independence can be achieved through reporting lines and organizational separation rather than dedicated staff — but the principle stands.

Bias testing is non-negotiable. The CFPB has made clear that AI used in lending must comply with fair lending laws — no “technology exception” exists. Document your disparate impact testing methodology, the protected classes tested, the results, and any remediation steps taken.

Layer 5: Approval and Change Management Records

What it is: Evidence that models were properly approved before deployment and that changes follow a controlled process.

Required documentation:

  • Pre-deployment approval records — who approved, when, based on what evidence, with what conditions
  • Change request logs — every modification to a production model, no matter how minor
  • Impact assessments for changes — does this change require re-validation?
  • Version control records — every model iteration preserved and traceable
  • Retirement/decommissioning records — when and why models were taken out of production

Implementation specifics:

  • Version-control every model iteration in a code repository (Git, not shared drives)
  • Require validation sign-off before production deployment
  • Maintain a change advisory board (CAB) or equivalent review process for Tier 1 models
  • Set automated drift detection thresholds at ±5% from baseline performance
  • Log all changes with timestamps, responsible parties, and business justification

Layer 6: Ongoing Monitoring and Performance Records

What it is: Continuous evidence that deployed models are performing as expected and haven’t degraded.

Monitoring framework:

Monitoring ActivityFrequencyDocumentation
Performance metrics trackingMonthly (Tier 1), Quarterly (Tier 2/3)Dashboard snapshots, trend analysis
Data drift analysisMonthlyInput distribution comparisons vs. training data
Outcomes analysisQuarterlyPredicted vs. actual outcomes comparison
Bias monitoringQuarterlyDisparate impact ratios across protected classes
BacktestingSemi-annually (Tier 1)Model predictions vs. realized outcomes
Escalation reviewsAs triggeredBreach of thresholds, documented response

What triggers a re-validation:

  • Performance degradation beyond pre-defined thresholds
  • Material change in input data sources or business use
  • Regulatory guidance updates affecting the model’s risk domain
  • More than 12 months since last validation (Tier 1)
  • Merger, acquisition, or significant organizational change

Document the absence of findings too. If monitoring shows the model is performing within parameters, that’s an artifact. Examiners want to see that you’re actively watching, not just that you catch problems.

Layer 7: Regulatory Reporting and Board-Level Oversight Records

What it is: Evidence that AI risk information reaches senior management and the board, and that they act on it.

Required artifacts:

  • Board/committee meeting minutes showing AI risk discussions
  • MRM reporting packages — aggregate risk view, model inventory summary, open issues, validation backlogs
  • Risk appetite statements that explicitly address AI/model risk
  • Escalation records — when issues were elevated to senior management and how they were resolved

Per the NACD’s 2025 Board Practices Survey, 62% of boards now hold regular AI discussions, but only 27% have formally incorporated AI governance into their committee charters. If you’re in the 73% without formal charter language — fix it. Examiners notice.

90-Day Implementation Roadmap

Building all seven layers from scratch takes time. Here’s a realistic roadmap with specific deliverables and owners.

Days 1-30: Foundation

DeliverableOwnerDependencies
Appoint AI compliance leadCRO / Chief Compliance OfficerBoard approval
Draft model inventory templateMRM team / ComplianceNone
Deploy enterprise-wide model surveyAI compliance leadTemplate complete
Draft risk tiering methodologyMRM teamRegulatory review of SR 11-7 / SB 205 criteria
Identify existing documentation gapsComplianceInventory survey in progress
Brief the board on AI documentation requirementsCROGAO report and SEC priorities as supporting materials

Days 31-60: Build

DeliverableOwnerDependencies
Complete model inventory (v1)MRM team + business unitsSurvey responses collected
Apply risk tiering to all inventoried modelsMRM teamTiering methodology approved
Collect existing development documentation for Tier 1 modelsModel developers / vendorsInventory complete
Establish validation schedule based on tieringMRM teamTiering complete
Implement version control for model code and configurationsEngineering / ITRepository access provisioned
Draft monitoring framework and threshold definitionsMRM teamTiering and performance baselines established

Days 61-90: Prove

DeliverableOwnerDependencies
Complete validation for highest-risk Tier 1 modelsMRM validators (independent)Development docs available
Build monitoring dashboard (even if manual/spreadsheet)MRM team + BI/analyticsMetrics defined
Conduct first bias/fair lending test on lending modelsMRM team + Fair Lending OfficerTesting methodology approved
Prepare first board-level MRM reporting packageAI compliance leadInventory, tiering, and validation status data
Document change management process and get sign-offMRM team + IT change managementCAB structure defined
Gap assessment: compare documentation against Annex IV (if EU operations)Legal + ComplianceEU AI Act requirements mapped
Conduct tabletop exam simulationInternal AuditAll Layer 1-7 artifacts assembled

So What? Why This Matters Now

The window between “regulators are talking about AI documentation” and “regulators are examining AI documentation” has closed. The SEC is already examining for it. Colorado is already enforcing it. The EU’s August 2026 deadline for high-risk system compliance is five months away. And the GAO has publicly told Congress that existing oversight frameworks have gaps — which usually means more regulation, not less.

The compliance teams that build audit-ready documentation now will have a 12-18 month head start on the ones scrambling to paper over gaps when the examiner shows up.

The ones that don’t? Look at the pattern: Citibank operated under OCC consent orders from 2020 through December 2025 — five years — for risk management deficiencies. JPMorgan Chase faced a $250 million OCC civil money penalty in March 2024. These were for existing risk management frameworks. AI adds entirely new documentation requirements on top. If your model risk management documentation is already thin, AI will break it.

Start with the inventory. Everything else builds from there.


Need a head start? The AI Risk Assessment Template & Guide gives you pre-built risk assessment templates, model inventory structures, and validation checklists designed for financial services teams. It’s the shortcut between “we should document this” and “here’s our exam-ready package.”

Frequently Asked Questions

What documentation do regulators expect for AI models in financial services?
Regulators expect a complete model inventory, risk assessments for each AI system, validation and testing evidence (including bias testing results), approval and change management records, and ongoing monitoring reports. The specific requirements come from OCC Bulletin 2011-12/Fed SR 11-7 for U.S. banks, EU AI Act Article 11 and Annex IV for firms operating in Europe, and Colorado SB 205 for deployers of high-risk AI in that state.
How do I build an AI model inventory from scratch?
Start by surveying every business unit — including procurement, IT, and vendor management — for any system that uses algorithms, machine learning, or automated decision-making. Capture the model name, owner, business purpose, risk tier, data inputs, vendor (if third-party), validation status, and last review date. Most organizations discover 2-3x more models than they expected. Update the inventory quarterly and require new model registration before deployment.
What's the difference between an AI compliance framework and an AI governance framework?
An AI governance framework defines the organizational structures, roles, and policies for overseeing AI — who makes decisions and how. An AI compliance framework is the documentation and evidence layer that proves you're meeting specific regulatory requirements. Governance is the 'how you operate' layer; compliance is the 'how you prove it' layer. You need both, and they should reference each other.
Rebecca Leung

Rebecca Leung

Rebecca Leung has 8+ years of risk and compliance experience across first and second line roles at commercial banks, asset managers, and fintechs. Former management consultant advising financial institutions on risk strategy. Founder of RiskTemplates.

Immaterial Findings ✉️

Weekly newsletter

Sharp risk & compliance insights practitioners actually read. Enforcement actions, regulatory shifts, and practical frameworks — no fluff, no filler.

Join practitioners from banks, fintechs, and asset managers. Delivered weekly.