AI Risk

The 7 Trustworthy AI Characteristics in NIST AI 100-1: What Compliance Teams Need to Know

April 22, 2026 Rebecca Leung
Table of Contents

TL;DR

  • NIST AI 100-1 defines seven trustworthiness characteristics: valid and reliable, safe, secure and resilient, explainable and interpretable, privacy-enhanced, fair with harmful bias managed, and accountable and transparent.
  • “Valid and reliable” is the foundation — without it, none of the others hold.
  • NIST explicitly acknowledges trade-offs between characteristics (privacy vs. accuracy, interpretability vs. performance) and requires organizations to document their balancing decisions.
  • Financial services regulators are using these characteristics as an assessment lens — they map directly to ECOA, FCRA, GLBA, and model governance obligations.

The seven trustworthy AI characteristics in NIST AI 100-1 are not a wishlist or a set of aspirational values. They’re the analytical framework the entire AI Risk Management Framework is built on, and increasingly, the vocabulary that regulators, examiners, and institutional AI governance committees use to evaluate whether an AI program is doing what it claims to do.

If you’ve been implementing NIST AI RMF — working through GOVERN, MAP, MEASURE, and MANAGE — the seven characteristics are the thread connecting all four functions. Every GOVERN policy, every MAP risk category, every MEASURE test, every MANAGE treatment decision can be traced back to one or more of these characteristics. Understanding them at depth changes how you implement the framework.

Why the Trustworthiness Characteristics Are the Architecture, Not an Appendix

NIST AI 100-1 structures the trustworthiness characteristics as a hierarchy, not a flat list. Valid and reliable sits at the base — it’s the necessary precondition. Accountable and transparent is shown as a cross-cutting property that spans the others. The remaining five occupy the space between: they represent specific dimensions of trustworthiness that can each be assessed, measured, and managed independently.

The practical implication: your AI governance program needs to address each of the seven characteristics for each deployed system. Not a checkbox confirming you’ve read the list, but documented evidence of what each characteristic means for that specific system, what tests or controls address it, and what residual risk exists.

The Seven Characteristics in Depth

1. Valid and Reliable

Valid and reliable is the base. An AI system is valid if it measures or predicts what it’s supposed to measure or predict. It’s reliable if it performs consistently across contexts, populations, and time.

Validity and reliability are where most traditional model validation work lives — out-of-sample testing, backtesting, holdout set performance, sensitivity analysis. The difference in NIST AI 100-1 is that validity extends beyond statistical accuracy. A model can be statistically accurate on average but invalid for specific subpopulations, in certain deployment contexts, or when inputs shift from training distribution. That’s where validity connects to fairness — a model that’s valid for majority-group applicants but invalid (unreliable at the subgroup level) for protected-class members is a fair lending problem, not just a performance gap.

For compliance teams: validity is the criterion that justifies every other claim about the system. If you’re claiming a model is fair, the claim is only as strong as your evidence of validity for the populations affected. If you’re claiming a model is safe, you need evidence that it’s valid under the conditions where it operates.

2. Safe

Safety in NIST AI 100-1’s context means the system doesn’t produce outcomes that cause unreasonable harm to people, institutions, or society — whether through intended operation or failure modes.

For most financial services AI, “safe” translates into: does the system, when operating as designed, produce outcomes that stay within acceptable harm boundaries? A credit model that accurately denies credit to statistically risky borrowers isn’t unsafe in a technical sense — but one that produces outputs that systematically exclude creditworthy borrowers based on protected characteristics creates real-world harm.

Safety also covers failure modes: what happens when the model receives out-of-distribution inputs, when data quality degrades, or when the system is manipulated? For AI systems making consequential decisions at scale, the failure mode analysis is as important as the performance evaluation.

3. Secure and Resilient

Secure means the system is protected against adversarial attacks — prompt injection for LLMs, adversarial examples for classification models, data poisoning during training, model inversion to extract training data. Resilient means it can maintain acceptable performance and recover from disruptions, including infrastructure failures, cyberattacks, and unexpected operating conditions.

In financial services, the security dimension of AI governance is where cybersecurity teams and risk teams often operate in silos — and it shows in exam findings. A model governance program that has rigorous performance testing but no adversarial robustness analysis is incomplete. The NIST AI 600-1 generative AI profile extends this with specific GenAI security risks — data poisoning, prompt injection, and training data extraction — that require different controls than traditional model security.

4. Explainable and Interpretable

NIST AI 100-1 draws a precise distinction here that’s operationally important: explainability is about the mechanisms — how the AI system produces its outputs. Interpretability is about the meaning — what those outputs signify in the deployment context.

A compliance officer doesn’t need a full description of the model’s feature weights to fulfill their obligations. They need interpretability: given this system’s output, what decision does it inform, what population does it affect, and what would a meaningful explanation to an affected individual look like?

This distinction matters enormously for adverse action notices. If a lending model produces a denial and the required ECOA/FCRA adverse action notice must explain the principal reasons for the denial, you need interpretability — the ability to translate the model’s output into a legally defensible human explanation. Explainability at the technical level is a means to that end, not the end itself.

The regulatorily mandated context for explainability in financial services isn’t NIST — it’s ECOA, FCRA, and the CFPB’s August 2024 reaffirmation that “there are no exceptions to the federal consumer financial protection laws for new technologies.” If your model can’t support an interpretable adverse action explanation, the explainability characteristic isn’t just a NIST goal — it’s a compliance requirement.

5. Privacy-Enhanced

Privacy-enhanced means the AI system is designed with appropriate protections for personal data across its lifecycle — training, inference, output, and storage. This covers data minimization, differential privacy techniques, access controls, and preventing unintended leakage of personal information in model outputs.

For AI systems in financial services, privacy connects to GLBA safeguard requirements, state privacy laws under frameworks like CCPA/CPRA, and sector-specific data handling requirements. The specific concern that differentiates AI privacy from standard data privacy is inference attack risk: a model trained on sensitive personal data can sometimes reveal information about individual training records through its outputs, even without direct access to the training data.

Privacy-enhanced also includes output privacy — particularly relevant for generative AI systems that might produce outputs containing personal information from training data or from user inputs fed back through retrieval systems.

6. Fair with Harmful Bias Managed

Fair with harmful bias managed is the characteristic where financial services compliance teams spend the most time — and for good reason. AI systems that produce outcomes with unjustified disparate impacts on protected classes create liability under ECOA, the Fair Housing Act, and state fair lending laws, regardless of intent.

NIST’s framing is careful: the characteristic is not “unbiased” — NIST acknowledges that statistical bias and fairness-related bias are different concepts, and that eliminating all statistical disparities may not be achievable or desirable. The goal is managing harmful bias: disparities that produce unjustified discriminatory effects, disproportionate harm to vulnerable groups, or outcomes inconsistent with the system’s stated purpose.

The AI fair lending compliance framework covers the specific test methodologies — disparate impact analysis, counterfactual fairness testing, adverse action rate disparities by protected class — that implement this characteristic in practice. The CFPB’s position, reiterated in 2024, is that algorithmic decisioning is subject to the same disparate impact standards as traditional underwriting, and that an institution’s choice to use an automated model can itself constitute a discriminatory policy if it produces discriminatory outcomes.

7. Accountable and Transparent

Accountable and transparent is the characteristic NIST frames as cross-cutting — it applies across the others and across the entire AI system lifecycle. Accountability means clear ownership: who is responsible for the system’s outcomes, who can override it, who has authority to decommission it, and who is accountable to affected stakeholders. Transparency means appropriate disclosure — not necessarily full technical documentation, but sufficient disclosure for operators, users, and oversight bodies to understand what the system does and how it behaves.

In financial services, transparency requirements have regulatory teeth. Model documentation standards under OCC and Federal Reserve guidance require that models be documented to a level where someone unfamiliar with the system can evaluate its purpose, methodology, assumptions, limitations, and monitoring status. “The vendor built it and it’s proprietary” is not a defensible governance position for a model making material business decisions.

Accountability also means incident accountability — when an AI system produces harmful outputs, there is a clear path for affected parties to raise concerns, for the institution to investigate, and for remediation to occur. That accountability structure is a MANAGE function deliverable.

Trade-offs NIST Explicitly Acknowledges

One of the more honest and useful aspects of NIST AI 100-1 is its acknowledgment that the seven characteristics create genuine tensions:

Trade-offWhat it means in practice
Privacy vs. accuracyDifferential privacy protections reduce information available for training, typically reducing model accuracy
Interpretability vs. performanceSimpler, more interpretable models generally underperform complex ensembles on accuracy metrics
Fairness interventions vs. accuracyPost-processing or in-processing fairness constraints often reduce overall accuracy while improving subgroup parity
Safety constraints vs. capabilityGuardrails on AI outputs (especially generative) restrict capabilities that may have legitimate uses

NIST doesn’t provide an algorithm for resolving these trade-offs — and it’s right not to. The resolution is context-specific. A consumer lending model subject to ECOA should weight the fairness characteristic more heavily than a back-office document summarization tool. A medical AI making diagnostic recommendations should weight safety above efficiency. The GOVERN function should establish the context-specific balancing principles; AI governance documentation should record how those principles were applied for each system.

How Regulators Map the Characteristics to Exam Expectations

The characteristics are not just conceptual — they map to specific regulatory obligations that financial services AI governance programs are already subject to:

CharacteristicPrimary Regulatory Touchpoint
Valid and reliableSR 11-7 / OCC model validation; OCC 2026-13 interagency guidance
SafeOperational risk management; consumer protection obligations
Secure and resilientFFIEC cybersecurity; operational resilience requirements
Explainable and interpretableECOA/Reg B adverse action; FCRA; CFPB supervisory guidance
Privacy-enhancedGLBA Safeguards; CCPA/CPRA; state privacy laws
Fair with harmful bias managedECOA; Fair Housing Act; UDAAP; state AG enforcement
Accountable and transparentModel governance documentation; SR 11-7 third-line oversight

OCC, Federal Reserve, and FDIC examiners reviewing AI programs in 2026 don’t require institutions to cite NIST chapter and verse. But the questions they ask — “how do you know the model is producing consistent outcomes for protected-class members,” “what happens when the model is wrong,” “who owns this model and who can decommission it” — are the trustworthiness characteristics in operational form.

So What?

The seven characteristics give compliance teams something they’ve needed since AI governance became an exam topic: a principled, framework-grounded vocabulary for structured AI risk conversations. Not “do we have a model risk policy” but “can we demonstrate that this system is valid for the populations it serves, explainable enough to support adverse action notices, fair enough to pass a disparate impact test, and accountable enough to produce a clear incident response.”

The practical starting point is a gap assessment against each characteristic for each high-risk deployed system. For most teams, the gaps concentrate in three places: explainability/interpretability (adverse action capability), fairness (disparate impact analysis documentation), and accountability (incident response and override procedures). Those are also where examiner attention concentrates.

The NIST AI RMF implementation guide covers how the full framework integrates. For the production governance and ongoing monitoring that the seven characteristics require, the MANAGE function is where these standards become operational obligations.


The AI Risk Assessment Template & Guide includes assessment templates for each of the seven trustworthiness characteristics, pre-deployment checklists, and model inventory frameworks built for financial services teams demonstrating NIST AI 100-1 alignment.


Sources:

Frequently Asked Questions

What are the seven characteristics of trustworthy AI in NIST AI 100-1?
NIST AI 100-1 defines seven characteristics: (1) valid and reliable — the AI system performs as intended with sufficient accuracy and consistency; (2) safe — the system protects against unacceptable harm; (3) secure and resilient — the system resists adversarial attacks and recovers from disruptions; (4) explainable and interpretable — users and operators can understand how the system works and what its outputs mean; (5) privacy-enhanced — the system handles personal data with appropriate protections; (6) fair with harmful bias managed — outcomes don't produce unjustified discriminatory effects; and (7) accountable and transparent — there is clear ownership, governance, and disclosure appropriate to the deployment context.
Which trustworthy AI characteristic is the foundation for the others?
NIST AI 100-1 presents 'valid and reliable' as the base characteristic — a necessary condition for trustworthiness. A system that doesn't perform as intended can't meaningfully claim any of the other characteristics. You can't have a safe AI system that's consistently wrong, or a fair one whose outputs are unreliable. Validity and reliability are the floor, not just one item on a list.
What is the difference between explainability and interpretability in NIST AI 100-1?
NIST draws a clear distinction: explainability refers to the mechanisms underlying how the AI system operates — the technical 'why' behind how it produces outputs. Interpretability refers to the meaning of those outputs in the context of the system's intended purpose — what the output actually means for the decision or action at hand. A compliance team reviewing an AI credit decision needs interpretability: not just that the model weighted 40 features, but that the output means 'this applicant is likely to default at an 8% rate.' Explainability supports interpretability but is not the same thing.
How do regulators use the NIST trustworthy AI characteristics in financial services exams?
OCC, Fed, and FDIC examiners increasingly use NIST AI 100-1's trustworthiness characteristics as a reference lens when reviewing AI governance programs. Fair with bias managed maps to ECOA, Reg B, and disparate impact analysis. Explainable and interpretable maps to adverse action notice requirements under FCRA and ECOA — if a model can't explain why an applicant was denied, that's a compliance problem. Secure and resilient maps to operational resilience and cybersecurity expectations. Accountable and transparent maps to model governance documentation requirements. You don't need to cite NIST in your exam responses, but the characteristics are the vocabulary examiners are using.
What trade-offs does NIST AI 100-1 acknowledge between trustworthiness characteristics?
NIST explicitly acknowledges that the seven characteristics can create tensions. Privacy versus accuracy: adding differential privacy protections to a model typically reduces its predictive accuracy. Interpretability versus performance: simpler, more interpretable models (logistic regression) often underperform complex ones (gradient boosted trees) on accuracy metrics. Fairness versus predictive accuracy: interventions to reduce disparate impact may reduce overall accuracy. NIST's framework doesn't resolve these trade-offs — it requires organizations to acknowledge them explicitly and document the context-specific balancing decisions made, with appropriate governance sign-off.
Does NIST AI 100-1 apply to generative AI and LLMs?
The seven trustworthiness characteristics from NIST AI 100-1 apply to all AI systems, including generative AI. However, NIST AI 600-1 — the Generative AI Profile — extends the AI 100-1 framework with 12 risk categories specific to GenAI systems, including confabulation, data privacy, homogenization, and value chain complexity. For compliance teams deploying LLMs, AI 100-1's characteristics provide the underlying standard, and AI 600-1 translates them into GenAI-specific risk language. The two documents are designed to work together.
Rebecca Leung

Rebecca Leung

Rebecca Leung has 8+ years of risk and compliance experience across first and second line roles at commercial banks, asset managers, and fintechs. Former management consultant advising financial institutions on risk strategy. Founder of RiskTemplates.

Related Framework

AI Risk Assessment Template & Guide

Comprehensive AI model governance and risk assessment templates for financial services teams.

Immaterial Findings ✉️

Weekly newsletter

Sharp risk & compliance insights practitioners actually read. Enforcement actions, regulatory shifts, and practical frameworks — no fluff, no filler.

Join practitioners from banks, fintechs, and asset managers. Delivered weekly.