Skip to main content
Compliance

NIST 800-53 Rev 5: the 40 controls AI systems trip on

NIST 800-53 Rev 5 has 1,189 controls and enhancements. For AI systems, roughly forty of them carry most of the real findings. Here they are, named, with the failure mode for each.

The pattern

NIST 800-53 Rev 5 is vast. The Moderate baseline pulls 325 controls; the High baseline 421. Across real federal AI projects, findings concentrate in a predictable subset of roughly forty controls across seven families: AC, AU, CM, IA, SC, SI, and CA. If you harden these, you clear the vast majority of AI-specific ATO findings.

THE HIGH-IMPACT CONTROL FAMILIES FOR AI

For AI under 800-53 Rev 5: SA (acquisition) for AI procurement policy, SI (integrity) for model integrity and poisoning detection, RA (risk assessment) for AI-specific risk, and AU (audit) for inference logging. These four families drive the majority of AI-related findings.

AI systems do not fail 800-53 in novel ways. They fail it in the same ways every system fails it, plus a small set of model-layer specifics that the catalog was not originally written for.

High-Impact Control Families — AI Finding Frequency at FedRAMP Moderate

SI — System and Information Integrity
92%
AU — Audit and Accountability
88%
AC — Access Control
82%
CM — Configuration Management
75%
SA — System and Services Acquisition (AI policy)
68%
RA — Risk Assessment (AI risk)
60%
IA — Identification and Authentication
55%

Percentage = frequency with which findings in this family appear during AI system ATOs. SI and AU dominate because inference logging and model integrity documentation are commonly underprepared.

Access Control (AC)

ControlAI failure mode
AC-2 Account ManagementService accounts used for model API calls never rotated; dormant developer accounts retain model access.
AC-3 Access EnforcementRetrieval layer pulls from document store without enforcing per-user clearance. Model returns content the user should not see.
AC-4 Information Flow EnforcementPrompts can carry data across classification boundaries without labels. Model output crosses boundary in the opposite direction.
AC-6 Least PrivilegeFine-tuning, model upload, and inference share the same role. One compromised developer can modify production models.
AC-16 Security AttributesClassification labels do not travel with data into or out of the model. RAG retrieval loses labels; output is unlabeled.

Audit and Accountability (AU)

ControlAI failure mode
AU-2 Event LoggingNo audit event emitted on every inference. Cannot tell who asked what.
AU-3 Content of Audit RecordsLogs missing full prompt, full response, model version, or tool calls.
AU-6 Audit Review, Analysis, ReportingNobody actually reviews the logs. No alerting on anomalous prompt patterns.
AU-9 Protection of Audit InformationPrompt logs in plaintext. Log store not at the same impact level as the source data.
AU-12 Audit GenerationOrchestration layer logs but downstream model-endpoint log events are missing or not correlated.

Configuration Management (CM)

ControlAI failure mode
CM-2 Baseline ConfigurationNo baseline configuration for the model itself — which version, which parameters, which prompt templates.
CM-3 Configuration Change ControlVendor-managed model updates. Version drifts under the ATO. No change ticket.
CM-6 Configuration SettingsInference parameters (temperature, max tokens, top-p) not documented or parameterized per use case.
CM-8 System Component InventorySBOM missing model weights, training-data lineage, RAG index components.
CM-10 Software Usage RestrictionsOpen-weight models whose licenses prohibit certain federal use cases not tracked.

Identification and Authentication (IA)

ControlAI failure mode
IA-2 User Identification and AuthenticationModel endpoint accepts API keys without binding to a named user. No trail back to a human.
IA-5 Authenticator ManagementModel API keys long-lived, not rotated, sometimes in source repositories.
IA-8 Identification and Authentication (Non-Organizational Users)External agency users not authenticated via PIV/PIV-I when federal policy requires it.
IA-9 Service Identification and AuthenticationService-to-model authentication by IP allow-list instead of workload identity.

System and Communications Protection (SC)

ControlAI failure mode
SC-4 Information in Shared ResourcesModel session or cache leaks data across user boundaries.
SC-7 Boundary ProtectionModel endpoint exposed beyond the authorization boundary without explicit interconnection documentation.
SC-8 Transmission Confidentiality and IntegrityPrompts sent to internal model endpoints over unencrypted channels because "it is internal."
SC-12 Cryptographic Key Establishment and ManagementKeys for encrypting prompt logs managed outside the authorization boundary.
SC-13 Cryptographic ProtectionNon-FIPS-validated crypto used somewhere in the stack (often a third-party SDK).
SC-28 Protection of Information at RestPrompt/response logs and fine-tuned weights not encrypted at rest.

System and Information Integrity (SI)

ControlAI failure mode
SI-4 System MonitoringNo monitoring for anomalous prompt patterns, extraction attempts, or tool-call abuse.
SI-7 Software, Firmware, Information IntegrityModel weights loaded without hash verification. No integrity check on boot.
SI-10 Information Input ValidationNo input filtering. Prompt injection patterns reach the model unchallenged.
SI-15 Information Output FilteringNo DLP on model output. PII, CUI, classification markers leave the boundary.
SI-16 Memory ProtectionGPU memory not cleared between tenants on shared inference hardware.

Assessment, Authorization, Monitoring (CA)

ControlAI failure mode
CA-2 Control AssessmentsModel-layer controls not assessed because FedRAMP baseline inherited without extension for AI.
CA-5 Plan of Action and MilestonesAI-specific findings (hallucination rates, bias metrics) not tracked in POA&M.
CA-7 Continuous MonitoringNo continuous monitoring of model behavior, only of infrastructure.
CA-8 Penetration TestingPen-test does not include red-team against the model — prompt injection, extraction, jailbreak.

Personnel Security and Risk (PS, RA)

  • PS-3 Personnel Screening. Staff with access to CUI training data or prompt logs must meet the applicable screening level. Often overlooked for ML engineers who "only touch test data."
  • RA-5 Vulnerability Monitoring and Scanning. Extended for AI to include model-level vulnerabilities — known jailbreaks, model-card CVE-style disclosures, training-data poisoning research.
The trick is not learning 1,189 controls. It is knowing the forty where AI systems consistently fail and making sure your implementation says something specific about each.

How to prepare

For each of the forty controls above, your SSP should contain an implementation statement that specifically addresses the AI layer. Generic boilerplate copied from the FedRAMP template is not enough; the 3PAO will ask AI-specific questions. Concrete, model-layer answers are the difference between a clean assessment and a findings-heavy one.

Bottom line

Forty controls across seven families catch the overwhelming majority of AI-specific 800-53 findings. Write specific, AI-aware implementation statements for each. Your ATO package improves more from depth on these forty than from surface-level coverage of the remaining 285-plus.

Frequently asked questions

Why forty controls specifically?

Across real federal AI projects, findings concentrate in AC, AU, CM, IA, SC, SI, and CA — roughly forty controls. The remaining 285+ in the Moderate baseline apply but rarely drive unique AI findings.

Which control catches the most AI findings?

AU-3 (audit record content) and SI-10 (input validation) are perennial. Prompt/response logging and prompt-injection handling are where AI systems consistently fall short.

Does NIST 800-53 Rev 6 change this?

Rev 6 is being drafted in OSCAL-native form. AI-specific overlays are under development. The forty-control concentration pattern will likely hold; the specific parameter values and enhancements will evolve.

What about the NIST AI RMF?

AI RMF is a governance framework that complements 800-53. The two work together: AI RMF produces evidence that maps to specific 800-53 controls. Neither replaces the other.

Can I skip controls my platform inherits?

Inherited controls still need an implementation statement, typically short, referencing the parent FedRAMP package and describing any customer responsibilities that remain. They are not zero-effort.

What is the fastest path to 800-53 readiness for AI?

Author your SSP alongside the engineering. Write specific, model-layer implementation statements for the forty controls. Run your eval harness to generate evidence for CA-7, CA-8, SI-4, SI-10.

1 business day response

Writing 800-53 implementation statements?

We can help you draft AI-aware implementation statements for the forty controls that drive most AI findings.

Talk to usRead more insights →
UEI Y2JVCZXT9HP5CAGE 1AYQ0NAICS 541512SAM.GOV ACTIVE