HUD AI Guidance in Housing
In effect since
Overview
Fair Housing Act disparate impact standard applies to AI-driven tenant screening, lending algorithms, and property valuations. HUD 2023 disparate impact rule (reinstated) allows challenges to facially neutral AI practices with discriminatory effects. Meta 2022 settlement over AI ad targeting in housing is a key precedent. Disparate impact rule status under Trump administration should be monitored.
This is federal agency guidance interpreting existing statutes and rules.
Who this applies to
This regulation applies to both companies that build AI products and companies that use AI tools from other vendors.
AI categories covered
- Housing
- Financial services AI
- Consumer-facing AI
Specific AI use cases:
- Credit scoring and risk assessment
- Customer profiling and segmentation
What this requires you to do
Bias testing required
Perform bias testing. Test your AI systems for discriminatory impact across protected classes.
Impact assessment required
Complete an impact assessment. Document the potential risks and effects of your AI system on affected people.
Record-keeping required
Maintain records. Keep documentation of your AI systems, decisions made, and compliance activities.
Enforcement and penalties
Existing Fair Housing Act penalties.
Legislative history
How this law got here
- effective
HUD disparate impact rule framework applied to algorithmic housing practices
Source
Read the full text
https://www.hud.gov/program_offices/fair_housing_equal_opp
Last verified: April 9, 2026
Always verify current language and amendments at the official source.
This rule references NIST AI RMF practices. See the federal NIST AI RMF entry for context and source links.
Other Federal (HUD) regulations
Explore more rules in the same jurisdiction that may apply to your AI systems.
Want to know what else applies to your company?
Run a free XIRA scan to see all regulations that match your states and AI tools.