Read the full report
This fact sheet collects the recommendations from Chapter 4: “Housing Regulators” of the joint report from Governing for Impact (GFI) and the Center for American Progress, “Taking Further Agency Action on AI: How Agencies Can Deploy Existing Statutory Authorities To Regulate Artificial Intelligence.” The chapter notes how the U.S. Department of Housing and Urban Development (HUD) and other housing regulators should consider addressing potential artificial intelligence (AI) risks to housing fairness and discrimination using existing statutory authorities in the Fair Housing Act (FHA) and the Dodd-Frank Wall Street Reform and Consumer Protection Act. The goal of these recommendations is to provoke a generative discussion about the following proposals, rather than outline a definitive executive action agenda. This menu of potential recommendations to address AI demonstrates that there are more options for agencies to explore beyond their current work and that agencies should immediately utilize existing authorities to address AI.
Read the full chapter
Fair Housing Act
Based on FHA authorities, HUD could take the following action:
- Update the “Fair Housing Advertising” guidelines—a separate document from the newly released advertising guidance—elucidating Section 804(c)’s prohibition against discrimination in the advertisement of housing opportunities in the context of online advertising that relies on algorithmic tools or data, as required by the 2023 “Executive Order on the Safe, Secure, and Trustworthy Development and Use of Artificial Intelligence” and consistent with the recent HUD guidance on advertising through digital platforms.1 Such guidance would be consistent with the U.S. Department of Justice’s (DOJ) settlement with Facebook, which targeted similar practices,2 and can specifically highlight practices that lead to housing advertisements being steered away from protected communities.3 Furthermore, the guidance should specify that companies providing advertising services using AI technologies are liable. The guidelines should mirror the responsibilities and liabilities outlined in HUD’s recent guidance.4
Dodd-Frank Act
Based on this authority, the Federal Housing Finance Agency (FHFA) should take the following actions:
- Continue the rulemaking process on the proposed automated valuation model (AVM) rule but also include its application to all mortgage lenders—specifically nonbanks, given that more than half of annual residential real estate loans were made by nonbanks in 2022.5 Furthermore, the rule should include specific minimum standards for each proposed goal, potentially incorporating the National Institute of Standards and Technology (NIST) AI guidelines6 or relevant minimum standards developed in response to the minimum risk management practices anticipated by the Office of Management and Budget (OMB) M-24-10 memorandum on “Advancing Governance, Innovation, and Risk Management for Agency Use of Artificial Intelligence.”7
- Specify, through the proposed AVM rule or additional rulemaking, that companies using AVMs must disclose their use to customers and allow customers to request nonautomated appraisals or seek valuation from alternative AVMs. The FHFA can do so using its broad authority in Section 1125 to “account for any other such factor that the agencies … determine to be appropriate.”8 This would align with the statute’s purpose to “ensure a high level of confidence in [AVMs],” “protect against the manipulation of data,” and “avoid conflict of interest.”9