Authors’ note: For this report, the authors use the definition of artificial intelligence (AI) from the 2020 National Defense Authorization Act, which established the National Artificial Intelligence Initiative.1 This definition was also used by the 2023 “Executive Order on the Safe, Secure, and Trustworthy Development and Use of Artificial Intelligence.”2 Similarly, this report makes repeated reference to “Appendix I: Purposes for Which AI is Presumed to be Safety-Impacting and Rights-Impacting” of the 2024 OMB M-24-10 memo, “Advancing Governance, Innovation, and Risk Management for Agency Use of Artificial Intelligence.”3
Read the fact sheet
The fact sheet lists all of the recommendations detailed in this chapter of the report.
The U.S. Department of Education has several existing statutory authorities to consider in addressing the challenges and opportunities of artificial intelligence (AI). Governing for Impact (GFI) and the Center for American Progress have extensively researched these existing authorities in consultation with numerous subject matter experts. However, the goal is to provoke a generative discussion about the following proposals, rather than outline a definitive executive action agenda. Each potential recommendation will require further vetting before agencies act. Even if additional AI legislation is needed, this menu of potential recommendations to address AI demonstrates that there are more options for agencies to explore beyond their current work and that agencies should immediately utilize existing authorities to address AI.
The Department of Education has already started to consider the risks and opportunities of AI, resulting in its 2023 “Artificial Intelligence and the Future of Teaching and Learning” report4 and its various taskings in the 2023 “Executive Order on the Safe, Secure, and Trustworthy Development and Use of Artificial Intelligence.”5 Artificial intelligence may be able to provide new kinds of education opportunities,6 but it also comes with risks, including those related to student surveillance and discipline, biased plagiarism detection, student loan misinformation, and discrimination in admissions.
The Department of Education’s authorities to address AI include titles VI and IX of the Civil Rights Act, the Americans with Disabilities Act (ADA), and the Higher Education Act (HEA).7 These statutes can be used to address impermissible discrimination using AI technology and provide various requirements for contractors servicing student loans.
AI risks and opportunities
Equal access to education is paramount to social and economic advancement,8 and educators increasingly rely on AI and related technologies to advance access to education.9 However, the use of AI in education may also cause unintended harm.10
There are several areas of potential concern in using AI in education, including:
- Student surveillance and discipline: School software and online systems track various school interactions, performance, and other information.11 In many cases, AI tools, which may train on datasets with racial bias,12 are then deployed to “predict” future misconduct, track academic performance, and develop personalized learning programs.13 The May 2023 Department of Education report, “Artificial Intelligence and the Future of Teaching and Learning,” notes: “Examples of discrimination from algorithmic bias are on the public’s mind, such as … an exam monitoring system that may unfairly identify some groups of students for disciplinary action.”14 The finalized Office of Management and Budget (OMB) AI guidance issued for federal agencies in March 2024—the OMB M-24-10 memorandum on “Advancing Governance, Innovation, and Risk Management for Agency Use of Artificial Intelligence”—lists “monitoring students online or in virtual-reality,” “recommending disciplinary interventions,” and “facilitating surveillance (whether online or in-person)” as uses of AI by federal agencies that should be presumed to be rights-impacting.15 The Center for Democracy and Technology (CDT) found that 88 percent of teachers report their schools use AI-powered software to monitor student activity online, and two-thirds of those teachers report these data are being used to discipline students.16 The same study found that 40 percent of schools monitor students’ personal devices.17
Alarmingly, this surveillance is even more harmful to LGBTQ+, Black, and brown students, who are already overrepresented in incidents of school discipline and law enforcement involvement.18 Additionally, Black and Latinx students are more likely to use school-issued devices than their peers and, therefore, are more likely to be subjected to increased monitoring and related negative consequences, such as discipline.19
- Plagiarism detection: Many schools use online platforms or software that purport to detect plagiarism in work submitted by students.20 However, these programs—equipped with AI-detecting capabilities—have been found to falsely conclude that students have used AI tools or otherwise plagiarized their work.21 Even when these tools work as intended, they fail to consider the many reasons why students may use assistance when submitting written work, especially as an accommodation for students with disabilities or those for whom English is a second language.22 The OMB M-24-10 AI memo lists “detecting student cheating or plagiarism” as an AI use by federal agencies that should be presumed to be rights-impacting.23
- Federal student loan misinformation: As the Consumer Financial Protection Bureau (CFPB) has reported, banking institutions increasingly rely on AI-generated chatbots to provide consumers with important information about their products and services.24 Given the changing nature of federal student loan repayment and the on-ramp period for resuming repayment,25 loan servicers’ usage of chatbots, especially any chatbots that utilize generative AI, could expose consumers to misleading information about their federal student loans due to their unreliability from hallucinations.26 Hallucinations are described by the National Institute of Standards and Technology (NIST) as “the production of confidently stated but erroneous or false content.”27
- Undergraduate admissions and scholarship awards: Half of the higher education institutions currently use AI technology to review incoming applications, including initial screenings of transcripts and recommendation letters.28 The OMB M-24-10 AI memo lists “influencing admissions processes,” “determining access to educational resources or programs,” and “determining eligibility for student aid” as uses of AI by federal agencies that should be presumed to be rights-impacting.29 Advocates have highlighted that the unrestrained use of AI in school admissions may replicate and favor past prejudices, as AI programs may take into account the demographics and characteristics of past students and give those same groups preferential treatment, often to the detriment of students whose communities have historically been pushed out of higher education.30 This is especially troublesome given the U.S. Supreme Court’s decisions in Students for Fair Admissions v. Harvard and Students for Fair Admissions v. University of North Carolina, which prohibited the use of race as a factor in college admissions outside of narrow circumstances.31
Current state
The Department of Education has a unique opportunity to regulate and place safeguards around the responsible use of AI in the classroom. The recent 2023 executive order on AI directs the department to develop policies that will ensure “safe, responsible, and nondiscriminatory” uses of AI in the education context, including by providing for human review of AI determinations where possible and developing an “AI toolkit” for education leaders.32 In May 2023, the department released its “Artificial Intelligence and the Future of Teaching and Learning” report.33
Notably, the OMB M-24-10 AI memo for agencies identifies several rights-impacting AI use cases in education, including: “detecting student cheating or plagiarism; influencing admissions processes; monitoring students online or in virtual-reality; projecting student progress or outcomes; recommending disciplinary interventions; determining access to educational resources or programs; determining eligibility for student aid or Federal education; or facilitating surveillance (whether online or in-person).”34 It also warns that AI-related “[decisions] blocking, removing, hiding, or limiting the reach of protected speech” may infringe on protected rights.35
Additional ongoing and existing agency initiatives are also likely to protect students and consumers from the potential harms of misaligned or inaccurate AI.36 For example, the Department of Education’s existing guidance relating to the Family Educational Rights and Privacy Act applies if protected records or information are collected, analyzed, or distributed using automated technologies.37 Moreover, a proposed rule to restrict misinformation in the advertising market for secondary degrees should also apply when institutions deploy automated algorithms to target those deceptive ads.38
Relevant statutory authorities
The Department of Education has already proposed several protections for students that target AI harms, including the many recommendations in the “Artificial Intelligence and the Future of Teaching and Learning” report.39 It has also taken on extensive litigation to enforce its existing protections.40 However, the department can take additional actions under its existing statutory authority to address the unique threats to students posed by AI technologies.
Title VI of the Civil Rights Act of 1964
Title VI protects individuals from discrimination based on race, color, or national origin in programs or activities that receive federal financial assistance. Title VI states:
No person in the United States shall, on the ground of race, color, or national origin, be excluded from participation in, be denied the benefits of, or be subjected to discrimination under any program or activity receiving Federal financial assistance.41
Title VI applies to all the operations of a school district or college that receives federal funding, including its admissions, financial aid, recruiting, academic programs, student discipline, grading, class assignment, physical education, athletics, and housing.42 Past guidance has clarified that Title VI also covers national origin-minority groups, including when students cannot speak or understand English because of that identity.43
Recommendation
Given this authority, the Department of Education could take the following action:
- Issue guidance under Title VI explaining that 34 C.F.R. Part 100 applies to discrimination enabled by AI or other generative technology. Specifically, this guidance would include examples of impermissible discrimination using AI technology, including disproportionate discipline for students of color, students with disabilities, or students for whom English is not their first language.
Title IX
Title IX protects individuals from discrimination based on sex in education programs or activities that receive federal financial assistance. Title IX states:
No person in the United States shall, on the basis of sex, be excluded from participation in, be denied the benefits of, or be subjected to discrimination under any education program or activity receiving Federal financial assistance.44
Title IX includes protections from sex discrimination, including: “sexual harassment; the failure to provide equal athletic opportunity; sex-based discrimination in a school’s science, technology, engineering, and math (STEM) courses and programs; and discrimination based on pregnancy.”45 An interpretative notice has explicitly recognized sexual orientation and gender identity as a protected class under Title IX.46
Recommendation
Given this authority, the Department of Education should consider the following action:
- Issue guidance specifying that, under 34 C.F.R. 106.31(b), using AI or other automated technologies, including generative AI, may violate Title IX if it results in sex discrimination. This includes discriminatory surveillance of students because of their sex, disparate discipline resulting from that surveillance, or the filtering out of appropriate internet content because of discriminatory or imprecise AI internet monitoring.
Americans with Disabilities Act, Section 504 of the Rehabilitation Act, and the Individuals with Disabilities Education Act
Section 504 prohibits the discrimination of individuals with disabilities in programs and activities that receive federal financial assistance from the Department of Education. Section 504 provides:
No otherwise qualified individual with a disability in the United States … shall, solely by reason of her or his disability, be excluded from the participation in, be denied the benefits of, or be subjected to discrimination under any program or activity receiving Federal financial assistance.47
The Americans with Disabilities Act, as amended, expanded the definition of covered disabilities under Section 504, emphasizing its wide applicability.48
The Individuals with Disabilities Education Act (IDEA) provides federal funds to educational agencies to assist students with disabilities.49 Each participating school must identify and evaluate students with disabilities and develop individualized educational programs (IEPs), including a statement of the services the school is providing the student.50 The IDEA grants the department the authority to promulgate regulations and guidance that govern participating schools.51
Under Section 504 of the ADA and the IDEA authority, the department has issued several guidance documents, including “dear colleague” letters and technical assistance documents, and has implemented regulations.52
Recommendation
Given this authority, the department could take the following action:
- Issue guidance explaining the ADA’s application to AI’s discriminatory effects in surveillance and discipline, including specific examples of possible discriminatory effects of programs that detect AI-generated work or cheating. The guidance could also address how some students may benefit from AI-assisted programs, which can constitute an accommodation in certain circumstances.
Higher Education Act
Under 20 U.S.C. § 1087(f), the Department of Education is authorized to enter into contracts with qualified institutions for loan servicers as long as they meet certain criteria, which the secretary can set.53 The department has promulgated several servicing-related regulations under its rulemaking authority under Title IV of the HEA54 and has published several guidance documents.55
Recommendation
Given this authority, the Department of Education can take the following action:
- Require that contracted servicers using AI-generated chatbots ensure that borrowers are receiving accurate information about their individual loans. This includes an option to speak with a human within a reasonable amount of time and incorporating any of the relevant minimum risk management practices for rights-impacting purposes developed through the OMB M-24-10 AI memo.56
Conclusion
Education is at the front lines of the AI revolution, and the Department of Education is already engaged in examining the challenges and opportunities of AI. Its existing statutory authority across various education levels provides a variety of options to address the potential challenges and opportunities of AI. GFI and CAP hope this chapter will offer thoughtful options to educators, administrators, and, ultimately, students as AI becomes commonplace in the education system.
Read the fact sheet
The fact sheet lists all of the recommendations detailed in this chapter of the report.