
American Privacy Rights Act (APRA) Must Include Online Civil Rights, Says the Federation of American Scientists
Require that personal data is processed safely and fairly
Washington, DC – June 27, 2024 – The Federation of American Scientists believes that the House Energy & Commerce Committee made the right decision in canceling its markup of the American Privacy Rights Act (APRA) today. Privacy laws are only effective if they include civil rights protections that ensure personal data is processed safely and fairly regardless of race, gender, sexuality, age, or other protected characteristics.
“We believe that the United States needs federal data privacy laws, and despite the clear shortcomings of the latest version of APRA, we commend Chair McMorris Rodgers and Ranking Member Pallone for their leadership and efforts in advancing comprehensive privacy legislation. We hope the conversation on privacy doesn’t end here, and that new attempts to pass privacy legislation include the critical civil rights protections necessary to build trust across all communities participating in our data-based society,” says Dan Correa, CEO of Federation of American Scientists.
“The latest draft of APRA not only discards basic protections, but also baseline considerations for AI and other algorithms that would prevent data-based discrimination. If this draft became law, it would also undermine hard-fought privacy protections currently in effect in several states,” says Clara Langevin, AI Policy Specialist on FAS’s Emerging Technologies team.
###
ABOUT FAS
The Federation of American Scientists (FAS) works to advance progress on a broad suite of contemporary issues where science, technology, and innovation policy can deliver dramatic progress, and seeks to ensure that scientific and technical expertise have a seat at the policymaking table. Established in 1945 by scientists in response to the atomic bomb, FAS continues to work on behalf of a safer, more equitable, and more peaceful world. More information at fas.org.
At this inflection point, the choice is not between speed and safety but between ungoverned acceleration and a calculated momentum that allows our strategic AI advantage to be both sustained and secured.
Improved detection could strengthen deterrence, but only if accompanying hazards—automation bias, model hallucinations, exploitable software vulnerabilities, and the risk of eroding assured second‑strike capability—are well managed.
A dedicated and properly resourced national entity is essential for supporting the development of safe, secure, and trustworthy AI to drive widespread adoption, by providing sustained, independent technical assessments and emergency coordination.
Congress should establish a new grant program, coordinated by the Cybersecurity and Infrastructure Security Agency, to assist state and local governments in addressing AI challenges.