AISN #22: The Landscape of US AI Legislation - Hearings, Frameworks, Bills, and Laws

post by aogara (Aidan O'Gara), Dan H (dan-hendrycks) · 2023-09-19T14:44:22.945Z · LW · GW · 0 comments

This is a link post for https://newsletter.safe.ai/p/the-landscape-of-us-ai-legislation

Contents

  Senator Schumer’s AI Insight Forum
  The Blumenthal-Hawley Framework
  Agencies Proposed to Govern Digital Platforms
  Deepfakes and Watermarking Legislation
  State and Local Laws Against AI Surveillance
  National AI Research Resource (NAIRR)
  Links
None
No comments

Welcome to the AI Safety Newsletter by the Center for AI Safety. We discuss developments in AI and AI safety. No technical background required.

This week we’re looking closely at AI legislative efforts in the United States, including:

Subscribe here to receive future versions.


Senator Schumer’s AI Insight Forum

The CEOs of more than a dozen major AI companies gathered in Washington on Wednesday for a hearing with the Senate. Organized by Democratic Majority Leader Chuck Schumer and a bipartisan group of Senators, this was the first of many hearings in their AI Insight Forum. 

After the hearing, Senator Schumer said, “I asked everyone in the room, ‘Is government needed to play a role in regulating AI?’ and every single person raised their hands.” Elon Musk, CEO of xAI, called the hearings “a great service to humanity.” 

Senator Josh Hawley raised concerns that despite the hearings, “nothing is advancing” in terms of legislation. Below, we’ll discuss several bills on AI policy which have been introduced to Congress, none of which have come to a vote. 

The Blumenthal-Hawley Framework

Senator Hawley recently introduced a framework for AI legislation alongside Senator Richard Blumenthal. The pair lead the Senate Judiciary Subcommittee on Privacy, Technology and the Law, and have hosted three hearings on AI policy over the last five months. 

The Blumenthal-Hawley framework recommends:

This framework was endorsed by Microsoft President Brad Smith in a hearing last week. In California, state Senator Scott Wiener introduced a similar bill of intent for state legislation on AI. The details of both proposals still need to be fleshed out into concrete policies for AI governance.

Agencies Proposed to Govern Digital Platforms

Whereas AI policy discussions are often conceptual and lack concrete legislative proposals, there are many concrete proposals to regulate digital platforms including social media. These bills could affect AI developers, and offer insights for those crafting AI legislation. 

In June, a bill that would create a federal agency to govern digital platforms was introduced by Democratic Senators Michael Bennet and Peter Welch. The agency would have “a broad mandate to promote the public interest” via methods including rules, civil penalties, hearings, investigations, and research. Senators Elizabeth Warren and Lindsey Graham introduced a similar bill in July requiring digital platforms to obtain a federal license each year. 

These bills would likely apply to AI companies, but many AI-specific policies are not included in these bills. For example, they do not require companies to disclose their training data, nor mandate red-teaming and evaluations before the release of new AI models. The federal agency could work on making rules about AI systems, but amending the bill itself to address AI-specific concerns could provide a clearer mandate for the new agency.

Despite public support for regulating AI and social media, there is no guarantee that these bills will become law. Senator Ted Cruz recently came out against AI regulation, and others may follow. Last year, no vote was held on two bills with broad bipartisan support to regulate technology companies after the industry spent $37 million lobbying against them.

Deepfakes and Watermarking Legislation

An AI-specific proposal put forth in several recent bills intends to combat deepfakes and AI-generated misinformation. By using AI to fabricate text, images, videos, and audio, scammers have run fake kidnapping scams and even stole $600,000 from a Chinese businessman. More powerful AI systems could soon be used for widespread misinformation campaigns of persuasion and deception.

Several recent bills intend to combat AI deepfakes with clearly visible notices of AI-generated content. The DEEP FAKES Accountability Act of 2019 first proposed the idea, and the AI Labeling Act of 2023 proposes a similar requirement where the primary responsibility for providing benchmarks would be placed on AI developers rather than downstream users. The REAL Political Advertisements Act is more narrowly targeted, mandating disclosure of AI-generated content only in political advertisements. 

Clearly visible notices of AI-generated content might be intrusive or unpleasing to viewers. Another potential solution is watermarking, the practice of embedding statistical patterns in AI outputs which are typically unnoticed by a consumer, but which can be detected using a watermark detection tool. The National Defense Authorization Act for 2024 directed DARPA to hold a prize competition for the development of AI watermarking techniques. 

State and Local Laws Against AI Surveillance

While federal laws tend to get the most attention, it's important to note that states and local governments are also playing a role in AI regulation, particularly against AI surveillance.

Alabama, Colorado, Washington, and Baltimore, Maryland have all passed laws restricting police use of facial recognition technology. Some have banned it outright, while others require a search warrant and prohibit its use in sensitive locations such as schools. 

These state and local initiatives highlight concerns about civil liberties and the willingness of states and cities to check the unbridled use of AI by law enforcement agencies.

National AI Research Resource (NAIRR)

The CREATE AI Act is a bipartisan bill which has been designed and developed with strong support over the last few years. It would establish a National AI Research Resource (NAIRR) to provide compute and data to AI researchers outside of industry.

The bill states that priority access will be given to research projects focused on “privacy, ethics, civil rights and civil liberties, safety, security, risk mitigation, and trustworthiness.” Today, only 2% of AI research focuses on key topics relevant for safety. But by targeting these neglected topics in AI safety, NAIRR could substantially boost the amount of research in the field.

Creating “testbeds'' for evaluating AI systems is another statutory requirement for NAIRR. They would be tasked with working with NIST to “develop a comprehensive catalog of open AI testbeds.” These could include evaluations of extreme risks such as dangerous capabilities and misalignment. Perhaps a future federal regulator could also benefit from access to a suite of AI evaluation testbeds. 

Decisions about which projects to support would be made by federal agencies, as well as an “operating entity” – likely a university or federally funded research and development center – chosen to run the day-to-day operations of the cluster. The work would be overseen by the National Science Foundation and expert advisory committees. 

See also: CAIS website, CAIS twitter, A technical safety research newsletter, An Overview of Catastrophic AI Risks, and our feedback form

Subscribe here to receive future versions.

0 comments

Comments sorted by top scores.