SENATE BILL REPORT

SB 6513

This analysis was prepared by non-partisan legislative staff for the use of legislative members in their deliberations. This analysis is not a part of the legislation nor does it constitute a statement of legislative intent.

As Reported by Senate Committee On:

State Government, Tribal Relations & Elections, February 7, 2020

Title: An act relating to restricting the use of deepfake audio or visual media in campaigns for elective office.

Brief Description: Restricting the use of deepfake audio or visual media in campaigns for elective office. [Revised for 1st Substitute: Restricting the use of synthetic media in campaigns for elective office.]

Sponsors: Senators Frockt, Hunt, Nguyen, Billig, Hasegawa, Kuderer and Wilson, C.

Brief History:

Committee Activity: State Government, Tribal Relations & Elections: 1/29/20, 2/07/20 [DPS, w/oRec].

Brief Summary of First Substitute Bill

  • Requires a disclosure when any manipulated audio or visual media of a candidate is used in an electioneering communication.

  • Creates a cause of action for candidates whose voices or likenesses appear in synthetic media distributed without disclosure.

  • Provides exceptions for parody and news reporting.

SENATE COMMITTEE ON STATE GOVERNMENT, TRIBAL RELATIONS & ELECTIONS

Majority Report: That Substitute Senate Bill No. 6513 be substituted therefor, and the substitute bill do pass.

Signed by Senators Hunt, Chair; Kuderer, Vice Chair; Hasegawa and Takko.

Minority Report: That it be referred without recommendation.

Signed by Senators Zeiger, Ranking Member; Muzzall, Assistant Ranking Member.

Staff: Samuel Brown (786-7470)

Background: Political Advertising. All political advertising must identify the sponsor of the advertisement. Political advertisements undertaken as independent expenditures or which are distributed within 60 days of an election must also disclose the five persons or entities making the largest aggregate contributions to the advertisement's sponsor of at least $1,000, and the top three individual contributors to any of the top five donors which are political or incidental committees.

A person cannot sponsor, with actual malice, a defamatory statement in political advertising that:

Deepfake Media. Advances in machine learning led to the development of technology where a person in an existing image or video is replaced with another's likeness and techniques to generate new synthetic audio of a person's speech based on past recordings of the person. Media created through machine learning to falsely depict a person's speech or conduct are known as deepfakes.

Summary of Bill (First Substitute): Use of Synthetic Media. Synthetic media is an image or audio or video recording of a person's appearance, speech, or conduct, that has been manipulated to create a realistic or false image, audio, or video that:

An electioneering communication which contains synthetic media may not be distributed without a disclosure. The disclosure must state that the media has been manipulated and:

Exceptions are provided for:

Cause of Action. A candidate whose voice or likeness appears in synthetic media distributed without the required disclosure within 60 days of an election may seek to enjoin distribution of the media and bring an action for general or special damages against the party distributing the media. A prevailing party may be awarded attorneys' fees and costs. The violation of the disclosure requirement must be established by clear and convincing evidence.

Other Provisions. The bill contains a severability clause.

EFFECT OF CHANGES MADE BY STATE GOVERNMENT, TRIBAL RELATIONS & ELECTIONS COMMITTEE (First Substitute): References to "deepfake media" are changed to "synthetic media." Synthetic media need not be distributed with intent to injure a candidate's reputation or deceive voters to trigger disclosure requirements. The disclosure requirement for synthetic media applies only to electioneering communications.

Appropriation: None.

Fiscal Note: Available.

Creates Committee/Commission/Task Force that includes Legislative members: No.

Effective Date: Ninety days after adjournment of session in which bill is passed.

Staff Summary of Public Testimony on Original Bill: The committee recommended a different version of the bill than what was heard. PRO: In something as important as our elections, people need to know what they are being shown to preserve the public trust. Having the imprimatur of a legal construct helpfully sets the boundaries for what is acceptable in political discourse. Myth and disinformation are being weaponized online against society. The bill should be broadened to cover any form of manipulated content. The 60-day limitation could be abused. It is fine to go after content creators, but gatekeeper platforms are turning a blind eye to this kind of conduct. Deepfakes are sophisticated but highly accessible. Anybody can make them. This bill would put pressure to prevent dissemination of deepfakes.

OTHER: Broadcasters suggest an amendment—under the federal Communications Act, once a legally qualified candidate buys advertising on a station, the station cannot refuse or censor the advertisement. Algorithms to determine if media has been manipulated are not always accurate. To the degree this law has a disclosure requirement, it is a good step forward, but the Public Disclosure Commission recommends caution on any restriction and has concerns about how to enforce the bill's requirement that an intent to deceive voters be proven.

Persons Testifying: PRO: Senator David Frockt, Prime Sponsor; Professor Hao Li, citizen; Professor Hany Farid, citizen. OTHER: Mark Allen, Washington State Association of Broadcasters; Vicki Christophersen, Internet Association; Anna Powell, CompTIA; Sean Flynn, Public Disclosure Commission.

Persons Signed In To Testify But Not Testifying: No one.