SENATE BILL REPORT
SB 6513
This analysis was prepared by non-partisan legislative staff for the use of legislative members in their deliberations. This analysis is not a part of the legislation nor does it constitute a statement of legislative intent. |
As Reported by Senate Committee On:
State Government, Tribal Relations & Elections, February 7, 2020
Title: An act relating to restricting the use of deepfake audio or visual media in campaigns for elective office.
Brief Description: Restricting the use of deepfake audio or visual media in campaigns for elective office. [Revised for 1st Substitute: Restricting the use of synthetic media in campaigns for elective office.]
Sponsors: Senators Frockt, Hunt, Nguyen, Billig, Hasegawa, Kuderer and Wilson, C.
Brief History:
Committee Activity: State Government, Tribal Relations & Elections: 1/29/20, 2/07/20 [DPS, w/oRec].
Brief Summary of First Substitute Bill |
|
SENATE COMMITTEE ON STATE GOVERNMENT, TRIBAL RELATIONS & ELECTIONS |
Majority Report: That Substitute Senate Bill No. 6513 be substituted therefor, and the substitute bill do pass.
Signed by Senators Hunt, Chair; Kuderer, Vice Chair; Hasegawa and Takko.
Minority Report: That it be referred without recommendation.
Signed by Senators Zeiger, Ranking Member; Muzzall, Assistant Ranking Member.
Staff: Samuel Brown (786-7470)
Background: Political Advertising. All political advertising must identify the sponsor of the advertisement. Political advertisements undertaken as independent expenditures or which are distributed within 60 days of an election must also disclose the five persons or entities making the largest aggregate contributions to the advertisement's sponsor of at least $1,000, and the top three individual contributors to any of the top five donors which are political or incidental committees.
A person cannot sponsor, with actual malice, a defamatory statement in political advertising that:
contains a false statement of material fact about a candidate for public office;
falsely represents that a candidate is an incumbent for the office sought; or
falsely indicates that a candidate has the support or endorsement of an organization.
Deepfake Media. Advances in machine learning led to the development of technology where a person in an existing image or video is replaced with another's likeness and techniques to generate new synthetic audio of a person's speech based on past recordings of the person. Media created through machine learning to falsely depict a person's speech or conduct are known as deepfakes.
Summary of Bill (First Substitute): Use of Synthetic Media. Synthetic media is an image or audio or video recording of a person's appearance, speech, or conduct, that has been manipulated to create a realistic or false image, audio, or video that:
would appear to a reasonable person of a real individual, but did not occur in reality; and
would cause a reasonable person to have a fundamentally different understanding of the content of the media than of the unaltered media.
An electioneering communication which contains synthetic media may not be distributed without a disclosure. The disclosure must state that the media has been manipulated and:
for visual media, be printed in at least the largest font size of other text in the media or a size easily readable for the average viewer;
for video media, appear for the duration of the video; and
for audio media, be read in a clearly spoken manner and a pitch easily heard by the average listener at the beginning and end of the audio, and at least every two minutes during the audio, if applicable.
Exceptions are provided for:
news media if the news presentation acknowledges that the media does not accurately reflect the speech or conduct of the candidate; and
satire and parody.
Cause of Action. A candidate whose voice or likeness appears in synthetic media distributed without the required disclosure within 60 days of an election may seek to enjoin distribution of the media and bring an action for general or special damages against the party distributing the media. A prevailing party may be awarded attorneys' fees and costs. The violation of the disclosure requirement must be established by clear and convincing evidence.
Other Provisions. The bill contains a severability clause.
EFFECT OF CHANGES MADE BY STATE GOVERNMENT, TRIBAL RELATIONS & ELECTIONS COMMITTEE (First Substitute): References to "deepfake media" are changed to "synthetic media." Synthetic media need not be distributed with intent to injure a candidate's reputation or deceive voters to trigger disclosure requirements. The disclosure requirement for synthetic media applies only to electioneering communications.
Appropriation: None.
Fiscal Note: Available.
Creates Committee/Commission/Task Force that includes Legislative members: No.
Effective Date: Ninety days after adjournment of session in which bill is passed.
Staff Summary of Public Testimony on Original Bill: The committee recommended a different version of the bill than what was heard. PRO: In something as important as our elections, people need to know what they are being shown to preserve the public trust. Having the imprimatur of a legal construct helpfully sets the boundaries for what is acceptable in political discourse. Myth and disinformation are being weaponized online against society. The bill should be broadened to cover any form of manipulated content. The 60-day limitation could be abused. It is fine to go after content creators, but gatekeeper platforms are turning a blind eye to this kind of conduct. Deepfakes are sophisticated but highly accessible. Anybody can make them. This bill would put pressure to prevent dissemination of deepfakes.
OTHER: Broadcasters suggest an amendment—under the federal Communications Act, once a legally qualified candidate buys advertising on a station, the station cannot refuse or censor the advertisement. Algorithms to determine if media has been manipulated are not always accurate. To the degree this law has a disclosure requirement, it is a good step forward, but the Public Disclosure Commission recommends caution on any restriction and has concerns about how to enforce the bill's requirement that an intent to deceive voters be proven.
Persons Testifying: PRO: Senator David Frockt, Prime Sponsor; Professor Hao Li, citizen; Professor Hany Farid, citizen. OTHER: Mark Allen, Washington State Association of Broadcasters; Vicki Christophersen, Internet Association; Anna Powell, CompTIA; Sean Flynn, Public Disclosure Commission.
Persons Signed In To Testify But Not Testifying: No one.