Artificial intelligence (AI) is defined and understood in different ways. The National Artificial Intelligence Initiative Act of 2020 (Act) defines AI as a machine-based system that can, for a given set of human-defined objectives, make predictions, recommendations, or decisions influencing real or virtual environments. The Act further describes AI systems as systems using machine- and human-based inputs to: (1) perceive real and virtual environments; (2) abstract such perceptions into models through analysis in an automated manner; and (3) use model inference to formulate options for information or action.
By Executive Order, the White House adopted the Act's definition of AI and defined related concepts, including defining "generative AI" as the class of AI models that emulate the structure and characteristics of input data in order to generate derived synthetic content, which may include images, videos, audio, text, and other digital content.
Required Documentation.
Before a generative AI system or service, or a substantial modification to a generative AI system or service, is made publicly available for use, the developer of the system or service is required to post documentation regarding the data used to train the generative AI system or service.
The required documentation must include a high-level summary of the datasets used in the development of the generative AI system or service, including:
Developers must comply with this requirement by January 1, 2026, for generative AI systems or services, or substantial modifications of such systems or services, that are released on or after January 1, 2022.
A developer is not required to post documentation regarding the data used to train a generative AI system or service that:
Enforcement.
The Attorney General is required to enforce these requirements, and a developer who is found in violation is liable for a civil penalty of $5,000 per day of being in violation, to be collected in a civil action filed by the Attorney General. Before bringing a civil action, the Attorney General is required to notify a developer of the alleged violation if the Attorney General determines that a cure is possible. If the developer fails to cure the violation within 45 days after receiving notice, the Attorney General may bring a civil action without further notice.
Definitions.
"Developer" is defined as "a person, partnership, state or local government agency, or corporation primarily engaged in designing, coding, producing, or substantially modifying an artificial intelligence system or service intended for commercial distribution or public use." For the purposes of this definition, "members of the public" does not include an affiliate or a hospital's medical staff member. This definition excludes individuals or entities that develop artificial intelligence systems or services solely for internal use or research purposes or those that utilize third-party artificial intelligence systems via application programming interface without substantial modification.
"Generative artificial intelligence system or service" is defined as "an artificial intelligence system or service that can generate derived synthetic content, such as text, images, video, and audio, that emulates the structure and characteristics of the artificial intelligence system's or service's training data."
The following additional terms are defined: aggregate consumer information; artificial intelligence; security and integrity; substantially modifies; synthetic data generation; train a generative artificial intelligence system or service; and types of data points.
The substitute bill:
(In support) Artificial Intelligence has quickly increasing capabilities, and there are accounts of AI being used in harmful ways. People who deploy AI models know very little about them, and AI is not a very regulated medium. This bill will result in much-needed transparency requirements and give the public better insight into the data used to train generative AI models, including whether the data includes personal information. The idea that requiring disclosures about AI models forces companies to disclose trade secrets is a false narrative, driven by fears of liability. Many AI models use the same datasets, but the tools vary because of the ingenuity used in building the models. This bill is based on a California law, and the text was debated and negotiated in California. There are people who are worried that once they get started into a career, AI may take their job. Studies show that image generation services have had issues with racial bias and stereotyping. This bill will help address the copyright issues associated with generative AI tools that produce data verbatim. This bill is narrower than people think, because it does not apply to in-house applications of AI, and it only applies to generative AI. This bill will promote industry growth and public safety. Other states have enacted similar bills. Washington should be a leader in the AI industry, while balancing that leadership with responsibility.
(Opposed) There should not be a private right of action in this bill. Washington should wait to legislate on AI until the AI Task Force has issued recommendations. California has a full-time legislature, so they can continue to work on bills throughout the year. Stakeholders, including the AI Task Force, were not given enough time to weigh in on the bill draft.
(Other) A health marketplace is important for innovation. Some of the language in the bill may result in unclear expectations. Training data for AI models is like a nutritional label on food; it is not a recipe, but it gives information in a context that people can understand. The bill should be amended to not encourage spurious litigation.
(In support) Representative Clyde Shavers, prime sponsor; Jai Jaisimha, Transparency Coalition.ai; Zachary McKay, Lake Washington High School; Mihai-Alexandru Radu; and Nathan Guan.