The California State Senate has recently passed legislation aimed at regulating the development and training of advanced AI models. The Safe and Secure Innovation for Frontier Artificial Intelligence Models Act, designed to prevent the misuse of AI for malicious purposes, has sparked significant debate. Developers argue that this bill could hinder innovation, while supporters believe it is a necessary step to ensure AI safety.
The Drive for AI Regulation
Introduced by State Sen. Scott Weiner in February, SB 1047 is currently under review in the Assembly. Recent amendments clarified several points of confusion but also ensured that the law would primarily target the largest AI developers, such as Anthropic, Google DeepMind, OpenAI, and Meta. The Assembly is set to vote on the bill in August.
SB 1047 seeks to make large-scale AI developers accountable for ensuring their models do not pose critical threats. These threats include creating weapons of mass destruction, causing significant financial damage through cyberattacks, or autonomous actions resulting in severe harm. Developers must incorporate “kill switches” to disable their models if necessary.
The legislation applies to AI models trained using more than 10^26 integer or floating-point operations (FLOPS), costing over $100 million. This targets the largest and most costly AI models, using the same metric referenced in the Biden Administration’s AI Executive Order. Developers must submit annual compliance certifications and report AI safety incidents to a newly established Frontier Model Division within California’s Department of Technology.
Balancing Safety and Innovation
The discussion around SB 1047 highlights the challenge of regulating advanced AI without stifling innovation. The bill is designed to prevent potential harm while not imposing excessive burdens on developers. It requires developers to implement basic safety measures and does not hold them strictly liable for damages unless they fail to take reasonable precautions or falsely report model capabilities.
AI safety advocates support the bill, though some argue it doesn’t go far enough in protecting the public. They insist that AI, like any other industry, needs regulation to ensure safety. However, developers worry that the bill might limit open foundation model development and fine-tuning. Foundation models are large-scale AI systems capable of various tasks, while open models allow users to modify algorithmic parameters.
Industry Concerns and Resistance
Opposition to SB 1047 includes TechNet, a network of tech companies such as Anthropic, Apple, Google, Meta, and OpenAI, as well as the California Chamber of Commerce. Critics argue that developers cannot foresee all potential misuse of their models and that state regulations might conflict with future federal rules, creating additional burdens.
Despite these concerns, the immediate impact on innovation seems limited, as the number of companies affected in the short term is small. Much of the private sector’s worry is about future regulatory burdens on smaller companies.
Transitioning from Voluntary to Mandatory Regulation
SB 1047’s shift from voluntary self-regulation to mandatory legal liability has sparked significant debate. Federal AI regulation has so far relied on voluntary compliance and agency guidelines. For example, President Biden’s summit with tech executives and the National Institute of Standards and Technology’s AI risk management framework are voluntary.
Soft law mechanisms are practical when government resources and expertise are limited. Policymakers often struggle to keep up with technological advancements, leading to slower regulatory processes. As SB 1047 moves from soft law to legal liability, the opposition may signal future challenges for similar transitions at the federal level.
Crafting Future-Proof Legislation
SB 1047 attempts to create forward-looking legislation by using broader definitions to avoid overly prescriptive risk mitigation. However, this has allowed opponents to speculate about potential harsh enforcement. Consequently, a clause that would have covered more efficient future models was removed, reducing the bill’s applicability to small businesses and researchers.
Ultimately, SB 1047 is likely to remain contentious as it progresses through legislative channels. The debate around this bill underscores the complexity of creating agile, effective AI regulation. As broader regulatory discussions continue across the United States, policymakers must balance innovation with safety.
This episode highlights the difficulty of crafting legislation that keeps pace with technological advancements while ensuring public safety. As AI technology continues to evolve, the challenge for regulators will be to create policies that protect society without stifling innovation.