Read the text below.
California lawmakers are considering legislation that would require artificial intelligence companies to test their systems and add safety measures so they can’t be potentially manipulated to wipe out the state’s electric grid or help build chemical weapons—scenarios that experts say could be possible in the future as technology evolves at warp speed.
The first-of-its-kind bill aims to reduce risks created by AI. It is fiercely opposed by tech companies, including Meta, the parent company of Facebook and Instagram, and Google. They say the regulations take aim at developers and instead should be focused on those who use and exploit the AI systems for harm.
Democratic state Sen. Scott Wiener, who authors the bill, said the proposal would provide reasonable safety standards by preventing “catastrophic harms” from extremely powerful AI models that may be created in the future. The requirements would only apply to systems that cost more than $100 million in computing power to train. No current AI models have hit that as of July.
“This is not about smaller AI models,” Wiener said at a recent legislative hearing. “This is about incredibly large and powerful models that, as far as we know, do not exist today but will exist in the near future.”
Democratic Gov. Gavin Newsom has touted California as an early AI adopter and regulator, saying the state could soon deploy generative AI tools to address highway congestion, make roads safer and provide tax guidance. At the same time, his administration is considering new rules against AI discrimination in hiring practices. He declined to comment on the bill but has warned that overregulation could put the state in a “perilous position.”
The proposal, supported by some of the most renowned AI researchers, would also create a new state agency to oversee developers and provide best practices. The state attorney general also would be able to pursue legal actions in case of violations.
A growing coalition of tech companies argues the requirements would discourage companies from developing large AI systems or keeping their technology open-source.
This article was provided by The Associated Press.