The One Big Beautiful Bill Act would ban states from regulating AI

Share

Buried in the Republican budget bill is a proposal that will radically change how artificial intelligence develops in the U.S., according to both its supporters and critics. The provision would ban states from regulating AI for the next decade.

Opponents say the moratorium is so broadly written that states wouldn’t be able to enact protections for consumers affected by harmful applications of AI, like discriminatory employment tools, deepfakes, and addictive chatbots.

Instead, consumers would have to wait for Congress to pass its own federal legislation to address those concerns. Currently it has no draft of such a bill. If Congress fails to act, consumers will have little recourse until the end of the decade-long ban, unless they decide to sue companies responsible for alleged harms.

SEE ALSO:

AI has entered the therapy session — and it’s recording you

Proponents of the proposal, which include the Chamber of Commerce, say that it will ensure America’s global dominance in AI by freeing small and large companies from what they describe as a burdensome patchwork of state-by-state regulations.

But many say the provision’s scope, scale, and timeline is without precedent — and a big gift to tech companies, including ones that donated to President Donald Trump.

This week, a coalition of 77 advocacy organizations, including Common Sense Media, Fairplay, and the Center For Humane Technology, called on congressional leadership to jettison the provision from the GOP-led budget.

“By wiping out all existing and future state AI laws without putting new federal protections in place, AI companies would get exactly what they want: no rules, no accountability, and total control,” the coalition wrote in an open letter.

Some states already have AI-related laws on the books. In Tennessee, for example, a state law known as the ELVIS Act was written to prevent the impersonation of a musician’s voice using AI. Republican Sen. Marsha Blackburn, who represents Tennessee in Congress, recently hailed the act’s protections and said a moratorium on regulation can’t come before a federal bill.

Other states have drafted legislation to address specific emerging concerns, particularly related to youth safety. California has two bills that would place guardrails on AI companion platforms, which advocates say are currently not safe for teens.

One of the bills specifically outlaws high-risk uses of AI, including “anthropomorphic chatbots that offer companionship” to children and will likely lead to emotional attachment or manipulation.

Camille Carlton, policy director at the Center for Humane Technology, says that while remaining competitive amidst greater regulation may be a valid concern for smaller AI companies, states are not proposing or passing expansive restrictions that would fundamentally hinder them. Nor are they targeting companies’ ability to innovate in areas that would make America truly world-leading, like in health care, security, and the sciences. Instead, they are focused on key areas of safety, like fraud and privacy. They’re also tailoring bills to cover larger companies or offering tiered responsibilities appropriate to a company’s size.

Historically, tech companies have lobbied against certain state regulations, arguing that federal legislation would be preferable, Carlton says. But then they lobby Congress to water down or kill their own regulatory bills too, she notes.

Arguably, that’s why Congress hasn’t passed any major encompassing consumer protections related to digital technology in the decades since the internet became ascendant, Carlton says. She adds that consumers may see the same pattern play out with AI, too.

Some experts are particularly worried that a hands-off approach to regulating AI will only repeat what happened when social media companies first operated without much interference. They say that came at the cost of youth mental health.

Gaia Bernstein, a tech policy expert and professor at the Seton Hall University School of Law, says that states have increasingly been at the forefront of regulating social media and tech companies, particularly with regard to data privacy and youth safety. Now they’re doing the same for AI.

Bernstein says that in order to protect kids from excessive screen time and other online harms, states also need to regulate AI, because of how frequently the technology is used in algorithms. Presumably, the moratorium would prohibit states from doing so.

“Most protections are coming from the states. Congress has largely been unable to do anything,” Bernstein says. “If you’re saying that states cannot do anything, then it’s very alarming, because where are any protections going to come from?”

Source : The One Big Beautiful Bill Act would ban states from regulating AI