US President Joe Biden has signed an ambitious executive order on artificial intelligence that seeks to balance the needs of cutting-edge technology companies with national security and consumer rights, creating an early set of guardrails that could be fortified by legislation and global agreements.
Before signing the order, Mr Biden said AI is driving change at “warp speed” and carries tremendous potential as well as perils.
“AI is all around us,” Mr Biden said.
“To realise the promise of AI and avoid the risk, we need to govern this technology.”
The order is an initial step that is meant to ensure that AI is trustworthy and helpful, rather than deceptive and destructive.
The order – which will probably need to be augmented by congressional action – seeks to steer how AI is developed so that companies can profit without putting public safety in jeopardy.
Using the Defence Production Act, the order requires leading AI developers to share safety test results and other information with the government.
The National Institute of Standards and Technology is to create standards to ensure AI tools are safe and secure before public release.
The Commerce Department is to issue guidance to label and watermark AI-generated content to help differentiate between authentic interactions and those generated by software.
The extensive order touches on matters of privacy, civil rights, consumer protections, scientific research and worker rights.
White House chief of staff Jeff Zients recalled Mr Biden giving his staff a directive when formulating the order to move with urgency.
“We can’t move at a normal government pace,” Mr Zients said the Democratic president told him.
“We have to move as fast, if not faster, than the technology itself.”
In Mr Biden’s view, the government was late to address the risks of social media and now US youth are grappling with related mental health issues.
AI has the positive ability to accelerate cancer research, model the impacts of climate change, boost economic output and improve government services among other benefits.
But it could also warp basic notions of truth with false images, deepen racial and social inequalities and provide a tool to scammers and criminals.
With the European Union nearing final passage of a sweeping law to rein in AI harms and Congress still in the early stages of debating safeguards, the Biden administration is “stepping up to use the levers it can control”, said digital rights advocate Alexandra Reeve Givens, president of the Centre for Democracy & Technology.
“That’s issuing guidance and standards to shape private sector behaviour and leading by example in the federal government’s own use of AI.”
The order builds on voluntary commitments already made by technology companies.
It is part of a broader strategy that administration officials say also includes congressional legislation and international diplomacy, a sign of the disruptions already caused by the introduction of new AI tools such as ChatGPT that can generate text, images and sounds.
The guidance within the order is to be implemented and fulfilled over the range of 90 days to 365 days.
Last Thursday, Mr Biden gathered his aides in the Oval Office to review and finalise the executive order, a 30-minute meeting that stretched to 70 minutes, despite other pressing matters, including the mass shooting in Maine, the Israel-Hamas war and the selection of a new House speaker.
Mr Biden was profoundly curious about the technology in the months of meetings that led up to drafting the order.
His science advisory council focused on AI at two meetings and his cabinet discussed it at two meetings.
The president also pressed tech executives and civil society advocates about the technology’s capabilities at multiple gatherings.
“He was as impressed and alarmed as anyone,” deputy White House chief of staff Bruce Reed said in an interview.
“He saw fake AI images of himself, of his dog. He saw how it can make bad poetry. And he’s seen and heard the incredible and terrifying technology of voice cloning, which can take three seconds of your voice and turn it into an entire fake conversation.”
The issue of AI was seemingly inescapable for Mr Biden.
At Camp David one weekend, he relaxed by watching the Tom Cruise film Mission: Impossible – Dead Reckoning Part One.
The film’s villain is a sentient and rogue AI known as “the Entity” that sinks a submarine and kills its crew in the movie’s opening minutes.
“If he hadn’t already been concerned about what could go wrong with AI before that movie, he saw plenty more to worry about,” said Mr Reed, who watched the film with the president.
Governments around the world have raced to establish protections, some of them tougher than Mr Biden’s directives.
After more than two years of deliberation, the EU is putting the final touches on a comprehensive set of regulations that targets the riskiest applications with the tightest restrictions.
China, a key AI rival to the US, has also set some rules.
UK Prime Minister Rishi Sunak hopes to carve out a prominent role for Britain as an AI safety hub at a summit starting on Wednesday that US vice president Kamala Harris plans to attend.
And on Monday, officials from the Group of Seven major industrial nations agreed to a set of AI safety principles and a voluntary code of conduct for AI developers.
Finally, a word on the future.
AI brings huge opportunities to transform our public services, but also considerable risks.
On Thursday I set out how we will best manage these risks ahead of the world’s first #AISafetySummit – which the UK is hosting next week. pic.twitter.com/CG43SnnYYX— Rishi Sunak (@RishiSunak) October 27, 2023
The US, particularly its West Coast, is home to many of the leading developers of cutting-edge AI technology, including tech giants Google, Meta and Microsoft, and AI-focused start-ups such as OpenAI, maker of ChatGPT.
The White House took advantage of that industry weight earlier this year when it secured commitments from those companies to implement safety mechanisms as they build new AI models.
But the White House also faced significant pressure from Democratic allies, including labour and civil rights groups, to make sure its policies reflected their concerns about AI’s real-world harms.
Suresh Venkatasubramanian, a former Biden administration official who helped craft principles for approaching AI, said one of the biggest challenges within the federal government has been what to do about law enforcement’s use of AI tools, including at US borders.
“These are all places where we know that the use of automation is very problematic, with facial recognition, drone technology,” Venkatasubramanian said.
Facial recognition technology has been shown to perform unevenly across racial groups, and has been tied to mistaken arrests.
While the EU’s forthcoming AI law is set to ban real-time facial recognition in public, Mr Biden’s order appears to simply ask for federal agencies to review how they are using AI in the criminal justice system, falling short of the stronger language sought by some activists.
The American Civil Liberties Union is among the groups that met with the White House to try to ensure “we’re holding the tech industry and tech billionaires accountable” so that algorithmic tools “work for all of us and not just a few”, said ReNika Moore, director of the ACLU’s racial justice programme, who attended Monday’s signing.
After seeing the text of the order, Moore applauded how it addressed discrimination and other AI harms in workplaces and housing, but said the administration “essentially kicks the can down the road” in protecting people from law enforcement’s growing use of the technology.