Āé¶¹“«Ć½

Technology

Biden executive order: How the US is trying to tame AI

US president Joe Biden has announced an executive order that establishes ambitious guidelines on safety and security for artificial intelligence, but it will still need political will to put regulatory teeth and resources behind it

By Jeremy Hsu

30 October 2023

US president Joe Biden announced new guidelines for the safe development of AI

US president Joe Biden announced new guidelines for the safe development of AI

AFP via Getty Images

An executive order on artificial intelligence issued by US president Joe Biden aims to show leadership in regulating AI safety and security – but most of the follow-through will require action from US lawmakers and the voluntary goodwill of tech companies.

µž¾±»å±š²Ō’s directs a wide array of US government agencies to develop guidelines for testing and using AI systems, including having the National Institute of Standards and Technology set benchmarks for ā€œred team testingā€ to probe for potential AI vulnerabilities prior to public release.

ā€œThe language in this executive order and in the White House’s discussion of it suggests an interest in being seen as the most aggressive and proactive in addressing AI regulation,ā€ says at Cornell University in New York.

It is probably ā€œno coincidenceā€ that µž¾±»å±š²Ō’s executive order came out just before the UK government convened its own AI summit, says Kreps. But she cautioned that the executive order alone will not have much impact unless the US Congress can produce bipartisan legislation and resources to back it up – something that she sees as unlikely during the 2024 US presidential election year.

This follows a trend of non-binding actions by the Biden administration on AI. For example, last year the administration issued a and it recently solicited voluntary pledges from major companies developing AI, says at the University of Bologna, Italy.

One potentially impactful part of µž¾±»å±š²Ō’s executive order covers foundation models – large AI models trained on huge datasets – if they pose ā€œa serious risk to national security, national economic security, or national public health and safetyā€. The order uses another piece of legislation called the Defense Production Act to require companies developing such AIs to notify the federal government about the training process and share the results of all red team safety testing.

Free newsletter

Sign up to The Weekly

The best of Āé¶¹“«Ć½, including long-reads, culture, podcasts and news, each week.

Āé¶¹“«Ć½. Science news and long reads from expert journalists, covering developments in science, technology, health and the environment on the website and the magazine.

Such AIs could include OpenAI’s GPT-3.5 and GPT-4 models, which are behind ChatGPT, Google’s PaLM 2 model, which supports the company’s Bard AI chatbot, and Stability AI’s Stable Diffusion model, which generates images. ā€œIt would force companies that have been very closed-off about how their models work to crack open their black boxes,ā€ says Hine.

But Hine said ā€œthe devil is in the detailsā€ when it comes to how the US government defines which foundation models pose a ā€œserious riskā€. Similarly, Kreps questioned the ā€œqualifiers and ambiguitiesā€ of the executive order’s wording; the document is unclear about how it defines ā€œfoundation modelā€ and who determines what qualifies as a threat.

The US also still lacks the type of strong data protection laws seen in the European Union and China. Similar laws could support AI regulations, says Hine. She pointed out that China has focused on implementing ā€œtargeted, vertical laws addressing specific aspects of AIā€, such as generative AIs or facial recognition use. The European Union, on the other hand, has been working to create political consensus among its members on a broad horizontal approach covering all aspects of AI.

ā€œ[The US] has the [AI] development chops, but it doesn’t have much concrete regulation to stand on,ā€ says Hine. ā€œWhat it does have is strong statements about ā€˜AI with democratic values’ and agreements to cooperate with allied countries.ā€

Topics:

Sign up to our weekly newsletter

Receive a weekly dose of discovery in your inbox. We'll also keep you up to date with Āé¶¹“«Ć½ events and special offers.

Sign up
Piano Exit Overlay Banner Mobile Piano Exit Overlay Banner Desktop