US president Joe Biden announced new guidelines for the safe development of AI AFP via Getty Images
An executive order on artificial intelligence issued by US president Joe Biden aims to show leadership in regulating AI safety and security ā but most of the follow-through will require action from US lawmakers and the voluntary goodwill of tech companies.
µž¾±»å±š²Ōās directs a wide array of US government agencies to develop guidelines for testing and using AI systems, including having the National Institute of Standards and Technology set benchmarks for āred team testingā to probe for potential AI vulnerabilities prior to public release.
āThe language in this executive order and in the White Houseās discussion of it suggests an interest in being seen as the most aggressive and proactive in addressing AI regulation,ā says at Cornell University in New York.
Advertisement
It is probably āno coincidenceā that µž¾±»å±š²Ōās executive order came out just before the UK government convened its own AI summit, says Kreps. But she cautioned that the executive order alone will not have much impact unless the US Congress can produce bipartisan legislation and resources to back it up ā something that she sees as unlikely during the 2024 US presidential election year.
This follows a trend of non-binding actions by the Biden administration on AI. For example, last year the administration issued a and it recently solicited voluntary pledges from major companies developing AI, says at the University of Bologna, Italy.
One potentially impactful part of µž¾±»å±š²Ōās executive order covers foundation models ā large AI models trained on huge datasets ā if they pose āa serious risk to national security, national economic security, or national public health and safetyā. The order uses another piece of legislation called the Defense Production Act to require companies developing such AIs to notify the federal government about the training process and share the results of all red team safety testing.
Free newsletter
Sign up to The Weekly
The best of Āé¶¹“«Ć½, including long-reads, culture, podcasts and news, each week.

Such AIs could include OpenAIās GPT-3.5 and GPT-4 models, which are behind ChatGPT, Googleās PaLM 2 model, which supports the companyās Bard AI chatbot, and Stability AIās Stable Diffusion model, which generates images. āIt would force companies that have been very closed-off about how their models work to crack open their black boxes,ā says Hine.
But Hine said āthe devil is in the detailsā when it comes to how the US government defines which foundation models pose a āserious riskā. Similarly, Kreps questioned the āqualifiers and ambiguitiesā of the executive orderās wording; the document is unclear about how it defines āfoundation modelā and who determines what qualifies as a threat.
The US also still lacks the type of strong data protection laws seen in the European Union and China. Similar laws could support AI regulations, says Hine. She pointed out that China has focused on implementing ātargeted, vertical laws addressing specific aspects of AIā, such as generative AIs or facial recognition use. The European Union, on the other hand, has been working to create political consensus among its members on a broad horizontal approach covering all aspects of AI.
ā[The US] has the [AI] development chops, but it doesnāt have much concrete regulation to stand on,ā says Hine. āWhat it does have is strong statements about āAI with democratic valuesā and agreements to cooperate with allied countries.ā
Topics:


