HomePoliticsThe White House’s Voluntary...

The White House’s Voluntary Framework for Ensuring Safe, Secure, and Trustworthy AI — AI: The Washington Report

Welcome to this week’s issue of AI: The Washington Report, a joint undertaking of Mintz and its government affairs affiliate, ML Strategies. The accelerating advances in artificial intelligence (“AI”) and the practical, legal, and policy issues AI creates have exponentially increased the federal government’s interest in AI and its implications. In these weekly reports, we hope to keep our clients and friends abreast of that Washington-focused set of potential legislative, executive, and regulatory activities.

In this issue, we discuss the White House’s July 21, 2023 announcement of a voluntary framework for ensuring safe, secure, and trustworthy AI (“July 21 Framework”). Our key takeaways are:

  1. Leaders of seven major technology companies, including Amazon, Anthropic, Google, Inflection, Meta, Microsoft, and OpenAI, met with President Biden to affirm their commitment to the July 21 Framework.
  2. The July 21 Framework consists of eight concrete commitments, each falling under one of three guiding principles: safety, security, and trust.
  3. The announcement of this framework reaffirms the key role of the executive branch in guiding the development of comprehensive AI regulation. The Biden administration intends to issue an executive order on AI in the near future. 

The White House’s Voluntary Framework for “Ensuring Safe, Secure, and Trustworthy AI”

On July 21, 2023, President Biden met with the leaders of seven major technology companies to secure their commitment to a voluntary framework for responsible AI innovation. “These commitments are real,” asserted Biden, “and they’re concrete.” The pronouncement of this framework (“July 21 Framework”), along with the announcement of a forthcoming executive order on AI, signals the resolve of the White House to maintain the strong role that the executive branch has played in guiding the development of American AI regulation. Absent substantive AI legislation from Congress, the frameworks and initiatives on autonomous systems released by the executive branch are, and will continue to be, the most concrete guidance from the federal government to businesses on AI development.

Companies representing the cutting edge of AI development, including Amazon, Anthropic, Google, Inflection, Meta, Microsoft, and OpenAI, agreed to the adoption of the July 21 Framework for AI risk management. Through these commitments, the White House aims to “make the most of AI’s potential” by “encouraging this industry to uphold the highest standards to ensure that innovation doesn’t come at the expense of Americans’ rights and safety.” This strategy of seeking to establish guardrails to encourage the development of AI within circumscribed boundaries is consistent with executive branch actions going back to the Obama administration, as well as more nascent efforts being discussed in Congress.

The voluntary framework consists of eight commitments, each falling under one of three guiding principles: safety, security, and trust.


  1. Rigorously test major AI models to ensure that they are resilient, do not promote bias, and do not pose an existential risk to society.[1]
  2. Promote the establishment of and participate in a forum or information-sharing mechanism through which companies can “develop, advance, and adopt shared standards and best practices for frontier AI safety.”[2]


  1. Designate unreleased model weights, or parameters that AI models use to make decisions, as core intellectual property protected by cybersecurity and insider threat detection systems.
  2. Establish systems that incentivize third parties to discover and report vulnerabilities in AI systems.


  1. Commit to labeling AI-generated content, except content that is “readily distinguishable from reality or that is designed to be readily recognizable as generated by a company’s AI system.”
  2. Report the capabilities, limitations, and risks of AI models to users.
  3. Support research on countering the risks posed by AI systems.
  4. Develop cutting-edge AI systems to help address major social issues, such as climate change and cancer detection.

Context of and Reaction to the July 21 Framework

As discussed in the inaugural issue of this newsletter, since the closing months of the Obama administration, regulatory efforts on AI in the United States have largely been the prerogative of the executive branch.

However, the past few months have seen Congress wade into the question of AI regulation, with members in the House and Senate releasing targeted regulatory proposals and bills that would establish study groups on AI regulation. Senate Majority Leader Chuck Schumer’s (D-NY) SAFE Innovation in the AI Age and Representative Ted Lieu’s (D-CA-36) National AI Commission Act each promise to lay the groundwork for robust, comprehensive AI legislation. In spite of this flurry of legislative activity, even the most optimistic estimates place the passage of comprehensive legislation to regulate AI at next year, if not the next session of Congress.

Given the rapid pace of technological development, and the level of alarm expressed by experts at the potential of AI to disrupt the economy, perpetuate bias, and promote societal risk, business leaders have been expressing the need for immediate guidance on AI development. Absent comprehensive legislation to regulate AI from Congress, business leaders have turned to executive branch efforts on AI. On July 19, top AI firms and research institutions published an open letter to Congressional leadership urging the funding of the National AI Research Resource (“NAIRR”), arguing that the body “would transform the U.S. research ecosystem and facilitate the partnerships needed to address societal-level problems.”

Within this context, the White House’s July 21 Framework can be seen as an attempt to provide AI developers with concrete guidance in the absence of enacted AI regulation. In a speech delivered following the announcement of the framework, President Biden asserted that the framework’s commitments are “going to help…the industry fulfill its fundamental obligation to Americans to develop safe, secure, and trustworthy technologies that benefit society and uphold our values and our shared values.”

Leaders from the seven technology companies that publically committed to the July 21 Framework lauded the administration’s efforts to guide AI development. Microsoft President Brad Smith asserted that “the voluntary commitments address the risks presented by advanced AI models and promote the adoption of specific practices…that will propel the whole ecosystem forward.” Kent Walker, President of Global Affairs of Google and Alphabet, hailed the July 21 Framework as “a milestone in bringing the industry together to ensure that AI helps everyone.”

But despite the enthusiasm for this framework from key industry leaders, President Biden has indicated that this and other non-binding initiatives are insufficient, and must be followed up by enforceable AI legislation. “Realizing the promise of AI by managing the risk is going to require some new laws, regulations, and oversight,” said Biden. To ease the pathway for these “new laws” on AI, the White House has announced the future release of an executive order “that will ensure the federal government is doing everything in its power to advance safe, secure, and trustworthy AI and manage its risks to individuals and society.”

Conclusion: Reasserting the Executive Branch’s Role in Developing AI Regulation

As the 118th Congress has rapidly produced a slate of AI proposals, experts surveying the field of AI regulation have understandably turned their attention towards the legislative branch. The announcement of the July 21 Framework, along with the open letter on the NAIRR, are reminders to all those interested in the development of AI regulation in the United States to pay attention to executive branch efforts as well.

As we have discussed in previous editions of this newsletter, the executive branch has accumulated experience in developing AI risk-managing frameworks, establishing bodies to oversee AI development, and collaborating with industry and academia to build regulatory competency regarding autonomous systems. Any comprehensive regulation on AI is likely to draw from the executive branch’s work on AI regulation, including the Blueprint for an AI Bill of Rights, the AI Risk Management Framework, and the forthcoming National Priorities for Artificial Intelligence.

As the executive and legislative branches work towards the development of comprehensive AI regulation, we will continue to monitor, analyze, and issue reports.



[1] Specifically, this commitment entails the use of “red-teaming,” a strategy whereby an entity designates a team to emulate the behavior of an adversary attempting to break or exploit the entity’s technological systems. As the red team discovers vulnerabilities, the entity patches them, making their technological systems resilient to actual adversaries.
[2] On July 26, 2023, four of the companies that agreed to the July 21 Framework (Microsoft, Anthropic, Google, and OpenAI) announced the launch of such a body. The “Frontier Model Forum” is “a new industry body focused on ensuring safe and responsible development of frontier AI models.”


Subscribe To Viewpoints

Source link

Most Popular


Please enter your comment!
Please enter your name here

More from Author

Read Now

Selloff Gains Momentum as September Losses Mount: Markets Wrap

Stock Market Today: Dow, S&P Live Updates for September 28  Bloomberg10-year Treasury yield reaches level not seen in more than 15 years  CNBCJapan's 10-year bond yield hits decade high on US yield surge  NasdaqView Full Coverage on Google News Source link

3-team tie? 4-team tie?! Chaotic playoff scenarios, explained

As we head toward the end of the regular season on Sunday, several postseason races are poised to come down to the wire, and there are a number of absolutely chaotic scenarios still in play. When two teams finish in a tie for one spot, be it...

A key US government surveillance tool should face new limits, a divided privacy oversight board says

The FBI and other government agencies should be required to get court approval before reviewing the communications of U.S. citizens collected through a secretive foreign surveillance program, a sharply divided privacy oversight board recommended on Thursday. The recommendation came in a report from a three-member Democratic...

Britney Spears receives welfare check after posting knife video

By Deirdre Durkan-simonds and Adam Levy For Dailymail.com ...

Epic Games Is Cutting About 900 Jobs, or 16% of Staff

Business Of Sports If the only thing you know about sports is who wins and who loses, you are missing the highest stakes action of all. The business owners that power this multibillion dollar industry are changing, and a new era of the business of...

The S&P 500 is brushing up against ‘the mother of all trend lines.’ What happens next could make or break the market.

After what’s shaping up to the worst month for stocks so far this year, some investors have been zeroing in on a chart indicating a showdown is coming for the S&P 500 SPX. Here’s the chart...

Michael Gambon, Dumbledore in the ‘Harry Potter’ Films, Dies at 82

Michael Gambon, who played Professor Dumbledore in the “Harry Potter” films and was widely hailed as one of the greatest British actors, has died. He was 82.Mr. Gambon’s family confirmed his death in a brief statement issued on Thursday through a public relations company. “Michael...

Colin Kaepernick’s humble Jets letter is another reminder to the NFL

Colin Kaepernick will not give up, and he is not going away.On Tuesday, his friend and rap superstar J. Cole posted to social media a letter he says Kaepernick wrote to New York Jets general manager Joe Douglas in which Kaepernick asks the quarterback-needy franchise to...

Micron Earnings Top Views, But Memory-Chip Maker Falls On Mixed Outlook

Memory-chip maker Micron Technology (MU) isn't out of the woods amid a cyclical industry downturn. MU stock fell early Thursday after Micron reported fiscal fourth-quarter profit and revenue that topped estimates while guidance came in mixed. X The memory chip maker reported August quarter earnings after the market...

Israel’s Supreme Court convenes to decide on law that could determine Netanyahu’s fate | CNN

Editor’s Note: A version of this story appears in CNN’s Meanwhile in the Middle East newsletter, a three-times-a-week look inside the region’s biggest stories....

Raspberry Pi 5 finally revealed alongside major upgrades – Dexerto

Joel Loynds  ❘   Published: 2023-09-28T11:27:46  ❘   Updated: 2023-09-28T11:28:00 Raspberry Pi has finally launched the Pi 5, its next major single-board computer. It has been four years since its last major release. The Raspberry Pi 5 is ready to hit the market, as the company prepares to...