OpenAI, Anthropic To Share AI Models With US Government

Both OpenAI and Anthropic have signed deals with the United States government for research, testing and evaluation of their artificial intelligence models.

The US AI Safety Institute announced “agreements that enable formal collaboration on AI safety research, testing and evaluation with both Anthropic and OpenAI.”

Essentially the agreement will let the US government access major new AI models before their general release, in order to help improve their safety. This is a core goal of both the British and American AI Safety Institutes.

AI safety

In April 2024 both the United Kingdom and United States had signed a landmark agreement to work together on testing advanced artificial intelligence (AI).

That agreement saw the UK and US AI Safety Institutes pledge to work seamlessly with each other, partnering on research, safety evaluations, and guidance for AI safety.

It comes after last year’s AI Safety Summit in the UK, where big name companies including Amazon, Google, Facebook parent Meta Platforms, Microsoft and ChatGPT developer OpenAI all agreed to voluntary safety testing for AI systems, resulting in the so called ‘Bletchley Declaration.’

That agreement was backed by the EU and 10 countries including China, Germany, France, Japan, the UK and the US.

OpenAI, Anthropic agreement

Now according to the US AI Safety Institute, each company’s Memorandum of Understanding establishes the framework for it “to receive access to major new models from each company prior to and following their public release. The agreements will enable collaborative research on how to evaluate capabilities and safety risks, as well as methods to mitigate those risks.”

“Safety is essential to fueling breakthrough technological innovation,” said Elizabeth Kelly, director of the US AI Safety Institute. “With these agreements in place, we look forward to beginning our technical collaborations with Anthropic and OpenAI to advance the science of AI safety.”

“These agreements are just the start, but they are an important milestone as we work to help responsibly steward the future of AI,” said Kelly.

Additionally, the US AI Safety Institute plans to provide feedback to Anthropic and OpenAI on potential safety improvements to their models, in close collaboration with its partners at the UK AI Safety Institute.

Tom Jowitt

Tom Jowitt is a leading British tech freelancer and long standing contributor to Silicon UK. He is also a bit of a Lord of the Rings nut...

Recent Posts

TSMC Begins 4nm Chip Production In Arizona

TSMC begins production of advanced 4nm chips at Arizona plant as US seeks to bring…

26 mins ago

China Chip Imports Surge Ahead Of New Export Controls

China's semiconductor imports grow by double-digits in 2024 ahead of new US export controls that…

57 mins ago

US Rules Divide World To Conquer China’s AI

New US export controls divide world into three tiers as outgoing administration seeks to cut…

1 hour ago

Apple Board Advises Against Plan To End Diversity Programmes

Apple board advises investors to vote against shareholder proposal to end diversity programmes as Meta,…

2 hours ago

Technology Secretary Calls Online Safety Act ‘Unsatisfactory’

Technology secretary Peter Kyle admits Online Safety Act falls short on protection from social harm,…

2 hours ago

Blue Origin Aborts Test Flight Minutes Before Launch

Jeff Bezos' Blue Origin cancels New Glenn certification flight at last minute due to unspecified…

10 hours ago