AI Safety Institute Launches AI Model Safety Testing Tool Platform

Business developers can use Inspect to test AI models before public release.

Ben Wodecki

May 17, 2024

3 Min Read
AI safety illustration
Alamy

This article originally appeared in AI Business.

The UK’s AI Safety Institute has launched a new platform allowing businesses to test their AI models before launching them publicly.

The platform, named Inspect, is a software library designed to assess AI model capabilities, scoring them on areas like reasoning and autonomous abilities.

There’s an absence of safety testing tools available to developers today. MLCommons unveiled a large language model-focused benchmark for safety testing last month.

Inspect was built to fill the gap, launching in open source so anyone can use it to test their AI models.

Businesses can use Inspect to evaluate prompt engineering for their AI models and external tool usage. The tool also contains evaluation datasets containing labeled samples so developers can examine in detail the data being used to test the model.

It’s designed to be easy to use, with explainers for running the various tests provided throughout, including if a model is hosted in a cloud environment like AWS Bedrock.

The decision to open source the testing tool would enable developers worldwide to conduct more effective AI evaluations, according to the Safety Institute.

“As part of the constant drumbeat of UK leadership on AI safety, I have cleared the AI Safety Institute’s testing platform to be open sourced,” said Michelle Donelan, UK technology secretary. “The reason I am so passionate about this and why I have open sourced Inspect, is because of the extraordinary rewards we can reap if we grip the risks of AI.”

The Safety Institute said it plans to develop open source testing tools beyond Inspect in the future. The agency will be working on related projects with its US counterpart after it penned a joint working agreement in April.

“Successful collaboration on AI safety testing means having a shared, accessible approach to evaluations and we hope Inspect can be a building block for AI Safety Institutes, research organizations and academia,” said Ian Hogarth, the AI Safety Institute’s chair. “We hope to see the global AI community using Inspect to not only carry out their own model safety tests but to help adapt and build upon the open source platform so we can produce high-quality evaluations across the board.”

The success of the Safety Institute's new platform can only be measured by the number of companies who have already committed to using the testing tool, according to Amanda Brock, CEO of OpenUK.

“With the UK's slow position on regulating, this platform simply has to be successful for the UK to have a place in the future of AI,” Brock said. “All eyes will now be on South Korea and the next Safety Summit to see how this is received by the world.”

“The ability of Inspect to evaluate a wide range of AI capabilities and provide a safety score empowers organizations, big and small, to not only harness AI's potential but also ensure it is used responsibly and safely,” said Veera Siivonen, Saidot’s chief commercial officer. “This is a step towards democratizing AI safety, a move that will undoubtedly drive innovation while safeguarding against the risks associated with advanced AI systems.”

About the Author

Ben Wodecki

AI Business

Ben Wodecki is assistant editor at AI Business, a publication dedicated to the latest trends in artificial intelligence.

Subscribe to the Data Center Knowledge Newsletter
Get analysis and expert insight on the latest in data center business and technology delivered to your inbox daily.

You May Also Like