OpenAI will show how models do on hallucination tests

0
4


Sam Altman Co-founder and CEO of OpenAI speaks during the Italian Tech Week 2024 at OGR Officine Grandi Riparazioni on September 25, 2024 in Turin, Italy. 

Stefano Guidi | Getty Images News | Getty Images

OpenAI on Wednesday announced a new “safety evaluations hub,” a webpage where it will publicly display artificial intelligence models’ safety results and how they perform on tests for hallucinations, jailbreaks and harmful content, such as “hateful content or illicit advice.”

OpenAI said it used the safety evaluations “internally as one part of our decision making about model safety and deployment,” and that while system cards release safety test results when a model is launched, OpenAI will from now on “share metrics on an ongoing basis.”

“We will update the hub periodically as part of our ongoing company-wide effort to communicate more proactively about safety,” OpenAI wrote on the webpage, adding that the safety evaluations hub does not reflect the full safety efforts and metrics and instead shows a “snapshot.”

The news comes after CNBC reported earlier Wednesday that tech companies that are leading the way in artificial intelligence are prioritizing products over research, according to industry experts who are sounding the alarm about safety.

CNBC reached out to OpenAI and other AI labs mentioned in the story well before it was published.

Read more CNBC news on OpenAI

OpenAI recently sparked some online controversy for not running certain safety evaluations on the final version of its o1 AI model.

In a recent interview with CNBC, Johannes Heidecke, OpenAI’s head of safety systems, said the company ran its preparedness evaluations on near-final versions of the o1 model, and that minor variations to the model that took place after those tests wouldn’t have contributed to significant jumps in its intelligence or reasoning and thus wouldn’t require additional evaluations.

Still, Heidecke acknowledged in the interview that OpenAI missed an opportunity to more clearly explain the difference.

Meta, which was also mentioned in CNBC’s reporting on AI safety and research, also made an announcement Wednesday.

The company’s Fundamental AI Research team released new joint research with the Rothschild Foundation Hospital and an open dataset for advancing molecular discovery.

“By making our research widely available, we aim to provide easy access for the AI community and help foster an open ecosystem that accelerates progress, drives innovation, and benefits society as a whole, including our national research labs,” Meta wrote in a blog post announcing the research advancements.


LEAVE A REPLY

Please enter your comment!
Please enter your name here