LOGO

OpenAI May Loosen AI Safeguards Amidst Competition

April 15, 2025
OpenAI May Loosen AI Safeguards Amidst Competition

OpenAI Updates its AI Safety Framework

OpenAI has recently revised its Preparedness Framework – the internal system utilized for evaluating the safety of artificial intelligence models. This framework guides the implementation of necessary protective measures throughout the stages of development and deployment.

The update details that OpenAI reserves the right to “adjust” its safety standards should a rival AI development laboratory introduce a “high-risk” system lacking comparable safety features.

Increased Competitive Pressures and Safety Concerns

This modification acknowledges the escalating competitive pressures faced by commercial AI developers striving for rapid model deployment. OpenAI has faced accusations of prioritizing speed over safety, and of delays in publishing comprehensive reports detailing its safety testing procedures.

Recently, twelve former OpenAI employees submitted a legal brief supporting Elon Musk’s case against OpenAI. They argued that the company might be incentivized to further reduce safety measures if its planned corporate restructuring is finalized.

Policy Adjustments and Risk Assessment

Anticipating potential criticism, OpenAI asserts that any policy adjustments would be considered carefully. The company maintains it would maintain safeguards at “a level more protective” than currently implemented.

According to a blog post released on Tuesday, OpenAI stated, “If another frontier AI developer releases a high-risk system without comparable safeguards, we may adjust our requirements.” However, they would first confirm a genuine shift in the risk landscape, publicly announce the adjustment, and verify it doesn’t substantially elevate the overall risk of significant harm.

Automated Evaluations and Development Speed

The revised Preparedness Framework also highlights a growing reliance on automated evaluations to accelerate product development. While human-led testing isn’t being entirely abandoned, OpenAI has developed “a growing suite of automated evaluations” designed to maintain pace with a quicker release schedule.

However, some reports suggest a different reality. The Financial Times reported that testers were given less than a week to conduct safety checks on an upcoming major model – a significantly shorter timeframe than previous releases.

Sources cited by the publication also allege that many of OpenAI’s safety tests are now performed on older model versions, rather than those ultimately released to the public.

OpenAI has publicly refuted claims of compromising on safety standards.

Model Risk Categorization

Further changes to OpenAI’s framework concern the categorization of models based on their inherent risk levels. This includes models capable of concealing their functionalities, circumventing safeguards, resisting deactivation, and even self-replication.

OpenAI will now assess models against two key thresholds: “high” capability and “critical” capability.

A model exhibiting “high” capability is defined as one that could “amplify existing pathways to severe harm.” Those demonstrating “critical” capability are considered to “introduce unprecedented new pathways to severe harm.”

“Systems reaching high capability must have safeguards that sufficiently minimize the associated risk of severe harm before deployment,” OpenAI explained in its blog post. “Systems reaching critical capability also require safeguards that sufficiently minimize associated risks during development.”

Framework Update Timeline

These updates represent the first revisions to OpenAI’s Preparedness Framework since 2023.

#OpenAI#AI safety#artificial intelligence#AI competition#AI safeguards#high-risk AI