Technology

OpenAI’s latest AI models have a new safeguard to prevent biorisks

Openai says it has published a new system to monitor the latest models of thinking about artificial intelligence, O3 and O4-MINI, for demands related to biological and chemical threats. The regime aims to prevent models from advising a person who can direct someone regarding the implementation of possible harmful attacks, according to the Openai Safety Report.

O3 and O4-MINI represents a meaningful increase in the previous models of Openai, as the company says, and thus new risk in the hands of bad actors. According to Openai’s internal standards, O3 is more skilled in answering questions about creating certain types of biological threats in particular. For this reason-and reduce the other risks-OpenAi the new monitoring system, which the company describes as “monitoring of safety thinking.”

The screen, which is trained on request, is operated on Openai content policies, at the head of O3 and O4-MINI. It is designed to determine the claims related to biological and chemical risks and to direct models to refuse advice on these topics.

To create a foundation, Openai had spent a red difference about 1000 hours in a sign of “unsafe” conversations related to the O3 and O4-MINI. During a test in which Openai simulated “the logic of blocking” to monitor its safety, the models refused to respond to the risky demands of 98.7 % of the time, according to Openai.

Openai admits that his test did not explain people who may try new claims after being banned by the screen, which is why the company says it will continue to be partially dependent on human monitoring.

O3 and O4-MINI do not cross the “high risk” threshold for biomedics, according to the company. However, compared to O1 and GPT-4, Openai says that early versions of O3 and O4-MINI have proven to be more useful in answering questions about the development of biological weapons.

The graph from the O3 and O4-MINI system (screenshot: Openai)

The company actively follows how its models can facilitate malicious users to develop chemical and biological threats, according to the recently updated alert framework of Openai.

Openai is increasingly dependent on automated systems to mitigate the risk of their models. For example, to prevent the original GPT-4O from creating sexual assault on children (CSAM), Openai says it uses a screen similar screen to those that the company published for O3 and O4-MINI.

However, many researchers raised Oblay’s concerns that do not give priority to safety as much as it should. Metr, one of the company’s red partners, said he had little time to test O3 on a criterion for deception. Meanwhile, Openai decided not to issue a safety report for the GPT-4.1 model, which was launched earlier this week.

Don’t miss more hot News like this! Click here to discover the latest in Technology news!

2025-04-16 21:12:00

Related Articles

Back to top button