Physical Address
304 North Cardinal St.
Dorchester Center, MA 02124
Physical Address
304 North Cardinal St.
Dorchester Center, MA 02124
On Friday, Openai launched a new AI model “Thinking”, O3-MINI, newer in the company Q Family Models of Thinking.
Openai The first examination of the model in December Besides a more capable system called O3, but launch comes in a pivotal moment of the company, whose ambitions – and challenges – are growing day by day.
Openai fights the perception that it waives the land of artificial intelligence Chinese companies such as DepsicThe Openai allei may have stolen its IP address. He was trying that The beach of its relationship with Washington As is one at the same time The ambitious data center projectAnd It is also said that he sets the foundation For one of the largest financing rounds in history.
Which leads us to O3-MINI. Openai shows its new model as “strong” and “affordable”.
“The launch of the day represents an important step towards expanding the possibility of advanced artificial intelligence in the service of our mission,” Openai spokesman told Techcrunch.
Unlike most large language models, thinking models such as O3-MINI are completely achieved by the facts of the right to provide results. This helps them Avoid This is usually a journey to models. These thinking models take a little longer to reach solutions, but the comparison is that they tend to be more reliable-and if not perfect-in areas such as physics.
O3-MINI is set for STEM problems, specifically for programming, mathematics and science. Openai claims that the model is equal to the O1, O1 and O1-MINI family, in terms of capabilities, but it works faster and costs less.
The company claimed that external laboratories prefer the O3-MINI answers over those in O1-MINI more than half the time. O3-MINI seems to have also committed 39 % of “major errors” on “difficult realistic questions” in A/b tests. Video of O1-MINI, and produced “more clear” responses with 24 % faster answers.
O3-Mini will be available to all users across Chatgpt Starting on Friday, users who pay for Openai’s Chatgpt Plus and team plans will get up to 150 inquiries per day. Chatgpt PRO will receive unlimited access, and O3-MINI will come to Chatgpt Enterprise and Chatgpt Edu customers per week. (There is no word in Chatgpt government yet).
Users with distinct plans can determine the O3-MINI using the ChatGPT drop-down menu. Free users can click or click the new “Reason” button in the chat bar, or get the “Reglerate” Chatgpt.
Starting on Friday, O3-MINI will also be available via API from Openai to identify developers, but at first you will not get support for image analysis. Devs can choose the “thinking voltage” level (low, medium or high) to get the O3-MINI “to think hard” based on their needs of their use and use.
The price of O3-MINI is $ 0.55 per million temporary stored input codes and $ 4.40 per million output symbols, as a million icons are equivalent to approximately 750,000 words. This is 63 % cheaper than O1-MINI, competitive with the prices of the thinking model in Deepsek R1. Deepseek receives $ 0.14 per million temporary stored input codes and $ 2.19 per million output symbols to reach R1 through its application programming interface.
In ChatGPT, O3-MINI has been assigned to an average thinking voltage, which Openai says, “A balanced comparison between speed and accuracy.” Users paid will have the option to choose “O3-MINI-He
Regardless of the version of the O3-Mini Chatgpt users, the model will work with the Search to find updated answers with links to relevant web sources. Openai warns that the job is a “initial model” because it works to integrate research through thinking models.
“While O1 remains our broader general thinking model, O3-MINI provides a specialized alternative to technical domains that require accuracy and speed,” Openai wrote in a Blog Blog on Friday. “The O3-MINI version is another step on the Openai mission to push the limits of effective intelligence in terms of cost.”
O3-MINI is the most powerful Openai model yet, and it does not jump the R1’s Deepseek thinking in each standard.
The O3-MINI surpasses the R1 over AIME 2024, a test that measures the extent of the models well-understanding and responding to complex instructions-but only with a high thinking voltage. It also outperforms R1 on the program that focuses on programming (by 0.1 Point), but again, only with high thinking voltage. In low thinking voltage, O3-MINI R1 lags behind GPQA Diamond, which tests models with physics at the level of PhD, biology and chemistry.
To be fair, O3-MINI answers many inquiries at a low cost of competitiveness. In the post, Openai compares its performance with the O1 family:
“Through low-thinking voltage, O3-MINI achieves a similar performance with O1-MINI, while O3-MINI achieves a comparative performance with O1,” he writes Openai. “O3-MINI matches with an average thinking voltage O1 in mathematics, coding and science with faster responses. At the same time, with high thinking voltage, O3-MINI outperforms both O1-MINI and O1.”
It should be noted that the O3-MINI performance feature is minimal in some areas. In AIME 2024, O3-MINI O1 wins only 0.3 Celsius when he is set on high thinking voltage. And on GPQA Diamond, O3-MINI does not exceed the degree of high thinking.
Openai asserts that O3-MINI or more safely from the O1 family, thanks to the efforts of the red intersection and the “trading compatibility” methodology, which makes models “think” about Openai’s safety policy while responding to inquiries. According to the company, O3-MINI “greatly exceeds” one of the leading models in Openai. GPT-4OOn “Safety Challenges and Equipment of Evanses.”
Techcrunch has a news message focused on artificial intelligence! Subscribe here To get it in your inbox every Wednesday.