OpenAI's latest paper: o3 achieves gold medal with 395.64 points under the strict rules of IOI 2024

Feb. 12 (Bloomberg) -- With the influence of Chinese AI companyOpenAI Opened up the secrets of O-series reinforcement learning.

OpenAI's latest paper: o3 achieves gold medal with 395.64 points under the strict rules of IOI 2024

Today (February 12), OpenAI released research on the use of inference models in competitive programmingpaperThe report, Competitive Programming with Large Reasoning Models, puts out the results of OpenAI's three inference models: o1, o1-ioi, and o3 in the IOI (International Olympiad in Informatics) and CodeForces (the world's leading online programming competition).

The paper shows that in IOI 2024, o3 scored 395.64 points under strict rules to reach the gold medal achievement and performed on par with elite human competitors at CodeForces.

The paper also mentions that China's DeepSeek-R1 and Kimi k1.5 have shown through independent research that the combined performance of the models in mathematical problem solving and programming challenges can be significantly improved using the chain-of-thinking learning (COT) approach. r1, k1.5 are new inference models that were released simultaneously by DeepSeek and Kimi on January 20th.

The paper compares the performance of general-purpose reasoning models with systems optimized for specific domains in competitive programming through the performance improvement of large language models trained by reinforcement learning (RL) on complex coding and reasoning tasks. The findings show that adding reinforcement learning training computation and test-time computation can significantly improve model performance to approach that of the world's top human players, and that these models will unlock new application experiences in AI applications in science, coding, math, and other fields.

statement:The content of the source of public various media platforms, if the inclusion of the content violates your rights and interests, please contact the mailbox, this site will be the first time to deal with.
Information

Beanbag proposes a new sparse model architecture, UltraMem, which reduces inference cost by up to 83% compared to MoE.

2025-2-12 20:44:56

Information

OpenAI promises its AI models are uncensored and unblocked from viewpoints, guaranteeing freedom of thought

2025-2-13 11:09:19

Search