{"id":25578,"date":"2024-12-24T09:25:31","date_gmt":"2024-12-24T01:25:31","guid":{"rendered":"https:\/\/www.1ai.net\/?p=25578"},"modified":"2024-12-24T09:25:31","modified_gmt":"2024-12-24T01:25:31","slug":"openai-o3-ai%e6%a8%a1%e5%9e%8b%e7%aa%81%e7%a0%b4%e6%8e%a8%e7%90%86%e6%9e%81%e9%99%90%e8%83%8c%e5%90%8e%ef%bc%9a%e6%88%90%e6%9c%ac%e9%a3%99%e5%8d%87%ef%bc%8c%e6%b2%a1%e9%92%9e%e8%83%bd","status":"publish","type":"post","link":"https:\/\/www.1ai.net\/en\/25578.html","title":{"rendered":"OpenAI o3 AI model breaks through the reasoning limit behind: the cost soared, no \"money ability\" can not afford to use"},"content":{"rendered":"<p>TechCrunch, a technology media outlet, published a blog post yesterday (December 23) arguing that <a href=\"https:\/\/www.1ai.net\/en\/tag\/openai\" title=\"[View articles tagged with [OpenAI]]\" target=\"_blank\" >OpenAI<\/a> of <a href=\"https:\/\/www.1ai.net\/en\/tag\/o3\" title=\"_other organiser\" target=\"_blank\" >o3<\/a> While the model has achieved glowing results in tests such as ARC-AGI.<strong>However, the high computational cost behind it makes it difficult to popularize it in practical applications in the short term.<\/strong><\/p>\n<p><img loading=\"lazy\" decoding=\"async\" class=\"alignnone size-full wp-image-25579\" title=\"a65ea4ebj00soz59o0022d000gh00dap\" src=\"https:\/\/www.1ai.net\/wp-content\/uploads\/2024\/12\/a65ea4ebj00soz59o0022d000gh00dap.jpg\" alt=\"a65ea4ebj00soz59o0022d000gh00dap\" width=\"593\" height=\"478\" \/><\/p>\n<p><strong>o3 Performance<\/strong><\/p>\n<p>One of the new features of o3 is the ability to adjust the inference time, which is categorized into three computation levels: low, medium, and high; the higher the computation level, the better the task execution performance of o3.<\/p>\n<p>And o3 is a major step towards that goal for OpenAI. o3 scored 87.5% on the high compute setting and 75.7% on the low compute setting in the ARC-AGI benchmarks, tripling the performance of o1.<\/p>\n<p>In EpochAI's Frontier Math benchmark, o3 solved 25.2% (no other model exceeded 2%), setting a new record.<\/p>\n<p><strong>o3 Excessive modeling costs<\/strong><\/p>\n<p>Fran\u00e7ois Chollet, creator of the ARC-AGI benchmarks, wrote in a blog post that OpenAI's o3 model, while an important breakthrough in AI, is too expensive.<\/p>\n<p>According to the performance icons from the ARC-AGI test, the high-scoring version of o3 uses more than $1,000 (currently about Rs. 7,303) worth of computational resources per task, the o1 model uses about $5 of computational resources per task, and the o1-mini uses only a few cents.<\/p>\n<p>This means that while OpenAI achieved a high score of nearly 88%, it consumed more than 170 times the computational resources, and the high-computing version of o3 cost more than $10,000 (currently about $73,033) to invoke resources for the entire test, making the cost of using the o3 model affordable only to well-heeled organizations and individuals.<\/p>\n<p>The high computational cost of o3 models makes them more suitable for complex problems, such as long-term strategic decisions, rather than small, everyday problems; more efficient AI inference chips and more cost-effective AI chips may be the key to lowering the cost of using o3 models in the future.<\/p>","protected":false},"excerpt":{"rendered":"<p>TechCrunch, a technology media outlet, published a blog post yesterday (December 23), arguing that although OpenAI's o3 model has achieved impressive results in tests such as ARC-AGI, the high computational cost behind it makes it difficult to popularize it in practical applications in the short term. o3 Performance One of the new features of o3 is the ability to adjust the inference time, which is categorized into three computation levels: low, medium, and high, and the higher the computation level, the better the task execution performance of o3. o3 is an important step towards this goal for OpenAI. In the ARC-AGI benchmark, o3 scored 87.5% on the high compute setting and 75.7% on the low compute setting, tripling the performance of o1. In the E<\/p>","protected":false},"author":1,"featured_media":0,"comment_status":"closed","ping_status":"closed","sticky":false,"template":"","format":"standard","meta":{"footnotes":""},"categories":[146],"tags":[5275,190],"collection":[],"class_list":["post-25578","post","type-post","status-publish","format-standard","hentry","category-news","tag-o3","tag-openai"],"acf":[],"_links":{"self":[{"href":"https:\/\/www.1ai.net\/en\/wp-json\/wp\/v2\/posts\/25578","targetHints":{"allow":["GET"]}}],"collection":[{"href":"https:\/\/www.1ai.net\/en\/wp-json\/wp\/v2\/posts"}],"about":[{"href":"https:\/\/www.1ai.net\/en\/wp-json\/wp\/v2\/types\/post"}],"author":[{"embeddable":true,"href":"https:\/\/www.1ai.net\/en\/wp-json\/wp\/v2\/users\/1"}],"replies":[{"embeddable":true,"href":"https:\/\/www.1ai.net\/en\/wp-json\/wp\/v2\/comments?post=25578"}],"version-history":[{"count":0,"href":"https:\/\/www.1ai.net\/en\/wp-json\/wp\/v2\/posts\/25578\/revisions"}],"wp:attachment":[{"href":"https:\/\/www.1ai.net\/en\/wp-json\/wp\/v2\/media?parent=25578"}],"wp:term":[{"taxonomy":"category","embeddable":true,"href":"https:\/\/www.1ai.net\/en\/wp-json\/wp\/v2\/categories?post=25578"},{"taxonomy":"post_tag","embeddable":true,"href":"https:\/\/www.1ai.net\/en\/wp-json\/wp\/v2\/tags?post=25578"},{"taxonomy":"collection","embeddable":true,"href":"https:\/\/www.1ai.net\/en\/wp-json\/wp\/v2\/collection?post=25578"}],"curies":[{"name":"wp","href":"https:\/\/api.w.org\/{rel}","templated":true}]}}