{"id":28158,"date":"2025-02-06T11:27:38","date_gmt":"2025-02-06T03:27:38","guid":{"rendered":"https:\/\/www.1ai.net\/?p=28158"},"modified":"2025-02-06T11:31:08","modified_gmt":"2025-02-06T03:31:08","slug":"%e8%ae%ad%e7%bb%83%e6%88%90%e6%9c%ac%e4%b8%8d%e5%88%b0-50-%e7%be%8e%e5%85%83%ef%bc%8c%e7%a0%94%e7%a9%b6%e4%ba%ba%e5%91%98%e6%89%93%e9%80%a0%e5%87%ba%e5%aa%b2%e7%be%8e-openai-o1-%e7%9a%84%e6%8e%a8","status":"publish","type":"post","link":"https:\/\/www.1ai.net\/en\/28158.html","title":{"rendered":"For Less Than $50 to Train, Researchers Build an Inference Model That Rivals OpenAI o1"},"content":{"rendered":"<p>February 6, 2011 - A study released on Friday shows that artificial intelligence researchers at Stanford University and the University of Washington spent less than 50 U.S. dollars (note: the current cost is about 364 yuan) on cloud computing to successfully train an artificial intelligence researcher with \"reasoning\" capabilities.<a href=\"https:\/\/www.1ai.net\/en\/tag\/%e4%ba%ba%e5%b7%a5%e6%99%ba%e8%83%bd%e6%a8%a1%e5%9e%8b\" title=\"_Other Organiser\" target=\"_blank\" >Artificial Intelligence Model<\/a>.<\/p>\n<p><img loading=\"lazy\" decoding=\"async\" class=\"alignnone size-full wp-image-28159\" title=\"b44fec88j00sr8s8u0030d000v900b8p\" src=\"https:\/\/www.1ai.net\/wp-content\/uploads\/2025\/02\/b44fec88j00sr8s8u0030d000v900b8p.jpg\" alt=\"b44fec88j00sr8s8u0030d000v900b8p\" width=\"1125\" height=\"404\" \/><\/p>\n<p>The model is named s1.<strong>Performing at the top of math and programming aptitude tests with OpenAI's o1 and DeepSeek's r1.<a href=\"https:\/\/www.1ai.net\/en\/tag\/%e6%8e%a8%e7%90%86%e6%a8%a1%e5%9e%8b\" title=\"[View articles tagged with [inference model]]\" target=\"_blank\" >inference model<\/a>Similar levels<\/strong>. Currently, the s1 model and the data and code used for its training<a href=\"https:\/\/github.com\/simplescaling\/s1\">Open sourced on GitHub<\/a>.<\/p>\n<p>s1 The team stated that<strong>They created the AI model through \"distillation\" technology.<\/strong>The technology is designed to extract the \"reasoning\" ability of an AI model by training it to learn the answers of another AI model. The researchers revealed that s1 was distilled from Google's reasoning model Gemini 2.0 Flash Thinking Experimental. Last month, researchers at the University of California, Berkeley, used the same distillation method to create an AI reasoning model at a cost of about $450.<\/p>\n<p>The emergence of models like s1 also raises questions about the commoditization of AI models -- if someone can replicate a multi-million dollar model at relatively low cost, where is the \"moat\" for large tech companies?<\/p>\n<p>Unsurprisingly, the big AI labs aren't happy about this, with OpenAI, for example, previously accusing DeepSeek of improperly accessing its API data for model distillation.<\/p>\n<p>s1 researchers wanted to find the easiest way to achieve strong inference performance and \"test-time scaling\" (i.e., allowing AI models to think more before answering a question), and these are some of the breakthroughs in OpenAI's o1.<\/p>\n<p>The paper by s1 shows that<strong>A method called supervised fine-tuning (SFT) can be used, which can distill the inference model using a relatively small data set<\/strong>In SFT. In SFT, AI models are explicitly instructed to mimic certain behaviors in the dataset.SFT is more cost-effective than the large-scale reinforcement learning approach that DeepSeek uses to train its R1 models.<\/p>\n<p>Google offers free access to the Gemini 2.0 Flash Thinking Experimental model through its Google AI Studio platform, with daily usage limits. However, its terms prohibit reverse engineering the model to develop services that compete with Google's own AI offerings.<\/p>\n<p>S1 is based on a small, off-the-shelf, free AI model provided by Qwen, Alibaba's Chinese AI lab. To train s1, the<strong>The researchers created a dataset of just 1,000 carefully curated questions, along with the answers to those questions and the \"thinking\" process behind each answer as given by the Google Gemini 2.0 Flash Thinking Experimental.<\/strong><\/p>\n<p>The researchers said.<strong>After training s1 (which took less than 30 minutes using 16 Nvidia H100 GPUs)<\/strong>, s1 has performed well in certain AI benchmarks. Niklas Muennighoff, a Stanford researcher involved in the project, told TechCrunch that it currently costs about $20 to rent these computing resources.<\/p>\n<p>The researchers used a clever trick to get s1 to check its work and extend its \"thinking\" time: they made it \"wait\". The paper shows that adding the word \"wait\" to s1's reasoning helped the model get slightly more accurate answers.<\/p>","protected":false},"excerpt":{"rendered":"<p>Feb. 6 (Bloomberg) -- Artificial intelligence researchers at Stanford University and the University of Washington have successfully trained an AI model with \"reasoning\" capabilities for less than $50 (note: currently about 364 yuan) in cloud computing, according to a study released on Friday. The model, called s1, performed similarly to top reasoning models such as OpenAI's o1 and DeepSeek's r1 in tests of mathematical and programming ability. The s1 model and the data and code used for its training are currently open source on GitHub. The s1 team said they created the AI model through a \"distillation\" technique, which is designed to train a model to learn the answers of another AI model.<\/p>","protected":false},"author":1,"featured_media":0,"comment_status":"closed","ping_status":"closed","sticky":false,"template":"","format":"standard","meta":{"footnotes":""},"categories":[146],"tags":[599,5023],"collection":[],"class_list":["post-28158","post","type-post","status-publish","format-standard","hentry","category-news","tag-599","tag-5023"],"acf":[],"_links":{"self":[{"href":"https:\/\/www.1ai.net\/en\/wp-json\/wp\/v2\/posts\/28158","targetHints":{"allow":["GET"]}}],"collection":[{"href":"https:\/\/www.1ai.net\/en\/wp-json\/wp\/v2\/posts"}],"about":[{"href":"https:\/\/www.1ai.net\/en\/wp-json\/wp\/v2\/types\/post"}],"author":[{"embeddable":true,"href":"https:\/\/www.1ai.net\/en\/wp-json\/wp\/v2\/users\/1"}],"replies":[{"embeddable":true,"href":"https:\/\/www.1ai.net\/en\/wp-json\/wp\/v2\/comments?post=28158"}],"version-history":[{"count":0,"href":"https:\/\/www.1ai.net\/en\/wp-json\/wp\/v2\/posts\/28158\/revisions"}],"wp:attachment":[{"href":"https:\/\/www.1ai.net\/en\/wp-json\/wp\/v2\/media?parent=28158"}],"wp:term":[{"taxonomy":"category","embeddable":true,"href":"https:\/\/www.1ai.net\/en\/wp-json\/wp\/v2\/categories?post=28158"},{"taxonomy":"post_tag","embeddable":true,"href":"https:\/\/www.1ai.net\/en\/wp-json\/wp\/v2\/tags?post=28158"},{"taxonomy":"collection","embeddable":true,"href":"https:\/\/www.1ai.net\/en\/wp-json\/wp\/v2\/collection?post=28158"}],"curies":[{"name":"wp","href":"https:\/\/api.w.org\/{rel}","templated":true}]}}