{"id":45940,"date":"2025-11-14T11:18:10","date_gmt":"2025-11-14T03:18:10","guid":{"rendered":"https:\/\/www.1ai.net\/?p=45940"},"modified":"2025-11-14T11:18:10","modified_gmt":"2025-11-14T03:18:10","slug":"%e6%96%b0%e6%b5%aa%e5%be%ae%e5%8d%9a%e5%8f%91%e5%b8%83%e5%85%b6%e9%a6%96%e4%b8%aa%e5%bc%80%e6%ba%90%e5%a4%a7%e6%a8%a1%e5%9e%8b-vibethinker-1-5b%ef%bc%8c%e5%b0%8f%e6%a8%a1%e5%9e%8b%e6%8c%91%e6%88%98","status":"publish","type":"post","link":"https:\/\/www.1ai.net\/en\/45940.html","title":{"rendered":"Sina Weibo released its first large open source model, VibeThinker-1.5B, small models challenging a huge parametric rival"},"content":{"rendered":"<p>November 14th.<a href=\"https:\/\/www.1ai.net\/en\/tag\/%e6%96%b0%e6%b5%aa\" title=\"_Other Organiser\" target=\"_blank\" >New Wave<\/a><a href=\"https:\/\/www.1ai.net\/en\/tag\/%e5%be%ae%e5%8d%9a\" title=\"[Sees articles with [Webb] labels]\" target=\"_blank\" >Weibo<\/a>It's the first of its kind<a href=\"https:\/\/www.1ai.net\/en\/tag\/%e5%bc%80%e6%ba%90\" title=\"[View articles tagged with [open source]]\" target=\"_blank\" >Open Source<\/a><a href=\"https:\/\/www.1ai.net\/en\/tag\/%e5%a4%a7%e6%a8%a1%e5%9e%8b\" title=\"[View articles tagged with [large models]]\" target=\"_blank\" >Large Model<\/a> VibeThinker-1.5B, called \u201cSmall models can also be intelligent\u201d\u3002<\/p>\n<p>1AI HAS THE FOLLOWING OFFICIAL DESCRIPTION:<\/p>\n<p>NOW THAT MOST OF THE MOST POWERFUL MODELS IN THE INDUSTRY ARE MORE THAN 1T, AND EVEN 2T-SCALE MODELS HAVE EMERGED, IS IT THAT ONLY A MEGA-PARAMETER MODEL HAS HIGH INTELLIGENCE? IS ONLY A SMALL NUMBER OF TECHNOLOGY GIANTS CAPABLE OF MAKING LARGE MODELS<\/p>\n<p>VibeThinker-1.5B, which is exactly the negative answer given by Weibo AI to this question, proves that small models can also have high IQ. This means that making the most powerful models no longer depends primarily on pushing up parameters, as conventionally thought, but can also be done through sophisticated algorithm design\u3002<\/p>\n<p>The model has only 1.5B (1.5 billion) parameters and, after training in the innovative \u201cspectrum to signal principle\u201d (SSP) methodology developed by microblogging AI R &amp; D, its effects are subversive: VibeThinker's performance on three difficult mathematical test sets: AIME24, AIME25 and HMT25, surpassing the model DeepSeek-R1-0120 (model size 671B) with 400 times the amount of the parameter<strong>MiniMax-M1 on 456B is close to or equivalent to<\/strong>; in the LiveCodeBench v6 (Programmatic Test Series), successful tracking of models with tens of times the number of parameters, such as the Magistral-Medium-2506 version of the deep-thinking model of the lead AI firm Minstral.AI in Europe\u3002<\/p>\n<p><img loading=\"lazy\" decoding=\"async\" class=\"alignnone size-full wp-image-45941\" title=\"f3af879cj00t5p5560042d000v90gtp\" src=\"https:\/\/www.1ai.net\/wp-content\/uploads\/2025\/11\/f3af879cj00t5p5560042d000v900gtp.jpg\" alt=\"f3af879cj00t5p5560042d000v90gtp\" width=\"1125\" height=\"605\" \/><\/p>\n<p>VibeThinker is powerful not by stacking of parameters, but rather by the SSP training concept developed by microblogging researchers, which encourages models to explore all possible solution paths, rather than focusing on the right rate, at the learning stage; and then, to optimize the model ' s performance to the extreme, using enhanced learning to optimize the efficiency of the strategy and to pinpoint the right path\u3002<\/p>\n<p><strong>The single \"Post-Training\" cost of the model is less than $8000<\/strong>In contrast, DeepSeek-R1 and MiniMax-M1 cost after-training at $290,000 and $530,000, respectively, a reduction of several dozen times\u3002<\/p>\n<p>The Open Source of VibeThinker-1.5B, designed to provide a new, high-value, research and development path for medium-sized enterprises and higher education research teams with limited global computing resources, allows everyone to train the front-line large models rather than be left out as before, which is critical to technological progress in industry\u3002<\/p>\n<p>Github: https:\/\/github.com\/Weiboai\/VibeThinker<\/p>\n<p>HuggingFace: https:\/\/huggingface.co\/Weiboai\/VibeThinker-1.5B<\/p>\n<p>Arxiv: https:\/\/arxiv.org\/pdf\/2511.06221<\/p>\n<p>ModelScope: https:\/\/www.modelscope.cn\/models\/WeiboAI\/VibeThinker-1.5B<\/p>","protected":false},"excerpt":{"rendered":"<p>On November 14, New Wave Weibo released its first large open source model, VibeThinker-1.5B, called \u201cSmall models can also be intelligent\u201d. 1AI is accompanied by the following official presentation: Are the most powerful models in the industry at present more than 1T, or even 2T-scale models, that have a high level of intelligence? Is only a small number of technology giants capable of making large models? VibeThinker-1.5B, which is exactly the negative answer given by Weibo AI to this question, proves that small models can also have high IQs. This means that making the most powerful models no longer depends primarily on pushing up parameters, as conventionally thought, but can also be done through sophisticated algorithm design. This model is only 1.<\/p>","protected":false},"author":1,"featured_media":0,"comment_status":"closed","ping_status":"closed","sticky":false,"template":"","format":"standard","meta":{"footnotes":""},"categories":[146],"tags":[216,219,2644,7859],"collection":[],"class_list":["post-45940","post","type-post","status-publish","format-standard","hentry","category-news","tag-216","tag-219","tag-2644","tag-7859"],"acf":[],"_links":{"self":[{"href":"https:\/\/www.1ai.net\/en\/wp-json\/wp\/v2\/posts\/45940","targetHints":{"allow":["GET"]}}],"collection":[{"href":"https:\/\/www.1ai.net\/en\/wp-json\/wp\/v2\/posts"}],"about":[{"href":"https:\/\/www.1ai.net\/en\/wp-json\/wp\/v2\/types\/post"}],"author":[{"embeddable":true,"href":"https:\/\/www.1ai.net\/en\/wp-json\/wp\/v2\/users\/1"}],"replies":[{"embeddable":true,"href":"https:\/\/www.1ai.net\/en\/wp-json\/wp\/v2\/comments?post=45940"}],"version-history":[{"count":0,"href":"https:\/\/www.1ai.net\/en\/wp-json\/wp\/v2\/posts\/45940\/revisions"}],"wp:attachment":[{"href":"https:\/\/www.1ai.net\/en\/wp-json\/wp\/v2\/media?parent=45940"}],"wp:term":[{"taxonomy":"category","embeddable":true,"href":"https:\/\/www.1ai.net\/en\/wp-json\/wp\/v2\/categories?post=45940"},{"taxonomy":"post_tag","embeddable":true,"href":"https:\/\/www.1ai.net\/en\/wp-json\/wp\/v2\/tags?post=45940"},{"taxonomy":"collection","embeddable":true,"href":"https:\/\/www.1ai.net\/en\/wp-json\/wp\/v2\/collection?post=45940"}],"curies":[{"name":"wp","href":"https:\/\/api.w.org\/{rel}","templated":true}]}}