{"id":34314,"date":"2025-04-28T22:08:40","date_gmt":"2025-04-28T14:08:40","guid":{"rendered":"https:\/\/www.1ai.net\/?p=34314"},"modified":"2025-04-29T11:15:20","modified_gmt":"2025-04-29T03:15:20","slug":"%e9%98%bf%e9%87%8c%e9%80%9a%e4%b9%89%e5%8d%83%e9%97%ae-qwen3-%e7%b3%bb%e5%88%97%e6%a8%a1%e5%9e%8b%e6%9c%89%e6%9c%9b%e4%bb%8a%e6%97%a5%e5%8f%91%e5%b8%83%ef%bc%8c%e6%94%af%e6%8c%81-256k-%e4%b8%8a","status":"publish","type":"post","link":"https:\/\/www.1ai.net\/en\/34314.html","title":{"rendered":"Ali Tongyi's Qwen3 series of models is expected to be released today, supporting 256K context lengths"},"content":{"rendered":"<p>April 28 News.<a href=\"https:\/\/www.1ai.net\/en\/tag\/%e9%98%bf%e9%87%8c\" title=\"[View articles tagged with [Ali]]\" target=\"_blank\" >Ali<\/a><a href=\"https:\/\/www.1ai.net\/en\/tag\/%e9%80%9a%e4%b9%89%e5%8d%83%e9%97%ae\" title=\"[View articles tagged with [Tongyi Thousand Questions]]\" target=\"_blank\" >Thousand Questions on Tongyi<\/a>In a post on X, Open Source Leader Yang Jun Lin hinted that the Qwen3 model is expected to be released today.<\/p>\n<p><img loading=\"lazy\" decoding=\"async\" class=\"alignnone size-full wp-image-34315\" title=\"f88caa6bj00svgm1a000fd000di006hp\" src=\"https:\/\/www.1ai.net\/wp-content\/uploads\/2025\/04\/f88caa6bj00svgm1a000fd000di006hp.jpg\" alt=\"f88caa6bj00svgm1a000fd000di006hp\" width=\"486\" height=\"233\" \/><\/p>\n<p>It's worth mentioning that the Qwen3 model collection has been launched on ModelScope, Alibaba's open-source AI modeling community, and is now offline, including Qwen3-4B-Base, Qwen3-1.7B, Qwen3-0.6B, and Qwen3-30B-A3B-Base. All models are under the Apache License 2.0. Although the official announcement has not yet been released, but combined with the naming rules and the logic of the previous generation of technology, it can be speculated that its technical path and positioning direction.<\/p>\n<p>In this release, Qwen3-4B, Qwen3-1.7B and Qwen3-0.6B are directly named by the parameter scale, which corresponds to 4 billion, 1.7 billion and 600 million parameters, respectively. This kind of suffix-free naming method may indicate that they are single-architecture dense models without adopting the Mixed-Mode Expert (MoE) design, which is presumed to focus on lightweight application scenarios. On the other hand, Qwen3-30B-A3B-Base is a MoE architectural base model that dynamically invokes 3 billion (3B) parameters out of 30 billion (30B) total parameters to process tasks.<\/p>\n<p>According to 1AI, since August 2023, AliCloud has successively open-sourced four generations of models, including Qwen, Qwen1.5, Qwen2, and Qwen2.5, which encompass the full sizes of 0.5B, 1.5B, 3B, 7B, 14B, 32B, 72B, and 110B, and the full range of large languages, multimodalities, math, and code.<\/p>","protected":false},"excerpt":{"rendered":"<p>On April 28th, Ali Tongyi Qianqian open source head Yang Lin Junyan hinted in a post on X that Qwen3 models are expected to be released today. It is worth mentioning that the Qwen3 series of model collections has been briefly unveiled in Alibaba's AI model open source community ModelScope and then taken offline, including Qwen3-4B-Base, Qwen3-1.7B, Qwen3-0.6B, and Qwen3-30B-A3B-Base. All models are under the Apache License 2.0. Although the official announcement has not yet been released, but combined with the naming rules and the logic of the previous generation of technology, it can be speculated that its technical path and positioning direction. Among the models released this time<\/p>","protected":false},"author":1,"featured_media":0,"comment_status":"closed","ping_status":"closed","sticky":false,"template":"","format":"standard","meta":{"footnotes":""},"categories":[146],"tags":[331,1759],"collection":[],"class_list":["post-34314","post","type-post","status-publish","format-standard","hentry","category-news","tag-331","tag-1759"],"acf":[],"_links":{"self":[{"href":"https:\/\/www.1ai.net\/en\/wp-json\/wp\/v2\/posts\/34314","targetHints":{"allow":["GET"]}}],"collection":[{"href":"https:\/\/www.1ai.net\/en\/wp-json\/wp\/v2\/posts"}],"about":[{"href":"https:\/\/www.1ai.net\/en\/wp-json\/wp\/v2\/types\/post"}],"author":[{"embeddable":true,"href":"https:\/\/www.1ai.net\/en\/wp-json\/wp\/v2\/users\/1"}],"replies":[{"embeddable":true,"href":"https:\/\/www.1ai.net\/en\/wp-json\/wp\/v2\/comments?post=34314"}],"version-history":[{"count":0,"href":"https:\/\/www.1ai.net\/en\/wp-json\/wp\/v2\/posts\/34314\/revisions"}],"wp:attachment":[{"href":"https:\/\/www.1ai.net\/en\/wp-json\/wp\/v2\/media?parent=34314"}],"wp:term":[{"taxonomy":"category","embeddable":true,"href":"https:\/\/www.1ai.net\/en\/wp-json\/wp\/v2\/categories?post=34314"},{"taxonomy":"post_tag","embeddable":true,"href":"https:\/\/www.1ai.net\/en\/wp-json\/wp\/v2\/tags?post=34314"},{"taxonomy":"collection","embeddable":true,"href":"https:\/\/www.1ai.net\/en\/wp-json\/wp\/v2\/collection?post=34314"}],"curies":[{"name":"wp","href":"https:\/\/api.w.org\/{rel}","templated":true}]}}