{"id":29620,"date":"2025-02-26T11:27:23","date_gmt":"2025-02-26T03:27:23","guid":{"rendered":"https:\/\/www.1ai.net\/?p=29620"},"modified":"2025-02-26T11:27:23","modified_gmt":"2025-02-26T03:27:23","slug":"%e9%98%bf%e9%87%8c%e4%b8%87%e7%9b%b8%e8%a7%86%e9%a2%91%e7%94%9f%e6%88%90%e5%a4%a7%e6%a8%a1%e5%9e%8b%e5%ae%a3%e5%b8%83%e5%bc%80%e6%ba%90%ef%bc%9a8-2gb-%e6%98%be%e5%ad%98%e5%b0%b1%e8%83%bd%e8%b7%91","status":"publish","type":"post","link":"https:\/\/www.1ai.net\/en\/29620.html","title":{"rendered":"Ali Wanphase Video Generation Big Model Announced Open Source: 8.2GB Video Memory Can Run, Tests Surpass Sora"},"content":{"rendered":"<p><a href=\"https:\/\/www.1ai.net\/en\/tag\/%e9%98%bf%e9%87%8c%e4%ba%91\" title=\"_Other Organiser\" target=\"_blank\" >Alibaba Cloud<\/a>announced on February 25th that its visual life<strong>into a base model<a href=\"https:\/\/www.1ai.net\/en\/tag\/%e4%b8%87%e7%9b%b8\" title=\"[Sees articles with [many] labels]\" target=\"_blank\" >gazillion-phase<\/a> 2.1 (Wan) Open Source.<\/strong><\/p>\n<p><img loading=\"lazy\" decoding=\"async\" class=\"alignnone size-full wp-image-29622\" title=\"5e25399fj00ss9tj1009rd000u000kup\" src=\"https:\/\/www.1ai.net\/wp-content\/uploads\/2025\/02\/5e25399fj00ss9tj1009rd000u000kup.jpg\" alt=\"5e25399fj00ss9tj1009rd000u000kup\" width=\"1080\" height=\"750\" \/><\/p>\n<p>This open source uses the most lenient\u00a0<strong>Apache 2.0 protocol<\/strong>The newest version of the program is the \"14B\" and \"1.3B\" parameter specifications, which are open source and support both text-generated video and graph-generated video tasks, and can be downloaded from Github, HuggingFace, and the Magic Hitch community.<\/p>\n<p>It is reported that the 14B Wanphase model excels in command following, complex motion generation, physical modeling, and text and video generation in the review set VBench.<strong>With a total score of 86.22%, Manphase 2.1 outperforms domestic and international models such as Sora, Luma, and Pika.<\/strong>Version 1.3B not only outperforms larger open-source models and even approaches some closed-source models, but also runs on consumer graphics cards, claiming \"<strong>Generate 480P video with only 8.2GB of video memory<\/strong>\" for secondary model development and academic research.<\/p>\n<p><img loading=\"lazy\" decoding=\"async\" class=\"alignnone size-full wp-image-29621\" title=\"b026c831j00ss9tjl0092d000u000ghp\" src=\"https:\/\/www.1ai.net\/wp-content\/uploads\/2025\/02\/b026c831j00ss9tjl0092d000u000ghp.jpg\" alt=\"b026c831j00ss9tjl0092d000u000ghp\" width=\"1080\" height=\"593\" \/><\/p>\n<p>In terms of algorithm design, Maxthon has developed efficient causal 3D VAE and scalable pre-training strategies based on the mainstream DiT architecture and the Flow Matching paradigm for linear noise trajectories. Taking 3D VAE as an example, in order to efficiently support the encoding and decoding of videos of arbitrary length, Maxthon implements a feature caching mechanism in the causal convolution module of 3D VAE, which replaces the direct end-to-end coding and decoding process of long videos, and realizes the efficient coding and decoding of unlimited-length 1080P videos. In addition, by advancing the spatial downsampling compression, the memory footprint of 29% during inference is further reduced without performance loss.<\/p>\n<p>The experimental results of the Wanxiang team show that in the 14 main dimensions and 26 sub-dimension tests of motion quality, visual quality, style and multi-targeting<strong>The 10,000 phases have achieved industry-leading performance and five firsts.<\/strong>.<\/p>\n<p data-vmark=\"cb41\">With open source address:<\/p>\n<ul class=\"medium-size list-paddingleft-2\">\n<li>\n<p data-vmark=\"42ea\"><strong>Github:<\/strong><a href=\"https:\/\/github.com\/Wan-Video\" target=\"_blank\" rel=\"noopener\"><span class=\"link-text-start-with-http\">https:\/\/github.com\/Wan-Video<\/span><\/a><\/p>\n<\/li>\n<li>\n<p data-vmark=\"d75a\"><strong>HuggingFace:<\/strong><a href=\"https:\/\/huggingface.co\/Wan-AI\" target=\"_blank\" rel=\"noopener\"><span class=\"link-text-start-with-http\">https:\/\/huggingface.co\/Wan-AI<\/span><\/a><\/p>\n<\/li>\n<li>\n<p data-vmark=\"53b3\"><strong>Magic Match Community:<\/strong><a href=\"https:\/\/modelscope.cn\/organization\/Wan-AI\" target=\"_blank\" rel=\"noopener\"><span class=\"link-text-start-with-http\">https:\/\/modelscope.cn\/organization\/Wan-AI<\/span><\/a><\/p>\n<\/li>\n<\/ul>","protected":false},"excerpt":{"rendered":"<p>On February 25th, Aliyun announced the visual generation of a base model of 2.1 (Wan) open source. This time, using the most liberal Apache 2.0 protocol, 14B and 1.3B parameters specifications, with all the reasoning codes and weights open, while supporting the Vincent video and graphic video missions, the global developers can download from Github, HuggingFace and the magic community. It was described that the 14B phase model was prominent in terms of command compliance, complex motion generation, physical modelling, text video generation, and that in VBench, the score was 2.1 for 10,000 years, outnumbering domestic and foreign models such as Sora, Luma, Pika, for a total score of 86.22%. 1.3<\/p>","protected":false},"author":1,"featured_media":0,"comment_status":"closed","ping_status":"closed","sticky":false,"template":"","format":"standard","meta":{"footnotes":""},"categories":[146],"tags":[5808,2958,334],"collection":[],"class_list":["post-29620","post","type-post","status-publish","format-standard","hentry","category-news","tag-5808","tag-2958","tag-334"],"acf":[],"_links":{"self":[{"href":"https:\/\/www.1ai.net\/en\/wp-json\/wp\/v2\/posts\/29620","targetHints":{"allow":["GET"]}}],"collection":[{"href":"https:\/\/www.1ai.net\/en\/wp-json\/wp\/v2\/posts"}],"about":[{"href":"https:\/\/www.1ai.net\/en\/wp-json\/wp\/v2\/types\/post"}],"author":[{"embeddable":true,"href":"https:\/\/www.1ai.net\/en\/wp-json\/wp\/v2\/users\/1"}],"replies":[{"embeddable":true,"href":"https:\/\/www.1ai.net\/en\/wp-json\/wp\/v2\/comments?post=29620"}],"version-history":[{"count":0,"href":"https:\/\/www.1ai.net\/en\/wp-json\/wp\/v2\/posts\/29620\/revisions"}],"wp:attachment":[{"href":"https:\/\/www.1ai.net\/en\/wp-json\/wp\/v2\/media?parent=29620"}],"wp:term":[{"taxonomy":"category","embeddable":true,"href":"https:\/\/www.1ai.net\/en\/wp-json\/wp\/v2\/categories?post=29620"},{"taxonomy":"post_tag","embeddable":true,"href":"https:\/\/www.1ai.net\/en\/wp-json\/wp\/v2\/tags?post=29620"},{"taxonomy":"collection","embeddable":true,"href":"https:\/\/www.1ai.net\/en\/wp-json\/wp\/v2\/collection?post=29620"}],"curies":[{"name":"wp","href":"https:\/\/api.w.org\/{rel}","templated":true}]}}