{"id":29843,"date":"2025-03-01T11:07:22","date_gmt":"2025-03-01T03:07:22","guid":{"rendered":"https:\/\/www.1ai.net\/?p=29843"},"modified":"2025-03-01T11:07:22","modified_gmt":"2025-03-01T03:07:22","slug":"%e6%9c%80%e5%bc%ba%e5%bc%80%e6%ba%90%e8%a7%86%e9%a2%91%e6%a8%a1%e5%9e%8b%ef%bc%9f%e6%9c%ac%e5%9c%b0%e9%83%a8%e7%bd%b2%e9%98%bf%e9%87%8c%e7%9a%84%e4%b8%87%e8%b1%a12-1%ef%bc%88wan2-1%ef%bc%89%e6%96%87","status":"publish","type":"post","link":"https:\/\/www.1ai.net\/en\/29843.html","title":{"rendered":"The strongest open source video model? Local Deployment of Ali's Wanxiang 2.1 (Wan2.1) Vincent Video"},"content":{"rendered":"<p>Following Tencent's hybridization, Ali also announced<a href=\"https:\/\/www.1ai.net\/en\/tag\/%e5%bc%80%e6%ba%90\" title=\"[View articles tagged with [open source]]\" target=\"_blank\" >Open Source<\/a>Under<a href=\"https:\/\/www.1ai.net\/en\/tag\/%e8%a7%86%e9%a2%91%e7%94%9f%e6%88%90%e6%a8%a1%e5%9e%8b\" title=\"_Other Organiser\" target=\"_blank\" >Video Generation Model<\/a>: Universal Phase 2.1 (<a href=\"https:\/\/www.1ai.net\/en\/tag\/wan2-1\" title=\"_Other Organiser\" target=\"_blank\" >Wan2.1<\/a>), claiming SOTA-level performance, with highlights including:<\/p>\n<p>1, better than existing open source models, and even \"comparable to some closed source models\".<\/p>\n<p>2\u3001It is the first video model that can generate Chinese and English text at the same time.<\/p>\n<p>3. Supports consumer GPUs, of which the T2V-1.3B model requires only 8.19GB of video memory.<\/p>\n<p><img loading=\"lazy\" decoding=\"async\" class=\"alignnone size-full wp-image-29848\" title=\"6a8a24c2j00ssfce00016d000io007lm\" src=\"https:\/\/www.1ai.net\/wp-content\/uploads\/2025\/03\/6a8a24c2j00ssfce00016d000io007lm.jpg\" alt=\"6a8a24c2j00ssfce00016d000io007lm\" width=\"672\" height=\"273\" \/><\/p>\n<p>Currently 10,000 Phase 2.1 is available through<a href=\"https:\/\/www.1ai.net\/en\/tag\/comfyui\" title=\"_Other Organiser\" target=\"_blank\" >ComfyUI<\/a>Local deployment, here's how:<\/p>\n<p>I. Installation of the necessary tools<\/p>\n<p>Please keep your \"network free\" and make sure you have Python, Git, and the latest version of ComfyUI installed (if not, please refer to the second half of this post: ComfyUI Latest Version Installation for instructions on how to install it).<\/p>\n<p>After updating ComfyUI to the latest version, you can see the message that 10,000 phases 2.1 has been supported in the main interface:<\/p>\n<p><img loading=\"lazy\" decoding=\"async\" class=\"alignnone size-full wp-image-29846\" title=\"3286d452j00ssfce0000zd000bi0090m\" src=\"https:\/\/www.1ai.net\/wp-content\/uploads\/2025\/03\/3286d452j00ssfce0000zd000bi0090m.jpg\" alt=\"3286d452j00ssfce0000zd000bi0090m\" width=\"414\" height=\"324\" \/><\/p>\n<p>2. Download the model<\/p>\n<p>Next, download the model in four parts.<\/p>\n<p>1. Open the URL:<\/p>\n<p>https:\/\/huggingface.co\/Comfy-Org\/Wan_2.1_ComfyUI_repackaged\/tree\/main\/split_files\/diffusion_models<\/p>\n<p><img loading=\"lazy\" decoding=\"async\" class=\"alignnone size-full wp-image-29844\" title=\"a06a0b12j00ssfce1006xd000u000gsm\" src=\"https:\/\/www.1ai.net\/wp-content\/uploads\/2025\/03\/a06a0b12j00ssfce1006xd000u000gsm.jpg\" alt=\"a06a0b12j00ssfce1006xd000u000gsm\" width=\"1080\" height=\"604\" \/><\/p>\n<p>Please download the above models according to your needs, and all of them have a download capacity of more than 80GB.<\/p>\n<p>2. Download the following file and put it into the ComfyUI\/models\/text_encoders directory<\/p>\n<p>https:\/\/huggingface.co\/Comfy-Org\/Wan_2.1_ComfyUI_repackaged\/resolve\/main\/split_files\/text_encoders\/umt5_xxl_fp8_e4m3fn_scaled.safetensors?download=true<\/p>\n<p>3. Download the following file and put it into the ComfyUI\/models\/clip_vision directory<\/p>\n<p>https:\/\/huggingface.co\/Comfy-Org\/Wan_2.1_ComfyUI_repackaged\/resolve\/main\/split_files\/clip_vision\/clip_vision_h.safetensors?download=true<\/p>\n<p>4. Download the following file and put it into the ComfyUI\/models\/vae directory<\/p>\n<p>https:\/\/huggingface.co\/Comfy-Org\/Wan_2.1_ComfyUI_repackaged\/resolve\/main\/split_files\/vae\/wan_2.1_vae.safetensors?download=true<\/p>\n<p>III. Downloading workflows<\/p>\n<p>Next go to the workflow page and download the three workflows at the following URLs:<\/p>\n<p>https:\/\/huggingface.co\/Comfy-Org\/Wan_2.1_ComfyUI_repackaged\/tree\/main\/example%20workflows_Wan2.1<\/p>\n<p><img loading=\"lazy\" decoding=\"async\" class=\"alignnone size-full wp-image-29845\" title=\"b9a0b42ej00ssfce1006ad000u000gvm\" src=\"https:\/\/www.1ai.net\/wp-content\/uploads\/2025\/03\/b9a0b42ej00ssfce1006ad000u000gvm.jpg\" alt=\"b9a0b42ej00ssfce1006ad000u000gvm\" width=\"1080\" height=\"607\" \/><\/p>\n<p>Drag the workflow to the ComfyUI interface to load it, and you'll see something like the following:<\/p>\n<p><img loading=\"lazy\" decoding=\"async\" class=\"alignnone size-full wp-image-29847\" title=\"0863481dj00ssfce1002dd000u000a7m\" src=\"https:\/\/www.1ai.net\/wp-content\/uploads\/2025\/03\/0863481dj00ssfce1002dd000u000a7m.jpg\" alt=\"0863481dj00ssfce1002dd000u000a7m\" width=\"1080\" height=\"367\" \/><\/p>\n<p>IV. Basic Functional Tests<\/p>\n<p>By this point the installation is complete, so let's first test the Vincennes video (using the T2V-14B).<\/p>\n<p>Cue in: a big-breasted Victoria's Secret model in a tiny red bikini with a smile on a Mediterranean beach, a few yachts in the sea in the distance, a blue sky with a few white clouds<\/p>\n<p><img loading=\"lazy\" decoding=\"async\" class=\"alignnone size-full wp-image-29849\" title=\"862191edg00ssfce103hgd000go009mm\" src=\"https:\/\/www.1ai.net\/wp-content\/uploads\/2025\/03\/862191edg00ssfce103hgd000go009mm.gif\" alt=\"862191edg00ssfce103hgd000go009mm\" width=\"600\" height=\"346\" \/><\/p>\n<p>Cue word: a big-breasted Victoria's Secret model in a tiny red bikini on a Mediterranean beach, smiling and holding up a sign that says \"welcome.\"<\/p>\n<p><img loading=\"lazy\" decoding=\"async\" class=\"alignnone size-full wp-image-29850\" title=\"c1bdd655g00ssfce103p4d000go009mm\" src=\"https:\/\/www.1ai.net\/wp-content\/uploads\/2025\/03\/c1bdd655g00ssfce103p4d000go009mm.gif\" alt=\"c1bdd655g00ssfce103p4d000go009mm\" width=\"600\" height=\"346\" \/><\/p>\n<p>Cue word: a big-breasted Victoria's Secret model in a tiny red bikini on a Mediterranean beach, smiling and holding up a sign that says \"I love China.\"<\/p>\n<p><img loading=\"lazy\" decoding=\"async\" class=\"alignnone size-full wp-image-29851\" title=\"64571886g00ssfce103bad000go009mm\" src=\"https:\/\/www.1ai.net\/wp-content\/uploads\/2025\/03\/64571886g00ssfce103bad000go009mm.gif\" alt=\"64571886g00ssfce103bad000go009mm\" width=\"600\" height=\"346\" \/><\/p>\n<p>Continuing to test the Tucson video (using the I2V-14B-480P), use this image:<\/p>\n<p><img loading=\"lazy\" decoding=\"async\" class=\"alignnone size-full wp-image-29852\" title=\"96584451j00ssfce3014ud000n400xsm\" src=\"https:\/\/www.1ai.net\/wp-content\/uploads\/2025\/03\/96584451j00ssfce3014ud000n400xsm.jpg\" alt=\"96584451j00ssfce3014ud000n400xsm\" width=\"832\" height=\"1216\" \/><\/p>\n<p>Cue word: a girl laughing and talking in a car with a vehicle moving outside the window<\/p>\n<p><img loading=\"lazy\" decoding=\"async\" class=\"alignnone size-full wp-image-29853\" title=\"c9cc9998g00ssfce101jtd0009j00dwm\" src=\"https:\/\/www.1ai.net\/wp-content\/uploads\/2025\/03\/c9cc9998g00ssfce101jtd0009j00dwm.gif\" alt=\"c9cc9998g00ssfce101jtd0009j00dwm\" width=\"343\" height=\"500\" \/><\/p>\n<p>ComfyUI official video raw video (using T2V-14B), but kijai big brother made the workflow has been supported, here just generated video, modify the characters in the screen.<\/p>\n<p>Cue word: a cyborg girl sitting in a car<\/p>\n<p><img loading=\"lazy\" decoding=\"async\" class=\"alignnone size-full wp-image-29854\" title=\"a74da3bfg00ssfce10414d000go008cm\" src=\"https:\/\/www.1ai.net\/wp-content\/uploads\/2025\/03\/a74da3bfg00ssfce10414d000go008cm.gif\" alt=\"a74da3bfg00ssfce10414d000go008cm\" width=\"600\" height=\"300\" \/><\/p>\n<p>Note: kijai's workflows and models are available as a separate download, see the link to the code page at the bottom of the tweet for details.<\/p>\n<p>Also note that since Manphase 2.1 involves multiple models, please make sure you have selected the correct model in the ComfyUI dialog box before generating:<\/p>\n<p><img loading=\"lazy\" decoding=\"async\" class=\"alignnone size-full wp-image-29855\" title=\"9b315f28j00ssfce0000gd000f9004sm\" src=\"https:\/\/www.1ai.net\/wp-content\/uploads\/2025\/03\/9b315f28j00ssfce0000gd000f9004sm.jpg\" alt=\"9b315f28j00ssfce0000gd000f9004sm\" width=\"549\" height=\"172\" \/><\/p>\n<p>V. Preliminary test conclusions<\/p>\n<p>1, from the point of view of the effect, Ali Wanphase 2.1 is not inferior to the Tencent hybrid, and the degree of compliance with the cue word is also better.<\/p>\n<p>2\u3001It is true that Chinese can be generated, but only some cases can get the correct result, need to draw cards.<\/p>\n<p>3, the speed of operation, in the 4090 running Vincennes video (T2V-14B model), about 8 minutes to generate 5 seconds of 480P video (no quantization and other technologies), the use of T2V-1.3B is as little as 2 minutes or so, but at the expense of the effect. Graph-generated video (I2V) is even faster, generating 5 seconds of video in about 3 minutes on the 4090.<\/p>\n<p>4, Vince video can be selected according to the hardware 1.3B and 14B model, Tucson video can only choose 14B model, the resolution can be selected 480P and 720P.<\/p>\n<p>Web sites covered by the article<\/p>\n<p>The official code page for Manphase 2.1:<\/p>\n<p>https:\/\/github.com\/Wan-Video\/Wan2.1<\/p>\n<p>The official support page for Manphase 2.1 ComfyUI:<\/p>\n<p>https:\/\/blog.comfy.org\/p\/wan21-video-model-native-support<\/p>\n<p>Manphase 2.1 ComfyUI code page (kijai):<\/p>\n<p>https:\/\/github.com\/kijai\/ComfyUI-WanVideoWrapper<\/p>","protected":false},"excerpt":{"rendered":"<p>Following the Tencent mixed yuan, Ali also announced its open source video generation model: Wan phase 2.1 (Wan2.1), said it has SOTA-level performance, its highlights include: 1, better than the existing open source model, and can even \"with some closed source model comparable to. 2, is the first video model that can simultaneously generate Chinese and English text. 3, support for consumer-grade GPUs, of which the T2V-1.3B model requires only 8.19GB of video memory. At present, WanPhase 2.1 can be deployed locally through ComfyUI, the following method: First, the installation of the necessary tools Please keep the \"network open\", and make sure that you have installed Python, Git, and the latest version of ComfyUI (if not, the installation method refer to the latter part of this article.) ComfyUI latest<\/p>","protected":false},"author":1,"featured_media":0,"comment_status":"closed","ping_status":"closed","sticky":false,"template":"","format":"standard","meta":{"footnotes":""},"categories":[149,144],"tags":[1989,5846,219,5847,460],"collection":[],"class_list":["post-29843","post","type-post","status-publish","format-standard","hentry","category-jiaocheng","category-baike","tag-comfyui","tag-wan2-1","tag-219","tag-5847","tag-460"],"acf":[],"_links":{"self":[{"href":"https:\/\/www.1ai.net\/en\/wp-json\/wp\/v2\/posts\/29843","targetHints":{"allow":["GET"]}}],"collection":[{"href":"https:\/\/www.1ai.net\/en\/wp-json\/wp\/v2\/posts"}],"about":[{"href":"https:\/\/www.1ai.net\/en\/wp-json\/wp\/v2\/types\/post"}],"author":[{"embeddable":true,"href":"https:\/\/www.1ai.net\/en\/wp-json\/wp\/v2\/users\/1"}],"replies":[{"embeddable":true,"href":"https:\/\/www.1ai.net\/en\/wp-json\/wp\/v2\/comments?post=29843"}],"version-history":[{"count":0,"href":"https:\/\/www.1ai.net\/en\/wp-json\/wp\/v2\/posts\/29843\/revisions"}],"wp:attachment":[{"href":"https:\/\/www.1ai.net\/en\/wp-json\/wp\/v2\/media?parent=29843"}],"wp:term":[{"taxonomy":"category","embeddable":true,"href":"https:\/\/www.1ai.net\/en\/wp-json\/wp\/v2\/categories?post=29843"},{"taxonomy":"post_tag","embeddable":true,"href":"https:\/\/www.1ai.net\/en\/wp-json\/wp\/v2\/tags?post=29843"},{"taxonomy":"collection","embeddable":true,"href":"https:\/\/www.1ai.net\/en\/wp-json\/wp\/v2\/collection?post=29843"}],"curies":[{"name":"wp","href":"https:\/\/api.w.org\/{rel}","templated":true}]}}