{"id":45222,"date":"2025-10-27T11:31:13","date_gmt":"2025-10-27T03:31:13","guid":{"rendered":"https:\/\/www.1ai.net\/?p=45222"},"modified":"2025-10-27T11:31:13","modified_gmt":"2025-10-27T03:31:13","slug":"%e7%be%8e%e5%9b%a2%e5%8f%91%e5%b8%83%e5%bc%80%e6%ba%90-longcat-video-%e8%a7%86%e9%a2%91%e7%94%9f%e6%88%90%e6%a8%a1%e5%9e%8b%ef%bc%8c%e5%8f%af%e7%a8%b3%e5%ae%9a%e8%be%93%e5%87%ba-5-%e5%88%86%e9%92%9f","status":"publish","type":"post","link":"https:\/\/www.1ai.net\/en\/45222.html","title":{"rendered":"America releases open source LongCat-Video video generation model, which stabilizes 5 minute content"},"content":{"rendered":"<p>October 27th message, this morning<a href=\"https:\/\/www.1ai.net\/en\/tag\/%e7%be%8e%e5%9b%a2\" title=\"[Sees articles with [American] labels]\" target=\"_blank\" >Meituan (Japanese company)<\/a> LongCat team release and<a href=\"https:\/\/www.1ai.net\/en\/tag\/%e5%bc%80%e6%ba%90\" title=\"[View articles tagged with [open source]]\" target=\"_blank\" >Open Source<\/a> LongCat-Video <a href=\"https:\/\/www.1ai.net\/en\/tag\/%e8%a7%86%e9%a2%91%e7%94%9f%e6%88%90%e6%a8%a1%e5%9e%8b\" title=\"_Other Organiser\" target=\"_blank\" >Video Generation Model<\/a>I DON'T KNOW. ACCORDING TO OFFICIAL PRESENTATIONS, IT ACHIEVES OPEN-SOURCE SOTA (THE MOST ADVANCED LEVEL) WITH A UNIFIED MODEL FOR THE ENGINEERING AND GRAPHIC VIDEO BASE MISSION AND IS BASED ON PRE-TRAINING IN ORIGINAL VIDEO-RENEWAL TASKS<strong>Minute long video coherent generation<\/strong>,Safeguard<strong>Consistency of time series and soundness of physical motion across frames<\/strong>There are significant advantages in the area of long video generation\u3002<\/p>\n<p><img loading=\"lazy\" decoding=\"async\" class=\"alignnone size-full wp-image-45223\" title=\"d62800439j00t4rtqk00a2d000ufup\" src=\"https:\/\/www.1ai.net\/wp-content\/uploads\/2025\/10\/d6280439j00t4rtqk00a2d000u000fup.jpg\" alt=\"d62800439j00t4rtqk00a2d000ufup\" width=\"1080\" height=\"570\" \/><\/p>\n<p>According to the presentation, in recent years, the World Model has allowed artificial intelligence to truly understand, predict and even reconstruct the real world, thus being seen as the core engine leading to the next generation of intelligence. As an intelligent system capable of modelling physical patterns, temporal evolution and scene logic, the \u201cworld model\u201d gives artificial intelligence the ability to \u201csee\u201d the nature of the world's operations. And the video generation model<strong>It's about to become a key path to building a world model<\/strong>:: Compression of many forms of knowledge, such as geometry, semantics and physics, through video-generated missions, enabling artificial intelligence to simulate, push and even preview the operation of the real world in digital space\u3002<\/p>\n<p>As a base for multifunctional unified video generation based on Diffusion Transformer (DiT) architecture, LongCat-Video innovation achieves task segregation by \"number of condition frames\" - Vincent video does not require frame, frame reference 1 graphic video input, video continuation is based on multi-spectrum pre-sequenced content, birth supports three core tasks and does not require additional model adaptation to form a \"manual \/ graphic \/ video continuation\" full mission closed loop\u3002<\/p>\n<ul>\n<li>Vincent video: generated<strong>\u00a0<\/strong><strong>720p, 30fps<\/strong><strong>\u00a0<\/strong>High-resolution video, yes<strong>Exactly decipher the details of objects, characters, scenes, styles in text<\/strong>, SEMANTIC UNDERSTANDING AND VISUAL PRESENTATION UP TO THE OPEN SOURCE SOTA LEVEL\u3002<\/li>\n<li>Graphic video: Strictly preserves the main attributes, background relationships and overall style of the reference image, dynamic processes conform to physical laws, supports multiple types of input, such as detailed instructions, concise descriptions, empty commands, and excellent content consistency and dynamic naturality\u3002<\/li>\n<li>Video continuation: Video continuation is the core differentialization capability of LongCat Video, which provides primary technical support for long video generation based on multiple frame conditions\u3002<\/li>\n<\/ul>\n<p>Based on video-renewal mission pre-training, Block-Causual Attention mechanism and GRPO post-training, LongCat-Video can steadily export 5-minute long video with no loss of quality, known as industry \u201ctop\u201d level\u3002<\/p>\n<p>1AI attaches the relevant links below:<\/p>\n<ul>\n<li>GitHub: https:\/\/github.com\/metuan-longcat\/LongCat-Video<\/li>\n<li>Hugging Face: https:\/\/huggingface.co\/meituan-longcat\/LongCat-Video<\/li>\n<li>Project Page: https:\/\/meituan-longcat.github.io\/LongCat-Video\/<\/li>\n<\/ul>","protected":false},"excerpt":{"rendered":"<p>October 27th, this morning, the LongCat team released and opened the LongCat-Video Video Generation Model. According to the official presentation, it achieves open-source SOTA (the most advanced level) with a unified model in the context of the Ventian, Graphical video base mission, and is pre-trained in the use of raw video continuation missions, achieves consistent generation of long-minute video, ensures consistency of time series across frames and soundness of physical motion, and has a significant advantage in the area of long video generation. According to the presentation, in recent years, the World Model has allowed artificial intelligence to truly understand, predict and even reconstruct the real world, thus being seen as the core engine leading to the next generation of intelligence. As an intelligent system capable of modelling physical patterns, temporal evolution and scene logic, \"World Moot\"<\/p>","protected":false},"author":1,"featured_media":0,"comment_status":"closed","ping_status":"closed","sticky":false,"template":"","format":"standard","meta":{"footnotes":""},"categories":[146],"tags":[219,4871,460],"collection":[],"class_list":["post-45222","post","type-post","status-publish","format-standard","hentry","category-news","tag-219","tag-4871","tag-460"],"acf":[],"_links":{"self":[{"href":"https:\/\/www.1ai.net\/en\/wp-json\/wp\/v2\/posts\/45222","targetHints":{"allow":["GET"]}}],"collection":[{"href":"https:\/\/www.1ai.net\/en\/wp-json\/wp\/v2\/posts"}],"about":[{"href":"https:\/\/www.1ai.net\/en\/wp-json\/wp\/v2\/types\/post"}],"author":[{"embeddable":true,"href":"https:\/\/www.1ai.net\/en\/wp-json\/wp\/v2\/users\/1"}],"replies":[{"embeddable":true,"href":"https:\/\/www.1ai.net\/en\/wp-json\/wp\/v2\/comments?post=45222"}],"version-history":[{"count":0,"href":"https:\/\/www.1ai.net\/en\/wp-json\/wp\/v2\/posts\/45222\/revisions"}],"wp:attachment":[{"href":"https:\/\/www.1ai.net\/en\/wp-json\/wp\/v2\/media?parent=45222"}],"wp:term":[{"taxonomy":"category","embeddable":true,"href":"https:\/\/www.1ai.net\/en\/wp-json\/wp\/v2\/categories?post=45222"},{"taxonomy":"post_tag","embeddable":true,"href":"https:\/\/www.1ai.net\/en\/wp-json\/wp\/v2\/tags?post=45222"},{"taxonomy":"collection","embeddable":true,"href":"https:\/\/www.1ai.net\/en\/wp-json\/wp\/v2\/collection?post=45222"}],"curies":[{"name":"wp","href":"https:\/\/api.w.org\/{rel}","templated":true}]}}