{"id":19031,"date":"2024-08-31T09:24:03","date_gmt":"2024-08-31T01:24:03","guid":{"rendered":"https:\/\/www.1ai.net\/?p=19031"},"modified":"2024-08-31T09:24:03","modified_gmt":"2024-08-31T01:24:03","slug":"%e6%99%ba%e8%b0%b1ai%e5%8f%91%e5%b8%83glm-4-plus%ef%bc%9a%e5%aa%b2%e7%be%8egpt-4%ef%bc%8c%e9%a6%96%e5%88%9bc%e7%ab%af%e8%a7%86%e9%a2%91%e9%80%9a%e8%af%9d%e5%8a%9f%e8%83%bd","status":"publish","type":"post","link":"https:\/\/www.1ai.net\/en\/19031.html","title":{"rendered":"Zhipu AI releases GLM-4-Plus: comparable to GPT-4, the first C-end video call function"},"content":{"rendered":"<p><a href=\"https:\/\/www.1ai.net\/en\/tag\/%e6%99%ba%e8%b0%b1ai\" title=\"[SEES ARTICLES WITH [INTELLIGENCE AI] LABELS]\" target=\"_blank\" >Zhipu AI<\/a>Recently released its latest base<a href=\"https:\/\/www.1ai.net\/en\/tag\/%e5%a4%a7%e6%a8%a1%e5%9e%8b\" title=\"[View articles tagged with [large models]]\" target=\"_blank\" >Large Model<\/a>GLM-4-Plus, which demonstrates powerful visual capabilities comparable to OpenAI GPT-4, was announced to be open for use on August 30.<\/p>\n<p><strong>Major update highlights:<\/strong><\/p>\n<ul>\n<li><a href=\"https:\/\/www.1ai.net\/en\/tag\/%e8%af%ad%e8%a8%80%e5%9f%ba%e7%a1%80%e6%a8%a1%e5%9e%8b\" title=\"[See articles with labels]\" target=\"_blank\" >Language Basic Model<\/a>GLM-4-Plus: has achieved a qualitative leap in language parsing, instruction execution and long text processing capabilities, and continues to maintain its leading position in international competition.<\/li>\n<li><a href=\"https:\/\/www.1ai.net\/en\/tag\/%e6%96%87%e7%94%9f%e5%9b%be%e6%a8%a1%e5%9e%8b\" title=\"[Sees articles with tags]\" target=\"_blank\" >Wenshengtu Model<\/a>CogView-3-Plus: Performance comparable to the industry-leading MJ-V6 and FLUX models.<\/li>\n<li>Image\/video understanding model GLM-4V-Plus: Not only does it excel in image understanding, it also has video understanding capabilities based on time series analysis. This model will soon be launched on the open platform bigmodel.cn and will become the first general video understanding model API in China.<\/li>\n<li><a href=\"https:\/\/www.1ai.net\/en\/tag\/%e8%a7%86%e9%a2%91%e7%94%9f%e6%88%90%e6%a8%a1%e5%9e%8b\" title=\"_Other Organiser\" target=\"_blank\" >Video Generation Model<\/a>CogVideoX: After the release and open source of version 2B, version 5B was also officially open sourced to the outside world, with significantly improved performance, making it the leader among current open source video generation models.<\/li>\n<li>The cumulative downloads of Zhipu open source models have exceeded 20 million times, making a significant contribution to the prosperity and development of the open source community.<\/li>\n<\/ul>\n<p><img loading=\"lazy\" decoding=\"async\" class=\"alignnone size-full wp-image-19032\" title=\"a8dc381aj00sj26eh004yd000nw008xm\" src=\"https:\/\/www.1ai.net\/wp-content\/uploads\/2024\/08\/a8dc381aj00sj26eh004yd000nw008xm.jpg\" alt=\"a8dc381aj00sj26eh004yd000nw008xm\" width=\"860\" height=\"321\" \/><\/p>\n<p>GLM-4-Plus performs well in multiple key areas. In terms of language ability, the model has reached international leading levels in terms of comprehension, instruction following, and long text processing, and its performance is comparable to GPT-4 and Llama3.1 with 405B parameters. It is particularly worth mentioning that GLM-4-Plus improves the reasoning effect of long texts through a precise long and short text data mixing strategy.<\/p>\n<p><img loading=\"lazy\" decoding=\"async\" class=\"alignnone size-full wp-image-19033\" title=\"64ca6df2j00sj26eh004vd000nq009tm\" src=\"https:\/\/www.1ai.net\/wp-content\/uploads\/2024\/08\/64ca6df2j00sj26eh004vd000nq009tm.jpg\" alt=\"64ca6df2j00sj26eh004vd000nq009tm\" width=\"854\" height=\"353\" \/><\/p>\n<p>In the field of visual intelligence, GLM-4V-Plus demonstrates excellent image and video understanding capabilities. It not only has time perception capabilities, but also can process and understand complex video content. It is worth noting that the model will be launched on the Zhipu open platform and become the first general video understanding model API in China, providing powerful tools for developers and researchers.<\/p>\n<p><img loading=\"lazy\" decoding=\"async\" class=\"alignnone size-full wp-image-19035\" title=\"7f563270j00sj26eg004td000no006xm\" src=\"https:\/\/www.1ai.net\/wp-content\/uploads\/2024\/08\/7f563270j00sj26eg004td000no006xm.jpg\" alt=\"7f563270j00sj26eg004td000no006xm\" width=\"852\" height=\"249\" \/><\/p>\n<p>For example, if you give it a video like this and ask it what the player in green does throughout the video, it can accurately describe the actions taken by the player and tell you exactly at which second the highlight of the video is:<\/p>\n<p><img loading=\"lazy\" decoding=\"async\" class=\"alignnone size-full wp-image-19034\" title=\"9b5747fdj00sj26eh001xd000n40089m\" src=\"https:\/\/www.1ai.net\/wp-content\/uploads\/2024\/08\/9b5747fdj00sj26eh001xd000n40089m.jpg\" alt=\"9b5747fdj00sj26eh001xd000n40089m\" width=\"832\" height=\"297\" \/><\/p>\n<p>Screenshot from official<\/p>\n<p>CogView-3-Plus has already approached the current best models such as MJ-V6 and FLUX in terms of image performance. At the same time, the video generation model CogVideoX has launched a more powerful version 5B, which is considered to be the best choice among the current open source video generation models.<\/p>\n<p><img loading=\"lazy\" decoding=\"async\" class=\"alignnone size-full wp-image-19036\" title=\"6ca6f9bcj00sj26eh005ld000nw009fm\" src=\"https:\/\/www.1ai.net\/wp-content\/uploads\/2024\/08\/6ca6f9bcj00sj26eh005ld000nw009fm.jpg\" alt=\"6ca6f9bcj00sj26eh005ld000nw009fm\" width=\"860\" height=\"339\" \/><\/p>\n<p>THE MOST PROMISING IS THE IMMINENT INTRODUCTION OF THE \"VIDEO CALL\" FEATURE, THE FIRST IN THE COUNTRY TO BE OPEN TO THE C-END. THIS FUNCTION CUTS ACROSS THE THREE MAIN MODES OF TEXT, AUDIO AND VIDEO AND IS CAPABLE OF REAL-TIME REASONING. USERS CAN HAVE FLUID DIALOGUE WITH AI AND REACT QUICKLY EVEN WITH FREQUENT INTERRUPTIONS\u3002<\/p>\n<p>As long as the camera is turned on, AI can see and understand what the user sees, and accurately execute voice commands.<\/p>\n<p>The video call function was launched on August 30, and will be first opened to some Qingyan users, while also accepting external applications.<\/p>\n<p>Reference: https:\/\/mp.weixin.qq.com\/s\/Ww8njI4NiyH7arxML0nh8w<\/p>","protected":false},"excerpt":{"rendered":"<p>Smart Spectrum AI recently released its latest base big model GLM-4-Plus, demonstrating powerful visual capabilities comparable to OpenAI GPT-4, and announced that it will be open for use on August 30th. Main update highlights: Linguistic base model GLM-4-Plus:It has achieved a qualitative leap in language parsing, instruction execution and long text processing capabilities, and continues to maintain its leading position in international competition. Vincennes model CogView-3-Plus: Performance is comparable to the industry's top MJ-V6 and FLUX models. The image\/video understanding model GLM-4V-Plus: not only excels in image understanding, but also has the ability to understand video based on time series analysis. The model will soon be available on the open platform bigmodel.cn and<\/p>","protected":false},"author":1,"featured_media":0,"comment_status":"closed","ping_status":"closed","sticky":false,"template":"","format":"standard","meta":{"footnotes":""},"categories":[146],"tags":[216,1245,379,460,4212],"collection":[],"class_list":["post-19031","post","type-post","status-publish","format-standard","hentry","category-news","tag-216","tag-1245","tag-ai","tag-460","tag-4212"],"acf":[],"_links":{"self":[{"href":"https:\/\/www.1ai.net\/en\/wp-json\/wp\/v2\/posts\/19031","targetHints":{"allow":["GET"]}}],"collection":[{"href":"https:\/\/www.1ai.net\/en\/wp-json\/wp\/v2\/posts"}],"about":[{"href":"https:\/\/www.1ai.net\/en\/wp-json\/wp\/v2\/types\/post"}],"author":[{"embeddable":true,"href":"https:\/\/www.1ai.net\/en\/wp-json\/wp\/v2\/users\/1"}],"replies":[{"embeddable":true,"href":"https:\/\/www.1ai.net\/en\/wp-json\/wp\/v2\/comments?post=19031"}],"version-history":[{"count":0,"href":"https:\/\/www.1ai.net\/en\/wp-json\/wp\/v2\/posts\/19031\/revisions"}],"wp:attachment":[{"href":"https:\/\/www.1ai.net\/en\/wp-json\/wp\/v2\/media?parent=19031"}],"wp:term":[{"taxonomy":"category","embeddable":true,"href":"https:\/\/www.1ai.net\/en\/wp-json\/wp\/v2\/categories?post=19031"},{"taxonomy":"post_tag","embeddable":true,"href":"https:\/\/www.1ai.net\/en\/wp-json\/wp\/v2\/tags?post=19031"},{"taxonomy":"collection","embeddable":true,"href":"https:\/\/www.1ai.net\/en\/wp-json\/wp\/v2\/collection?post=19031"}],"curies":[{"name":"wp","href":"https:\/\/api.w.org\/{rel}","templated":true}]}}