{"id":22021,"date":"2024-10-26T09:54:21","date_gmt":"2024-10-26T01:54:21","guid":{"rendered":"https:\/\/www.1ai.net\/?p=22021"},"modified":"2024-10-26T09:54:21","modified_gmt":"2024-10-26T01:54:21","slug":"%e6%99%ba%e8%b0%b1%e6%b8%85%e8%a8%80%e4%b8%8a%e7%ba%bf%e6%83%85%e6%84%9f%e8%af%ad%e9%9f%b3%e6%a8%a1%e5%9e%8b-glm-4-voice%ef%bc%9a%e5%8f%af%e7%90%86%e8%a7%a3%e6%83%85%e6%84%9f%ef%bc%8c%e6%9c%89","status":"publish","type":"post","link":"https:\/\/www.1ai.net\/en\/22021.html","title":{"rendered":"Wisdom Spectrum Clear Speech Launches Emotional Speech Model GLM-4-Voice: Understanding Emotions, Emotional Expression and Empathy"},"content":{"rendered":"<p><img loading=\"lazy\" decoding=\"async\" class=\"alignnone size-full wp-image-22022\" title=\"0c2b0f2dj00slxx940092d000rs00gtp\" src=\"https:\/\/www.1ai.net\/wp-content\/uploads\/2024\/10\/0c2b0f2dj00slxx940092d000rs00gtp.jpg\" alt=\"0c2b0f2dj00slxx940092d000rs00gtp\" width=\"1000\" height=\"605\" \/><\/p>\n<p>Smart Spectrum Announces the Launch of GLM-4-Voice End-to-End Emotion<a href=\"https:\/\/www.1ai.net\/en\/tag\/%e8%af%ad%e9%9f%b3%e6%a8%a1%e5%9e%8b\" title=\"[View articles tagged with [speech modeling]]\" target=\"_blank\" >Voice Model<\/a>. Officially, its<strong>Ability to understand emotions, emotional expression, emotional empathy<\/strong>The program, which is self-adjusting, supports multiple languages and dialects, and features lower latency and the ability to interrupt at any time, is available to users now at \"<a href=\"https:\/\/www.1ai.net\/en\/tag\/%e6%99%ba%e8%b0%b1%e6%b8%85%e8%a8%80\" title=\"Look at the article that contains the tag\" target=\"_blank\" >Zhipu Qingyan<\/a>\"Experience on the App.<\/p>\n<p>The GLM-4-Voice is described as having the following features:<\/p>\n<ul>\n<li><strong>Emotional expression and emotional resonance:<\/strong>Voices have different emotions and subtleties, such as happy, sad, angry, and scared.<\/li>\n<li><strong>Adjust the speed of speech:<\/strong>In the same round of conversation, you can ask the TA to speak faster or slower.<\/li>\n<li><strong>Interrupt at any time and enter instructions flexibly:<\/strong>Adjust the content and style of voice output based on real-time user commands to support more flexible dialog interactions.<\/li>\n<li><strong>Multi-language and multi-dialect support:<\/strong>At present, GLM-4-Voice supports Chinese and English voices as well as dialects from all over China, and is especially good at Cantonese, Chongqing and Beijing.<\/li>\n<li><strong>Combined with video calling, you can see and talk:<\/strong>A video calling feature will be available soon.<\/li>\n<\/ul>\n<p>In addition, AutoGLM is equipped with phone use capability, which allows it to simulate human operation of a cell phone by receiving simple text\/voice commands. It is not limited to simple task scenarios or API calls, nor does it require users to manually build complex and cumbersome workflows, and its operation logic is similar to that of humans.<\/p>\n<p>GLM-4-Voice is open-sourced in the same period, and is officially called the first open-sourced end-to-end multimodal model of Smart Spectrum.IT Home with address:<\/p>\n<p>Code Repository:<\/p>\n<ul>\n<li>https:\/\/github.com\/THUDM\/GLM-4-Voice<\/li>\n<\/ul>","protected":false},"excerpt":{"rendered":"<p>Wisdom Spectrum announced the launch of GLM-4-Voice end-to-end emotional voice model. Officially, GLM-4-Voice is able to understand emotions, express and resonate emotions, self-adjust its speech rate, support multiple languages and dialects, have lower latency, and can be interrupted at any time, which can be experienced by users on the \"Wisdom Spectrum Clear Speech\" App from now on. According to the introduction, GLM-4-Voice has the following features: Emotional expression and emotional resonance: the voice has different emotions and subtle changes, such as happy, sad, angry, scared, etc. Adjust the speed of speech: In the same round of conversation, you can ask the TA to speak faster or slower. Interrupt at any time and input commands flexibly: adjust the content and style of voice output according to real-time user commands, supporting more flexible dialog interaction. Multi-language and multi-dialect support<\/p>","protected":false},"author":1,"featured_media":0,"comment_status":"closed","ping_status":"closed","sticky":false,"template":"","format":"standard","meta":{"footnotes":""},"categories":[146],"tags":[822,1875],"collection":[],"class_list":["post-22021","post","type-post","status-publish","format-standard","hentry","category-news","tag-822","tag-1875"],"acf":[],"_links":{"self":[{"href":"https:\/\/www.1ai.net\/en\/wp-json\/wp\/v2\/posts\/22021","targetHints":{"allow":["GET"]}}],"collection":[{"href":"https:\/\/www.1ai.net\/en\/wp-json\/wp\/v2\/posts"}],"about":[{"href":"https:\/\/www.1ai.net\/en\/wp-json\/wp\/v2\/types\/post"}],"author":[{"embeddable":true,"href":"https:\/\/www.1ai.net\/en\/wp-json\/wp\/v2\/users\/1"}],"replies":[{"embeddable":true,"href":"https:\/\/www.1ai.net\/en\/wp-json\/wp\/v2\/comments?post=22021"}],"version-history":[{"count":0,"href":"https:\/\/www.1ai.net\/en\/wp-json\/wp\/v2\/posts\/22021\/revisions"}],"wp:attachment":[{"href":"https:\/\/www.1ai.net\/en\/wp-json\/wp\/v2\/media?parent=22021"}],"wp:term":[{"taxonomy":"category","embeddable":true,"href":"https:\/\/www.1ai.net\/en\/wp-json\/wp\/v2\/categories?post=22021"},{"taxonomy":"post_tag","embeddable":true,"href":"https:\/\/www.1ai.net\/en\/wp-json\/wp\/v2\/tags?post=22021"},{"taxonomy":"collection","embeddable":true,"href":"https:\/\/www.1ai.net\/en\/wp-json\/wp\/v2\/collection?post=22021"}],"curies":[{"name":"wp","href":"https:\/\/api.w.org\/{rel}","templated":true}]}}