{"id":25139,"date":"2024-12-15T06:38:19","date_gmt":"2024-12-14T22:38:19","guid":{"rendered":"https:\/\/www.1ai.net\/?p=25139"},"modified":"2024-12-14T23:41:39","modified_gmt":"2024-12-14T15:41:39","slug":"%e9%98%b6%e8%b7%83%e6%98%9f%e8%be%b0%e6%8e%a8%e5%87%ba%e5%9b%bd%e5%86%85%e9%a6%96%e4%b8%aa%e5%8d%83%e4%ba%bf%e5%8f%82%e6%95%b0%e7%ab%af%e5%88%b0%e7%ab%af%e8%af%ad%e9%9f%b3%e5%a4%a7%e6%a8%a1%e5%9e%8b","status":"publish","type":"post","link":"https:\/\/www.1ai.net\/en\/25139.html","title":{"rendered":"Step-Star launches China's first end-to-end speech grand model with hundreds of billions of parameters \" Step-1o\""},"content":{"rendered":"<p><a href=\"https:\/\/www.1ai.net\/en\/tag\/%e9%98%b6%e8%b7%83%e6%98%9f%e8%be%b0\" title=\"[View articles tagged with [Step Star]]\" target=\"_blank\" >Step Star<\/a>Announcing the launch of Step-1o's 100 billion parameter end-to-end in public<a href=\"https:\/\/www.1ai.net\/en\/tag\/%e8%af%ad%e9%9f%b3%e5%a4%a7%e6%a8%a1%e5%9e%8b\" title=\"[View articles tagged with [Voice Megamodel]]\" target=\"_blank\" >Voice big model<\/a>The \"allegedly\"<strong>The first end-to-end speech macromodel with 100 billion parameters in China\".<\/strong><\/p>\n<p><img loading=\"lazy\" decoding=\"async\" class=\"alignnone size-full wp-image-25140\" title=\"64622ab4p00sohq8m0006d0009n0041p\" src=\"https:\/\/www.1ai.net\/wp-content\/uploads\/2024\/12\/64622ab4p00sohq8m0006d0009n0041p.png\" alt=\"64622ab4p00sohq8m0006d0009n0041p\" width=\"347\" height=\"145\" \/><\/p>\n<p>According to Step Star, the traditional voice model uses a cascade program, the user input voice information needs to be converted into text, and then converted into voice output, this process will not only reduce the transmission efficiency, but also in the process of loss of information, including emotions, resulting in the voice model to extend the slow response, the quality of the answer and the level of intelligence is insufficient, the expression of emotion is empty stereotypes of the shortcomings. However, an end-to-end voice solution that<strong>Integration of speech understanding and generation is possible<\/strong>, raising the upper limits of the model's IQ and EQ.<\/p>\n<p>1AI learned from the official presentation that<strong>Step-1o supports mixed forms of input and output such as voice and text.<\/strong>It can respond quickly and interrupt at any time, and it also understands and imitates in depth vocal characteristics such as timbre, rhythm, dialect, and individualized habits of spoken expression;<\/p>\n<p>Step-1o is able to continuously improve the quality of its responses through self-learning and imitation, both by providing professional advice on problem solving and as a companion providing high emotional value. In addition, Step-1o has inherited the ability to create a large model of the Step-Star language.<\/p>\n<p>Step-1o will be connected to the Leapfrog App terminal in the near future, Step-Star revealed.<strong>Provide real-time voice call service for users<\/strong>.<\/p>","protected":false},"excerpt":{"rendered":"<p>Step-1o announced the launch of Step-1o end-to-end speech model with hundreds of billions of parameters, which is said to be \"the first end-to-end speech model with hundreds of billions of parameters in China\". According to Step-1o, the traditional speech model adopts the cascade program, the user input voice information needs to be converted into text, and then converted into voice output, this process will not only reduce the transmission efficiency, but also in the process of loss of information, including emotions, resulting in the speech model to extend the slow response, the answer to the quality of the level of intelligence is insufficient, the expression of emotion empty stereotypes of the shortcomings. However, an end-to-end speech solution can realize the integration of speech understanding and generation, and improve the upper limit of the model IQ and emotional intelligence. 1AI learned from the official introduction that Step-1o supports input and output in mixed forms such as speech and text, and can quickly<\/p>","protected":false},"author":1,"featured_media":0,"comment_status":"closed","ping_status":"closed","sticky":false,"template":"","format":"standard","meta":{"footnotes":""},"categories":[146],"tags":[4061,1893],"collection":[],"class_list":["post-25139","post","type-post","status-publish","format-standard","hentry","category-news","tag-4061","tag-1893"],"acf":[],"_links":{"self":[{"href":"https:\/\/www.1ai.net\/en\/wp-json\/wp\/v2\/posts\/25139","targetHints":{"allow":["GET"]}}],"collection":[{"href":"https:\/\/www.1ai.net\/en\/wp-json\/wp\/v2\/posts"}],"about":[{"href":"https:\/\/www.1ai.net\/en\/wp-json\/wp\/v2\/types\/post"}],"author":[{"embeddable":true,"href":"https:\/\/www.1ai.net\/en\/wp-json\/wp\/v2\/users\/1"}],"replies":[{"embeddable":true,"href":"https:\/\/www.1ai.net\/en\/wp-json\/wp\/v2\/comments?post=25139"}],"version-history":[{"count":0,"href":"https:\/\/www.1ai.net\/en\/wp-json\/wp\/v2\/posts\/25139\/revisions"}],"wp:attachment":[{"href":"https:\/\/www.1ai.net\/en\/wp-json\/wp\/v2\/media?parent=25139"}],"wp:term":[{"taxonomy":"category","embeddable":true,"href":"https:\/\/www.1ai.net\/en\/wp-json\/wp\/v2\/categories?post=25139"},{"taxonomy":"post_tag","embeddable":true,"href":"https:\/\/www.1ai.net\/en\/wp-json\/wp\/v2\/tags?post=25139"},{"taxonomy":"collection","embeddable":true,"href":"https:\/\/www.1ai.net\/en\/wp-json\/wp\/v2\/collection?post=25139"}],"curies":[{"name":"wp","href":"https:\/\/api.w.org\/{rel}","templated":true}]}}