{"id":20883,"date":"2024-09-30T10:33:11","date_gmt":"2024-09-30T02:33:11","guid":{"rendered":"https:\/\/www.1ai.net\/?p=20883"},"modified":"2024-09-30T10:33:11","modified_gmt":"2024-09-30T02:33:11","slug":"%e4%b8%ad%e5%9b%bd%e7%94%b5%e4%bf%a1-ai-%e7%a0%94%e7%a9%b6%e9%99%a2%e5%ae%8c%e6%88%90%e9%a6%96%e4%b8%aa%e5%85%a8%e5%9b%bd%e4%ba%a7%e5%8c%96%e4%b8%87%e5%8d%a1%e4%b8%87%e5%8f%82%e5%a4%a7%e6%a8%a1","status":"publish","type":"post","link":"https:\/\/www.1ai.net\/en\/20883.html","title":{"rendered":"China Telecom AI Research Institute Completes the First Fully Localized Wankawansen Large Model Training, and TeleChat2-115B Is Open-Sourced to the Public"},"content":{"rendered":"<p><img loading=\"lazy\" decoding=\"async\" class=\"alignnone size-full wp-image-20884\" title=\"a0490a73j00skltoy007nd000je00gip\" src=\"https:\/\/www.1ai.net\/wp-content\/uploads\/2024\/09\/a0490a73j00skltoy007nd000je00gip.jpg\" alt=\"a0490a73j00skltoy007nd000je00gip\" width=\"698\" height=\"594\" \/><\/p>\n<p>Sept. 28, \"<a href=\"https:\/\/www.1ai.net\/en\/tag\/%e4%b8%ad%e5%9b%bd%e7%94%b5%e4%bf%a1\" title=\"[View articles tagged with [China Telecom]]\" target=\"_blank\" >China Telecom<\/a><a href=\"https:\/\/www.1ai.net\/en\/tag\/%e4%ba%ba%e5%b7%a5%e6%99%ba%e8%83%bd%e7%a0%94%e7%a9%b6%e9%99%a2\" title=\"[Sees articles with labels]\" target=\"_blank\" >Artificial Intelligence Research Institute<\/a>\"The official public number announced that China Telecom Artificial Intelligence Research Institute (hereinafter referred to as TeleAI) successfully completed the National<strong>the first trillion-parameter large model trained on a fully localized Wanka cluster.<\/strong>and officially open to the public<a href=\"https:\/\/www.1ai.net\/en\/tag\/%e5%bc%80%e6%ba%90\" title=\"[View articles tagged with [open source]]\" target=\"_blank\" >Open Source<\/a>The First 100 Billion Parameter Large Model Trained on Fully Domesticated Wanka Cluster and Domestic Deep Learning Framework -- Star<strong>Tatsu semantic macromodel TeleChat 2-115B.<\/strong><\/p>\n<p>Officials say the scientific research marks the<strong>Domestic large model training to truly realize the full localization of alternative<\/strong>The company has formally entered a new stage of independent innovation, safety and control of the national production.<\/p>\n<p>TeleChat2-115B is based on China Telecom's self-developed Tianyi Cloud \"Xiyang Integrated Intelligent Computing Service Platform\" and the artificial intelligence company's \"Star Ocean AI Platform\". According to reports, under the premise of guaranteeing the training accuracy, it uses a variety of optimization means to improve the model training efficiency and stability, and achieves the GPU equivalent arithmetic computation efficiency of more than 93%, and the effective training time of the model accounts for more than 98%.<\/p>\n<p>For ultra-large parameter model training, TeleAI employs a<strong>Lots of miniatures<\/strong>Scaling is performed to verify the effectiveness of different model structures. Meanwhile, in terms of data allocation, based on the feedback of the experimental results of the small model, the regression prediction model is used to get the better data allocation.<\/p>\n<p>For Post-Training, TeleAI first synthesized a large amount of Q&amp;A data for math, code, and logical reasoning for the first phase of SFT (supervised fine-tuning) model training.<\/p>\n<p>Second, it adopts an iterative updating strategy that uses models to enhance the complexity of instructions and expand the diversity of cue word data, improves the quality of answers through model synthesis and manual annotation, and utilizes rejection sampling to obtain high-quality SFT data and representative data of the RM (Reward Model), which are used for SFT training and DPO (Preference Alignment) training as well as iterative modeling effects.<\/p>\n<p><strong>with open source address<\/strong><\/p>\n<p>GitHub:<\/p>\n<ul>\n<li>https:\/\/github.com\/Tele-AI\/TeleChat2<\/li>\n<\/ul>\n<p>Gitee:<\/p>\n<ul>\n<li>https:\/\/gitee.com\/Tele-AI\/tele-chat2<\/li>\n<\/ul>\n<p>ModelScope:<\/p>\n<ul>\n<li>https:\/\/modelscope.cn\/models\/TeleAI\/TeleChat2-115B<\/li>\n<\/ul>\n<p>Modelers:<\/p>\n<ul>\n<li>https:\/\/modelers.cn\/models\/TeleAI\/TeleChat2-115B<\/li>\n<\/ul>","protected":false},"excerpt":{"rendered":"<p>On September 28th, the official public number of \"China Telecom Artificial Intelligence Research Institute\" announced that China Telecom Artificial Intelligence Research Institute (hereinafter referred to as TeleAI) has successfully completed the first trillion-parameter large model based on the fully localized WANKA cluster training in China, and formally open-sourced the first trillion-parameter large model -- Star Semantic Large Model TeleChat2-115B, which was trained based on the fully localized WANKA cluster and the deep learning framework of China. TeleChat2-115B, the first 100 billion parameter large model trained on a fully localized Wanka cluster and a domestic deep learning framework, has been officially open-sourced to the public. According to the official statement, this scientific research achievement marks that the training of the domestic large model has truly achieved the replacement of all localization, and has formally entered into a new stage of independent innovation, safety and control of the national production. TeleChat2-115B is based on China Telecom's self-developed Tianyi Cloud \"Xiyang Integrated Intelligent Computing Service Platform\" and artificial intelligence company \"Xinghai AI Platform\".<\/p>","protected":false},"author":1,"featured_media":0,"comment_status":"closed","ping_status":"closed","sticky":false,"template":"","format":"standard","meta":{"footnotes":""},"categories":[148,146],"tags":[3419,4523,219],"collection":[],"class_list":["post-20883","post","type-post","status-publish","format-standard","hentry","category-headline","category-news","tag-3419","tag-4523","tag-219"],"acf":[],"_links":{"self":[{"href":"https:\/\/www.1ai.net\/en\/wp-json\/wp\/v2\/posts\/20883","targetHints":{"allow":["GET"]}}],"collection":[{"href":"https:\/\/www.1ai.net\/en\/wp-json\/wp\/v2\/posts"}],"about":[{"href":"https:\/\/www.1ai.net\/en\/wp-json\/wp\/v2\/types\/post"}],"author":[{"embeddable":true,"href":"https:\/\/www.1ai.net\/en\/wp-json\/wp\/v2\/users\/1"}],"replies":[{"embeddable":true,"href":"https:\/\/www.1ai.net\/en\/wp-json\/wp\/v2\/comments?post=20883"}],"version-history":[{"count":0,"href":"https:\/\/www.1ai.net\/en\/wp-json\/wp\/v2\/posts\/20883\/revisions"}],"wp:attachment":[{"href":"https:\/\/www.1ai.net\/en\/wp-json\/wp\/v2\/media?parent=20883"}],"wp:term":[{"taxonomy":"category","embeddable":true,"href":"https:\/\/www.1ai.net\/en\/wp-json\/wp\/v2\/categories?post=20883"},{"taxonomy":"post_tag","embeddable":true,"href":"https:\/\/www.1ai.net\/en\/wp-json\/wp\/v2\/tags?post=20883"},{"taxonomy":"collection","embeddable":true,"href":"https:\/\/www.1ai.net\/en\/wp-json\/wp\/v2\/collection?post=20883"}],"curies":[{"name":"wp","href":"https:\/\/api.w.org\/{rel}","templated":true}]}}