{"id":6575,"date":"2024-03-28T14:39:56","date_gmt":"2024-03-28T06:39:56","guid":{"rendered":"https:\/\/www.1ai.net\/?p=6575"},"modified":"2024-03-28T14:39:56","modified_gmt":"2024-03-28T06:39:56","slug":"%e5%85%83%e8%b1%a1%e5%a4%a7%e6%a8%a1%e5%9e%8b%e5%bc%80%e6%ba%9030%e6%ac%be%e9%87%8f%e5%8c%96%e7%89%88%e6%9c%ac-%e5%8f%af%e6%9b%b4%e4%bd%8e%e6%88%90%e6%9c%ac%e9%83%a8%e7%bd%b2","status":"publish","type":"post","link":"https:\/\/www.1ai.net\/en\/6575.html","title":{"rendered":"Yuanxiang&#039;s open source model has 30 quantitative versions that can be deployed at a lower cost"},"content":{"rendered":"<p><a href=\"https:\/\/www.1ai.net\/en\/tag\/%e5%85%83%e8%b1%a1\" title=\"[Sees articles with [earth] labels]\" target=\"_blank\" >Yuanxiang<\/a><a href=\"https:\/\/www.1ai.net\/en\/tag\/%e5%a4%a7%e6%a8%a1%e5%9e%8b\" title=\"[View articles tagged with [large models]]\" target=\"_blank\" >Large Model<\/a><a href=\"https:\/\/www.1ai.net\/en\/tag\/%e5%bc%80%e6%ba%90\" title=\"[View articles tagged with [open source]]\" target=\"_blank\" >Open Source<\/a>30 quantitative versions have been released, supporting quantitative reasoning of mainstream frameworks such as vLLM and llama.cpp, and are unconditionally free for commercial use.<\/p>\n<p>The model capabilities and inference performance before and after quantization were evaluated. Taking the quantized version of XVERSE-13B-GPTQ-Int4 as an example, the model weights were compressed by 72% after quantization, the total throughput was increased by 1.5 times, and the capacity of 95% was retained.<\/p>\n<p class=\"article-content__img\"><img loading=\"lazy\" decoding=\"async\" class=\"alignnone size-full wp-image-6576\" src=\"https:\/\/www.1ai.net\/wp-content\/uploads\/2024\/03\/6384723181632147173094353.jpg\" alt=\"\" width=\"1000\" height=\"627\" \/><\/p>\n<p>Developers can choose models with different reasoning frameworks and data accuracy based on their skills, hardware and software configurations, and specific needs. If local resources are limited, you can directly call the API service of the Yuanxiang large model (chat.xverse.cn).<\/p>\n<p>In general, the open source quantitative version of the Yuanxiang large model provides a convenient and fast deployment method, and different frameworks and precision models can be selected for deployment and inference according to needs.<\/p>\n<p><strong>Download the Yuanxiang large model:<\/strong><\/p>\n<ul>\n<li>Hugging Face: https:\/\/huggingface.co\/xverse<\/li>\n<li>ModelScope Magic: https:\/\/modelscope.cn\/organization\/xverse<\/li>\n<li>Github:https:\/\/github.com\/xverse-ai<\/li>\n<\/ul>","protected":false},"excerpt":{"rendered":"<p>Meta-Elephant Big Model has open-sourced 30 quantized versions, supporting quantized reasoning for mainstream frameworks such as vLLM and llama.cpp, and is unconditionally free for commercial use. Evaluate the model capability and reasoning performance before and after quantization. Take the quantization version of XVERSE-13B-GPTQ-Int4 for example, after quantization, the model weight is compressed by 72%, and the total throughput is increased by 1.5 times, while retaining the capability of 95%. Developers can choose models with different inference frameworks and data accuracy according to their skills, hardware and software configurations and specific needs. If local resources are limited, they can directly call the API service of Meta-Elephant Big Model (chat.xverse.cn). Overall, the open-source quantitative version of the Meta-Elephant Big Model provides a convenient and fast deployment method, which can be rooted in the<\/p>","protected":false},"author":1,"featured_media":0,"comment_status":"open","ping_status":"open","sticky":false,"template":"","format":"standard","meta":{"footnotes":""},"categories":[146],"tags":[1936,216,219],"collection":[],"class_list":["post-6575","post","type-post","status-publish","format-standard","hentry","category-news","tag-1936","tag-216","tag-219"],"acf":[],"_links":{"self":[{"href":"https:\/\/www.1ai.net\/en\/wp-json\/wp\/v2\/posts\/6575","targetHints":{"allow":["GET"]}}],"collection":[{"href":"https:\/\/www.1ai.net\/en\/wp-json\/wp\/v2\/posts"}],"about":[{"href":"https:\/\/www.1ai.net\/en\/wp-json\/wp\/v2\/types\/post"}],"author":[{"embeddable":true,"href":"https:\/\/www.1ai.net\/en\/wp-json\/wp\/v2\/users\/1"}],"replies":[{"embeddable":true,"href":"https:\/\/www.1ai.net\/en\/wp-json\/wp\/v2\/comments?post=6575"}],"version-history":[{"count":0,"href":"https:\/\/www.1ai.net\/en\/wp-json\/wp\/v2\/posts\/6575\/revisions"}],"wp:attachment":[{"href":"https:\/\/www.1ai.net\/en\/wp-json\/wp\/v2\/media?parent=6575"}],"wp:term":[{"taxonomy":"category","embeddable":true,"href":"https:\/\/www.1ai.net\/en\/wp-json\/wp\/v2\/categories?post=6575"},{"taxonomy":"post_tag","embeddable":true,"href":"https:\/\/www.1ai.net\/en\/wp-json\/wp\/v2\/tags?post=6575"},{"taxonomy":"collection","embeddable":true,"href":"https:\/\/www.1ai.net\/en\/wp-json\/wp\/v2\/collection?post=6575"}],"curies":[{"name":"wp","href":"https:\/\/api.w.org\/{rel}","templated":true}]}}