{"id":27962,"date":"2025-01-31T14:24:18","date_gmt":"2025-01-31T06:24:18","guid":{"rendered":"https:\/\/www.1ai.net\/?p=27962"},"modified":"2025-01-31T14:24:18","modified_gmt":"2025-01-31T06:24:18","slug":"%e8%8b%b1%e4%bc%9f%e8%be%be%ef%bc%9adeepseek-r1-%e6%a8%a1%e5%9e%8b%e7%8e%b0%e5%b7%b2%e4%b8%8a%e7%ba%bf-nim-%e5%be%ae%e6%9c%8d%e5%8a%a1%e5%b9%b3%e5%8f%b0","status":"publish","type":"post","link":"https:\/\/www.1ai.net\/en\/27962.html","title":{"rendered":"NVIDIA: DeepSeek-R1 Model Now Live on NIM Microservices Platform"},"content":{"rendered":"<p>January 31st.<a href=\"https:\/\/www.1ai.net\/en\/tag\/%e8%8b%b1%e4%bc%9f%e8%be%be\" title=\"Look at the article with the label\" target=\"_blank\" >Nvidia<\/a>Announced today,<a href=\"https:\/\/www.1ai.net\/en\/tag\/deepseek\" title=\"[View articles tagged with [DeepSeek]]\" target=\"_blank\" >DeepSeek<\/a>The -R1 model is now available as an NVIDIA NIM Microservices Preview at build.nvidia.com.The DeepSeek-R1 NIM microservice can deliver up to 3,872 tokens per second on a single NVIDIA HGX H200 system.<\/p>\n<p><img loading=\"lazy\" decoding=\"async\" class=\"alignnone size-full wp-image-27963\" title=\"c508c76dj00sqxwet0017d000v100e8p\" src=\"https:\/\/www.1ai.net\/wp-content\/uploads\/2025\/01\/c508c76dj00sqxwet0017d000v100e8p.jpg\" alt=\"c508c76dj00sqxwet0017d000v100e8p\" width=\"1117\" height=\"512\" \/><\/p>\n<p>It was stated that developers can test and experiment with application programming interfaces (APIs), the<strong>The interface is expected to be available soon as a downloadable NIM microservice<\/strong>This is part of the NVIDIA AI Enterprise software platform.<\/p>\n<p>DeepSeek-R1 NIM microservices simplify deployment by supporting industry standard APIs. Organizations can maximize security and data privacy by running NIM microservices on their preferred accelerated computing infrastructure. By combining NVIDIA AI Foundry with NVIDIA NeMo software, organizations will also be able to create customized DeepSeek-R1 NIM microservices for dedicated AI agents.<\/p>\n<p>1AI notes that to make it easier for organizations of all sizes to deploy AI services, Nvidia launched NIM (Nvidia Inference Microservices) cloud-native microservices in March 2024.<\/p>\n<p>NIM is a suite of cloud-native microservices optimized to reduce time-to-market and simplify the deployment of generative AI models anywhere in the cloud, data centers, and GPU-accelerated workstations. It extends the developer pool by abstracting the complexity of AI model development and production packaging using industry-standard APIs.<\/p>","protected":false},"excerpt":{"rendered":"<p>On January 31st, Ying Weidar announced today that the DeepSeek-R1 model is now available as a NVIDIA NIM microservice preview on the bild.nvidia.com. DeepSeek-R1 NIM microservices can provide up to 3,872 tokens per second on a single NVIDIA HGX H200 system. It was described that developers could test and test the Application Programming Interface (API), which was expected to be made available shortly as a downloadable NIM microservice as part of the NVIDIA AI Enterprise software platform. DeepSeek-R1 NIM Microservices Through Support to Industry Standards A<\/p>","protected":false},"author":1,"featured_media":0,"comment_status":"closed","ping_status":"closed","sticky":false,"template":"","format":"standard","meta":{"footnotes":""},"categories":[146],"tags":[3606,239],"collection":[],"class_list":["post-27962","post","type-post","status-publish","format-standard","hentry","category-news","tag-deepseek","tag-239"],"acf":[],"_links":{"self":[{"href":"https:\/\/www.1ai.net\/en\/wp-json\/wp\/v2\/posts\/27962","targetHints":{"allow":["GET"]}}],"collection":[{"href":"https:\/\/www.1ai.net\/en\/wp-json\/wp\/v2\/posts"}],"about":[{"href":"https:\/\/www.1ai.net\/en\/wp-json\/wp\/v2\/types\/post"}],"author":[{"embeddable":true,"href":"https:\/\/www.1ai.net\/en\/wp-json\/wp\/v2\/users\/1"}],"replies":[{"embeddable":true,"href":"https:\/\/www.1ai.net\/en\/wp-json\/wp\/v2\/comments?post=27962"}],"version-history":[{"count":0,"href":"https:\/\/www.1ai.net\/en\/wp-json\/wp\/v2\/posts\/27962\/revisions"}],"wp:attachment":[{"href":"https:\/\/www.1ai.net\/en\/wp-json\/wp\/v2\/media?parent=27962"}],"wp:term":[{"taxonomy":"category","embeddable":true,"href":"https:\/\/www.1ai.net\/en\/wp-json\/wp\/v2\/categories?post=27962"},{"taxonomy":"post_tag","embeddable":true,"href":"https:\/\/www.1ai.net\/en\/wp-json\/wp\/v2\/tags?post=27962"},{"taxonomy":"collection","embeddable":true,"href":"https:\/\/www.1ai.net\/en\/wp-json\/wp\/v2\/collection?post=27962"}],"curies":[{"name":"wp","href":"https:\/\/api.w.org\/{rel}","templated":true}]}}