{"id":1250,"date":"2023-11-14T11:01:43","date_gmt":"2023-11-14T03:01:43","guid":{"rendered":"https:\/\/www.1ai.net\/?p=1250"},"modified":"2023-11-14T11:01:43","modified_gmt":"2023-11-14T03:01:43","slug":"%e8%8b%b1%e4%bc%9f%e8%be%be%e6%8e%a8%e5%87%ba-ai-%e9%a2%86%e5%9f%9f%e9%a1%b6%e5%b0%96%e6%96%b0%e8%8a%af%e7%89%87-hgx-h200%ef%bc%9a%e4%bb%a5%e6%9b%b4%e9%ab%98%e6%80%a7%e8%83%bd%e6%bb%a1%e8%b6%b3","status":"publish","type":"post","link":"https:\/\/www.1ai.net\/en\/1250.html","title":{"rendered":"NVIDIA launches HGX H200, a new AI chip that meets the needs of generative AI with higher performance"},"content":{"rendered":"<p><a href=\"https:\/\/www.1ai.net\/en\/tag\/%e8%8b%b1%e4%bc%9f%e8%be%be\" title=\"Look at the article with the label\" target=\"_blank\" >Nvidia<\/a>The company announced Monday local time the launch of a new generation of its AI efforts<span class=\"spamTxt\">Top<\/span><a href=\"https:\/\/www.1ai.net\/en\/tag\/%e8%8a%af%e7%89%87\" title=\"[Sees articles with [chips] labels]\" target=\"_blank\" >chip<\/a> -- HGX H200. <a href=\"https:\/\/www.1ai.net\/en\/tag\/gpu\" title=\"_OTHER ORGANISER\" target=\"_blank\" >GPU<\/a> Based on its popular predecessor, the H100, it realized the<strong> 1.4x memory bandwidth increase and 1.8x memory capacity increase<\/strong>,<strong>Dramatically improved its ability to handle intensive generative AI work<\/strong>.<\/p>\n<p class=\"article-content__img\"><img loading=\"lazy\" decoding=\"async\" class=\"alignnone size-full wp-image-1251\" title=\"202311140829471179_0\" src=\"https:\/\/www.1ai.net\/wp-content\/uploads\/2023\/11\/202311140829471179_0.jpg\" alt=\"202311140829471179_0\" width=\"1000\" height=\"561\" \/><\/p>\n<p>However, there are still questions about the availability of these new chips in the market, especially given the supply constraints of the H100. NVIDIA has not yet provided a definitive answer to this question. NVIDIA has stated that<strong>First H200 chips to be released in Q2 2024<\/strong>NVIDIA is currently working with \"global system manufacturers and cloud service providers\" to ensure chip availability. NVIDIA spokeswoman Kristin Uchiyama would not comment on production numbers.<\/p>\n<p>With the exception of memory, the H200 is largely identical to the H100 in every other respect. But the memory improvements make it a meaningful upgrade. This new GPU is<strong><span class=\"spamTxt\">The first<\/span>Increased its memory bandwidth to 4.8TB per second with the new, faster HBM3e memory specification<\/strong>This is a significant increase from the H100's 3.35TB.<strong>Total memory capacity has also been increased from 80GB to 141GB.<\/strong>.<\/p>\n<p>The integration of faster and higher-capacity HBM memory is designed to accelerate the performance of compute-intensive tasks, including generating AI models and high-performance computing applications, while optimizing GPU utilization and efficiency,\" said Ian Buck, vice president of high-performance computing products at NVIDIA, in a video demonstration today.<\/p>\n<p>The H200 is also designed to be compatible with systems that already support the H100. NVIDIA states that<strong>Cloud Service Providers need not make any changes when joining H200<\/strong>.. The cloud services divisions of Amazon, Google, Microsoft and Oracle will be among the first to offer new GPUs next year.<\/p>\n<p>The new chip is set to sell for a lot of money once it's launched. While NVIDIA hasn't announced its price, CNBC reports that the previous-generation H100s were estimated to sell for between $25,000 and $40,000 each, running<span class=\"spamTxt\">Top<\/span>Thousands of them are needed for a horizontal system, Uchiyama said, adding that pricing is determined by NVIDIA's partners.<\/p>\n<p>With AI companies still desperately looking for H100 chips, this announcement from NVIDIA is significant.<strong>NVIDIA's chips are seen as efficiently handling the huge amounts of data needed to generate image tools and large language models<span class=\"spamTxt\">optimal<\/span>option<\/strong>. These chips were valuable enough that companies used them as collateral for loans. Who owned the H100 became a hot topic in Silicon Valley, with startups even collaborating to share any availability of these chips.<\/p>\n<p><strong>Uchiyama says H200 launch won't affect H100 production<\/strong>. \"You're going to see us increase overall supply this year, and we're in the process of long-term supply purchases,\" Uchiyama wrote in an email to The Verge.<\/p>\n<p>Looking ahead to next year, it may be an even more auspicious time for GPU buyers. The Financial Times reported in August that<strong>NVIDIA plans to triple H100 production by 2024<\/strong>The goal is to produce up to 2 million next year, compared to about 500,000 in 2023. The goal is to produce as many as 2 million next year, compared to about 500,000 in 2023. But with demand for generative AI still as strong today as it was at the beginning of the year, demand is likely to only get bigger - and that doesn't even include the hotter new chips NVIDIA is rolling out.<\/p>","protected":false},"excerpt":{"rendered":"<p>ON MONDAY, THE LOCAL TIME ANNOUNCED THE LAUNCH OF A NEW GENERATION OF TOP CHIPS FOR ITS AI WORK - HGX H200. THE NEW GPU, BUILDING ON ITS POPULAR PREDECESSOR PRODUCT H100, ACHIEVED 1.4 TIMES THE RAM BANDWIDTH INCREASE AND 1.8 TIMES THE RAM INCREASE, SIGNIFICANTLY IMPROVING ITS ABILITY TO HANDLE INTENSIVE AI WORK. HOWEVER, THE CURRENT MARKET STILL QUESTIONS THE AVAILABILITY OF THESE NEW CHIPS, ESPECIALLY GIVEN H100 SUPPLY RESTRICTIONS. IT HAS NOT YET PROVIDED A DEFINITIVE RESPONSE TO THIS QUESTION. ACCORDING TO IN WEIDA, THE FIRST H200 CHIPS WILL BE RELEASED IN THE SECOND QUARTER OF 2024 AND ARE CURRENTLY WORKING WITH GLOBAL SYSTEM MANUFACTURERS AND CLOUD SERVICE PROVIDERS TO ENSURE THEIR AVAILABILITY. MR. K<\/p>","protected":false},"author":1,"featured_media":0,"comment_status":"open","ping_status":"open","sticky":false,"template":"","format":"standard","meta":{"footnotes":""},"categories":[146],"tags":[415,238,239],"collection":[],"class_list":["post-1250","post","type-post","status-publish","format-standard","hentry","category-news","tag-gpu","tag-238","tag-239"],"acf":[],"_links":{"self":[{"href":"https:\/\/www.1ai.net\/en\/wp-json\/wp\/v2\/posts\/1250","targetHints":{"allow":["GET"]}}],"collection":[{"href":"https:\/\/www.1ai.net\/en\/wp-json\/wp\/v2\/posts"}],"about":[{"href":"https:\/\/www.1ai.net\/en\/wp-json\/wp\/v2\/types\/post"}],"author":[{"embeddable":true,"href":"https:\/\/www.1ai.net\/en\/wp-json\/wp\/v2\/users\/1"}],"replies":[{"embeddable":true,"href":"https:\/\/www.1ai.net\/en\/wp-json\/wp\/v2\/comments?post=1250"}],"version-history":[{"count":0,"href":"https:\/\/www.1ai.net\/en\/wp-json\/wp\/v2\/posts\/1250\/revisions"}],"wp:attachment":[{"href":"https:\/\/www.1ai.net\/en\/wp-json\/wp\/v2\/media?parent=1250"}],"wp:term":[{"taxonomy":"category","embeddable":true,"href":"https:\/\/www.1ai.net\/en\/wp-json\/wp\/v2\/categories?post=1250"},{"taxonomy":"post_tag","embeddable":true,"href":"https:\/\/www.1ai.net\/en\/wp-json\/wp\/v2\/tags?post=1250"},{"taxonomy":"collection","embeddable":true,"href":"https:\/\/www.1ai.net\/en\/wp-json\/wp\/v2\/collection?post=1250"}],"curies":[{"name":"wp","href":"https:\/\/api.w.org\/{rel}","templated":true}]}}