{"id":14206,"date":"2024-06-28T09:17:33","date_gmt":"2024-06-28T01:17:33","guid":{"rendered":"https:\/\/www.1ai.net\/?p=14206"},"modified":"2024-06-28T09:17:33","modified_gmt":"2024-06-28T01:17:33","slug":"90-270-%e4%ba%bf%e5%8f%82%e6%95%b0%ef%bc%8c%e8%b0%b7%e6%ad%8c%e5%8f%91%e5%b8%83-gemma-2-%e5%bc%80%e6%ba%90-ai-%e6%a8%a1%e5%9e%8b%ef%bc%9a%e6%80%a7%e8%83%bd%e5%8a%9b%e5%8e%8b%e5%90%8c%e7%ba%a7%e3%80%81","status":"publish","type":"post","link":"https:\/\/www.1ai.net\/en\/14206.html","title":{"rendered":"Google releases Gemma 2 open source AI model with 9\/27 billion parameters: performance is better than its peers, and can be run on a single A100\/H100 GPU"},"content":{"rendered":"<p data-vmark=\"a0d3\"><a href=\"https:\/\/www.1ai.net\/en\/tag\/%e8%b0%b7%e6%ad%8c\" title=\"[View articles tagged with [Google]]\" target=\"_blank\" >Google<\/a>The company released yesterday<a href=\"https:\/\/blog.google\/technology\/developers\/google-gemma-2\/\" target=\"_blank\" rel=\"noopener\">Press release<\/a>The release is open to researchers and developers around the world <a href=\"https:\/\/www.1ai.net\/en\/tag\/gemma-2\" title=\"[See articles with [Gemma 2] labels]\" target=\"_blank\" >Gemma 2<\/a> <a href=\"https:\/\/www.1ai.net\/en\/tag\/%e5%a4%a7%e8%af%ad%e8%a8%80%e6%a8%a1%e5%9e%8b\" title=\"[View articles tagged with [large language model]]\" target=\"_blank\" >Large Language Model<\/a>,<strong>There are two sizes, 9 billion parameters (9B) and 27 billion parameters (27B).<\/strong><\/p>\n<p data-vmark=\"6b86\"><img loading=\"lazy\" decoding=\"async\" class=\"alignnone size-full wp-image-14207\" title=\"9b1a398e-c05c-4882-90a6-185496c40029\" src=\"https:\/\/www.1ai.net\/wp-content\/uploads\/2024\/06\/9b1a398e-c05c-4882-90a6-185496c40029.jpg\" alt=\"9b1a398e-c05c-4882-90a6-185496c40029\" width=\"1024\" height=\"427\" \/><\/p>\n<p data-vmark=\"1ad9\">The Gemma 2 large language model offers higher inference performance, greater efficiency, and significant advances in security compared to the first generation.<\/p>\n<p data-vmark=\"6e51\">In a press release, Google said the Gemma 2-27B model rivals the performance of mainstream models at twice the scale and requires only a single NVIDIA H100 ensor Core GPU or TPU host to achieve this performance, significantly reducing deployment costs.<\/p>\n<p data-vmark=\"7258\">The Gemma 2-9B model outperforms Llama 3 8B and other open source models of similar size. Google also plans to release the Gemma 2 model with 2.6 billion parameters in the coming months, which is better suited for smartphone AI scenarios.<\/p>\n<p data-vmark=\"4ce9\">Google says it redesigned the overall architecture for Gemma 2 to achieve superior performance and inference efficiency.<\/p>\n<p data-vmark=\"4ce9\">The main features of Gemma 2 are as follows:<\/p>\n<h3 data-vmark=\"f150\">Excellent performance:<\/h3>\n<p data-vmark=\"e42a\">The 27B version offers the best performance in its size class and is even more competitive than models twice its size.The 9B version also leads its class in performance, outperforming the Llama 3 8B and other open models in its size class.<\/p>\n<h3 data-vmark=\"5a4b\">Efficiency and cost:<\/h3>\n<p data-vmark=\"831c\">27B Gemma 2 models can run inference efficiently at full precision on a single Google Cloud TPU host, NVIDIA A100 80GB Tensor Core GPU, or NVIDIA H100 Tensor Core GPU, dramatically reducing costs while maintaining high performance. This makes AI deployments easier to achieve and more budget-friendly.<\/p>\n<h3 data-vmark=\"b7c1\">Cross-Hardware Rapid Reasoning<\/h3>\n<p data-vmark=\"946b\">Gemma 2 is optimized to run at blazing speeds on a wide range of hardware, from powerful gaming laptops and high-end desktops to cloud-based setups.<\/p>\n<p data-vmark=\"514a\">Trying full-precision Gemma 2 in Google AI Studio on a CPU with\u00a0<span class=\"link-text-start-with-http\">Gemma.cpp<\/span>\u00a0Unlock local performance with a quantized version of NVIDIA RTX or GeForce RTX, or try it out on a NVIDIA RTX or GeForce RTX-equipped home computer with Hugging Face Transformers.<\/p>","protected":false},"excerpt":{"rendered":"<p>Google issued a press release yesterday, issuing Gemma 2 large-language models for researchers and developers around the world, totalling 9 billion parameters (9B) and 27 billion parameters (27B). Gemma 2 Large-language models are more deduced and more efficient than the first generation and have made significant progress in safety. Google stated in its press release that the Gemma 2-27B model performed twice as well as the mainstream model, and that it could only be achieved by an English-led H100 endor Core GPU or TPU host, thus significantly reducing deployment costs. Gemma 2-9B model better than the open source of Llama 3 8B and other similar scales<\/p>","protected":false},"author":1,"featured_media":0,"comment_status":"open","ping_status":"open","sticky":false,"template":"","format":"standard","meta":{"footnotes":""},"categories":[146],"tags":[3252,706,713,281],"collection":[],"class_list":["post-14206","post","type-post","status-publish","format-standard","hentry","category-news","tag-gemma-2","tag-706","tag-ai","tag-281"],"acf":[],"_links":{"self":[{"href":"https:\/\/www.1ai.net\/en\/wp-json\/wp\/v2\/posts\/14206","targetHints":{"allow":["GET"]}}],"collection":[{"href":"https:\/\/www.1ai.net\/en\/wp-json\/wp\/v2\/posts"}],"about":[{"href":"https:\/\/www.1ai.net\/en\/wp-json\/wp\/v2\/types\/post"}],"author":[{"embeddable":true,"href":"https:\/\/www.1ai.net\/en\/wp-json\/wp\/v2\/users\/1"}],"replies":[{"embeddable":true,"href":"https:\/\/www.1ai.net\/en\/wp-json\/wp\/v2\/comments?post=14206"}],"version-history":[{"count":0,"href":"https:\/\/www.1ai.net\/en\/wp-json\/wp\/v2\/posts\/14206\/revisions"}],"wp:attachment":[{"href":"https:\/\/www.1ai.net\/en\/wp-json\/wp\/v2\/media?parent=14206"}],"wp:term":[{"taxonomy":"category","embeddable":true,"href":"https:\/\/www.1ai.net\/en\/wp-json\/wp\/v2\/categories?post=14206"},{"taxonomy":"post_tag","embeddable":true,"href":"https:\/\/www.1ai.net\/en\/wp-json\/wp\/v2\/tags?post=14206"},{"taxonomy":"collection","embeddable":true,"href":"https:\/\/www.1ai.net\/en\/wp-json\/wp\/v2\/collection?post=14206"}],"curies":[{"name":"wp","href":"https:\/\/api.w.org\/{rel}","templated":true}]}}