{"id":5352,"date":"2024-03-12T09:33:12","date_gmt":"2024-03-12T01:33:12","guid":{"rendered":"https:\/\/www.1ai.net\/?p=5352"},"modified":"2024-03-12T09:33:12","modified_gmt":"2024-03-12T01:33:12","slug":"%e8%b0%b7%e6%ad%8c%e5%8f%91%e5%b8%83gemini-1-5-pro%e6%a8%a1%e5%9e%8b%e6%8a%80%e6%9c%af%e6%8a%a5%e5%91%8a-%e9%95%bf%e6%96%87%e8%ae%b0%e5%bf%86%e8%83%bd%e5%8a%9b%e6%83%8a%e4%ba%ba","status":"publish","type":"post","link":"https:\/\/www.1ai.net\/en\/5352.html","title":{"rendered":"Google releases Gemini 1.5 Pro model technical report with amazing long text memory ability"},"content":{"rendered":"<p><a href=\"https:\/\/www.1ai.net\/en\/tag\/google\" title=\"[View articles tagged with [Google]]\" target=\"_blank\" >Google<\/a>Published<a href=\"https:\/\/www.1ai.net\/en\/tag\/gemini\" title=\"[View articles tagged with [Gemini]]\" target=\"_blank\" >Gemini<\/a>1.5 Pro modeling of<a href=\"https:\/\/www.1ai.net\/en\/tag\/%e6%8a%80%e6%9c%af%e6%8a%a5%e5%91%8a\" title=\"[Sees articles with [technical report] labels]\" target=\"_blank\" >Technical Report<\/a>This report describes the model architecture, training data and infrastructure, long text reviews, and general-purpose competency reviews of Gemini 1.5 Pro, a high-performance, multimodal, hybrid expert model that processes and analyzes information from millions of Token, including long documents and hours of video and audio content.<\/p>\n<p>Gemini1.5Pro demonstrates almost perfect memory recall in long-form information retrieval tasks, refreshing the technological heights in the fields of long-form document quizzing, long-form video quizzing and automatic speech recognition, surpassing its predecessor, Gemini1.0Ultra.In terms of predicting the next Token, Gemini1.5Pro makes significant progress, reaching an accuracy rate of 991 TP3T or more when dealing with tasks with more than 10 million Token tasks with an accuracy rate of over 99%, a huge leap forward.<\/p>\n<p class=\"article-content__img\"><img loading=\"lazy\" decoding=\"async\" class=\"alignnone size-full wp-image-5353\" src=\"https:\/\/www.1ai.net\/wp-content\/uploads\/2024\/03\/6384576268991971561313404.png\" alt=\"\" width=\"670\" height=\"509\" \/><\/p>\n<p>Paper address:https:\/\/arxiv.org\/pdf\/2403.05530.pdf<\/p>\n<p>In addition, the Gemini 1.5Pro demonstrated an amazing ability to learn to translate English into Kalamang at a level comparable to a human being when exposed to the Kalamang Grammar Manual, even though Kalamang is a niche language with fewer than 200 speakers worldwide.<\/p>\n<p>In summary, Gemini 1.5Pro performs well in processing long-form information across multiple media formats, not only surpassing its predecessor in terms of technical performance, but also demonstrating an amazing translation capability that opens up new possibilities for the development of multimodal hybrid expert models.<\/p>","protected":false},"excerpt":{"rendered":"<p>Google has released a technical report on the Gemini1.5Pro model, which describes the model architecture, training data and infrastructure, long text reviews, and general-purpose capability reviews of Gemini1.5Pro, a high-performance, multimodal, hybrid expert model that processes and analyzes information from millions of Token, including long-form documents and hours of video and audio content. Gemini1.5Pro demonstrates near-perfect memory recall for long-form information retrieval tasks, breaking new ground in the areas of long-form document quizzing, long-form video quizzing, and automatic speech recognition, surpassing its predecessor, Gemini1.0Ultra. In terms of predicting the next Token, Gemini1.5Pro is the most efficient multimodal hybrid expert model.<\/p>","protected":false},"author":1,"featured_media":0,"comment_status":"open","ping_status":"open","sticky":false,"template":"","format":"standard","meta":{"footnotes":""},"categories":[146],"tags":[436,308,1633],"collection":[],"class_list":["post-5352","post","type-post","status-publish","format-standard","hentry","category-news","tag-gemini","tag-google","tag-1633"],"acf":[],"_links":{"self":[{"href":"https:\/\/www.1ai.net\/en\/wp-json\/wp\/v2\/posts\/5352","targetHints":{"allow":["GET"]}}],"collection":[{"href":"https:\/\/www.1ai.net\/en\/wp-json\/wp\/v2\/posts"}],"about":[{"href":"https:\/\/www.1ai.net\/en\/wp-json\/wp\/v2\/types\/post"}],"author":[{"embeddable":true,"href":"https:\/\/www.1ai.net\/en\/wp-json\/wp\/v2\/users\/1"}],"replies":[{"embeddable":true,"href":"https:\/\/www.1ai.net\/en\/wp-json\/wp\/v2\/comments?post=5352"}],"version-history":[{"count":0,"href":"https:\/\/www.1ai.net\/en\/wp-json\/wp\/v2\/posts\/5352\/revisions"}],"wp:attachment":[{"href":"https:\/\/www.1ai.net\/en\/wp-json\/wp\/v2\/media?parent=5352"}],"wp:term":[{"taxonomy":"category","embeddable":true,"href":"https:\/\/www.1ai.net\/en\/wp-json\/wp\/v2\/categories?post=5352"},{"taxonomy":"post_tag","embeddable":true,"href":"https:\/\/www.1ai.net\/en\/wp-json\/wp\/v2\/tags?post=5352"},{"taxonomy":"collection","embeddable":true,"href":"https:\/\/www.1ai.net\/en\/wp-json\/wp\/v2\/collection?post=5352"}],"curies":[{"name":"wp","href":"https:\/\/api.w.org\/{rel}","templated":true}]}}