{"id":23233,"date":"2024-11-17T08:59:19","date_gmt":"2024-11-17T00:59:19","guid":{"rendered":"https:\/\/www.1ai.net\/?p=23233"},"modified":"2024-11-17T08:59:19","modified_gmt":"2024-11-17T00:59:19","slug":"%e8%b0%b7%e6%ad%8cgemini-exp-1114%e6%a8%aa%e7%a9%ba%e5%87%ba%e4%b8%96%ef%bc%81%e9%a6%96%e6%88%98%e7%a2%be%e5%8e%8bgpt-4%ef%bc%8c%e5%a4%9a%e9%a1%b9%e8%83%bd%e5%8a%9b%e8%af%84%e6%b5%8b%e7%99%bb%e9%a1%b6","status":"publish","type":"post","link":"https:\/\/www.1ai.net\/en\/23233.html","title":{"rendered":"Google Gemini Exp 1114 came out of nowhere! Crushing GPT-4 in the first battle and topping many capability reviews to shake the industry!"},"content":{"rendered":"<p><a href=\"https:\/\/www.1ai.net\/en\/tag\/%e8%b0%b7%e6%ad%8c\" title=\"[View articles tagged with [Google]]\" target=\"_blank\" >Google<\/a>DeepMind's latest<a href=\"https:\/\/www.1ai.net\/en\/tag\/gemini\" title=\"[View articles tagged with [Gemini]]\" target=\"_blank\" >Gemini<\/a>The experimental version (Exp1114) has achieved impressive results on the Chatbot Arena platform. After more than a week of community testing and accumulating more than 6,000 votes, this new model outperforms the competition by a significant margin, showing amazing strength in several key areas.<\/p>\n<p><img loading=\"lazy\" decoding=\"async\" class=\"alignnone size-full wp-image-23234\" title=\"cfa6b43bj00sn2ld4002jd000rs00m4m\" src=\"https:\/\/www.1ai.net\/wp-content\/uploads\/2024\/11\/cfa6b43bj00sn2ld4002jd000rs00m4m.jpg\" alt=\"cfa6b43bj00sn2ld4002jd000rs00m4m\" width=\"1000\" height=\"796\" \/><\/p>\n<p>In terms of overall scoring, Gemini-Exp-1114 tied for first place with GPT-4-latest with an outstanding score of over 40 points, surpassing the previously leading GPT-4-preview version. What is even more amazing is that the model fully topped the core areas of math, complex prompts, and creative writing, demonstrating great overall strength.<\/p>\n<p>Specifically, the Gemini-Exp-1114's progress is impressive: the<\/p>\n<p>Jumped from No. 3 to No. 1 in the overall rankings<\/p>\n<p>Moved from 3rd to 1st place on the Math Proficiency Assessment<\/p>\n<p>Complex cue processing climbs from #4 to #1<\/p>\n<p>Creative writing performance improved from 2nd to 1st place<\/p>\n<p>Visual processing power also tops the list<\/p>\n<p>Programming level also improved from 5th to 3rd place<\/p>\n<p>Google AI Studio has officially launched this new version for users to actually experience. However, the community has also expressed concerns about some specific issues, such as whether the 1000 token limit still exists, and how to deal with practical application issues such as extra-long text output.<\/p>\n<p>Industry analysts believe that this breakthrough shows that Google's long-term investment in AI is starting to reap results. Interestingly, the model maintains its 4th place ranking in style control, which may suggest that the development team primarily used new post-training methods rather than making changes to the pre-trained model.<\/p>\n<p>It has been suggested that this could herald the arrival of Gemini2 and that Google is becoming significantly more competitive in the large modeling space.<\/p>","protected":false},"excerpt":{"rendered":"<p>Google DeepMind's latest Gemini Experiment (Exp.1114) has achieved remarkable results on Chatbot Arena. After more than a week of community testing, the cumulative data of more than 6,000 votes shows that the new model, with a significant advantage over competition, has shown remarkable strength in many key areas. On the overall score, Gemini-Exp-1114 ranked first on top of GPT-4-latest with an outstanding score above 40, surpassing the previous leading GPT-4-preview version. Even more striking is the fact that the model has reached its full peak in core areas such as mathematics, complex tips and creative writing, and has shown a very strong combined strength. Specifically, Gemini-Exp-1114<\/p>","protected":false},"author":1,"featured_media":0,"comment_status":"closed","ping_status":"closed","sticky":false,"template":"","format":"standard","meta":{"footnotes":""},"categories":[146],"tags":[436,281],"collection":[],"class_list":["post-23233","post","type-post","status-publish","format-standard","hentry","category-news","tag-gemini","tag-281"],"acf":[],"_links":{"self":[{"href":"https:\/\/www.1ai.net\/en\/wp-json\/wp\/v2\/posts\/23233","targetHints":{"allow":["GET"]}}],"collection":[{"href":"https:\/\/www.1ai.net\/en\/wp-json\/wp\/v2\/posts"}],"about":[{"href":"https:\/\/www.1ai.net\/en\/wp-json\/wp\/v2\/types\/post"}],"author":[{"embeddable":true,"href":"https:\/\/www.1ai.net\/en\/wp-json\/wp\/v2\/users\/1"}],"replies":[{"embeddable":true,"href":"https:\/\/www.1ai.net\/en\/wp-json\/wp\/v2\/comments?post=23233"}],"version-history":[{"count":0,"href":"https:\/\/www.1ai.net\/en\/wp-json\/wp\/v2\/posts\/23233\/revisions"}],"wp:attachment":[{"href":"https:\/\/www.1ai.net\/en\/wp-json\/wp\/v2\/media?parent=23233"}],"wp:term":[{"taxonomy":"category","embeddable":true,"href":"https:\/\/www.1ai.net\/en\/wp-json\/wp\/v2\/categories?post=23233"},{"taxonomy":"post_tag","embeddable":true,"href":"https:\/\/www.1ai.net\/en\/wp-json\/wp\/v2\/tags?post=23233"},{"taxonomy":"collection","embeddable":true,"href":"https:\/\/www.1ai.net\/en\/wp-json\/wp\/v2\/collection?post=23233"}],"curies":[{"name":"wp","href":"https:\/\/api.w.org\/{rel}","templated":true}]}}