{"id":1734,"date":"2023-12-06T09:34:41","date_gmt":"2023-12-06T01:34:41","guid":{"rendered":"https:\/\/www.1ai.net\/?p=1734"},"modified":"2023-12-06T09:34:41","modified_gmt":"2023-12-06T01:34:41","slug":"35%e4%ba%bf%e5%85%83%ef%bc%81%e5%bc%80%e6%ba%90%e7%b1%bbchatgpt%e5%b9%b3%e5%8f%b0mistral-ai%ef%bc%8c%e5%86%8d%e8%8e%b7%e5%b7%a8%e9%a2%9d%e8%9e%8d%e8%b5%84","status":"publish","type":"post","link":"https:\/\/www.1ai.net\/en\/1734.html","title":{"rendered":"3.5 billion yuan! Mistral AI, an open source ChatGPT platform, receives another huge round of financing"},"content":{"rendered":"<p>December 6, Bloomberg News.<a href=\"https:\/\/www.1ai.net\/en\/tag\/%e5%bc%80%e6%ba%90%e7%b1%bbchatgpt%e5%b9%b3%e5%8f%b0\" title=\"[Sees articles with labels of [open source-type ChatGPT platform]\" target=\"_blank\" >Open source ChatGPT platform<\/a><a href=\"https:\/\/www.1ai.net\/en\/tag\/mistral\" title=\"[See article with [Mistral] label]\" target=\"_blank\" >Mistral<\/a>\u00a0AI gets 450 million euros (nearly 3.5 billion dollars)<a href=\"https:\/\/www.1ai.net\/en\/tag\/%e8%9e%8d%e8%b5%84\" title=\"[View articles tagged with [financing]]\" target=\"_blank\" >Financing<\/a>, valued at nearly $2 billion ($14.2 billion). The current investment by NVIDIA, Salesforce and others.<\/p>\n<p>Mistral AI's open-source big language model Mistral7B majors in<strong>features such as small parameters, low energy consumption, and high performance, and allows for commercialization of the<\/strong>. Support for generating text\/code, data fine-tuning, summarizing content, etc. currently has 4500 stars on github.<\/p>\n<p>It is worth mentioning that<strong>Mistral AI had secured a $113 million seed round without releasing any products<\/strong>This is the first time in the history of European technology that<span class=\"spamTxt\">maximum<\/span>One of the seed rounds of funding for the<\/p>\n<p>Open source address:https:\/\/github.com\/mistralai\/mistral-src<\/p>\n<p>Help file:https:\/\/docs.mistral.ai\/<\/p>\n<p>API Interface:https:\/\/docs.mistral.ai\/api<\/p>\n<p class=\"article-content__img\"><img loading=\"lazy\" decoding=\"async\" class=\"alignnone size-full wp-image-1737\" title=\"2023120608572428200\" src=\"https:\/\/www.1ai.net\/wp-content\/uploads\/2023\/12\/2023120608572428200.jpg\" alt=\"2023120608572428200\" width=\"554\" height=\"310\" \/><\/p>\n<p>Compared to the meta-universe, ChatGPT, which just celebrated its 1st birthday, has withstood multiple tests such as commercial landing and user audience, and has driven a large number of tech companies to participate in generative AI change.<\/p>\n<p>Currently, there are two main camps, closed source and open source. Llama hits the ground running at Meta<span class=\"spamTxt\">First<\/span>After the gun.<strong>A large number of outstanding companies have emerged in the field of open source large language modeling, such as Writer, Baichuan Intelligence, Together.ai, Mistral AI, etc.<\/strong>, while gaining recognition in the capital markets. These vendors are also convinced that open source is one of the shortcuts to AGI for large models.<\/p>\n<p>Mistral AI was introduced to the \"AIGC Open Community\" back in June of this year, and I was very impressed with it at that time.<strong>Since no products have been released, the official website has only 3 sentences:.<\/strong>We are assembling a world-class technical team to develop the<span class=\"spamTxt\">most<\/span>of generative AI models.<\/p>\n<p class=\"article-content__img\"><img loading=\"lazy\" decoding=\"async\" class=\"alignnone size-full wp-image-1735\" title=\"2023120608572428211\" src=\"https:\/\/www.1ai.net\/wp-content\/uploads\/2023\/12\/2023120608572428211.jpg\" alt=\"2023120608572428211\" width=\"554\" height=\"166\" \/><\/p>\n<p>Mistral AI's official website content from June this year<\/p>\n<p>We operate in Europe and are based in Paris, France. If you have extensive research and development experience in the AI field, please contact us.<\/p>\n<p>At that time, with these three words, it raised $113 million in seed round financing, valued at $260 million. Usually this kind of enterprise either rub a wave of heat to get money, and then casually change the model to sit and wait for death.<\/p>\n<p>Either that or it's a sweeper-level tech bull that's famous as soon as it hits the ground running. Judging from the results of this financing, Mistral AI belongs to the latter does have two tricks up its sleeve.<\/p>\n<p>Mistral AI's three co-founders, Timoth\u00e9e Lacroix, Guillaume Lample, and Arthur Mensch, are well known for their big factory resumes and successful projects, as well as being university alumni.<\/p>\n<p>Timoth\u00e9e and Guillaume worked in Meta's AI research department and led the development of LLaMA, the originator of the ChatGPT-like open-source model.Arthur worked at DeepMind, Google's AI research lab.<\/p>\n<p>Products.<strong>Mistral AI's Mistral 7B, launched on September 27 of this year, is the current<span class=\"spamTxt\">Strongest<\/span>Open source large language model that outperforms Llama213B in all benchmarks<\/strong>; outperforms or is comparable to Llama134B on many benchmarks; performance on code tests is comparable to CodeLlama7B.<\/p>\n<p class=\"article-content__img\"><img loading=\"lazy\" decoding=\"async\" class=\"alignnone size-full wp-image-1738\" title=\"2023120608572428212\" src=\"https:\/\/www.1ai.net\/wp-content\/uploads\/2023\/12\/2023120608572428212.jpg\" alt=\"2023120608572428212\" width=\"554\" height=\"180\" \/><\/p>\n<p>In order for the model to reason faster and with less energy consumption, the<strong>Mistral AI uses two main mechanisms, grouped query attention and sliding window attention, respectively<\/strong>.<\/p>\n<p>Grouped query attention is an improvement on the standard attention mechanism that reduces computational complexity by grouping queries. In the Transformer model, the attention mechanism typically involves three sets of vectors for queries, keys, and values.<\/p>\n<p>In the standard self-attentive mechanism, each query is matched against all keys, which leads to a huge computational burden when the sequence is long.<\/p>\n<p class=\"article-content__img\"><img loading=\"lazy\" decoding=\"async\" class=\"alignnone size-full wp-image-1736\" title=\"2023120608572428213\" src=\"https:\/\/www.1ai.net\/wp-content\/uploads\/2023\/12\/2023120608572428213.jpg\" alt=\"2023120608572428213\" width=\"554\" height=\"243\" \/><\/p>\n<p>And grouped query attention works by combining multiple queries into a single group. The query vectors of each group then interact with only a portion of the key vectors instead of all of them, making the overall efficiency very efficient.<\/p>\n<p>Sliding window attention is a technique used in sequence processing tasks to limit the scope of the attention mechanism and reduce the amount of computation. In this approach, instead of computing attention for the entire sequence, the attention of each element is limited to elements within a window in its neighborhood.<\/p>\n<p>In this way, each part of the model only needs to process the information within the window, thus reducing the number of elements involved in each attentional computation.<\/p>\n<p>This not only reduces the computational requirements, but also limits the scope of the model's context and helps the model to focus on localized information.<\/p>","protected":false},"excerpt":{"rendered":"<p>December 6, Bloomberg news, open source class ChatGPT platform Mistral AI received 450 million euros (nearly 3.5 billion yuan) in financing, valued at nearly 2 billion U.S. dollars (14.2 billion yuan). This time by NVIDIA, Salesforce and other investment. Mistral AI's open source large language model Mistral7B focuses on features such as small parameters, low energy consumption, high performance, and allows commercialization. It supports generating text\/code, data fine-tuning, summarizing content, etc. It currently has 4,500 stars in github. It is worth mentioning that Mistral AI had received a $113 million seed round without releasing any product, which is one of the largest seed rounds in European tech history. Open source address:https.<\/p>","protected":false},"author":1,"featured_media":0,"comment_status":"open","ping_status":"open","sticky":false,"template":"","format":"standard","meta":{"footnotes":""},"categories":[146],"tags":[559,560,421],"collection":[],"class_list":["post-1734","post","type-post","status-publish","format-standard","hentry","category-news","tag-mistral","tag-chatgpt","tag-421"],"acf":[],"_links":{"self":[{"href":"https:\/\/www.1ai.net\/en\/wp-json\/wp\/v2\/posts\/1734","targetHints":{"allow":["GET"]}}],"collection":[{"href":"https:\/\/www.1ai.net\/en\/wp-json\/wp\/v2\/posts"}],"about":[{"href":"https:\/\/www.1ai.net\/en\/wp-json\/wp\/v2\/types\/post"}],"author":[{"embeddable":true,"href":"https:\/\/www.1ai.net\/en\/wp-json\/wp\/v2\/users\/1"}],"replies":[{"embeddable":true,"href":"https:\/\/www.1ai.net\/en\/wp-json\/wp\/v2\/comments?post=1734"}],"version-history":[{"count":0,"href":"https:\/\/www.1ai.net\/en\/wp-json\/wp\/v2\/posts\/1734\/revisions"}],"wp:attachment":[{"href":"https:\/\/www.1ai.net\/en\/wp-json\/wp\/v2\/media?parent=1734"}],"wp:term":[{"taxonomy":"category","embeddable":true,"href":"https:\/\/www.1ai.net\/en\/wp-json\/wp\/v2\/categories?post=1734"},{"taxonomy":"post_tag","embeddable":true,"href":"https:\/\/www.1ai.net\/en\/wp-json\/wp\/v2\/tags?post=1734"},{"taxonomy":"collection","embeddable":true,"href":"https:\/\/www.1ai.net\/en\/wp-json\/wp\/v2\/collection?post=1734"}],"curies":[{"name":"wp","href":"https:\/\/api.w.org\/{rel}","templated":true}]}}