{"id":6464,"date":"2024-03-28T09:28:53","date_gmt":"2024-03-28T01:28:53","guid":{"rendered":"https:\/\/www.1ai.net\/?p=6464"},"modified":"2024-03-28T09:28:53","modified_gmt":"2024-03-28T01:28:53","slug":"%e5%8f%b7%e7%a7%b0%e5%85%a8%e7%90%83%e6%9c%80%e5%bc%ba%e5%bc%80%e6%ba%90-ai-%e6%a8%a1%e5%9e%8b%ef%bc%8cdbrx-%e7%99%bb%e5%9c%ba%ef%bc%9a1320-%e4%ba%bf%e5%8f%82%e6%95%b0%ef%bc%8c%e8%af%ad%e8%a8%80","status":"publish","type":"post","link":"https:\/\/www.1ai.net\/en\/6464.html","title":{"rendered":"DBRX is the world&#039;s most powerful open source AI model: 132 billion parameters, language understanding, programming capabilities, etc. are better than GPT-3.5"},"content":{"rendered":"<p data-vmark=\"f605\">Startups <a href=\"https:\/\/www.1ai.net\/en\/tag\/databricks\" title=\"[See articles with [Databricks] labels]\" target=\"_blank\" >Databricks<\/a> Recently announced the launch of<a href=\"https:\/\/www.1ai.net\/en\/tag\/%e5%bc%80%e6%ba%90\" title=\"[View articles tagged with [open source]]\" target=\"_blank\" >Open Source<\/a> <a href=\"https:\/\/www.1ai.net\/en\/tag\/ai%e6%a8%a1%e5%9e%8b\" title=\"[View articles tagged with [AI models]]\" target=\"_blank\" >AI Models<\/a> DBRX claims to be the world&#039;s most powerful open source large-scale language model to date, more powerful than Meta&#039;s Llama 2.<\/p>\n<p data-vmark=\"2109\"><img loading=\"lazy\" decoding=\"async\" class=\"alignnone size-full wp-image-6465\" title=\"00913eec-49cd-4d7b-a5a6-ad51449ab066\" src=\"https:\/\/www.1ai.net\/wp-content\/uploads\/2024\/03\/00913eec-49cd-4d7b-a5a6-ad51449ab066.jpg\" alt=\"00913eec-49cd-4d7b-a5a6-ad51449ab066\" width=\"660\" height=\"235\" \/><\/p>\n<p data-vmark=\"f1a1\">DBRX uses a transformer architecture, contains 132 billion parameters, and consists of 16 expert networks. Each inference uses 4 of the expert networks, activating 36 billion parameters.<\/p>\n<p data-vmark=\"cadb\">Databricks introduced in a company blog post that in terms of language understanding, programming, mathematics and logic, it compares with mainstream open source models such as Meta&#039;s Llama 2-70B, France&#039;s MixtralAI&#039;s Mixtral, and Musk&#039;s xAI&#039;s Grok-1.<strong>DBRX wins in both categories.<\/strong><\/p>\n<p data-vmark=\"4594\"><img loading=\"lazy\" decoding=\"async\" class=\"alignnone size-full wp-image-6466\" title=\"a522ab11-f579-49a2-8a42-b31912e1c725\" src=\"https:\/\/www.1ai.net\/wp-content\/uploads\/2024\/03\/a522ab11-f579-49a2-8a42-b31912e1c725.jpg\" alt=\"a522ab11-f579-49a2-8a42-b31912e1c725\" width=\"1440\" height=\"855\" \/><\/p>\n<p>Figure 1: DBRX outperforms existing open source models in language understanding (MMLU), programming (HumanEval), and mathematics (GSM8K).<\/p>\n<p data-vmark=\"b0d6\">exist<strong>Language Comprehension<\/strong>In terms of performance, DBRX&#039;s score is 73.7%, which is higher than GPT-3.5&#039;s 70.0%, Llama 2-70B&#039;s 69.8%, Mixtral&#039;s 71.4%, and Grok-1&#039;s 73.0%.<\/p>\n<table width=\"100%\">\n<tbody>\n<tr class=\"firstRow\">\n<td>\n<p data-vmark=\"e6ad\">Model<\/p>\n<\/td>\n<td>\n<p data-vmark=\"6246\">DBRX Instruct<\/p>\n<\/td>\n<td>\n<p data-vmark=\"6566\">Mixtral Instruct<\/p>\n<\/td>\n<td>\n<p data-vmark=\"043d\">Mixtral Base<\/p>\n<\/td>\n<td>\n<p data-vmark=\"331c\">LLaMA2-70B Chat<\/p>\n<\/td>\n<td>\n<p data-vmark=\"5f64\">LLaMA2-70B Base<\/p>\n<\/td>\n<td>\n<p data-vmark=\"91c9\">Grok-1<sup>1<\/sup><\/p>\n<\/td>\n<\/tr>\n<tr>\n<td>\n<p data-vmark=\"d41c\">Open LLM Leaderboard<sup>2<\/sup><\/p>\n<p data-vmark=\"3298\">(Avg of next 6 rows)<\/p>\n<\/td>\n<td>\n<p data-vmark=\"43bf\">74.5%<\/p>\n<\/td>\n<td>\n<p data-vmark=\"e760\">72.7%<\/p>\n<\/td>\n<td>\n<p data-vmark=\"9350\">68.4%<\/p>\n<\/td>\n<td>\n<p data-vmark=\"3a9e\">62.4%<\/p>\n<\/td>\n<td>\n<p data-vmark=\"57c3\">67.9%<\/p>\n<\/td>\n<td>\n<p data-vmark=\"352e\">\u2014<\/p>\n<\/td>\n<\/tr>\n<tr>\n<td>\n<p data-vmark=\"88f5\">ARC-challenge 25-shot<\/p>\n<\/td>\n<td>\n<p data-vmark=\"0bb7\">68.9%<\/p>\n<\/td>\n<td>\n<p data-vmark=\"4a8d\">70.1%<\/p>\n<\/td>\n<td>\n<p data-vmark=\"328f\">66.4%<\/p>\n<\/td>\n<td>\n<p data-vmark=\"1172\">64.6%<\/p>\n<\/td>\n<td>\n<p data-vmark=\"d4de\">67.3%<\/p>\n<\/td>\n<td>\n<p data-vmark=\"d70e\">\u2014<\/p>\n<\/td>\n<\/tr>\n<tr>\n<td>\n<p data-vmark=\"3765\">HellaSwag 10-shot<\/p>\n<\/td>\n<td>\n<p data-vmark=\"8ec0\">89.0%<\/p>\n<\/td>\n<td>\n<p data-vmark=\"6ba8\">87.6%<\/p>\n<\/td>\n<td>\n<p data-vmark=\"58f1\">86.5%<\/p>\n<\/td>\n<td>\n<p data-vmark=\"8014\">85.9%<\/p>\n<\/td>\n<td>\n<p data-vmark=\"ab83\">87.3%<\/p>\n<\/td>\n<td>\n<p data-vmark=\"3161\">\u2014<\/p>\n<\/td>\n<\/tr>\n<tr>\n<td>\n<p data-vmark=\"ac99\">MMLU 5-shot<\/p>\n<\/td>\n<td>\n<p data-vmark=\"4d0c\">73.7%<\/p>\n<\/td>\n<td>\n<p data-vmark=\"1d14\">71.4%<\/p>\n<\/td>\n<td>\n<p data-vmark=\"225b\">71.9%<\/p>\n<\/td>\n<td>\n<p data-vmark=\"90ce\">63.9%<\/p>\n<\/td>\n<td>\n<p data-vmark=\"1031\">69.8%<\/p>\n<\/td>\n<td>\n<p data-vmark=\"7466\">73.0%<\/p>\n<\/td>\n<\/tr>\n<tr>\n<td>\n<p data-vmark=\"7205\">Truthful QA 0-shot<\/p>\n<\/td>\n<td>\n<p data-vmark=\"7bc3\">66.9%<\/p>\n<\/td>\n<td>\n<p data-vmark=\"8c6a\">65.0%<\/p>\n<\/td>\n<td>\n<p data-vmark=\"b1eb\">46.8%<\/p>\n<\/td>\n<td>\n<p data-vmark=\"40c6\">52.8%<\/p>\n<\/td>\n<td>\n<p data-vmark=\"e89d\">44.9%<\/p>\n<\/td>\n<td>\n<p data-vmark=\"7e52\">\u2014<\/p>\n<\/td>\n<\/tr>\n<tr>\n<td>\n<p data-vmark=\"9133\">WinoGrande 5-shot<\/p>\n<\/td>\n<td>\n<p data-vmark=\"13b5\">81.8%<\/p>\n<\/td>\n<td>\n<p data-vmark=\"bdde\">81.1%<\/p>\n<\/td>\n<td>\n<p data-vmark=\"210b\">81.7%<\/p>\n<\/td>\n<td>\n<p data-vmark=\"b19a\">80.5%<\/p>\n<\/td>\n<td>\n<p data-vmark=\"5881\">83.7%<\/p>\n<\/td>\n<td>\n<p data-vmark=\"d58f\">\u2014<\/p>\n<\/td>\n<\/tr>\n<tr>\n<td>\n<p data-vmark=\"708b\">GSM8k CoT 5-shot maj@1<sup>3<\/sup><\/p>\n<\/td>\n<td>\n<p data-vmark=\"bac5\">66.9%<\/p>\n<\/td>\n<td>\n<p data-vmark=\"046f\">61.1%<\/p>\n<\/td>\n<td>\n<p data-vmark=\"9431\">57.6%<\/p>\n<\/td>\n<td>\n<p data-vmark=\"15c5\">26.7%<\/p>\n<\/td>\n<td>\n<p data-vmark=\"c9a7\">54.1%<\/p>\n<\/td>\n<td>\n<p data-vmark=\"508b\">62.9% (8-shot)<\/p>\n<\/td>\n<\/tr>\n<tr>\n<td>\n<p data-vmark=\"e178\">Gauntlet v0.3<sup>4<\/sup><\/p>\n<p data-vmark=\"629f\">(Avg of 30+ diverse tasks)<\/p>\n<\/td>\n<td>\n<p data-vmark=\"e5f5\">66.8%<\/p>\n<\/td>\n<td>\n<p data-vmark=\"3327\">60.7%<\/p>\n<\/td>\n<td>\n<p data-vmark=\"d3d1\">56.8%<\/p>\n<\/td>\n<td>\n<p data-vmark=\"e028\">52.8%<\/p>\n<\/td>\n<td>\n<p data-vmark=\"f613\">56.4%<\/p>\n<\/td>\n<td>\n<p data-vmark=\"5b5e\">\u2014<\/p>\n<\/td>\n<\/tr>\n<tr>\n<td>\n<p data-vmark=\"67e0\">HumanEval<sup>5<\/sup><\/p>\n<p data-vmark=\"a763\">0-Shot, pass@1<\/p>\n<p data-vmark=\"311e\">(Programming)<\/p>\n<\/td>\n<td>\n<p data-vmark=\"c6dc\">70.1%<\/p>\n<\/td>\n<td>\n<p data-vmark=\"bb87\">54.8%<\/p>\n<\/td>\n<td>\n<p data-vmark=\"7726\">40.2%<\/p>\n<\/td>\n<td>\n<p data-vmark=\"d3fe\">32.2%<\/p>\n<\/td>\n<td>\n<p data-vmark=\"1318\">31.0%<\/p>\n<\/td>\n<td>\n<p data-vmark=\"847c\">63.2%<\/p>\n<\/td>\n<\/tr>\n<\/tbody>\n<\/table>\n<p data-vmark=\"9e61\">exist<strong>Programming skills<\/strong>In terms of performance, DBRX scored 70.1%, far exceeding GPT-3.5&#039;s 48.1%, and higher than Llama 2-70B&#039;s 32.3%, Mixtral&#039;s 54.8%, and Grok-1&#039;s 63.2%.<\/p>\n<table width=\"100%\">\n<tbody>\n<tr class=\"firstRow\">\n<td>\n<p data-vmark=\"13b2\">Model<\/p>\n<\/td>\n<td>\n<p data-vmark=\"a0a6\">DBRX<br \/>\nInstruct<\/p>\n<\/td>\n<td>\n<p data-vmark=\"de26\">GPT-3.5<sup>7<\/sup><\/p>\n<\/td>\n<td>\n<p data-vmark=\"214c\">GPT-4<sup>8<\/sup><\/p>\n<\/td>\n<td width=\"41\">\n<p data-vmark=\"dc04\">Claude 3 Haiku<\/p>\n<\/td>\n<td>\n<p data-vmark=\"29f9\">Claude 3 Sonnet<\/p>\n<\/td>\n<td>\n<p data-vmark=\"ed91\">Claude 3 Opus<\/p>\n<\/td>\n<td>\n<p data-vmark=\"bab4\">Gemini 1.0 Pro<\/p>\n<\/td>\n<td>\n<p data-vmark=\"c019\">Gemini 1.5 Pro<\/p>\n<\/td>\n<td>\n<p data-vmark=\"2a60\">Mistral Medium<\/p>\n<\/td>\n<td>\n<p data-vmark=\"face\">Mistral Large<\/p>\n<\/td>\n<\/tr>\n<tr>\n<td>\n<p data-vmark=\"60b5\">MT Bench (Inflection corrected, n=5)<\/p>\n<\/td>\n<td>\n<p data-vmark=\"52d0\">8.39 \u00b1 0.08<\/p>\n<\/td>\n<td>\n<p data-vmark=\"ea92\">\u2014<\/p>\n<\/td>\n<td>\n<p data-vmark=\"8631\">\u2014<\/p>\n<\/td>\n<td width=\"41\">\n<p data-vmark=\"15bc\">8.41 \u00b1 0.04<\/p>\n<\/td>\n<td>\n<p data-vmark=\"7bcb\">8.54 \u00b1 0.09<\/p>\n<\/td>\n<td>\n<p data-vmark=\"f7a4\">9.03 \u00b1 0.06<\/p>\n<\/td>\n<td>\n<p data-vmark=\"aac5\">8.23 \u00b1 0.08<\/p>\n<\/td>\n<td>\n<p data-vmark=\"b6f7\">\u2014<\/p>\n<\/td>\n<td>\n<p data-vmark=\"1e0f\">8.05 \u00b1 0.12<\/p>\n<\/td>\n<td>\n<p data-vmark=\"feeb\">8.90 \u00b1 0.06<\/p>\n<\/td>\n<\/tr>\n<tr>\n<td>\n<p data-vmark=\"e340\">MMLU 5-shot<\/p>\n<\/td>\n<td>\n<p data-vmark=\"1376\">73.7%<\/p>\n<\/td>\n<td>\n<p data-vmark=\"38e9\">70.0%<\/p>\n<\/td>\n<td>\n<p data-vmark=\"9bbe\">86.4%<\/p>\n<\/td>\n<td width=\"41\">\n<p data-vmark=\"27fa\">75.2%<\/p>\n<\/td>\n<td>\n<p data-vmark=\"c1eb\">79.0%<\/p>\n<\/td>\n<td>\n<p data-vmark=\"3758\">86.8%<\/p>\n<\/td>\n<td>\n<p data-vmark=\"2103\">71.8%<\/p>\n<\/td>\n<td>\n<p data-vmark=\"dc96\">81.9%<\/p>\n<\/td>\n<td>\n<p data-vmark=\"5607\">75.3%<\/p>\n<\/td>\n<td>\n<p data-vmark=\"ce18\">81.2%<\/p>\n<\/td>\n<\/tr>\n<tr>\n<td>\n<p data-vmark=\"3070\">HellaSwag 10-shot<\/p>\n<\/td>\n<td>\n<p data-vmark=\"b505\">89.0%<\/p>\n<\/td>\n<td>\n<p data-vmark=\"1c6a\">85.5%<\/p>\n<\/td>\n<td>\n<p data-vmark=\"6275\">95.3%<\/p>\n<\/td>\n<td width=\"41\">\n<p data-vmark=\"585d\">85.9%<\/p>\n<\/td>\n<td>\n<p data-vmark=\"3e5a\">89.0%<\/p>\n<\/td>\n<td>\n<p data-vmark=\"1917\">95.4%<\/p>\n<\/td>\n<td>\n<p data-vmark=\"ee61\">84.7%<\/p>\n<\/td>\n<td>\n<p data-vmark=\"f68c\">92.5%<\/p>\n<\/td>\n<td>\n<p data-vmark=\"2203\">88.0%<\/p>\n<\/td>\n<td>\n<p data-vmark=\"a51d\">89.2%<\/p>\n<\/td>\n<\/tr>\n<tr>\n<td>\n<p data-vmark=\"3bc9\">HumanEval 0-Shot<br \/>\npass@1<br \/>\n(Programming)<\/p>\n<\/td>\n<td>\n<p data-vmark=\"b1c5\">70.1%<\/p>\n<p data-vmark=\"bccb\">temp=0, N=1<\/p>\n<\/td>\n<td>\n<p data-vmark=\"77d0\">48.1%<\/p>\n<\/td>\n<td>\n<p data-vmark=\"2069\">67.0%<\/p>\n<\/td>\n<td width=\"41\">\n<p data-vmark=\"b87e\">75.9%<\/p>\n<\/td>\n<td>\n<p data-vmark=\"f854\">73.0%<\/p>\n<\/td>\n<td>\n<p data-vmark=\"c1c4\">84.9%<\/p>\n<\/td>\n<td>\n<p data-vmark=\"2540\">67.7%<\/p>\n<\/td>\n<td>\n<p data-vmark=\"1ff3\">71.9%<\/p>\n<\/td>\n<td>\n<p data-vmark=\"5bce\">38.4%<\/p>\n<\/td>\n<td>\n<p data-vmark=\"6f7f\">45.1%<\/p>\n<\/td>\n<\/tr>\n<tr>\n<td>\n<p data-vmark=\"eb15\">GSM8k CoT maj@1<\/p>\n<\/td>\n<td>\n<p data-vmark=\"5de2\">72.8% (5-shot)<\/p>\n<\/td>\n<td>\n<p data-vmark=\"66a5\">57.1% (5-shot)<\/p>\n<\/td>\n<td>\n<p data-vmark=\"559b\">92.0% (5-shot)<\/p>\n<\/td>\n<td width=\"41\">\n<p data-vmark=\"3f01\">88.9%<\/p>\n<\/td>\n<td>\n<p data-vmark=\"6445\">92.3%<\/p>\n<\/td>\n<td>\n<p data-vmark=\"8ada\">95.0%<\/p>\n<\/td>\n<td>\n<p data-vmark=\"4912\">86.5%<\/p>\n<p data-vmark=\"2fe5\">(maj1@32)<\/p>\n<\/td>\n<td>\n<p data-vmark=\"6331\">91.7% (11-shot)<\/p>\n<\/td>\n<td>\n<p data-vmark=\"4eeb\">66.7% (5-shot)<\/p>\n<\/td>\n<td>\n<p data-vmark=\"d2db\">81.0% (5-shot)<\/p>\n<\/td>\n<\/tr>\n<tr>\n<td>\n<p data-vmark=\"7be1\">WinoGrande 5-shot<\/p>\n<\/td>\n<td>\n<p data-vmark=\"038a\">81.8%<\/p>\n<\/td>\n<td>\n<p data-vmark=\"2083\">81.6%<\/p>\n<\/td>\n<td>\n<p data-vmark=\"a331\">87.5%<\/p>\n<\/td>\n<td width=\"41\">\n<p data-vmark=\"ac55\">\u2014<\/p>\n<\/td>\n<td>\n<p data-vmark=\"9284\">\u2014<\/p>\n<\/td>\n<td>\n<p data-vmark=\"e9ef\">\u2014<\/p>\n<\/td>\n<td>\n<p data-vmark=\"7271\">\u2014<\/p>\n<\/td>\n<td>\n<p data-vmark=\"a613\">\u2014<\/p>\n<\/td>\n<td>\n<p data-vmark=\"11fc\">88.0%<\/p>\n<\/td>\n<td>\n<p data-vmark=\"71c1\">86.7%<\/p>\n<\/td>\n<\/tr>\n<\/tbody>\n<\/table>\n<p data-vmark=\"fafe\">In mathematics, DBRX scored 66.9%, higher than GPT-3.5&#039;s 57.1%, and higher than Llama 2-70B&#039;s 54.1%, Mixtral&#039;s 61.1%, and Grok-1&#039;s 62.9%.<\/p>\n<p data-vmark=\"9398\"><img loading=\"lazy\" decoding=\"async\" class=\"alignnone size-full wp-image-6467\" title=\"39f6e358-5234-4579-a7df-4814c03846da\" src=\"https:\/\/www.1ai.net\/wp-content\/uploads\/2024\/03\/39f6e358-5234-4579-a7df-4814c03846da.jpg\" alt=\"39f6e358-5234-4579-a7df-4814c03846da\" width=\"1440\" height=\"895\" \/><\/p>\n<p data-vmark=\"84f6\">Databricks introduced that DBRX is a hybrid expert model (MoE) built on MegaBlocks research and open source projects, so it can output tokens extremely quickly per second. Databricks believes that this will pave the way for the most advanced open source model of MoE in the future.<\/p>","protected":false},"excerpt":{"rendered":"<p>The start-up company Databricks recently announced the launch of the open source AI model, DBRX, which claims to be the world ' s most powerful open source large-language model to date, stronger than Meta ' s Llama 2. DBRX uses the transformer structure, which contains 132.2 billion parameters and consists of 16 expert networks, using 4 of which each reason to activate 36 billion parameters. Databricks, in a company blog post, describes the comparison of Llama 2-70B of Meta, Mixtrali of France, and xA of Mask<\/p>","protected":false},"author":1,"featured_media":0,"comment_status":"open","ping_status":"open","sticky":false,"template":"","format":"standard","meta":{"footnotes":""},"categories":[146],"tags":[167,1906,219],"collection":[],"class_list":["post-6464","post","type-post","status-publish","format-standard","hentry","category-news","tag-ai","tag-databricks","tag-219"],"acf":[],"_links":{"self":[{"href":"https:\/\/www.1ai.net\/en\/wp-json\/wp\/v2\/posts\/6464","targetHints":{"allow":["GET"]}}],"collection":[{"href":"https:\/\/www.1ai.net\/en\/wp-json\/wp\/v2\/posts"}],"about":[{"href":"https:\/\/www.1ai.net\/en\/wp-json\/wp\/v2\/types\/post"}],"author":[{"embeddable":true,"href":"https:\/\/www.1ai.net\/en\/wp-json\/wp\/v2\/users\/1"}],"replies":[{"embeddable":true,"href":"https:\/\/www.1ai.net\/en\/wp-json\/wp\/v2\/comments?post=6464"}],"version-history":[{"count":0,"href":"https:\/\/www.1ai.net\/en\/wp-json\/wp\/v2\/posts\/6464\/revisions"}],"wp:attachment":[{"href":"https:\/\/www.1ai.net\/en\/wp-json\/wp\/v2\/media?parent=6464"}],"wp:term":[{"taxonomy":"category","embeddable":true,"href":"https:\/\/www.1ai.net\/en\/wp-json\/wp\/v2\/categories?post=6464"},{"taxonomy":"post_tag","embeddable":true,"href":"https:\/\/www.1ai.net\/en\/wp-json\/wp\/v2\/tags?post=6464"},{"taxonomy":"collection","embeddable":true,"href":"https:\/\/www.1ai.net\/en\/wp-json\/wp\/v2\/collection?post=6464"}],"curies":[{"name":"wp","href":"https:\/\/api.w.org\/{rel}","templated":true}]}}