{"id":32448,"date":"2025-04-06T15:41:48","date_gmt":"2025-04-06T07:41:48","guid":{"rendered":"https:\/\/www.1ai.net\/?p=32448"},"modified":"2025-04-06T15:41:48","modified_gmt":"2025-04-06T07:41:48","slug":"meta-%e5%8f%91%e5%b8%83-llama-4-%e7%b3%bb%e5%88%97-ai%e6%a8%a1%e5%9e%8b%ef%bc%8c%e5%bc%95%e5%85%a5%e6%b7%b7%e5%90%88%e4%b8%93%e5%ae%b6%e6%9e%b6%e6%9e%84%e6%8f%90%e5%8d%87%e6%95%88","status":"publish","type":"post","link":"https:\/\/www.1ai.net\/en\/32448.html","title":{"rendered":"Meta Releases Llama 4 Series of AI Models, Introduces \"Hybrid Expert Architecture\" to Improve Efficiency"},"content":{"rendered":"<p>April 6 News.<a href=\"https:\/\/www.1ai.net\/en\/tag\/meta\" title=\"[View articles tagged with [Meta]]\" target=\"_blank\" >Meta<\/a> Now releasing its latest <a href=\"https:\/\/www.1ai.net\/en\/tag\/llama-4\" title=\"[See article with [Llama 4] label]\" target=\"_blank\" >Llama 4<\/a> series <a href=\"https:\/\/www.1ai.net\/en\/tag\/ai%e6%a8%a1%e5%9e%8b\" title=\"[View articles tagged with [AI models]]\" target=\"_blank\" >AI Models<\/a>Contains <a href=\"https:\/\/www.1ai.net\/en\/tag\/llama\" title=\"_Other Organiser\" target=\"_blank\" >Llama<\/a> 4 Scout, Llama 4 Maverick and Llama 4 Behemoth.<strong>Meta reveals that the corresponding models are trained on \"large amounts of unlabeled text, image, and video data\" to give them \"broad visual comprehension capabilities.\"<\/strong>.<\/p>\n<p><img loading=\"lazy\" decoding=\"async\" class=\"alignnone size-full wp-image-32449\" title=\"5ce2b7f3j00suadcl00agd000v900hkp\" src=\"https:\/\/www.1ai.net\/wp-content\/uploads\/2025\/04\/5ce2b7f3j00suadcl00agd000v900hkp.jpg\" alt=\"5ce2b7f3j00suadcl00agd000v900hkp\" width=\"1125\" height=\"632\" \/><\/p>\n<p>Currently, Meta has uploaded Scout and Maverick to Hugging Face (visit here), while the Behemoth model is still in training. Scout can be run on a single NVIDIA H100 GPU, while Maverick requires the NVIDIA H100 DGX AI platform or an \"equivalent device\".<\/p>\n<p>Meta stated that<strong>The Llama 4 Series models are the first of the company's models to utilize the Mixed Expertise (MoE) architecture<\/strong>This architecture is more efficient at training and answering user queries, and the so-called \"hybrid expert architecture\" basically breaks down data processing tasks into subtasks and delegates them to smaller, specialized \"expert\" models.<\/p>\n<p>For example,<strong>Maverick has 400 billion parameters.<\/strong>The number of active parameters (the number of parameters roughly corresponds to the model's problem solving ability) is only 17 billion in 128 \"expert\" models.<strong>Scout has 17 billion active parameters<\/strong>The model has 16 \"expert\" models and a total of 109 billion parameters.<\/p>\n<p>It is worth noting, however, that none of the models in the Llama 4 series are truly \"inferential\" in the sense that OpenAI's o1 and o3-mini are. By way of comparison, \"inference models\" fact-check their answers and generally answer questions more reliably, but as a result they also take longer to answer than traditional \"non-inference\" models.<\/p>\n<p>Meta's internal tests show that the Maverick model is best suited for use in \"general purpose AI assistants and chat\" scenarios, outperforming models like OpenAI's GPT-4o and Google's Gemini 2.0 in creative writing, code generation, translation, inference, contextual summarization of long text, and image benchmarks. Gemini 2.0. However, Maverick still has some room for improvement over more powerful recent models such as Google's Gemini 2.5 Pro, Anthropic's Claude 3.7 Sonnet, and OpenAI's GPT-4.5.<\/p>\n<p>Scout's strengths are summarizing documents and reasoning based on a large code base. The model supports 10 million lexical elements (\"lexical elements\" represent fragments of the original text, e.g., the word \"fantastic\" can be split into \"fan\", \"tas\", and \"tic\"), so it can process \"up to millions of words at a time\", \"tas\" and \"tic\"), and can therefore process \"up to millions of words of text\" at a time.<\/p>\n<p>1AI notes that Meta further teased its Behemoth model, which according to the company<strong>Behemoth has 288 billion active parameters.<\/strong>Meta's internal benchmarks show Behemoth outperforming GPT-4.5, Claude 3.7, Sonnet, and Gemini 2.0 Pro, but not as well as Gemini 2.5 Pro, on a number of assessments that measure science, technology, engineering, and math (STEM) skills such as solving math problems. Sonnet and Gemini 2.0 Pro, but not as well as Gemini 2.5 Pro.<\/p>","protected":false},"excerpt":{"rendered":"<p>On April 6, Meta has released the latest Llama 4 series of AI models under its flag, which include Llama 4 Scott, Llama 4 Maverick and Llama 4 Behemoth, and Meta has revealed that the corresponding models have been trained in \u201cmany unmarked text, images and video data\u201d to provide them with \u201ca wide range of visual understanding\u201d. At present, Meta has mounted Scout and Maverick in a series of models to Hugging Face, while the Behemoth model is still in training. Of which scout can run on a single engweeda H100 GPU, while M<\/p>","protected":false},"author":1,"featured_media":0,"comment_status":"closed","ping_status":"closed","sticky":false,"template":"","format":"standard","meta":{"footnotes":""},"categories":[146],"tags":[167,184,6188,297],"collection":[],"class_list":["post-32448","post","type-post","status-publish","format-standard","hentry","category-news","tag-ai","tag-llama","tag-llama-4","tag-meta"],"acf":[],"_links":{"self":[{"href":"https:\/\/www.1ai.net\/en\/wp-json\/wp\/v2\/posts\/32448","targetHints":{"allow":["GET"]}}],"collection":[{"href":"https:\/\/www.1ai.net\/en\/wp-json\/wp\/v2\/posts"}],"about":[{"href":"https:\/\/www.1ai.net\/en\/wp-json\/wp\/v2\/types\/post"}],"author":[{"embeddable":true,"href":"https:\/\/www.1ai.net\/en\/wp-json\/wp\/v2\/users\/1"}],"replies":[{"embeddable":true,"href":"https:\/\/www.1ai.net\/en\/wp-json\/wp\/v2\/comments?post=32448"}],"version-history":[{"count":0,"href":"https:\/\/www.1ai.net\/en\/wp-json\/wp\/v2\/posts\/32448\/revisions"}],"wp:attachment":[{"href":"https:\/\/www.1ai.net\/en\/wp-json\/wp\/v2\/media?parent=32448"}],"wp:term":[{"taxonomy":"category","embeddable":true,"href":"https:\/\/www.1ai.net\/en\/wp-json\/wp\/v2\/categories?post=32448"},{"taxonomy":"post_tag","embeddable":true,"href":"https:\/\/www.1ai.net\/en\/wp-json\/wp\/v2\/tags?post=32448"},{"taxonomy":"collection","embeddable":true,"href":"https:\/\/www.1ai.net\/en\/wp-json\/wp\/v2\/collection?post=32448"}],"curies":[{"name":"wp","href":"https:\/\/api.w.org\/{rel}","templated":true}]}}