{"id":16019,"date":"2024-07-21T08:44:08","date_gmt":"2024-07-21T00:44:08","guid":{"rendered":"https:\/\/www.1ai.net\/?p=16019"},"modified":"2024-07-21T08:44:08","modified_gmt":"2024-07-21T00:44:08","slug":"iphone-15-%e4%b9%9f%e5%8f%af%e8%bf%90%e8%a1%8c%ef%bc%8chugging-face-%e6%8e%a8%e5%87%basmollm%e5%b0%8f%e8%af%ad%e8%a8%80-python-%e7%bc%96%e7%a8%8b%e6%a8%a1%e5%9e%8b","status":"publish","type":"post","link":"https:\/\/www.1ai.net\/en\/16019.html","title":{"rendered":"iPhone 15 can also run, Hugging Face launched &quot;SmolLM&quot; small language Python programming model"},"content":{"rendered":"<p>Nowadays, small language models are becoming popular, and many manufacturers have begun to launch &quot;<a href=\"https:\/\/www.1ai.net\/en\/tag\/%e5%b0%8f%e6%a8%a1%e5%9e%8b\" title=\"_Other Organiser\" target=\"_blank\" >Small Model<\/a>\u201d, this week <a href=\"https:\/\/www.1ai.net\/en\/tag\/hugging-face\" title=\"[See articles with [Hugging Face] label]\" target=\"_blank\" >Hugging Face<\/a> It announced the &quot;<a href=\"https:\/\/www.1ai.net\/en\/tag\/smollm\" title=\"_Other Organiser\" target=\"_blank\" >SmolLM<\/a>&quot;A family of small language models, including 135 million, 360 million, and 1.7 billion parameter models.<\/p>\n<p><img loading=\"lazy\" decoding=\"async\" class=\"alignnone size-full wp-image-16020\" title=\"000b9180-7851-4f5b-a684-1302a8b2a364\" src=\"https:\/\/www.1ai.net\/wp-content\/uploads\/2024\/07\/000b9180-7851-4f5b-a684-1302a8b2a364.png\" alt=\"000b9180-7851-4f5b-a684-1302a8b2a364\" width=\"990\" height=\"500\" \/><\/p>\n<p>According to reports, these models are said to be trained with carefully planned high-quality training data sets, and are said to be quite powerful in Python programming performance. The team pointed out that they focused on optimizing the amount of RAM required for the model, &quot;even on an iPhone 15 with 6GB of RAM.&quot;<\/p>\n<p>In terms of training, the Hugging Face team first created a dataset called SmolLM-Corpus (click here to access the dataset address), which mainly includes Python teaching content Python-Edu, Web education content FineWeb-Edu, and common sense content generated by the Mixtral-8x7B-Instruct-v0.1 and Cosmopedia v2 models, with a total token volume of 600 billion. After that, the Hugging Face team used the SmolLM-Corpus dataset to train the &quot;SmolLM&quot; small language model.<\/p>\n<p>The Hugging Face team benchmarked the SmolLM model they developed against other models with the same number of parameters. The SmolLM-135M surpassed other models with less than 200 million parameters in multiple tests. The SmolLM-360M performed better than all models with less than 500 million parameters, but was inferior to the MobileLLM-350M just announced by Meta in some projects. The SmolLM-1.7B model surpassed all models with less than 2 billion parameters, including Microsoft Phi-1.5, MobileLLM-1.5B and Qwen2.<\/p>\n<p><img loading=\"lazy\" decoding=\"async\" class=\"alignnone size-full wp-image-16021\" title=\"1ffb042e-324b-4f3a-9ff8-13943b6c3fb9\" src=\"https:\/\/www.1ai.net\/wp-content\/uploads\/2024\/07\/1ffb042e-324b-4f3a-9ff8-13943b6c3fb9.png\" alt=\"1ffb042e-324b-4f3a-9ff8-13943b6c3fb9\" width=\"1356\" height=\"856\" \/><\/p>","protected":false},"excerpt":{"rendered":"<p>Now that small language models are on the rise, and many vendors are starting to release \"small models\" for lightweight devices like cell phones, this week Hugging Face announced the \"SmolLM\" family of small language models, which includes 135 million. 360 million and 1.7 billion parameter models, 360 million and 1.7 billion parameter models. The models are said to be trained on a carefully curated, high-quality training dataset, and are said to be quite powerful in terms of Python programming performance, with the team noting that they focused on optimizing the amount of RAM usage required for the models to \"run even on an iPhone 15 with 6GB of RAM\". On the training side, the Hugging Face team started by building a SmolLM-Corpus<\/p>","protected":false},"author":1,"featured_media":0,"comment_status":"closed","ping_status":"closed","sticky":false,"template":"","format":"standard","meta":{"footnotes":""},"categories":[146],"tags":[384,3566,3631],"collection":[],"class_list":["post-16019","post","type-post","status-publish","format-standard","hentry","category-news","tag-hugging-face","tag-smollm","tag-3631"],"acf":[],"_links":{"self":[{"href":"https:\/\/www.1ai.net\/en\/wp-json\/wp\/v2\/posts\/16019","targetHints":{"allow":["GET"]}}],"collection":[{"href":"https:\/\/www.1ai.net\/en\/wp-json\/wp\/v2\/posts"}],"about":[{"href":"https:\/\/www.1ai.net\/en\/wp-json\/wp\/v2\/types\/post"}],"author":[{"embeddable":true,"href":"https:\/\/www.1ai.net\/en\/wp-json\/wp\/v2\/users\/1"}],"replies":[{"embeddable":true,"href":"https:\/\/www.1ai.net\/en\/wp-json\/wp\/v2\/comments?post=16019"}],"version-history":[{"count":0,"href":"https:\/\/www.1ai.net\/en\/wp-json\/wp\/v2\/posts\/16019\/revisions"}],"wp:attachment":[{"href":"https:\/\/www.1ai.net\/en\/wp-json\/wp\/v2\/media?parent=16019"}],"wp:term":[{"taxonomy":"category","embeddable":true,"href":"https:\/\/www.1ai.net\/en\/wp-json\/wp\/v2\/categories?post=16019"},{"taxonomy":"post_tag","embeddable":true,"href":"https:\/\/www.1ai.net\/en\/wp-json\/wp\/v2\/tags?post=16019"},{"taxonomy":"collection","embeddable":true,"href":"https:\/\/www.1ai.net\/en\/wp-json\/wp\/v2\/collection?post=16019"}],"curies":[{"name":"wp","href":"https:\/\/api.w.org\/{rel}","templated":true}]}}