{"id":44549,"date":"2025-10-12T16:45:29","date_gmt":"2025-10-12T08:45:29","guid":{"rendered":"https:\/\/www.1ai.net\/?p=44549"},"modified":"2025-10-12T16:45:35","modified_gmt":"2025-10-12T08:45:35","slug":"%e8%89%be%e4%bc%a6%e3%83%bb%e5%9b%be%e7%81%b5%e7%a0%94%e7%a9%b6%e6%89%80%e7%ad%89%e6%9c%80%e6%96%b0%e7%a0%94%e7%a9%b6%ef%bc%9a%e5%8f%aa%e9%9c%80-250-%e4%bb%bd%e6%81%b6%e6%84%8f%e6%96%87%e6%a1%a3","status":"publish","type":"post","link":"https:\/\/www.1ai.net\/en\/44549.html","title":{"rendered":"A RECENT STUDY BY THE ALAN TURING INSTITUTE: 250 MALICIOUS FILES ARE NEEDED TO BREAK THROUGH ANY SIZE"},"content":{"rendered":"<p>October 13th, according to TechXplore, the latest research by Anthropic, AI, and Alan Turing found that even the largest <a href=\"https:\/\/www.1ai.net\/en\/tag\/ai%e6%a8%a1%e5%9e%8b\" title=\"[View articles tagged with [AI models]]\" target=\"_blank\" >AI Models<\/a>Just about 250 malicious documents could be successfully invaded\u3002<\/p>\n<p><img loading=\"lazy\" decoding=\"async\" class=\"alignnone size-full wp-image-44550\" title=\"9634f9a7j00t40gai008pd000m809mm\" src=\"https:\/\/www.1ai.net\/wp-content\/uploads\/2025\/10\/9634f9a7j00t40gai008pd000m8009mp.jpg\" alt=\"9634f9a7j00t40gai008pd000m809mm\" width=\"800\" height=\"346\" \/><\/p>\n<p>Training data for large language models<strong>Mostly from open networks<\/strong>This allows them to accumulate a vast knowledge base and to generate natural languages, but is also exposed to the risk of data poisoning\u3002<\/p>\n<p>In the past, it had been widely felt that, as the size of the model grew, the risk would be diluted, as the proportion of poisoning data needed to remain constant. In other words, to contaminate a giant model<strong>We need a lot of malice<\/strong>I don't know. However, this study, published on the arXiv platform, subverts this assumption that the attackers need only<strong>Very few malicious documents<\/strong>It can cause serious damage\u3002<\/p>\n<p>In order to verify the difficulty of the attack, the team built a multi-model from zero<strong>\u00a0<\/strong><strong>600 million to 13 billion parameters<\/strong>Not at all. Each model is trained in clean, open data, but the researchers are separate<strong>Insert 100 to 500 malicious documents<\/strong>.<\/p>\n<p>Then the team tried to pass<strong>Adjusting the distribution or timing of malicious data<\/strong>To defend against attack and retest it in the model fine-tuning phase\u3002<\/p>\n<p>The result was shocking: the size of the model<strong>It barely worked<\/strong>I don't know. Only 250 malicious documents<strong>Successfully implanted \"back door\" in all models<\/strong>(NOTE: A HIDDEN MECHANISM THAT ALLOWS AI TO EXECUTE HARMFUL INSTRUCTIONS AFTER THEY ARE TRIGGERED). EVEN LARGE MODELS WITH 20 TIMES MORE TRAINED DATA THAN THE SMALLEST MODEL ARE UNABLE TO WITHSTAND ATTACK\u3002<strong>Adding additional clean data will neither dilute the risk nor prevent intrusion<\/strong>.<\/p>\n<p>ACCORDING TO RESEARCHERS, THIS MEANS THAT DEFENCE ISSUES ARE MORE URGENT THAN ANTICIPATED. INSTEAD OF BLINDLY PURSUING LARGER MODELS, THE AI FIELD SHOULD BE MORE APPROPRIATE<strong>Development of focused security mechanisms<\/strong>I don't know. The paper mentioned that: \u201cOur research shows that the difficulty of a large model being drugged by data into the back door does not increase with size, which points to the urgent need for more research on means of defence in the future.\u201d<\/p>\n<p>Link to paper:<\/p>\n<ul>\n<li><a href=\"https:\/\/arxiv.org\/abs\/2510.07192\">[2510.07192]<\/a><\/li>\n<\/ul>","protected":false},"excerpt":{"rendered":"<p>According to October 13, according to a foreign media source, TechXplore, recent studies by Anthropic, AI UK and Alan Turing found that even the largest AI model, with about 250 malicious files, could be successfully invaded. Training data on large-language models are mostly derived from open networks, which allow them to accumulate a vast knowledge base and generate natural languages, but are also exposed to the risk of data poisoning. In the past, it had been widely felt that, as the size of the model grew, the risk would be diluted, as the proportion of poisoning data needed to remain constant. That is to say, the contamination of mega-models requires a large number of malicious samples. However, this study, published on the arXiv platform, subverts this assumption that the attackers need only<\/p>","protected":false},"author":1,"featured_media":0,"comment_status":"closed","ping_status":"closed","sticky":false,"template":"","format":"standard","meta":{"footnotes":""},"categories":[146],"tags":[167],"collection":[],"class_list":["post-44549","post","type-post","status-publish","format-standard","hentry","category-news","tag-ai"],"acf":[],"_links":{"self":[{"href":"https:\/\/www.1ai.net\/en\/wp-json\/wp\/v2\/posts\/44549","targetHints":{"allow":["GET"]}}],"collection":[{"href":"https:\/\/www.1ai.net\/en\/wp-json\/wp\/v2\/posts"}],"about":[{"href":"https:\/\/www.1ai.net\/en\/wp-json\/wp\/v2\/types\/post"}],"author":[{"embeddable":true,"href":"https:\/\/www.1ai.net\/en\/wp-json\/wp\/v2\/users\/1"}],"replies":[{"embeddable":true,"href":"https:\/\/www.1ai.net\/en\/wp-json\/wp\/v2\/comments?post=44549"}],"version-history":[{"count":0,"href":"https:\/\/www.1ai.net\/en\/wp-json\/wp\/v2\/posts\/44549\/revisions"}],"wp:attachment":[{"href":"https:\/\/www.1ai.net\/en\/wp-json\/wp\/v2\/media?parent=44549"}],"wp:term":[{"taxonomy":"category","embeddable":true,"href":"https:\/\/www.1ai.net\/en\/wp-json\/wp\/v2\/categories?post=44549"},{"taxonomy":"post_tag","embeddable":true,"href":"https:\/\/www.1ai.net\/en\/wp-json\/wp\/v2\/tags?post=44549"},{"taxonomy":"collection","embeddable":true,"href":"https:\/\/www.1ai.net\/en\/wp-json\/wp\/v2\/collection?post=44549"}],"curies":[{"name":"wp","href":"https:\/\/api.w.org\/{rel}","templated":true}]}}