{"id":32339,"date":"2025-04-04T15:21:13","date_gmt":"2025-04-04T07:21:13","guid":{"rendered":"https:\/\/www.1ai.net\/?p=32339"},"modified":"2025-04-04T15:21:13","modified_gmt":"2025-04-04T07:21:13","slug":"%e8%b0%b7%e6%ad%8c%e6%8b%89%e5%93%8d%e9%80%9a%e7%94%a8%e4%ba%ba%e5%b7%a5%e6%99%ba%e8%83%bd%e8%ad%a6%e6%8a%a5%ef%bc%8c%e9%a6%96%e6%ac%a1%e5%85%ac%e5%bc%80-ai-%e5%ae%89%e5%85%a8%e9%98%b2%e5%be%a1","status":"publish","type":"post","link":"https:\/\/www.1ai.net\/en\/32339.html","title":{"rendered":"Google sounds general AI alarm, makes AI security defense blueprint public for the first time"},"content":{"rendered":"<p>April 4, 2011 - Technology media outlet WinBuzzer published a blog post yesterday, April 3, reporting that<a href=\"https:\/\/www.1ai.net\/en\/tag\/%e8%b0%b7%e6%ad%8c\" title=\"[View articles tagged with [Google]]\" target=\"_blank\" >Google<\/a>'s DeepMind's latest release of global <a href=\"https:\/\/www.1ai.net\/en\/tag\/agi\" title=\"_OTHER ORGANISER\" target=\"_blank\" >AGI<\/a>(General)<a href=\"https:\/\/www.1ai.net\/en\/tag\/%e4%ba%ba%e5%b7%a5%e6%99%ba%e8%83%bd\" title=\"[View articles tagged with [artificial intelligence]]\" target=\"_blank\" >AI<\/a>) security framework, calling for transnational protection mechanisms to be put in place before technology gets out of hand.<\/p>\n<p><img loading=\"lazy\" decoding=\"async\" class=\"alignnone size-full wp-image-32340\" title=\"98d35cfaj00su6n2h00aqd000v900n8p\" src=\"https:\/\/www.1ai.net\/wp-content\/uploads\/2025\/04\/98d35cfaj00su6n2h00aqd000v900n8p.jpg\" alt=\"98d35cfaj00su6n2h00aqd000v900n8p\" width=\"1125\" height=\"836\" \/><\/p>\n<p>DeepMind believes that AGI is on the verge of being realized and advocates immediate action. AGI may achieve human-level cognitive capabilities in the coming years, and its autonomous decision-making characteristics may accelerate breakthroughs in healthcare, education, and other fields, but the risks of misuse and misalignment of goals need to be guarded against.<\/p>\n<p>Google DeepMind releases \"A Technical AGI Safety and Assurance Approach\" whitepaper proposing a systematic approach to addressing the potential risks of general-purpose artificial intelligence (AGI).<\/p>\n<p>Citing a blog post, 1AI described the report as focusing on four major risk areas (misuse, misalignment, accidents, and structural risk).<strong>It is proposed to reduce hazards through safety mechanism design, transparent research and industry collaboration.<\/strong><\/p>\n<p>Target misalignment is one of the core risks of AGI. When AI adopts unconventional means to accomplish a task (e.g., hacking into the booking system to get a seat), it will deviate from the human intent. deepMind trains AI to identify the correct target through \"amplified supervision\" technology and uses AI self-assessment (e.g., debating mechanism) to improve the judgment in complex scenarios.<\/p>\n<p>DeepMind's proposed international security framework eschews abstract ethical discussions and focuses on practical issues in the rapid evolution of technology, including the formation of a transnational assessment body similar to the Nuclear Non-Proliferation Treaty and the establishment of a national AI risk monitoring center.<\/p>\n<p>Google DeepMind has proposed a three-pillar program of strengthening technical research, deploying early warning systems, and coordinating governance through international institutions, emphasizing the urgent need to limit dangerous capabilities such as AI cyber attacks.<\/p>\n<p>DeepMind's initiatives are not isolated. Competitor Anthropic warned in November 2024 of the need to curb AI failures within 18 months and set capacity thresholds to trigger protections; Meta launched its Cutting Edge AI Framework in February 2025 to stop making high-risk models publicly available.<\/p>\n<p>Security has been extended to hardware. NVIDIA 2025 launched the NeMo Guardrails microservices suite in January to intercept harmful outputs in real time, with current applications in healthcare, automotive, and other industries.<\/p>","protected":false},"excerpt":{"rendered":"<p>On April 4th, the technology media WinBuzzer published an article yesterday (3 April) reporting that DeepMind, under Google, recently released the Global AGI (General Artificial Intelligence) Safety Framework, calling for a transnational protection mechanism before technology went out of control. DeepMind thinks AGI is about to land and advocates immediate action. AGI may achieve human-level cognitive capacity in the years ahead, and its autonomous decision-making characteristics may accelerate breakthroughs in areas such as health, education and so forth, but the risks of abuse and misalignment also require vigilance. Google DeepMind released a white paper entitled \" Technical AGI Safety and Security Approach \" , proposing a systematic approach to addressing potential AGI risks. 1AI Citation of Boven<\/p>","protected":false},"author":1,"featured_media":0,"comment_status":"closed","ping_status":"closed","sticky":false,"template":"","format":"standard","meta":{"footnotes":""},"categories":[146],"tags":[151,204,6181,281],"collection":[],"class_list":["post-32339","post","type-post","status-publish","format-standard","hentry","category-news","tag-agi","tag-204","tag-6181","tag-281"],"acf":[],"_links":{"self":[{"href":"https:\/\/www.1ai.net\/en\/wp-json\/wp\/v2\/posts\/32339","targetHints":{"allow":["GET"]}}],"collection":[{"href":"https:\/\/www.1ai.net\/en\/wp-json\/wp\/v2\/posts"}],"about":[{"href":"https:\/\/www.1ai.net\/en\/wp-json\/wp\/v2\/types\/post"}],"author":[{"embeddable":true,"href":"https:\/\/www.1ai.net\/en\/wp-json\/wp\/v2\/users\/1"}],"replies":[{"embeddable":true,"href":"https:\/\/www.1ai.net\/en\/wp-json\/wp\/v2\/comments?post=32339"}],"version-history":[{"count":0,"href":"https:\/\/www.1ai.net\/en\/wp-json\/wp\/v2\/posts\/32339\/revisions"}],"wp:attachment":[{"href":"https:\/\/www.1ai.net\/en\/wp-json\/wp\/v2\/media?parent=32339"}],"wp:term":[{"taxonomy":"category","embeddable":true,"href":"https:\/\/www.1ai.net\/en\/wp-json\/wp\/v2\/categories?post=32339"},{"taxonomy":"post_tag","embeddable":true,"href":"https:\/\/www.1ai.net\/en\/wp-json\/wp\/v2\/tags?post=32339"},{"taxonomy":"collection","embeddable":true,"href":"https:\/\/www.1ai.net\/en\/wp-json\/wp\/v2\/collection?post=32339"}],"curies":[{"name":"wp","href":"https:\/\/api.w.org\/{rel}","templated":true}]}}