{"id":1872,"date":"2023-12-12T09:32:03","date_gmt":"2023-12-12T01:32:03","guid":{"rendered":"https:\/\/www.1ai.net\/?p=1872"},"modified":"2023-12-12T09:32:03","modified_gmt":"2023-12-12T01:32:03","slug":"%e9%ba%bb%e7%9c%81%e7%90%86%e5%b7%a5%e5%ad%a6%e9%99%a2%e5%ad%a6%e8%80%85%e5%8f%91%e5%b8%83%e5%85%b3%e4%ba%8e%e4%ba%ba%e5%b7%a5%e6%99%ba%e8%83%bd%e6%b2%bb%e7%90%86%e7%9a%84%e6%94%bf%e7%ad%96%e6%96%87","status":"publish","type":"post","link":"https:\/\/www.1ai.net\/en\/1872.html","title":{"rendered":"MIT scholars release policy paper on AI governance"},"content":{"rendered":"<p><a href=\"https:\/\/www.1ai.net\/en\/tag\/%e9%ba%bb%e7%9c%81%e7%90%86%e5%b7%a5%e5%ad%a6%e9%99%a2\" title=\"[Sees articles with labels]\" target=\"_blank\" >Massachusetts Institute of Technology<\/a>A group of<span class=\"spamTxt\">Leaders<\/span>A temporary committee of scholars has issued a series of<a href=\"https:\/\/www.1ai.net\/en\/tag\/%e4%ba%ba%e5%b7%a5%e6%99%ba%e8%83%bd\" title=\"[View articles tagged with [artificial intelligence]]\" target=\"_blank\" >AI<\/a>The main policy document is titled &quot;A Framework for the Governance of American Artificial Intelligence: Building a Safe and Prosperous Artificial Intelligence Industry,&quot; and suggests that existing U.S. government agencies could be expanded to oversee AI.<a href=\"https:\/\/www.1ai.net\/en\/tag\/ai%e5%b7%a5%e5%85%b7\" title=\"[SEE ARTICLES WITH [AI TOOL] LABELS]\" target=\"_blank\" >AI Tools<\/a>, and stressed the importance of defining the purpose of AI tools in order to develop appropriate regulatory provisions.<\/p>\n<p class=\"article-content__img\"><img loading=\"lazy\" decoding=\"async\" class=\"alignnone size-full wp-image-1873\" title=\"202311231146402911_4\" src=\"https:\/\/www.1ai.net\/wp-content\/uploads\/2023\/12\/202311231146402911_4.jpg\" alt=\"202311231146402911_4\" width=\"1000\" height=\"752\" \/><\/p>\n<p>Source Note: The image is generated by AI, and the image is authorized by Midjourney<\/p>\n<p>\u201cWe already have regulation and governance in this country for a lot of relatively high-risk things,\u201d said Dan Huttenlocher, dean of the MIT Schwarzman School of Computing. \u201cWe\u2019re not saying that\u2019s enough, but let\u2019s start with where we already have regulation, and that\u2019s a practical approach.\u201d<\/p>\n<p>The policy document emphasizes the importance of AI providers defining the purpose and intent of applications in advance. In this way, the regulatory system can clarify which existing regulations and regulators apply to specific AI tools. In addition, the document discusses situations where AI systems may exist in multiple layers, what technologists call a &quot;stack&quot; system, highlighting the complexity of developing responsibilities and supervision.<\/p>\n<p>The policy document not only involves existing institutions, but also proposes to enhance new regulatory capabilities. Among them, the document calls for audits of new AI tools, which can be initiated by the government, driven by users, or derived from legal liability lawsuits. The document also recommends the development of public auditing standards, which can be established by a non-profit entity similar to the Public Company Accounting Oversight Board (PCAOB) or a federal entity similar to the National Institute of Standards and Technology (NIST).<\/p>\n<p>In addition, the document considers the possibility of creating a new, government-approved &quot;self-regulatory organization&quot; (SRO) agency that would function similarly to the government-created Financial Industry Regulatory Authority FINRA. Such an agency, focused on AI, could accumulate domain-specific knowledge to remain flexible and responsive when interacting with the rapidly changing AI industry.<\/p>\n<p>The policy document states that there are specific legal issues that need to be addressed in the field of AI, such as copyright issues related to intellectual property. In addition, the committee recognizes that &quot;human-plus&quot; legal issues, that is, situations where AI has super-human capabilities, such as large-scale surveillance tools, may require special legal considerations.<\/p>\n<p>This series of policy documents covers the analysis of AI regulation issues from multiple disciplinary perspectives, demonstrating the interim committee&#039;s commitment to influencing policy making from a broad perspective, not just limited to technical issues. The committee emphasizes the importance of academic institutions&#039; expertise in the interaction between science, technology and society, and believes that policymakers need to think about the relationship between social systems and technology.<\/p>\n<p>The committee hopes to bridge the gap between those who are radical and those who are concerned about AI, and promote the healthy development of the AI industry by advocating that technological advancement be accompanied by appropriate regulation.<\/p>","protected":false},"excerpt":{"rendered":"<p>An interim committee of a group of leaders and scholars at MIT has released a series of policy papers on AI governance aimed at providing resources for U.S. policymakers to develop a better regulatory framework for AI. The main policy paper, titled \"A Framework for the Governance of Artificial Intelligence in the United States:Creating a Safe and Prosperous Artificial Intelligence Industry,\" suggests that AI tools can be regulated by expanding existing U.S. government agencies and emphasizes the importance of defining the purpose of AI tools in order to develop appropriate regulatory provisions. Image source note: Image generated by AI, image license provider Midjourney \"We as a country already regulate and govern a lot of relatively high-risk things,\" said Dan Huttenloch, dean of the Schwartzman School of Computing at the Massachusetts Institute of Technology, \"We're not saying that's already enough, but<\/p>","protected":false},"author":1,"featured_media":0,"comment_status":"open","ping_status":"open","sticky":false,"template":"","format":"standard","meta":{"footnotes":""},"categories":[146],"tags":[387,204,619],"collection":[],"class_list":["post-1872","post","type-post","status-publish","format-standard","hentry","category-news","tag-ai","tag-204","tag-619"],"acf":[],"_links":{"self":[{"href":"https:\/\/www.1ai.net\/en\/wp-json\/wp\/v2\/posts\/1872","targetHints":{"allow":["GET"]}}],"collection":[{"href":"https:\/\/www.1ai.net\/en\/wp-json\/wp\/v2\/posts"}],"about":[{"href":"https:\/\/www.1ai.net\/en\/wp-json\/wp\/v2\/types\/post"}],"author":[{"embeddable":true,"href":"https:\/\/www.1ai.net\/en\/wp-json\/wp\/v2\/users\/1"}],"replies":[{"embeddable":true,"href":"https:\/\/www.1ai.net\/en\/wp-json\/wp\/v2\/comments?post=1872"}],"version-history":[{"count":0,"href":"https:\/\/www.1ai.net\/en\/wp-json\/wp\/v2\/posts\/1872\/revisions"}],"wp:attachment":[{"href":"https:\/\/www.1ai.net\/en\/wp-json\/wp\/v2\/media?parent=1872"}],"wp:term":[{"taxonomy":"category","embeddable":true,"href":"https:\/\/www.1ai.net\/en\/wp-json\/wp\/v2\/categories?post=1872"},{"taxonomy":"post_tag","embeddable":true,"href":"https:\/\/www.1ai.net\/en\/wp-json\/wp\/v2\/tags?post=1872"},{"taxonomy":"collection","embeddable":true,"href":"https:\/\/www.1ai.net\/en\/wp-json\/wp\/v2\/collection?post=1872"}],"curies":[{"name":"wp","href":"https:\/\/api.w.org\/{rel}","templated":true}]}}