{"id":10163,"date":"2024-05-12T09:18:21","date_gmt":"2024-05-12T01:18:21","guid":{"rendered":"https:\/\/www.1ai.net\/?p=10163"},"modified":"2024-05-12T09:18:21","modified_gmt":"2024-05-12T01:18:21","slug":"anthropic%e5%b0%86%e5%85%81%e8%ae%b8%e6%9c%aa%e6%88%90%e5%b9%b4%e4%ba%ba%e5%9c%a8%e9%99%90%e5%88%b6%e8%8c%83%e5%9b%b4%e5%86%85%e4%bd%bf%e7%94%a8%e5%85%b6%e4%ba%ba%e5%b7%a5%e6%99%ba%e8%83%bd","status":"publish","type":"post","link":"https:\/\/www.1ai.net\/en\/10163.html","title":{"rendered":"Anthropic will allow minors to use its artificial intelligence within limits"},"content":{"rendered":"<p>Artificial Intelligence Startups <a href=\"https:\/\/www.1ai.net\/en\/tag\/anthropic\" title=\"[View articles tagged with [Anthropic]]\" target=\"_blank\" >Anthropic<\/a>\u00a0<span class=\"spamTxt\">up to date<\/span>The policy change has attracted attention. The company announced that it would allow<a href=\"https:\/\/www.1ai.net\/en\/tag\/%e6%9c%aa%e6%88%90%e5%b9%b4%e4%ba%ba\" title=\"[Sees articles with [minor] labels]\" target=\"_blank\" >Minors<\/a>The move to allow developers to use its generative artificial intelligence system, subject to specific safety requirements, has sparked widespread discussion and concern in the industry.<\/p>\n<p>Under Anthropic&#039;s new policy, teens and children will be able to use third-party applications powered by its artificial intelligence models under certain conditions. This move is seen as having potential benefits for education and personal problem solving. However, the challenge that comes with it is how to ensure that minors use artificial intelligence tools safely.<\/p>\n<p class=\"article-content__img\"><img loading=\"lazy\" decoding=\"async\" class=\"alignnone size-full wp-image-10164\" title=\"202310180948538535_0\" src=\"https:\/\/www.1ai.net\/wp-content\/uploads\/2024\/05\/202310180948538535_0.jpg\" alt=\"202310180948538535_0\" width=\"680\" height=\"357\" \/><\/p>\n<p>Anthropic lists safety measures that developers should take, such as age verification systems, content review, and educational resources. At the same time, the company said it will regularly audit apps for compliance, terminate accounts that violate regulations, and require developers to clearly state that they adhere to compliance requirements.<\/p>\n<p>The policy adjustment also reflects a trend across the industry. Competitors including Google and OpenAI are also exploring more use cases for generative AI for children. Last year, OpenAI worked with Common Sense Media to develop child-friendly AI guidelines, and Google renamed its chatbot Bard to Gemini, specifically for teenagers.<\/p>\n<p>However, the potential risks of generative AI have also attracted much attention. Some surveys show that some children have seen their peers use generative AI in a negative way. Therefore, guidance and supervision of children&#039;s use of generative AI has become particularly important.<\/p>\n<p>In this era full of opportunities and challenges, Anthropic&#039;s policy adjustment has triggered in-depth thinking and discussion on the use of artificial intelligence by minors. How to balance the benefits of using artificial intelligence while protecting the safety and privacy of minors will be a key issue for future development.<\/p>","protected":false},"excerpt":{"rendered":"<p>Artificial intelligence startup Anthropic's latest policy change has raised concerns. The company announced that it would allow minors to use its generative AI system, subject to specific safety requirements. The move has sparked widespread discussion and concern within the industry. Under Anthropic's new policy, teens and children will be able to use third-party apps powered by its AI models under specific conditions. This initiative is seen as potentially beneficial for education and personal problem solving. However, the challenge that comes with it is how to ensure that minors use AI tools safely. Anthropic lists safety measures that developers should take, such as age verification systems, content audits, and educational resources. At the same time, the company says it will regularly audit apps for compliance and end violations of the<\/p>","protected":false},"author":1,"featured_media":0,"comment_status":"open","ping_status":"open","sticky":false,"template":"","format":"standard","meta":{"footnotes":""},"categories":[146],"tags":[320,1208],"collection":[],"class_list":["post-10163","post","type-post","status-publish","format-standard","hentry","category-news","tag-anthropic","tag-1208"],"acf":[],"_links":{"self":[{"href":"https:\/\/www.1ai.net\/en\/wp-json\/wp\/v2\/posts\/10163","targetHints":{"allow":["GET"]}}],"collection":[{"href":"https:\/\/www.1ai.net\/en\/wp-json\/wp\/v2\/posts"}],"about":[{"href":"https:\/\/www.1ai.net\/en\/wp-json\/wp\/v2\/types\/post"}],"author":[{"embeddable":true,"href":"https:\/\/www.1ai.net\/en\/wp-json\/wp\/v2\/users\/1"}],"replies":[{"embeddable":true,"href":"https:\/\/www.1ai.net\/en\/wp-json\/wp\/v2\/comments?post=10163"}],"version-history":[{"count":0,"href":"https:\/\/www.1ai.net\/en\/wp-json\/wp\/v2\/posts\/10163\/revisions"}],"wp:attachment":[{"href":"https:\/\/www.1ai.net\/en\/wp-json\/wp\/v2\/media?parent=10163"}],"wp:term":[{"taxonomy":"category","embeddable":true,"href":"https:\/\/www.1ai.net\/en\/wp-json\/wp\/v2\/categories?post=10163"},{"taxonomy":"post_tag","embeddable":true,"href":"https:\/\/www.1ai.net\/en\/wp-json\/wp\/v2\/tags?post=10163"},{"taxonomy":"collection","embeddable":true,"href":"https:\/\/www.1ai.net\/en\/wp-json\/wp\/v2\/collection?post=10163"}],"curies":[{"name":"wp","href":"https:\/\/api.w.org\/{rel}","templated":true}]}}