{"id":13891,"date":"2024-06-24T09:17:29","date_gmt":"2024-06-24T01:17:29","guid":{"rendered":"https:\/\/www.1ai.net\/?p=13891"},"modified":"2024-06-24T09:17:29","modified_gmt":"2024-06-24T01:17:29","slug":"%e8%81%8a%e5%a4%a9%e6%9c%ba%e5%99%a8%e4%ba%ba%e8%83%a1%e8%af%b4%e5%85%ab%e9%81%93%ef%bc%9f%e7%89%9b%e6%b4%a5%e7%a0%94%e7%a9%b6%e4%ba%ba%e5%91%98%e5%88%a9%e7%94%a8%e8%af%ad%e4%b9%89%e7%86%b5%e6%9d%a5","status":"publish","type":"post","link":"https:\/\/www.1ai.net\/en\/13891.html","title":{"rendered":"Chatbots talking nonsense? Oxford researchers use semantic entropy to see through AI &quot;hallucinations&quot;"},"content":{"rendered":"<p data-vmark=\"96f4\">In recent years, artificial intelligence has boomed and<a href=\"https:\/\/www.1ai.net\/en\/tag\/%e8%81%8a%e5%a4%a9%e6%9c%ba%e5%99%a8%e4%ba%ba\" title=\"[View articles tagged with [chatbot]]\" target=\"_blank\" >Chatbots<\/a>Applications such as ChatGPT are gaining popularity, and people can get information from these chatbots (e.g., ChatGPT) through simple commands. However, these chatbots are still prone to \"<a href=\"https:\/\/www.1ai.net\/en\/tag\/ai%e5%b9%bb%e8%a7%89\" title=\"_OTHER ORGANISER\" target=\"_blank\" >AI Hallucinations<\/a>\"questions, i.e., providing wrong answers and sometimes even dangerous information.<\/p>\n<p data-vmark=\"4b4c\"><img loading=\"lazy\" decoding=\"async\" class=\"alignnone size-full wp-image-13892\" title=\"7e9b88b1-f235-496d-ad4d-07e24b2e321e\" src=\"https:\/\/www.1ai.net\/wp-content\/uploads\/2024\/06\/7e9b88b1-f235-496d-ad4d-07e24b2e321e.jpg\" alt=\"7e9b88b1-f235-496d-ad4d-07e24b2e321e\" width=\"940\" height=\"627\" \/><\/p>\n<p>Image source: Pexels<\/p>\n<p data-vmark=\"8426\">One of the reasons for the \"illusion\" is inaccurate training data, insufficient generalization and side effects of the data collection process. However, researchers at the University of Oxford have taken a different approach to the problem, and have developed a new method in<a href=\"https:\/\/www.nature.com\/articles\/s41586-024-07421-0?utm_medium=affiliate&amp;utm_source=commission_junction&amp;utm_campaign=CONR_PF018_ECOM_GL_PBOK_ALWYS_DEEPLINK&amp;utm_content=textlink&amp;utm_term=PID100046186&amp;CJEVENT=f452bca631be11ef83dbf95b0a18b8fa\" target=\"_blank\" rel=\"noopener\">The latest issue of Nature<\/a>on details a method they have newly developed for detecting large language models (<a href=\"https:\/\/www.1ai.net\/en\/tag\/llm\" title=\"[SEE ARTICLES WITH [LLM] LABELS]\" target=\"_blank\" >LLM<\/a>The problem of \"fabrication\" (i.e., arbitrarily generated incorrect information) in the context of the \"S\" (s).<\/p>\n<p data-vmark=\"36ff\">LLM generates answers by looking for specific patterns in the training data. But this doesn't always work, and just as humans can see animals in the shape of clouds, AI bots can find patterns that don't exist. However, while humans know that clouds are just shapes and that there are no giant elephants floating in the sky, LLM may see this as real and \"make up\" new technologies and other false information that doesn't exist.<\/p>\n<p data-vmark=\"3759\"><strong>Researchers at the University of Oxford used the concept of semantic entropy to determine whether the LLM is \"hallucinating\" through probability.<\/strong>Semantic entropy refers to situations where the same word has multiple meanings. Semantic entropy refers to situations where the same word has more than one meaning, e.g., \"desert\" can mean desert or abandonment of someone. When LLM uses such words, there may be confusion about the meaning of the expression.<strong>By detecting semantic entropy, the researchers aim to determine whether the output of the LLM is likely to be \"hallucinatory\".<\/strong><\/p>\n<p data-vmark=\"2d60\">The advantage of utilizing semantic entropy is that it can be used to quickly detect \"hallucination\" problems in LLM without additional supervised or reinforcement learning. Since the method does not rely on task-specific data, it can be applied even when the LLM is faced with a new task that it has never encountered before. This will greatly increase the user's trust in LLM, even when the AI encounters a problem or instruction for the first time.<\/p>\n<p data-vmark=\"02cd\">The research team said, \"Our approach helps users understand when they need to be cautious about the output of LLMs and opens up new horizons for LLM applications that would otherwise be limited by unreliability.\"<\/p>\n<p data-vmark=\"3441\">If semantic entropy proves to be an effective means of detecting \"illusions\", then we can use such tools to double-check the output of AI, making it a more reliable partner. However, IT House would like to remind that, just like humans are not infallible, LLM can still make mistakes even with the most advanced error detection tools. Therefore, it is still wise to always double-check the answers provided by chatbots such as ChatGPT.<\/p>","protected":false},"excerpt":{"rendered":"<p>In recent years, artificial intelligence (AI) has boomed and applications such as chatbots have become popular, allowing people to get information from these chatbots (e.g., ChatGPT) through simple commands. However, these bots are still prone to the problem of \"AI hallucinations,\" where they provide incorrect answers and sometimes even dangerous information. Pexels One of the causes of \"hallucinations\" is inaccurate training data, insufficient generalization capabilities, and side effects of the data collection process. However, researchers at the University of Oxford have taken a different approach, detailing a newly developed method for detecting \"fabrication\" (i.e., arbitrarily generated incorrect information) in Large Language Models (LLMs) in the latest issue of Nature. LLMs detect the problem of \"fabrication\" (i.e., arbitrarily generated incorrect information) by looking for<\/p>","protected":false},"author":1,"featured_media":0,"comment_status":"open","ping_status":"open","sticky":false,"template":"","format":"standard","meta":{"footnotes":""},"categories":[146],"tags":[3042,473,275],"collection":[],"class_list":["post-13891","post","type-post","status-publish","format-standard","hentry","category-news","tag-ai","tag-llm","tag-275"],"acf":[],"_links":{"self":[{"href":"https:\/\/www.1ai.net\/en\/wp-json\/wp\/v2\/posts\/13891","targetHints":{"allow":["GET"]}}],"collection":[{"href":"https:\/\/www.1ai.net\/en\/wp-json\/wp\/v2\/posts"}],"about":[{"href":"https:\/\/www.1ai.net\/en\/wp-json\/wp\/v2\/types\/post"}],"author":[{"embeddable":true,"href":"https:\/\/www.1ai.net\/en\/wp-json\/wp\/v2\/users\/1"}],"replies":[{"embeddable":true,"href":"https:\/\/www.1ai.net\/en\/wp-json\/wp\/v2\/comments?post=13891"}],"version-history":[{"count":0,"href":"https:\/\/www.1ai.net\/en\/wp-json\/wp\/v2\/posts\/13891\/revisions"}],"wp:attachment":[{"href":"https:\/\/www.1ai.net\/en\/wp-json\/wp\/v2\/media?parent=13891"}],"wp:term":[{"taxonomy":"category","embeddable":true,"href":"https:\/\/www.1ai.net\/en\/wp-json\/wp\/v2\/categories?post=13891"},{"taxonomy":"post_tag","embeddable":true,"href":"https:\/\/www.1ai.net\/en\/wp-json\/wp\/v2\/tags?post=13891"},{"taxonomy":"collection","embeddable":true,"href":"https:\/\/www.1ai.net\/en\/wp-json\/wp\/v2\/collection?post=13891"}],"curies":[{"name":"wp","href":"https:\/\/api.w.org\/{rel}","templated":true}]}}