{"id":7955,"date":"2024-04-14T09:13:20","date_gmt":"2024-04-14T01:13:20","guid":{"rendered":"https:\/\/www.1ai.net\/?p=7955"},"modified":"2024-04-14T09:13:20","modified_gmt":"2024-04-14T01:13:20","slug":"%e9%a9%ac%e6%96%af%e5%85%8b-xai-%e5%b1%95%e7%a4%ba%e9%a6%96%e4%b8%aa%e5%a4%9a%e6%a8%a1%e6%80%81%e6%a8%a1%e5%9e%8b-grok-1-5v%ef%bc%9a%e5%8f%af%e5%b0%86%e6%b5%81%e7%a8%8b%e5%9b%be%e8%bd%ac%e6%88%90-pyth","status":"publish","type":"post","link":"https:\/\/www.1ai.net\/en\/7955.html","title":{"rendered":"Musk xAI demonstrates the first multimodal model Grok-1.5V: can convert flowcharts into Python code"},"content":{"rendered":"<p data-vmark=\"d6f8\">Musk&#039;s artificial intelligence company <a href=\"https:\/\/www.1ai.net\/en\/tag\/xai\" title=\"[View articles tagged with [xAI]]\" target=\"_blank\" >xA<\/a> Launched in late March <a href=\"https:\/\/www.1ai.net\/en\/tag\/grok\" title=\"[See articles with [Grok] labels]\" target=\"_blank\" >Grok<\/a>-1.5 After the large language model,<strong>Recently launched the first<a href=\"https:\/\/www.1ai.net\/en\/tag\/%e5%a4%9a%e6%a8%a1%e6%80%81%e6%a8%a1%e5%9e%8b\" title=\"[View articles tagged with [multimodal model]]\" target=\"_blank\" >Multimodal Model<\/a> Grok-1.5 Vision.<\/strong><\/p>\n<p data-vmark=\"ab23\">xAI said it will soon invite early testers and existing Grok users to test Grok-1.5 Vision (Grok-1.5V), which can not only understand text, but also process the content in documents, charts, screenshots and photos.<\/p>\n<p data-vmark=\"fe80\">xAI said: &quot;Grok-1.5V is comparable to existing cutting-edge multimodal models in many areas such as multidisciplinary reasoning, document understanding, scientific graphs, table processing, screenshots and photos.&quot;<\/p>\n<p data-vmark=\"9c89\">In its official press release, xAI demonstrated seven Grok-1.5V cases, including converting flowchart sketches on a whiteboard into Python code, generating bedtime stories based on children&#039;s drawings, interpreting buzzwords, converting tables into CSV file format, and more.<\/p>\n<p data-vmark=\"71bf\"><img loading=\"lazy\" decoding=\"async\" class=\"alignnone size-full wp-image-7956\" title=\"875e0f40-9a63-4b6a-940e-36da59f6d9f6\" src=\"https:\/\/www.1ai.net\/wp-content\/uploads\/2024\/04\/875e0f40-9a63-4b6a-940e-36da59f6d9f6.jpg\" alt=\"875e0f40-9a63-4b6a-940e-36da59f6d9f6\" width=\"1024\" height=\"813\" \/><\/p>\n<p data-vmark=\"1712\">xAI also shared the running results of Grok-1.5V, which outperformed mainstream competitors such as GPT-4V, Claude 3Sonnet, Claude 3 Opus and Gemini Pro 1.5 in the RealWorldQA benchmark.<\/p>","protected":false},"excerpt":{"rendered":"<p>After launching the Grok-1.5 big language model in late March, Musk's AI company xAI recently reintroduced its first multimodal model, Grok-1.5 Vision. xAI says it will soon invite early beta testers and existing Grok users to test Grok-1.5 Vision (Grok-1.5V) to not only understand text, but also to processing content in documents, charts, screenshots, and photos,\" xAI said. \"Grok-1.5V rivals existing cutting-edge multimodal models in a number of areas, including multidisciplinary reasoning, document comprehension, scientific diagrams, table processing, screenshots, and photographs,\" says xAI. In an official press release, xAI demonstrated seven Grok-1.5V<\/p>","protected":false},"author":1,"featured_media":0,"comment_status":"open","ping_status":"open","sticky":false,"template":"","format":"standard","meta":{"footnotes":""},"categories":[146],"tags":[364,356,1096],"collection":[],"class_list":["post-7955","post","type-post","status-publish","format-standard","hentry","category-news","tag-grok","tag-xai","tag-1096"],"acf":[],"_links":{"self":[{"href":"https:\/\/www.1ai.net\/en\/wp-json\/wp\/v2\/posts\/7955","targetHints":{"allow":["GET"]}}],"collection":[{"href":"https:\/\/www.1ai.net\/en\/wp-json\/wp\/v2\/posts"}],"about":[{"href":"https:\/\/www.1ai.net\/en\/wp-json\/wp\/v2\/types\/post"}],"author":[{"embeddable":true,"href":"https:\/\/www.1ai.net\/en\/wp-json\/wp\/v2\/users\/1"}],"replies":[{"embeddable":true,"href":"https:\/\/www.1ai.net\/en\/wp-json\/wp\/v2\/comments?post=7955"}],"version-history":[{"count":0,"href":"https:\/\/www.1ai.net\/en\/wp-json\/wp\/v2\/posts\/7955\/revisions"}],"wp:attachment":[{"href":"https:\/\/www.1ai.net\/en\/wp-json\/wp\/v2\/media?parent=7955"}],"wp:term":[{"taxonomy":"category","embeddable":true,"href":"https:\/\/www.1ai.net\/en\/wp-json\/wp\/v2\/categories?post=7955"},{"taxonomy":"post_tag","embeddable":true,"href":"https:\/\/www.1ai.net\/en\/wp-json\/wp\/v2\/tags?post=7955"},{"taxonomy":"collection","embeddable":true,"href":"https:\/\/www.1ai.net\/en\/wp-json\/wp\/v2\/collection?post=7955"}],"curies":[{"name":"wp","href":"https:\/\/api.w.org\/{rel}","templated":true}]}}