{"id":13481,"date":"2024-04-07T13:31:20","date_gmt":"2024-04-07T17:31:20","guid":{"rendered":"https:\/\/www.predictiveanalyticsworld.com\/machinelearningtimes\/?p=13481"},"modified":"2024-05-04T11:50:15","modified_gmt":"2024-05-04T15:50:15","slug":"apple-researchers-develop-ai-that-can-see-and-understand-screen-context","status":"publish","type":"post","link":"https:\/\/www.predictiveanalyticsworld.com\/machinelearningtimes\/apple-researchers-develop-ai-that-can-see-and-understand-screen-context\/13481\/","title":{"rendered":"Apple researchers develop AI that can \u2018see\u2019 and understand screen context"},"content":{"rendered":"Originally published in Venture Beat, April 1, 2024 Apple researchers have developed a new artificial intelligence system that can understand ambiguous references to on-screen entities as well as conversational and background context, enabling more natural interactions with voice assistants, according to a paper published on Friday. The system, called ReALM (Reference Resolution As Language Modeling), leverages large language models to convert the complex task of reference resolution \u2014 including understanding references to visual elements on a screen \u2014 into a pure language modeling problem. This allows ReALM to achieve substantial performance gains compared to existing methods. To continue <a href=\"https:\/\/www.predictiveanalyticsworld.com\/machinelearningtimes\/apple-researchers-develop-ai-that-can-see-and-understand-screen-context\/13481\/\" class=\"more-link\">(more&hellip;)<\/a>","protected":false},"excerpt":{"rendered":"<p>Originally published in Venture Beat, April 1, 2024 Apple researchers have developed a new artificial intelligence system that can understand ambiguous references to on-screen entities as well as conversational and background context, enabling more natural interactions with voice assistants, according to a paper published on Friday. The system, called ReALM (Reference Resolution As Language Modeling), [&hellip;]<\/p>\n","protected":false},"author":78,"featured_media":0,"comment_status":"open","ping_status":"open","sticky":false,"template":"","format":"standard","meta":{"footnotes":"","_links_to":"","_links_to_target":""},"categories":[11,48],"tags":[],"class_list":["post-13481","post","type-post","status-publish","format-standard","hentry","category-industry-news","category-left-hand"],"_links":{"self":[{"href":"https:\/\/www.predictiveanalyticsworld.com\/machinelearningtimes\/wp-json\/wp\/v2\/posts\/13481","targetHints":{"allow":["GET"]}}],"collection":[{"href":"https:\/\/www.predictiveanalyticsworld.com\/machinelearningtimes\/wp-json\/wp\/v2\/posts"}],"about":[{"href":"https:\/\/www.predictiveanalyticsworld.com\/machinelearningtimes\/wp-json\/wp\/v2\/types\/post"}],"author":[{"embeddable":true,"href":"https:\/\/www.predictiveanalyticsworld.com\/machinelearningtimes\/wp-json\/wp\/v2\/users\/78"}],"replies":[{"embeddable":true,"href":"https:\/\/www.predictiveanalyticsworld.com\/machinelearningtimes\/wp-json\/wp\/v2\/comments?post=13481"}],"version-history":[{"count":1,"href":"https:\/\/www.predictiveanalyticsworld.com\/machinelearningtimes\/wp-json\/wp\/v2\/posts\/13481\/revisions"}],"predecessor-version":[{"id":13482,"href":"https:\/\/www.predictiveanalyticsworld.com\/machinelearningtimes\/wp-json\/wp\/v2\/posts\/13481\/revisions\/13482"}],"wp:attachment":[{"href":"https:\/\/www.predictiveanalyticsworld.com\/machinelearningtimes\/wp-json\/wp\/v2\/media?parent=13481"}],"wp:term":[{"taxonomy":"category","embeddable":true,"href":"https:\/\/www.predictiveanalyticsworld.com\/machinelearningtimes\/wp-json\/wp\/v2\/categories?post=13481"},{"taxonomy":"post_tag","embeddable":true,"href":"https:\/\/www.predictiveanalyticsworld.com\/machinelearningtimes\/wp-json\/wp\/v2\/tags?post=13481"}],"curies":[{"name":"wp","href":"https:\/\/api.w.org\/{rel}","templated":true}]}}