{"id":119243,"date":"2020-10-12T15:01:40","date_gmt":"2020-10-12T15:01:40","guid":{"rendered":"https:\/\/news.microsoft.com\/?p=439481"},"modified":"2020-10-12T15:01:40","modified_gmt":"2020-10-12T15:01:40","slug":"shrinking-the-data-desert-making-ai-systems-more-inclusive-of-people-with-disabilities","status":"publish","type":"post","link":"https:\/\/sickgaming.net\/blog\/2020\/10\/12\/shrinking-the-data-desert-making-ai-systems-more-inclusive-of-people-with-disabilities\/","title":{"rendered":"Shrinking the \u2018data desert\u2019: making AI systems more inclusive of people with disabilities"},"content":{"rendered":"<p>\u201cWe are in a data desert,\u201d said Mary Bellard, principal innovation architect lead at Microsoft who also oversees the AI for Accessibility program. \u201cThere\u2019s a lot of passion and energy around doing really cool things with AI and people with disabilities, but we don\u2019t have enough data.\u201d<\/p>\n<p>\u201cIt\u2019s like we have the car and the car is packed and ready to go, but there\u2019s no gas in it. We don\u2019t have enough data to power these ideas.\u201d<\/p>\n<p>To begin to shrink that data desert, Microsoft researchers have been working for the past year and a half to investigate and suggest ways to make AI systems more inclusive of people with disabilities. The company is also funding and collaborating with AI for Accessibility grantees to create or use more representative training datasets, such as ORBIT and the <a href=\"https:\/\/www.microsoft.com\/en-us\/research\/blog\/microsoft-ability-initiative-a-collaborative-quest-to-innovate-in-image-captioning-for-people-who-are-blind-or-with-low-vision\/\">Microsoft Ability Initiative<\/a> with University of Texas at Austin researchers.<\/p>\n<figure id=\"attachment_82377\" aria-describedby=\"caption-attachment-82377\" class=\"wp-caption alignleft\"><a href=\"https:\/\/www.sickgaming.net\/blog\/wp-content\/uploads\/2020\/10\/shrinking-the-data-desert-making-ai-systems-more-inclusive-of-people-with-disabilities-3.jpg\"><img loading=\"lazy\" decoding=\"async\" class=\"wp-image-82377 size-medium\" src=\"https:\/\/www.sickgaming.net\/blog\/wp-content\/uploads\/2020\/10\/shrinking-the-data-desert-making-ai-systems-more-inclusive-of-people-with-disabilities.jpg\" alt=\"Mary Bellard sits on outdoor steps\" width=\"300\" height=\"300\"><\/a><figcaption id=\"caption-attachment-82377\" class=\"wp-caption-text\">Mary Bellard, principal innovation architect lead at Microsoft who oversees the AI for Accessibility program. Photo provided by Bellard.<\/figcaption><\/figure>\n<p>Today, <a href=\"https:\/\/teamgleason.org\/\">Team Gleason<\/a> announced it is partnering with Microsoft on <a href=\"https:\/\/techcrunch.com\/2020\/10\/12\/microsoft-and-partners-aim-to-shrink-the-data-desert-limiting-accessible-ai\/\">Project Insight<\/a>, which will create an open dataset of facial imagery of people living with ALS to help advance innovation in computer vision and train those AI models more inclusively.<\/p>\n<p>It\u2019s an industry-wide problem that won\u2019t be solved by one project or organization alone, Microsoft says. But new collaborations are beginning to address the issue.<\/p>\n<p>A research roadmap on <a href=\"https:\/\/www.microsoft.com\/en-us\/research\/publication\/toward-fairness-in-ai-for-people-with-disabilities-a-research-roadmap\/\">AI Fairness and Disability<\/a> published by Microsoft Research and a workshop on <a href=\"https:\/\/www.microsoft.com\/en-us\/research\/publication\/disability-bias-and-ai\/\">Disability, Bias and AI<\/a> hosted last year with the AI Now Institute at New York University found a host of potential areas in which mainstream AI algorithms that aren\u2019t trained on inclusive data either don\u2019t work well for people with disabilities or can actively harm them.<\/p>\n<p>If a self-driving car\u2019s pedestrian detection algorithms haven\u2019t been shown examples of people who use wheelchairs or whose posture or gait is different due to advanced age, for example, they may not correctly identify those people as objects to avoid or estimate how much longer they need to safely cross a street, researchers noted.<\/p>\n<p>AI models used in hiring processes that try to read personalities or interpret sentiment from potential job candidates can misread cues and screen out qualified candidates with autism or who emote differently. Algorithms that read handwriting may not be able to cope with examples from people who have Parkinson\u2019s disease or tremors. Gesture recognition systems may be confused by people with amputated limbs or different body shapes.<\/p>\n<p>It\u2019s fairly common for some people with disabilities to be early adopters of intelligent technologies, yet they\u2019ve often not been adequately represented in the data that informs how those systems work, researchers say.<\/p>\n<p>\u201cWhen technologies are so desired by a community, they\u2019re often willing to tolerate a higher rate of errors,\u201d said Meredith Ringel Morris, senior principal researcher who manages the <a href=\"http:\/\/aka.ms\/msrability\">Microsoft Research Ability Team<\/a>. \u201cSo imperfect AI systems still have value, but they could provide so much more and work so much better if they were trained on more inclusive data.\u201d<\/p>\n<h2><strong>\u2018Pushing the state of the art\u2019<\/strong><\/h2>\n<p>Danna Gurari, an AI for Accessibility grantee and assistant professor at the University of Texas at Austin, had that goal in mind when she began developing the <a href=\"https:\/\/vizwiz.org\/\">VizWiz datasets<\/a>. They include tens of thousands of photographs and questions submitted by people who are blind or have low vision to an <a href=\"https:\/\/www.cs.cmu.edu\/~jbigham\/pubs\/pdfs\/2010\/vizwiz.pdf\">app originally developed by researchers at Carnegie Mellon University<\/a>.<\/p>\n<p>The questions run the gamut: What is the expiration date on this milk? What does this shirt say? Do my fingertips look blue? Do these clouds look stormy? Do the charcoal briquettes in this grill look ready? What does the picture on this birthday card look like?<\/p>\n<p>The app originally crowdsourced answers from people across the internet, but Gurari wondered if she could use the data to improve how computer vision algorithms interpret photos taken by people who are blind.<\/p>\n<figure id=\"attachment_82378\" aria-describedby=\"caption-attachment-82378\" class=\"wp-caption alignright\"><a href=\"https:\/\/www.sickgaming.net\/blog\/wp-content\/uploads\/2020\/10\/shrinking-the-data-desert-making-ai-systems-more-inclusive-of-people-with-disabilities-5.jpg\"><img loading=\"lazy\" decoding=\"async\" class=\"wp-image-82378 size-medium\" src=\"https:\/\/www.sickgaming.net\/blog\/wp-content\/uploads\/2020\/10\/shrinking-the-data-desert-making-ai-systems-more-inclusive-of-people-with-disabilities-1.jpg\" alt=\"Danna Guarari stands outside\" width=\"300\" height=\"300\"><\/a><figcaption id=\"caption-attachment-82378\" class=\"wp-caption-text\">AI for Accessibility grantee Danna Gurari, assistant professor at the University of Texas at Austin who developed the VizWiz dataset and directs the School of Information\u2019s Image and Video Computing Group.<\/figcaption><\/figure>\n<p>Many of those questions require reading text, such as determining how much of an over-the-counter medicine is safe to take. Computer vision research has often treated that as a separate problem, for example, from recognizing objects or trying to interpret low-quality photos. But successfully describing real-world photos requires an integrated approach, Gurari said.<\/p>\n<p>Moreover, computer vision algorithms typically learn from large image datasets of pictures downloaded from the internet. Most are taken by sighted people and reflect the photographer\u2019s interest, with items that are centered and in focus.<\/p>\n<p>But an algorithm that\u2019s only been trained on perfect images is likely to perform poorly in describing what\u2019s in a photo taken by a person who is blind; it may be blurry, off center or backlit. And sometimes the thing that person wants to know hinges on a detail that a person who is sighted might not think to label, such as whether a shirt is clean or dirty.<\/p>\n<p>\u201cOften it\u2019s not obvious what is meaningful to people, and that\u2019s why it\u2019s so important not just to design for \u2014 but design these technologies with \u2014 people who are in the blind and low vision community,\u201d said Gurari, who also directs the School of Information\u2019s Image and Video Computing Group at the University of Texas at Austin.<\/p>\n<p>Her team undertook the massive task of cleaning up the original VizWiz dataset to make it usable for training machine learning algorithms \u2014 removing inappropriate images, sourcing new labels, scrubbing personal information and even translating audio questions into text to remove the possibility that someone\u2019s voice could be recognized.<\/p>\n<p>Working with Microsoft funding and researchers, Gurari\u2019s team has developed a <a href=\"https:\/\/vizwiz.org\/tasks-and-datasets\/image-captioning\/\">new public dataset to train, validate and test image captioning algorithms<\/a>. It includes more than 39,000 images taken by blind and low vision participants and five possible captions for each. Her team is also working on algorithms that can recognize right off the bat when an <a href=\"https:\/\/vizwiz.org\/tasks-and-datasets\/image-quality-issues\/\">image someone has submitted is too blurry, obscured or poorly lit<\/a> and suggest how to try again.<\/p>\n<p>Earlier this year, Microsoft sponsored an open challenge to <a href=\"https:\/\/vizwiz.org\/workshops\/2020-workshop\/\">other industry and academic researchers<\/a> to test their image captioning algorithms on the VizWiz dataset. In one common evaluation metric, the top performing algorithm posted a 33% improvement over the prior state of the art.<\/p>\n<p>\u201cThis is really pushing the state of the art in captioning for the blind community forward,\u201d said Seeing AI lead engineer Shaikh, who is working with AI for Accessibility grantees and their datasets to develop potential improvements for the app.<\/p>\n","protected":false},"excerpt":{"rendered":"<p>\u201cWe are in a data desert,\u201d said Mary Bellard, principal innovation architect lead at Microsoft who also oversees the AI for Accessibility program. \u201cThere\u2019s a lot of passion and energy around doing really cool things with AI and people with disabilities, but we don\u2019t have enough data.\u201d \u201cIt\u2019s like we have the car and the [&hellip;]<\/p>\n","protected":false},"author":2,"featured_media":119244,"comment_status":"open","ping_status":"closed","sticky":false,"template":"","format":"standard","meta":{"footnotes":""},"categories":[49],"tags":[63,1102],"class_list":["post-119243","post","type-post","status-publish","format-standard","has-post-thumbnail","hentry","category-microsoft-news","tag-accessibility","tag-innovation-stories"],"_links":{"self":[{"href":"https:\/\/sickgaming.net\/blog\/wp-json\/wp\/v2\/posts\/119243","targetHints":{"allow":["GET"]}}],"collection":[{"href":"https:\/\/sickgaming.net\/blog\/wp-json\/wp\/v2\/posts"}],"about":[{"href":"https:\/\/sickgaming.net\/blog\/wp-json\/wp\/v2\/types\/post"}],"author":[{"embeddable":true,"href":"https:\/\/sickgaming.net\/blog\/wp-json\/wp\/v2\/users\/2"}],"replies":[{"embeddable":true,"href":"https:\/\/sickgaming.net\/blog\/wp-json\/wp\/v2\/comments?post=119243"}],"version-history":[{"count":0,"href":"https:\/\/sickgaming.net\/blog\/wp-json\/wp\/v2\/posts\/119243\/revisions"}],"wp:featuredmedia":[{"embeddable":true,"href":"https:\/\/sickgaming.net\/blog\/wp-json\/wp\/v2\/media\/119244"}],"wp:attachment":[{"href":"https:\/\/sickgaming.net\/blog\/wp-json\/wp\/v2\/media?parent=119243"}],"wp:term":[{"taxonomy":"category","embeddable":true,"href":"https:\/\/sickgaming.net\/blog\/wp-json\/wp\/v2\/categories?post=119243"},{"taxonomy":"post_tag","embeddable":true,"href":"https:\/\/sickgaming.net\/blog\/wp-json\/wp\/v2\/tags?post=119243"}],"curies":[{"name":"wp","href":"https:\/\/api.w.org\/{rel}","templated":true}]}}