{"id":2042,"date":"2018-01-15T07:47:13","date_gmt":"2018-01-15T07:47:13","guid":{"rendered":"http:\/\/labs.webfoundation.org\/id\/how-open-data-can-save-ai\/"},"modified":"2018-01-22T07:48:51","modified_gmt":"2018-01-22T07:48:51","slug":"how-open-data-can-save-ai","status":"publish","type":"post","link":"https:\/\/labs.webfoundation.org\/id\/how-open-data-can-save-ai\/","title":{"rendered":"How open data can save AI"},"content":{"rendered":"<p><img loading=\"lazy\" decoding=\"async\" class=\"alignnone size-full wp-image-2041\" src=\"http:\/\/labs.webfoundation.org\/wp-content\/uploads\/2018\/01\/Traffic_AI.jpg\" alt=\"\" width=\"1200\" height=\"675\" srcset=\"https:\/\/labs.webfoundation.org\/wp-content\/uploads\/2018\/01\/Traffic_AI.jpg 1200w, https:\/\/labs.webfoundation.org\/wp-content\/uploads\/2018\/01\/Traffic_AI-300x169.jpg 300w, https:\/\/labs.webfoundation.org\/wp-content\/uploads\/2018\/01\/Traffic_AI-768x432.jpg 768w, https:\/\/labs.webfoundation.org\/wp-content\/uploads\/2018\/01\/Traffic_AI-1024x576.jpg 1024w\" sizes=\"(max-width: 1200px) 100vw, 1200px\" \/><\/p>\n<p><i>This post originally appears on the <a href=\"https:\/\/webfoundation.org\/2018\/01\/how-open-data-can-save-ai\/\">Web Foundation website<\/a>\u00a0and was written by Web Foundation Policy Fellow Juan Ortiz Freuler &amp; Policy Director Craig Fagan.<\/i><\/p>\n<p>Artificial intelligence (AI) is reshaping our world \u2014 changing our economies (think Uber and driverless cars), our politics (think Cambridge Analytica and\u00a0<a href=\"https:\/\/www.newscientist.com\/article\/2142072-how-to-turn-facebook-into-a-weaponised-ai-propaganda-machine\/\">targeted political adverts<\/a>), and our societies (think\u00a0<a href=\"https:\/\/www.theguardian.com\/science\/2017\/nov\/03\/from-health-to-crimefighting-ai-has-brought-us-to-the-threshold-of-a-new-era\">predictive policing<\/a>). While AI systems offer a host of benefits, the risks cannot be ignored. Two recent\u00a0<a href=\"https:\/\/webfoundation.org\/research\/white-paper-series-opportunities-and-risks-in-emerging-technologies\/\">Web Foundation studies<\/a>\u00a0highlight both the risks and opportunities, looking at how\u00a0<a href=\"https:\/\/webfoundation.org\/docs\/2017\/07\/AI_Report_WF.pdf\">AI<\/a>\u00a0and\u00a0<a href=\"http:\/\/webfoundation.org\/docs\/2017\/07\/Algorithms_Report_WF.pdf\">algorithms<\/a>\u00a0are impacting low and middle-income countries.<\/p>\n<p>AI and algorithms require the same basic fuel to work: data. AI systems acquire their\u00a0<i>intelligence<\/i>\u00a0by using data to understand how past problems were solved, and applying this learning to decision-making. The availability of quality data is crucial for AI systems to learn and function effectively.<\/p>\n<h3><b>Data is the fuel that drives AI<\/b><\/h3>\n<p>Though AI has been around for decades, the discipline is going through a\u00a0<a href=\"https:\/\/en.wikipedia.org\/wiki\/AI_winter\">new spring<\/a>, with a surge of research and development triggered by a combination of greater data availability and lower-cost, more powerful computers. While existing data is being digitised, falling costs allow chips to be inserted into everyday objects \u2014 from exercise bands to public transport \u2014 generating entirely new sources of data. Giving a sense of the exponential growth in volume and varieties of data, some estimate that 90% of all the data in the world has been\u00a0<a href=\"https:\/\/www.sciencedaily.com\/releases\/2013\/05\/130522085217.htm\">created in the last two years<\/a>.<\/p>\n<h3><b>Garbage in, garbage out<\/b><\/h3>\n<p>But the data that is being produced and fed to AI systems is frequently incomplete, biased, or otherwise of poor quality. Unless AI algorithms are designed to be aware that they are being fed imperfect data and take steps to account for this, bad data will undermine the learning process and skew the outputs of AI systems. This can lead to devastating and often discriminatory outcomes, particularly as AI is increasingly used to make decisions in important areas of our lives.<\/p>\n<p>Better data is one way to address these risks. Offering data under an\u00a0<a href=\"https:\/\/opendatacharter.net\/resource\/definition-key-terms-charter-principles\/\">open license<\/a>\u00a0(i.e., one that allows data to be freely used, modified, and shared by anyone for any purpose) is an effective strategy towards achieving quality.<\/p>\n<p>Through our\u00a0<a href=\"http:\/\/opendatabarometer.org\/?_year=2016&amp;indicator=ODB\">Open Data Barometer<\/a>\u00a0\u2014 which assesses the openness and quality of government data \u2014 we are continuously monitoring whether government datasets are being proactively disclosed in key sectors, including education, health, the environment, and public expenditure. The latest edition suggests that most government datasets \u2014 more than in 9 in 10 \u2014 are still not open.<\/p>\n<h3><b>Open data to improve AI<\/b><\/h3>\n<p>One way to make more data available and to improve data quality is to push governments that use algorithms and AI systems for public service delivery to open up the data upon which these systems rely. All non-personally identifiable datasets used should be released in open formats. When datasets are considered too sensitive for release, appropriate metadata should be provided. Opening key datasets will help identify potential biases, lead to more competition between potential service providers, ensure better public services, and increase citizen trust in government.<\/p>\n<p>As governments adopt algorithms and AI systems to improve service delivery, we should take steps to ensure this is done in a transparent way that reassures citizens that these systems will produce fair outcomes, as well as higher quality services. Making the underlying data available is a first step towards public\u00a0<a href=\"https:\/\/poseidon01.ssrn.com\/delivery.php?ID=728119126066065068071126077003092022025045004018028059126002112101121117112011116086101020111102045051044087096072101083091115053087047021061120070103118025003080003089005041124026089074114001117006018097109028075088011077092023098118068030104071068078&amp;EXT=pdf\">understanding<\/a>\u00a0of how public service AI systems make decisions.<\/p>\n<p>If AI systems are to reshape the world, we must have the opportunity to shape them. And a wider group of people must have access to datasets necessary to build AI technologies. The Web Foundation will continue to explore these issues and how to tackle them.<\/p>\n<hr \/>\n<p><b>If you are interested in how governments use AI, algorithms or statistical models, please\u00a0<\/b><a href=\"https:\/\/docs.google.com\/forms\/d\/e\/1FAIpQLScsvOCysgI_jIEGH8472HCvEkGJUE6ykG8SVJ0EwxdJFbVrPw\/viewform\"><b>complete our 4-question survey<\/b><\/a><b>\u00a0(also available in\u00a0<\/b><a href=\"https:\/\/docs.google.com\/forms\/d\/e\/1FAIpQLSdnh8B9F1EVhBJyYnUu_AL6TAw9Tmx0liS0XVnpIq4yKDYOqQ\/viewform\"><b>Spanish<\/b><\/a><b>)\u00a0<\/b><b>to share your expertise and help shape our research on governments\u2019 use of data and algorithms.<\/b><\/p>\n<p>&nbsp;<\/p>\n<p><i>The Web Foundation recently\u00a0<\/i><a href=\"https:\/\/webfoundation.org\/research\/artificial-intelligence-starting-the-policy-dialogue-in-africa\/\"><i>published research<\/i><\/a><i>\u00a0mapping academic and commercial AI activity underway in various African countries. For updates on our research, follow us on Twitter at\u00a0<\/i><a href=\"http:\/\/twitter.com\/webfoundation\"><i>@webfoundation<\/i><\/a><i>\u00a0and sign up to our\u00a0<\/i><a href=\"https:\/\/webfoundation.us8.list-manage.com\/subscribe\/post?u=b3c8e6e91fe1905e99f8b59fd&amp;id=8516b77d8a\"><i>email newsletters<\/i><\/a><i>.<\/i><\/p>\n","protected":false},"excerpt":{"rendered":"<p>This post originally appears on the Web Foundation website\u00a0and was written by Web Foundation Policy Fellow Juan Ortiz Freuler &amp; Policy Director Craig Fagan. Artificial intelligence (AI) is reshaping our world \u2014 changing our economies (think Uber and driverless cars), our politics (think Cambridge Analytica and\u00a0targeted political adverts), and our societies (think\u00a0predictive policing). While AI [&hellip;]<\/p>\n","protected":false},"author":16,"featured_media":2041,"comment_status":"open","ping_status":"open","sticky":false,"template":"","format":"standard","meta":{"footnotes":""},"categories":[179,175,182],"tags":[605,606,607],"class_list":{"0":"post-2042","1":"post","2":"type-post","3":"status-publish","4":"format-standard","5":"has-post-thumbnail","7":"category-blog-id","8":"category-uncategorized-id","9":"category-web-foundation-id","10":"tag-ai-id","11":"tag-algorithms-id","12":"tag-shaping-id","13":"entry"},"_links":{"self":[{"href":"https:\/\/labs.webfoundation.org\/id\/wp-json\/wp\/v2\/posts\/2042"}],"collection":[{"href":"https:\/\/labs.webfoundation.org\/id\/wp-json\/wp\/v2\/posts"}],"about":[{"href":"https:\/\/labs.webfoundation.org\/id\/wp-json\/wp\/v2\/types\/post"}],"author":[{"embeddable":true,"href":"https:\/\/labs.webfoundation.org\/id\/wp-json\/wp\/v2\/users\/16"}],"replies":[{"embeddable":true,"href":"https:\/\/labs.webfoundation.org\/id\/wp-json\/wp\/v2\/comments?post=2042"}],"version-history":[{"count":0,"href":"https:\/\/labs.webfoundation.org\/id\/wp-json\/wp\/v2\/posts\/2042\/revisions"}],"wp:featuredmedia":[{"embeddable":true,"href":"https:\/\/labs.webfoundation.org\/id\/wp-json\/wp\/v2\/media\/2041"}],"wp:attachment":[{"href":"https:\/\/labs.webfoundation.org\/id\/wp-json\/wp\/v2\/media?parent=2042"}],"wp:term":[{"taxonomy":"category","embeddable":true,"href":"https:\/\/labs.webfoundation.org\/id\/wp-json\/wp\/v2\/categories?post=2042"},{"taxonomy":"post_tag","embeddable":true,"href":"https:\/\/labs.webfoundation.org\/id\/wp-json\/wp\/v2\/tags?post=2042"}],"curies":[{"name":"wp","href":"https:\/\/api.w.org\/{rel}","templated":true}]}}