{"id":40470,"date":"2014-03-03T12:20:29","date_gmt":"2014-03-03T12:20:29","guid":{"rendered":"http:\/\/www.transcend.org\/tms\/?p=40470"},"modified":"2015-05-05T22:11:01","modified_gmt":"2015-05-05T21:11:01","slug":"gigabytes-gone-wild","status":"publish","type":"post","link":"https:\/\/www.transcend.org\/tms\/2014\/03\/gigabytes-gone-wild\/","title":{"rendered":"Gigabytes Gone Wild"},"content":{"rendered":"<p><i>Big data has outpaced our legal system&#8217;s ability to control it \u2014 we need a new ethics for a new digital age.<\/i><\/p>\n<p>These days, everyone seems to be talking about \u201cbig data.\u201d Engineers, researchers, lawyers, executives and self-trackers all tout the surprising insights they can get from applying math to large data sets. The rhetoric of big data is often <a href=\"http:\/\/papers.ssrn.com\/sol3\/papers.cfm?abstract_id=2325537\"  target=\"_blank\">overblown, exaggerated and contradictory<\/a>, but there\u2019s an element of truth to the claim that data science is helping us to know more about our world, our society and ourselves.<\/p>\n<p>Data scientists use big data to deliver personalized ads to Internet users, to make better spell checkers and search engines, to predict weather patterns, perform medical research, learn about customers, set prices and plan traffic flow patterns. Big data can also fight crime, whether through the use of automated license-plate readers or, at least theoretically, through the collection of vast amounts of \u201cmetadata\u201d about our communications and associations by the National Security Agency.<\/p>\n<p>Big data allows us to know more, to predict and to influence others. This is its power, but it\u2019s also its danger. The entities that can harness the power of math applied to large sets of personal information can do things that used to be impossible. Many of these new uses are good, but some of them aren\u2019t. For example, if our \u201cpersonalized prices\u201d can be based on our race or sex, or if our college admissions are based on things like ZIP code or car ownership, we might want to think more deeply about the kinds of big decisions our big data can be used for. We\u2019re creating a society based on data, and we need to make sure that we create a society that we want to live in.<\/p>\n<p>The values we build or fail to build into our new digital structures will define us. If we don\u2019t balance the human values that we care about \u2014 such as privacy, confidentiality, transparency, identity and free choice \u2014 with the compelling uses of big data, our society risks abandoning them for the sake of mere innovation or expediency.<\/p>\n<p>We think the answer lies in a conversation about the ethics of big data. What should we allow it to do for us, and why? Big data has allowed the impossible to become possible, and it has outpaced our legal system\u2019s ability to control it. This is understandable, as our elected officials don\u2019t pass laws to regulate things that aren\u2019t possible. We need to talk about big data ethics, and we think four facts should guide our discussion.<\/p>\n<p><b>Big data ethics<\/b><\/p>\n<p>First, when we talk about decisions based upon personal data, we need to realize that privacy rules are necessary. Some people might argue that privacy is dead in an age of information, but nothing could be further from the truth. Privacy isn\u2019t just about keeping things hidden, it\u2019s about the rules we use to govern information. Look at the \u201cprivacy policies\u201d of even big data companies \u2014 these tell you not just what information gets collected about you, but how it is used and when it can be destroyed.<\/p>\n<p>Second, we need to realize that even shared personal information can be protected. When you go to see doctors or lawyers, you don\u2019t expect that the information you give them is theirs to use any way they want. The information is confidential: We confide in them so they can help us, and it\u2019s the promise of confidentiality that lets us trust them enough to tell them everything they need to know, even if it\u2019s embarrassing or sensitive. This essential trust is backed up by laws as well as professional rules of ethics. We don\u2019t think of this information as \u201cpublic\u201d or \u201cnonprivate,\u201d and we can think about much of the data gathered about us the same way, whether it\u2019s the websites we visit, the books we read or the places we go that our digital devices track automatically. Amazon, Apple or our ISP or mobile phone carrier might need to know this information to help us go about our days, but that doesn\u2019t mean this data is \u201cpublic\u201d or that it should be beyond our control.<\/p>\n<p>If we\u2019re constantly sorted and nudged by big-data-based decisions, we risk letting the powerful entities in our lives determine who we are before we even know ourselves.<\/p>\n<p>Third, big data requires transparency. If important decisions are being made about us based on an algorithm and data, we have a right to know how the algorithm works and what data is being used. It\u2019s outrageous that while big data has allegedly eliminated privacy, many of the ways it\u2019s used are themselves shrouded in secrecy. This has things entirely the wrong way around. If we\u2019re to build a society through decisions based upon data, we should know how they work, especially when those decisions will affect our daily lives, privacy and social opportunities.<\/p>\n<p>Finally, we should recognize that big data can compromise identity, our right to decide who we are. If we\u2019re constantly sorted and nudged by big-data-based decisions in areas from our choice of books to our voting habits, we risk letting the powerful entities in our lives determine who we are before we even know ourselves. We need to think imaginatively about the kinds of data inferences and data decisions we will allow. We must regulate or prohibit ones we find corrosive, threatening or offensive, just as we\u2019ve long protected decisions surrounding voting and contraception and prohibited invidious decisions made upon criteria such as race, sex or gender.<\/p>\n<p><b>A new framework<\/b><\/p>\n<p>How should we make sure that big data ethics gets built into our digital future? Law should certainly be part of the answer, and despite the claims of some technologists, law can work here. For example, the federal Fair Credit Reporting Act effectively regulates the credit reporting agencies\u2019 use of big data to generate consumer credit reports and calculate consumer credit scores. In fact, the FCRA has <a href=\"http:\/\/www.futureofprivacy.org\/wp-content\/uploads\/LEGAL-Hoofnagle-How-FCRA-Regulates-Big-Data.pdf\"  target=\"_blank\">regulated<\/a> growing uses of big data in this context since 1970. (Some kinds of big data are really old.) As big data\u2019s analytical tools become more common in our society, we should extend similar legal protections to other essential areas as well.<\/p>\n<p>But law alone cannot solve these problems. As a society, we need to talk about big data ethics. Are we comfortable using race or proxies for race to price goods or allocate government benefits such as school funding or welfare payments? What about using big data inferences to decide college admissions or lawsuits, to investigate crimes or impose criminal sentences? As scholars, we certainly have our own moral views on these questions (as do many data scientists), but if we\u2019re building a society in which data science is deployed more often, we need to talk as a society about what we will allow and what we won\u2019t. In this respect, the White House\u2019s <a href=\"http:\/\/www.whitehouse.gov\/blog\/2012\/03\/29\/big-data-big-deal\"  target=\"_blank\">initiative<\/a> to study the technological, legal and ethical implications of big data is a good first step. But we need to do more.<\/p>\n<p>We need to establish social norms for the use of data to make decisions about people, and for the rights that people have for understanding and disputing those decisions, just as we established norms for safe working conditions in the wake of the Industrial Revolution and norms for the allocation of government services and benefits at the dawn of the welfare state. When we do this, software designers and engineers need to be at the center of the conversation. Individual users certainly have responsibility to behave responsibly when their data is at stake, but users alone can\u2019t bear the whole burden. We need to build structures that encourage ethical data usage rather than merely incentivizing individual consumers into sharing as much as possible for as little as possible in return.<\/p>\n<p>We must build these structures, such as in-house ethicists or review boards, into government and private entities that use big data. Such proposals might seem far-fetched, but they are already starting to become widespread. For decades, university scientists wishing to perform experiments (whether physical or data-based) on human subjects have had to submit their research projects to <a href=\"http:\/\/www.stanfordlawreview.org\/online\/privacy-and-big-data\/consumer-subject-review-boards\"  target=\"_blank\">institutional review boards<\/a>, in-house panels that ensure that scientific tools are deployed ethically and for the benefit of human beings. And many leading corporations have started to take steps along these lines, such as the widespread growth of chief privacy officers as senior corporate executives or experiments such as Google\u2019s ethical review board or ethicists-in-residence. If we\u2019re building a data-science revolution, let\u2019s make sure it\u2019s a revolution we want \u2014 one that makes society better as well as making companies richer.<\/p>\n<p>Big data ethics first begins as a state of mind, before it becomes a set of mandates. While engineers in particular must embrace the idea of big data ethics, in an information society that cares about privacy, confidentiality, transparency and identity, we must all be part of the conversation, and part of the solution. Big data ethics is for everyone.<\/p>\n<p>__________________________<\/p>\n<p><i>Neil Richards is a law professor at Washington University in St. Louis and an internationally recognized expert in privacy and technology law. He is the author of \u201c<a href=\"http:\/\/papers.ssrn.com\/sol3\/papers.cfm?abstract_id=2239412\"  target=\"_blank\">The Dangers of Surveillance<\/a>\u201d and \u201cIntellectual Privacy\u201d (Oxford Press, forthcoming 2014). \u00a0<\/i><\/p>\n<p><i>Jonathan King is the vice president of cloud strategy and business development for <a href=\"http:\/\/www.centurylinkcloud.com\/\"  target=\"_blank\">CenturyLink Cloud<\/a>. He is the co-author of\u00a0<a href=\"http:\/\/papers.ssrn.com\/sol3\/papers.cfm?abstract_id=2325537\"  target=\"_blank\">Three Paradoxes of Big Data<\/a> and <a href=\"http:\/\/papers.ssrn.com\/sol3\/papers.cfm?abstract_id=2384174\"  target=\"_blank\">Big Data Ethics<\/a> (Wake Forest Law Review, forthcoming 2014).<\/i><\/p>\n<p><i>This op-ed is adapted from &#8220;<a target=\"_blank\" href=\"http:\/\/papers.ssrn.com\/sol3\/papers.cfm?abstract_id=2384174\" >Big Data Ethics<\/a>,&#8221; an essay by Neil M. Richards and Jonathan H. King in the Wake Forest Law Review (forthcoming 2014). You can access the paper online <a target=\"_blank\" href=\"http:\/\/papers.ssrn.com\/sol3\/papers.cfm?abstract_id=2384174\" >here<\/a>.<\/i><\/p>\n<p><a target=\"_blank\" href=\"http:\/\/america.aljazeera.com\/opinions\/2014\/2\/gigabytes-gone-wild.html\" >Go to Original \u2013 aljazeera.com<\/a><\/p>\n","protected":false},"excerpt":{"rendered":"<p>Big data has outpaced our legal system&#8217;s ability to control it \u2014 we need a new ethics for a new digital age.<\/p>\n","protected":false},"author":4,"featured_media":0,"comment_status":"open","ping_status":"open","sticky":false,"template":"","format":"standard","meta":{"footnotes":""},"categories":[62,60],"tags":[],"class_list":["post-40470","post","type-post","status-publish","format-standard","hentry","category-media","category-whistleblowing-surveillance"],"_links":{"self":[{"href":"https:\/\/www.transcend.org\/tms\/wp-json\/wp\/v2\/posts\/40470","targetHints":{"allow":["GET"]}}],"collection":[{"href":"https:\/\/www.transcend.org\/tms\/wp-json\/wp\/v2\/posts"}],"about":[{"href":"https:\/\/www.transcend.org\/tms\/wp-json\/wp\/v2\/types\/post"}],"author":[{"embeddable":true,"href":"https:\/\/www.transcend.org\/tms\/wp-json\/wp\/v2\/users\/4"}],"replies":[{"embeddable":true,"href":"https:\/\/www.transcend.org\/tms\/wp-json\/wp\/v2\/comments?post=40470"}],"version-history":[{"count":0,"href":"https:\/\/www.transcend.org\/tms\/wp-json\/wp\/v2\/posts\/40470\/revisions"}],"wp:attachment":[{"href":"https:\/\/www.transcend.org\/tms\/wp-json\/wp\/v2\/media?parent=40470"}],"wp:term":[{"taxonomy":"category","embeddable":true,"href":"https:\/\/www.transcend.org\/tms\/wp-json\/wp\/v2\/categories?post=40470"},{"taxonomy":"post_tag","embeddable":true,"href":"https:\/\/www.transcend.org\/tms\/wp-json\/wp\/v2\/tags?post=40470"}],"curies":[{"name":"wp","href":"https:\/\/api.w.org\/{rel}","templated":true}]}}