The Devil Is in the Defaults Ian Kerr*
Total Page:16
File Type:pdf, Size:1020Kb
The Devil Is in the Defaults Ian Kerr* Abstract This review essay explores the notion of “shifting defaults” as discussed by Mireille Hil- debrandt in her book Smart Technologies and the End(s) of Law. Although readers might mistakenly see defaults as a fleeting topic within Hildebrandt’s text, this review essay demonstrates their central significance, identifying and elaborating on fundamental char- acteristics of four categories of shifting defaults: (i) natural defaults; (ii) technological defaults; (iii) legal defaults; and (iv) normative defaults. The essay concludes by reframing Hildebrandt’s central question, considering whether the Rule of Law is itself merely our current legal default, and whether it could be legitimately displaced by smart technologies. I. Introduction Although it will surely cause discomfort to some, very few members of an increasingly tech-savvy legal profession would be surprised by claims that advances in robotics and artificial intelligence will revolutionize the delivery of legal services and transform the practice of law.1 But here is a truly disquieting thought: what if “modern law is dependent on and shares the affordances of the printing press . and, ultimately . under the next dominant information and communications infrastructure, law as we know it might be gone”?2 In her perspicacious new book, Smart Technologies and the End(s) of Law, Mireille Hildebrandt, philosopher, lawyer, and internationally renowned expert in intellectual mac- ramé, examines this and other core questions at the nexus of philosophy of technology and philosophy of law. In a stunning display of rigor and discipline, Hildebrandt weaves together a dazzling array of key sources from law, philosophy, computer science, psychol- ogy, sociology, anthropology, science and technology studies, and other fields, exercising some of the very best lateral thinking I have seen in the past decade. * Canada Research Chair in Ethics, Law and Technology, University of Ottawa, Faculty of Law. Contact: [email protected]. I would like to thank Dustin Moores for his tireless enthusiasm, his ability to find anything under the sun and for the high quality research assistance that he so regularly and reliably provides. Thanks also to Ryan Calo, Woodrow Hartzog and Jason Millar for insightful feedback on an earlier draft. Finally, thank you to Mireille Hildebrandt for her superb and inspiring scholarship, her many intellectual contributions to our community, and for the honor of including me to participate in this book forum with such eminent scholars in the field. 1 Richard Susskind & Daniel Susskind, The Future of the Professions: How Technology Will Transform the Work of Human Experts (2015); Julie Sobowale, How Artificial Intelligence is Transforming the Legal Profession, ABA J. (Apr. 1, 2016), http://www.abajournal.com/magazine/article/how_artificial_intelligence_is_ transforming_the_legal_profession. 2 Mireille Hildebrandt, Smart Technologies and the End(s) of Law 176 (2015). ISSN 2291-9732 92 Critical Analysis of Law 4:1 (2017) Commencing with a compelling day-in-the-life narrative, she illustrates how smart technologies will catalyze the coming shift from life online to an “onlife world”—“a trans- formative life world, situated beyond the increasingly artificial distinction between online and offline . thriv[ing] on artificial life forms, or at least on ‘things’ that seem to develop a life of their own” (8). As Hildebrandt goes on to explain, through autonomic computing and emerging techniques in machine learning, these “data-driven agents” will generate the capability to profile us, and to predict and preempt many of our day-to-day activities. Without our help (often, without our awareness), they will anticipate and serve our prefer- ences and, in some instances, reconfigure them. Imagining, as she does, a world in which we have delegated many of our daily decisions to machines, devices that we come to rely upon to anticipate our needs and accomplish our goals, Hildebrandt warns that smart technologies have the potential to threaten fundamental rights including: privacy, identity, autonomy, non-discrimination, due process, and the presumption of innocence. In 226 pages of erudite, urbane investigation, Hildebrandt offers an extremely so- bering synthesis that details the character of the threats that smart technologies pose to fundamental rights and to the end(s) of law. One intriguing thing about her book, howev- er, is that despite these threats, she believes that “[t]here is no way back, we can only move forward” (214). With no turning back Hildebrandt instead forewarns that the only thing able to save the Rule of Law as we know it is “legal protection by design.” These legal protections are not meant simply to forbid preemptive computing when it interferes with fundamental rights.3 Rather, Hildebrandt follows a well-entrenched regulate-by- design approach, recommending two mechanisms for legal reform as examples of how the legal protection by design method can be baked directly into the emerging infor- mation and communications infrastructure. These and other such mechanisms are meant to ensure that not only our fundamental rights, but also the very fabric of modern law, remain intact. To whet appetites, I will quickly mention her two primary examples, both of which are too complicated to fully develop in my brief description. After that, I will turn my focus to a recurring theme in the book: the shifting defaults in our technological and legal architectures. First, Hildebrandt recommends the creation of a legal obligation to provide the socio-technical infrastructure necessary to enshrine “counter-profiling” as a fundamental right. Performing an analogous role to “freedom of expression” in our inherited legal in- frastructure (which she argues was constructed as a result of the affordances of the 4 printing press), the right to counter-profile would entail that the objects of profiling be- 3 This is an approach that I had suggested in an earlier volume co-edited by Hildebrandt. See Ian Kerr, Prediction, Preemption, Presumption: The Path of Law After the Computational Turn, in Privacy, Due Process and the Computational Turn (Mireille Hildebrandt & Katja de Vries eds., 2013); see also Ian Kerr & Jessica Earle, Prediction, Preemption, Presumption: How Big Data Threatens Big Picture Privacy, 66 Stan. L. Rev. Online 65 (2013). 4 As Hildebrandt puts it: “In many ways, counter-profiling has the same function as the free press had in the era of books and newspapers. From the perspective of the Rule of Law, the right to freedom of information has horizontal effect and may include a positive obligation on the state to ensure that the right is effective” (223). Kerr — The Devil is in the Defaults 93 come agents of profiling, rendering them capable of conducting data mining operations on the behaviors of those that are in the business of profiling. It is crucial to understand, however, that this prescription is far more than merely sanctioning audits or permitting counter-profiling as an activity. It is a much bigger deal than that. Hildebrandt is recom- mending a legal requirement that we build a technological infrastructure that would allow everyone the continuous ability to watch the watchers. Such a change to our infrastructure would be a serious and contentious commitment to transparency. Second, Hildebrandt advocates the adoption of “morphological computing” as a means of diminishing dependence on central control models of intelligence. “Morphological computing refers to the embedded nature of cognition, which enables granular adaptive re- sponses by body parts based on the morphology of that particular part instead of being initiated by a control mechanism like the brain” (29). Although it is unclear whether this kind of computation would likewise become an actual legal requirement, to the extent that more morphological computation means less data-driven computation, dictating the use of such techniques accords with the prime directive of “big data space”: namely, data minimization.5 One of the things I admire about Hildebrandt’s book is the way in which it moves beyond Lessig’s famous watchword that “code is law.”6 Following on Joel Reidenberg’s classic work on the formulation of information policy rules through technology,7 Lessig was among the first technology lawyers to recognize as dangerous the strategy of using preemptive technologies as a means of regulating conduct.8 As Hildebrandt depicts the strategy, “If the idea is to influence behavior instead of addressing action, and if the means are interchangeable, why not opt for a means that is less visible, less contestable and thus more influential in getting people to behave in alignment with policy objec- tives?” (165). Hildebrandt opposes this strategy and provides several successful counter- examples throughout the book. However, Hildebrandt’s disquisition is not simply about the perils of such techno-regulation or how best to address them. In addition, Hilde- brandt calmly, meticulously, and persuasively offers a philosophically rich account that explains how smart technologies might alter the modality of law’s existence. Her book isn’t just about how to debug legal code in order to hold techno-regulation in check. Alt- hough she does offer proposals in that regard, her central purpose is to provide an ontological account of the nature of law through an examination of its philosophical en- tanglements with technology. It is a tall order and she rises to the occasion. Luddite jurisprudes, take notice: your understanding of the philosophy of law is impoverished and 5 According to Hildebrandt, morphological computing permits many of our transactions with data-driven agents to escape the storage and searchability of human behaviors, enabling smooth and safe interaction with humans based on the physical design of robots, instead of being based on the data-driven computations in their central processing units. 6 Lawrence Lessig, Code 2.0 (2d ed. 2006). 7 Joel R. Reidenberg, Lex Informatica: The Formulation of Information Policy Rules Through Technology, 76 Tex. L. Rev. 553 (1997).