PERSUASIVE TECHNOLOGY and HUMAN RIGHTS

Total Page:16

File Type:pdf, Size:1020Kb

PERSUASIVE TECHNOLOGY and HUMAN RIGHTS PERSUASIVE TECHNOLOGY and HUMAN RIGHTS LL.M Law and Technology Tilburg University Netherlands Panagiota Tokamani 2016 PERSUASIVE TECHNOLOGY and HUMAN RIGHTS Abstract The fast-paced development of technology throughout the last decades has been redounded, inter alia, to the emergence of a newly-introduced notion, stipulated as “persuasive technology”. Computing systems are designed in many cases having the intention to shape human attitudes or behaviors. Under these terms, the reasonable question arisen, from a legal perspective of viewing, is the following: to what extent may persuasive technology affect fundamental human rights established in European legal order and what the regulatory level of protection of these human rights is so for this possible effect to be mitigated. This paper aims at presenting the issues following this emerging question analyzing the legislative imprinting of the human right to privacy, the cornerstone for two inherent rights, the right to data protection and the right to autonomy. Conclusions are drawn with respect to the interaction of the aforementioned human rights with “persuasive technology” doctrine, examining the consistency of the privacy protective law, under the current legislative framework, with the policies used by persuasive computing systems’ technology. 2 TILBURG LAW SCHOOL Abstract Le développement rapide de la technologie au long des ces dernières décennies a activement contribué, inter alia, à l’émergence d’une notion récemment introduite, définie comme ‘technologie persuasive’. Des systèmes informatiques sont désignées, aux plusieurs cas, en ayant l’intention de formuler des attitudes ou comportements humains. Selon ces termes, la question raisonnable qui se pose, d’un point de vue legal, est la suivante: dans quelle mesure la technologie persuasive est-elle capable d’influencer les droits de l’homme fondamentaux, établis à l’ordre juridique européenne, et quel est le niveau de régulation de la protection de ces droits, afin que l’effet possible soit mitigé. Cette thèse vise à presenter les sujets qui suivent cette question émergente, en analysant l’empreinte legislative au droit à la vie privée, pierre angulaire de deux doits inhérents, le droit à la protection des données et le droit à l’autonomie personnelle. Les conclusions concernant l’interaction entre les droits humains ci-dessus et la doctrine de la ‘technologie persuasive’, sont tirées en examinant la consistence de la loi protectrice de la vie privée, dans le cadre législatif en vigueur, avec les pratiques utilisées dans la technologie des systèmes informatiques persuasives. 3 PERSUASIVE TECHNOLOGY and HUMAN RIGHTS Περίληψη Ο ραγδαίος ρυθμός ανάπτυξης της τεχνολογίας τις τελευταίες δεκαετίες έχει οδηγήσει, μεταξύ άλλων, στην εμφάνιση ενός νέο εισαχθέντος φαινομένου, οριζόμενο ως «τεχνολογία της πειθούς». Τα υπολογιστικά συστήματα σχεδιάζονται, σε πολλές περιπτώσεις, έχοντας την πρόθεση να διαμορφώσουν τις ανθρώπινες συμπεριφορές και τη στάση ζωής. Υπό αυτές τις συνθήκες, το εύλογο ερώτημα που ανακύπτει, από άποψη νομικής παρατήρησης, είναι το εξής: μέχρι ποίου βαθμού δύναται η «τεχνολογία της πειθούς» να επηρεάσει θεμελιώδη ανθρώπινα δικαιώματα, όπως αυτά θεσμοθετούνται στην Ευρωπαϊκή έννομη τάξη και ποιο είναι το κανονιστικό επίπεδο προστασίας ώστε να μετριαστεί αυτή η εν δυνάμει επιρροή. Η παρούσα εργασία θέτει ως στόχο την παρουσίαση των θεμάτων που απορρέουν από το ως άνω ανακύπτον ερώτημα αναλύοντας το νομοθετικό αποτύπωμα του ανθρώπινου δικαιώματος στην ιδιωτικότητα, ως ακρογωνιαίου λίθου δύο συνυφασμένων ανθρωπίνων δικαιωμάτων, του δικαιώματος στην προστασία των προσωπικών δεδομένων και του δικαιώματος στην αυτονομία. Καταληκτικά, θα προκύψουν συμπεράσματα όσον αφορά στην αλληλεπίδραση των ως άνω αναφερόμενων ανθρωπίνων δικαιωμάτων με το φαινόμενο της «τεχνολογίας της πειθούς» και στη συνέπεια του προστατευτικού νόμου για το δικαίωμα στην ιδιωτικότητα με τις στρατηγικές που ακολουθούνται από τα υπολογιστικά συστήματα της «τεχνολογίας της πειθούς». 4 TILBURG LAW SCHOOL Table of contents TABLE OF CONTENTS…………………………………………………………. 5 CHAPTER 1 I.Introduction……………………………………………………………………7 1.1. Abstracted description of the topic………………………………………7 1.2. Overview of the paper……………………………………………………10 CHAPTER 2 II. Persuasive Technology- A newly introduced notion……………………...13 2.1. Introduction to the term………………………………………………...13 2.1.1. Persuasion…………………………………………………………...14 2.1.2. Technology…………………………………………………………..15 2.1.3. Persuasive Technology……………………………………………...17 2.2. Captology…………………………………………………………………18 2.3. Practical examples of persuasive technology…………………………...20 2.4. Conclusion………………………………………………………………...23 CHAPTER 3 III. Right to Privacy………………………………………………………………24 3.1. The notion of privacy………………………………………………………24 3.2. The aspects of the right to privacy………………………………………...26 3.3. Conclusion…………………………………………………………………..28 CHAPTER 4 IV. Right to Privacy and Persuasive technology………………………………..30 4.1. The legal context…………………………………………………………....30 4.2. Persuasive technology and e-privacy: Challenges………………………..34 CHAPTER 5 V. Right to Data Protection and Persuasive Technology……………………….36 5.1. The legal context……………………………………………………………36 5.2. Data protection and persuasive technology……………………………….39 CHAPTER 6 VI. Right to Autonomy and Persuasive Technology…………………………….43 6.1. The context of autonomy…………………………………………………...43 6.2. The right to autonomy and persuasive technology……………………….45 CONCLUSION……………………………………………………………………49 5 PERSUASIVE TECHNOLOGY and HUMAN RIGHTS LIST OF SOURCES…………………………………………………………….52 List of books……………………………………………………………………..52 List of articles……………………………………………………………………53 Table of legislation………………………………………………………………55 Table of case law………………………………………………………………...56 Table of sites……………………………………………………………………..56 6 TILBURG LAW SCHOOL CHAPTER 1 Introduction 1.1 Abstracted description of the topic Persuasion is “an attempt to shape, reinforce or change behaviors, feelings or thoughts about an issue, object or action”1. This is the most general definition which is used in the fields of psychological analysis of human characteristics and conducts. People can persuade, artifacts may persuade and lately technology is considered to persuade. This thesis focuses on persuasive technologies, and especially computing systems, with the aim to analyze the meaning and the extent of this term, and to probe, under the spectrum of legal order, this newly-introduced subject and how it may affect specific fundamental human rights. The term “persuasive technology” triggers various questions about the deeper analysis of both the two compositing words of this phrase. How can persuasion be achieved through or with the technology? Which faculties of technology can be identified as persuasive? Who wields this power of technological influence? What is the distinctive characteristic of persuasion in order not to be confused with deception or coercion? All these questions are attempted to be answered in the first place as an introductory note, so the reader may be familiar with the issue in question before further going deeper to the examination of its legal aspects in the following chapters. Within the context of this introduction to the main issue, a newly-emerged science, that is captology, will be used as a component for the comprehension of “persuasive technology” in terms of persuasive computing systems. Captology is the study of the area in which computing, operating systems and persuasion intersect2 and it constitutes an acronym originated from the term “Computers As Persuasive Technology”. By definition, captology focuses on the persuasive “power” that merely computing systems may bear. Accordingly, the present paper will content itself to narrow the wide range of persuasive technologies to the extent of “computing/operating systems as persuasive tools”, focusing on the cases of persuasion originated from computing systems and based on the reviews and analysis coming from the field of captology. The most recognized and thorough definition of “persuasion” in the field of technological development could be the one interpreting persuasion as “the interaction 1 B. J. Fogg, “Persuasive Computers: Perspectives and Research directions”, (1998), CHI 98, Stanford University, p. 225 2 B. J. Foggs, Persuasive Technology, Using Computers to Change What We Think and Do, (2011, Kaufmann Publishers), p. 5 7 PERSUASIVE TECHNOLOGY and HUMAN RIGHTS of computing systems in order to change behaviors and attitudes”3.The key point of persuasion in the debated case is intentionality; persuasion is not resulted as an accidental consequence of the occurrence of sequel incidents. Persuasion needs to imply an intention of changing attitudes of persuaded subjects. More particularly, the designed effects which are accrued by the planned persuading process may force users of technological equipment to register with their personal data in myriads sites on internet under the pretext of urgency for identification, to accept the terms of a service in order to use it with the assumption of their own safety or to be motivated under various apps to foster a specific way of life which will be controlled in total by these apps per se4. These examples raise concerns related to the consequences of persuasive technology’s prevalence with regards to the respect or violation of fundamental human rights related within the content of these practical examples. “Technology” is the second compositing word of the term “persuasive technology”. Pursuant to the theory of captology,
Recommended publications
  • Technology and Behaviour Change, for Good and Evil
    Technology and Behaviour Change, for Good and Evil Adam N. Joinson & Lukasz Piwek Behavioural Research Lab Bristol Social Marketing Centre Bristol Business School University of the West of England (UWE) DRAFT: WORKING PAPER FOR SUBMISSION TO SOCIAL TRENDS INSTITUTE MEETING ‘Communication Technologies and Lifestyles’, MAY 2013, Barcelona. Not for distribution beyond STI workshop participants without permission of authors. For latest version, email: [email protected] ! 1! Since the first basic stone and bone tools were used by Plio-Pleistocene hominids over three million years ago (McPherron et al., 2010), the ability of humans to fashion tools has not only extended our capacity to conduct tasks, but may also have had a transformative impact on our own selves. For instance, tool development may have been a key facilitator in the development of a large, energy hungry brain in humans (Gibbons, 1998). In more recent times, technologies and inventions - ranging from tally marks as a pre-cursor to numbers, the number zero, writing, the printing press, the stirrup, and the computer - have transformed not only individual human abilities through an extension of physical capabilities (McLuhan, 1964), but also society through both the intended and unintended consequences of widespread adoption and use. In the present position paper we summarise the key ways in which technology influences behaviour, and propose ways in which the same technology can be utilized in order to achieve a social good. We then look at two technologies – smart technologies for self- monitoring and social media – and discuss potential ways in which our approach to understanding the use and impact of tools can help shape their application in changing behaviour.
    [Show full text]
  • The Ethics of Persuasion in Technology
    The Ethics of Persuasion in Technology Nathaniel Zinda February 21, 2019 Abstract This essay is concerned with the issue of persuasion in our information and communication technologies. More specifically, it is a rebuke of the common practice in Silicon Valley of designing products with the intention of maximizing user “engagement” - in other words, the time that users spend with the product. This article’s normative contention is that this practice (i.e., persuasive technology) degrades the user’s capacity for freedom and autonomy. The argument is divided into three parts: 1) the causes, 2) the characteristics, and 3) the ethical effects of persuasive technology. Part 1 argues that Silicon Valley’s widespread adoption of the ad-based business model has produced a competitive marketplace for users’ attention, whereby companies compete to extract as much “time spent” as possible. Part 2 outlines the two primary methods by which they do this, differentiating between persuasive design (the behavioral method) and algorithmic persuasion (the technological method). Part 3 contends that these methods are responsible for producing both functional and existential distraction amongst users, both of which undermine their sense of well-being and threaten their capacity for freedom and autonomy. i Introduction It is becoming increasingly clear that the information and communication technologies that have revolution- ized society are not on our side. This cynicism is well-deserved: from the rampant abuses of user data and privacy, to the abject failure of social networking sites to keep disinformation off their platforms, to the ef- fect that these oligopolies have had on the economics of journalism and other industries, this past decade has been marked by an impressive set of failures, abuses, and consequences that have duly marred the public’s optimism and trust.
    [Show full text]
  • The Ethics of Persuasion: Persuasive Technology and the Power Behind Design
    The Ethics of Persuasion: Persuasive Technology and the Power Behind Design A Research Paper submitted to the Department of Engineering and Society Presented to the Faculty of the School of Engineering and Applied Science University of Virginia • Charlottesville, Virginia In Partial Fulfillment of the Requirements for the Degree Bachelor of Science, School of Engineering Danielle Newman Spring 2020 On my honor as a University Student, I have neither given nor received unauthorized aid on this assignment as defined by the Honor Guidelines for Thesis-Related Assignments Advisor Kathryn A. Neeley, Associate Professor of STS, Department of Engineering and Society Introduction Throughout the last decade, there has been a significant increase in technology usage. The development of technology and different technological systems has allowed the general public to access more information, increase task production, and engage in more opportunities. Specifically, technology has opened up new opportunities to develop new information systems for influencing users. This has led to the development of Behavior Change Support Systems (BCSS), which is “a socio-technical information system with psychological and behavioral outcomes designed to form, alter or reinforce attitudes, and behaviors or an act of complying using coercion or deception” (Kulyk, p. 2), as a key construct for research in persuasive technology. Persuasive technology is a research field that studies how people are persuaded while interacting with computer technology. Although effective in its ability to treat patients, using persuasive systems and technology can become biased and cause ethical issues, based on the developer of the system. The introduction of design biases can potentially hinder patient results because people are susceptible to decision bias, which often makes it difficult for them to make self-beneficial choices.
    [Show full text]
  • Coercion and Deception in Persuasive Technologies
    Coercion and deception in persuasive technologies Timotheus Kampik Juan Carlos Nieves Helena Lindgren [email protected] [email protected] [email protected] (Corresponding author) Department of Computing Science Umeå University Sweden Abstract Technologies that shape human behavior are of high societal rele- vance, both when considering their current impact and their future potential. In information systems research and in behavioral psychol- ogy, such technologies are typically referred to as persuasive tech- nologies. Traditional definitions like the ones created by Fogg, and Harjumaa and Oinas-Kukkonen, respectively, limit the scope of per- suasive technology to non-coercive, non-deceptive technologies that are explicitly designed for persuasion. In this paper we analyze ex- isting technologies that blur the line between persuasion, deception, and coercion. Based on the insights of the analysis, we lay down an updated definition of persuasive technologies that includes coercive and deceptive forms of persuasion. Our definition also accounts for persuasive functionality that was not designed by the technology de- velopers. We argue that this definition will help highlight ethical and societal challenges related to technologies that shape human behavior and encourage research that solves problems with technology-driven persuasion. Finally, we suggest multidisciplinary research that can help address the challenges our definition implies. The suggestions we provide range from empirical studies to multi-agent system the- ory. 1 Introduction In scientific literature, the term persuasive technologies appeared first in 1998 in an abstract (special interest group announcement) by the behavioral psychologist Brian J. Fogg in the proceedings of the CHI Conference on Human Factors in Computing Systems.
    [Show full text]
  • Persuasive Technology: Development and Implementation of Personalized Technologies to Change Attitudes and Behaviours
    View metadata, citation and similar papers at core.ac.uk brought to you by CORE provided by Hanze UAS repository Peter W. de Vries, Thomas Van Rompay (Eds.) Persuasive Technology: Development and implementation of personalized technologies to change attitudes and behaviours 12th International Conference, PERSUASIVE 2017, Amsterdam, The Netherlands, April 4-6, 2017 Adjunct Proceedings Second edition: April 2017 12th International Conference on Persuasive Technology, PERSUASIVE 2017, Amsterdam, The Netherlands, April 4-6, 2017, Adjunct Proceedings Edited by Peter W. de Vries and Thomas van Rompay © Copyright of the complete adjunct proceedings is held by the Centre for eHealth & Wellbeing Research, Department of Psychology, Health and Technology, University of Twente, The Netherlands. Copyright of individual contributions is held by the author(s). Contact: Centre for eHealth & Wellbeing Research, Department of Psychology, Health and Technology, University of Twente, The Netherlands PO Box 217, 7500 AE, Enschede, The Netherlands https://www.utwente.nl/igs/ehealth/ Preface Persuasive Technology (PT) is an emerging, interdisciplinary research field, focusing on the design, development and evaluation of technologies aimed at creating awareness and inducing behavior change with the ultimate goal to increase wellbeing and quality of life. Academic researchers, designers, and practitioners from the social sciences and technological disciplines, as well as from the fields of health, healthcare, safety, sustainability, and ICT have developed this field in the preceding years, giving rise to a community which aims to ‘persuade’ people into adopting healthier lifestyles, behave more safely, and reduce consumption of renewable resources, to name a few examples. The ‘technology’ component in PT reflects usage of, amongst others, big data analytics, sensor technology for monitoring, personalized feedback and coaching, mHealth, data visualization techniques, serious gaming, and social media.
    [Show full text]
  • Designing Persuasive Health Behavior Change Interventions
    Chapter 11 Designing Persuasive Health Behavior Change Interventions Tuomas Lehto A b s t r a c t In the past decade, the utilization of various technologies to change indi- viduals’ health behaviors has been a rapidly expanding fi eld of interest. Examples of persuasive technologies can be found rather easily as there are a variety of Internet-, Web- and mobile-based systems and applications promoting healthier lifestyles. Still, the use of persuasive technology in the E-health arena is in its infancy. While the fi eld is expanding, it is evident that more research is needed to better determine how the systems affect users’ intended behaviors. This book chapter outlines several important perspectives in designing and developing persuasive health behavior change interventions. Furthermore, this chapter offers novel viewpoints, both theo- retical and practical, in designing and developing health behavior change interven- tions. In addition, useful underlying theories and design models are identifi ed and discussed. This type of knowledge may assist in building, deploying and evaluating behavior change support systems that are able to engage and retain large amounts of individuals, potentially enhancing population health and well-being. Keywords Health • Behavior change • Persuasive technology • Design 11.1 Introduction Changing people’s behavior is at the heart of health promotion. In the past decade, the use of technologies to persuade, motivate and activate individuals’ for health behavior change has been a quickly expanding fi eld of research. The T. Lehto Oulu Advanced Research on Software and Information Systems, Department of Information Processing Science , University of Oulu , Oulu , Finland e-mail: tuomas.lehto@oulu.fi N.
    [Show full text]
  • Designing Persuasion: Health Technology for Low-Income African American Communities
    Designing Persuasion: Health Technology for Low-Income African American Communities Andrea Grimes and Rebecca E. Grinter College of Computing, Georgia Institute of Technology, Atlanta, GA {agrimes,beki}@cc.gatech.edu Abstract. In the United States, African Americans face a disproportionate amount of diet-related health problems. For example, African American adults are 1.6 times more likely to have diabetes than their Caucasian counterparts. Individuals in low-income communities may face a greater risk because they typically have less access to healthy foods. Due to the significant diet-related problems within the African American community, public health researchers call for approaches to health promotion that take into account the relationship between culture and dietary habits. In this paper, we discuss three important considerations for the design of technologies that address the diet-related health disparities in low-income African American communities. These considerations include designing for cultural relevancy, modeling health behavior, and encouraging healthy behavior through the use of social psychological theories of persuasion. We use a game design example to illustrate how each of these considerations can be incorporated into the development of new technology. Keywords: Culture, Health, Low-Income, Nutrition, Persuasive Technology. 1 Introduction The United States faces a serious crisis as rates of diet-related diseases such as obesity continue to rise. Today 65% of people 20 and older are either overweight or obese [8]. This statistic reflects an increase of over 75% in obesity prevalence since 1991. While these statistics are sobering, the issue becomes even more alarming when looking at the disproportionate amount of diet-related health problems in minority groups such as the African American community.
    [Show full text]
  • Increasing Mental Health Care Access with Persuasive Technology for Social Good Tine Kolenik1 and Matjaž Gams2
    Increasing Mental Health Care Access with Persuasive Technology for Social Good Tine Kolenik1 and Matjaž Gams2 Abstract.1 The alarming trend of increasing mental health It has been recognized that the reasons for increasing of SAD problems and the global inability to find effective ways to address include a severe lack of mental health professionals and regulations them is hampering both individual and societal good. Barriers to [39] as well as unequal access to mental health care [9]. These access mental health care are many and high, ranging from socio- factors make the field ripe for technological and other scientific economic inequalities to personal stigmas. This gives technology, therapy-based interventions, especially as individuals with mental especially technology based in artificial intelligence, the opportunity to help alleviate the situation and offer unique health issues prefer therapies to medication [2]. solutions. The multi- and interdisciplinary research on persuasive Given the vast advances in behavioral sciences in regards to technology, which attempts to change behavior or attitudes without human decision-making and similar phenomena [34] as well as the deception or coercion, shows promise in improving well-being, arrival of digital technologies, AI and big data, many societal which results in increased equality and social good. This paper efforts have been put into creating technologies that would help, presents such systems with a brief overview of the field, and offers motivate, guide and affect people into bettering themselves and the general, technical and critical thoughts on the implementation as world around them in various domains, especially those pertaining well as impact. We believe that such technology can complement to SDGs.
    [Show full text]
  • Persuasive Technologies and the Right to Mental Liberty: the 'Smart'
    Persuasive technologies and the right to mental liberty: The ‘smart’ rehabilitation of criminal offenders Sjors Ligthart, Gerben Meynen and Thomas Douglas Forthcoming in Ienca, M., Pollicino, O., Liguori L., Andorno R., Stefanini E. eds, Cambridge Handbook of Information Technology, Life Sciences and Human Rights (Cambridge UK: Cambridge University Press, 2021). 1. Introduction Every day, millions of people use mobile phones, play video games and surf the Internet. It is thus important to determine how technologies like these change what people think and how they behave. This is a central issue in the study of persuasive technologies. ‘Persuasive technologies’—henceforth ‘PTs’—are digital technologies, such as mobile apps, video games and virtual reality systems, that are deployed for the explicit purpose of changing attitudes and/or behaviours, without using coercion, deception or extreme forms of psychological manipulation (such as hypnosis or indoctrination), and without exerting a direct (not psychologically mediated) physical effect on the brain.1 Typically PTs employ strategies such as prompting, information provision and encouragement. Since many important determinants of wellbeing, such as health, pollution, climate change and crime, have a large behavioural component, and since PTs are intended to change behaviour, some hold that PTs could potentially have large benefits, both for the persuaded individuals, and for society at large.2 For example, eco-feedback systems persuade car drivers to use less fuel by showing images of green leaves, and health apps might motivate users to exercise regularly, quit smoking, and stop drinking.3 The mobile app Sobriety Counter helps users to stop drinking alcohol by employing motivational techniques like visualising the money a user saves by not drinking, and providing health statistics about drinking.4 1 Fogg 2003, p.
    [Show full text]
  • Does Persuasive Technology Make Smartphones More Addictive? - an Empirical Study of Chinese University Students
    Does Persuasive Technology Make Smartphones More Addictive? - An Empirical Study of Chinese University Students Xiaowei Chen EECS School of Electrical Engineering and Computer Science KTH Royal Institute of Technology Stockholm, Sweden [email protected] ABSTRACT designers face and users’ neglected rights of acknowledge- ment were discussed. With the development of computer hardware, computers with persuasion have become more powerful and influential than ever. The latest trends show that Persuasive Technology KEYWORDS integrates with cutting-edge technologies, such as Natural Language Processing, Big Data, and Machine Learning algo- Persuasive Technology, Persuasive design principles, Smart- rithms. As persuasion is becoming increasingly intelligent phone addiction, HCI ethics and subtle, it is urgent to reflect on the dark sides of Per- suasive Technology. The study aims to investigate one of Persuasive Technology’s accusations, making smartphones 1 INTRODUCTION more addictive to its users. The study uses questionnaires and in-depth interviews Fogg was one of the first scholars who researched the over- to examine the impact of persuasive technologies on young lapping field of persuasion and computer technology. Fogg smartphone users. The participants of the study are 18 to 26 created the term "Captology" to study computers as persua- years old Chinese university students. Questionnaires were sive technologies. Since then, persuasive technologies were distributed through a university forum, student group chats, explored from multiple angles by academia and industries and Tencent Survey Service. Ten interviewees were sampled and have been integrated into various hardware and soft- randomly from the survey results. Eight interviewees shared ware products, affecting users’ healthcare, education, and their smartphone screen time for three consecutive weeks lifestyle.
    [Show full text]
  • Persuasive Technology
    Eleni Kyza • Pasi Karppinen • Evangelos Karapanos • Khin Than Win • Harri Oinas-Kukkonen (Eds.) Persuasive Technology 14th International Conference, PERSUASIVE 2019 Limassol, Cyprus, April 9–11, 2019 Adjunct Proceedings Editors Eleni A. Kyza Cyprus University of Technology Limassol, Cyprus Pasi Karppinen University of Oulu Oulu, Finland Evangelos Karapanos Cyprus University of Technology Limassol, Cyprus Khin Than Win University of Wollongong Wollongong, NSW, Australia Harri Oinas-Kukkonen University of Oulu Oulu, Finland Editorial Assistant Loukas Konstantinou Cyprus University of Technology Limassol, Cyprus ISBN 978-9963-697-37-3 (eBook) Persuasive Technology 2019 Adjunct Proceedings 14th International Conference on Persuasive Technology © Cyprus University of Technology Preface Persuasive technology is a vibrant interdisciplinary research field, focusing on the study, design, development, and evaluation of information technologies aimed at influencing people’s attitudes or behaviors through open and transparent means. The International Conference on Persuasive Technology series brings together researchers, designers, practitioners and business people from various disciplines and a wide variety of application domains. The research community seeks to facilitate healthier lifestyles, make people feel or behave more safely, reduce consumption of renewable resources, among other notable goals, by, for instance, designing software applications, monitoring through sensor technologies, analyzing obtained data, and providing various types of
    [Show full text]
  • Advancing Typology of Computer-Supported Influence: Moderation Effects in Socially Influencing Systems
    Advancing Typology of Computer-Supported Influence: Moderation Effects in Socially Influencing Systems The MIT Faculty has made this article openly available. Please share how this access benefits you. Your story matters. Citation Stibe, Agnis. “Advancing Typology of Computer-Supported Influence: Moderation Effects in Socially Influencing Systems.” Lecture Notes in Computer Science (2015): 253–264. © 2015 Springer International Publishing Switzerland As Published http://dx.doi.org/10.1007/978-3-319-20306-5_23 Publisher Springer, Cham Version Author's final manuscript Citable link http://hdl.handle.net/1721.1/108334 Terms of Use Article is made available in accordance with the publisher's policy and may be subject to US copyright law. Please refer to the publisher's site for terms of use. Advancing Typology of Computer-Supported Influence: Moderation Effects in Socially Influencing Systems Agnis Stibe MIT Media Lab, Cambridge, MA, USA [email protected] Abstract. Persuasive technologies are commonly engineered to change behav- ior and attitudes of users through persuasion and social influence without using coercion and deception. While earlier research has been extensively focused on exploring the concept of persuasion, the present theory-refining study aims to explain the role of social influence and its distinctive characteristics in the field of persuasive technology. Based on a list of notable differences, this study out- lines how both persuasion and social influence can be best supported through computing systems and introduces a notion of computer-moderated influence, thus extending the influence typology. The novel type of influence tends to be more salient for socially influencing systems, which informs designers to be mindful when engineering such technologies.
    [Show full text]