Computational Trust Model for Repeated Trust Games Quang-Vinh Dang, Claudia-Lavinia Ignat

Total Page:16

File Type:pdf, Size:1020Kb

Computational Trust Model for Repeated Trust Games Quang-Vinh Dang, Claudia-Lavinia Ignat Computational Trust Model for Repeated Trust Games Quang-Vinh Dang, Claudia-Lavinia Ignat To cite this version: Quang-Vinh Dang, Claudia-Lavinia Ignat. Computational Trust Model for Repeated Trust Games. The 15th IEEE International Conference on Trust, Security and Privacy in Computing and Commu- nications (TrustCom-16), Aug 2016, Tianjin, China. hal-01351250 HAL Id: hal-01351250 https://hal.inria.fr/hal-01351250 Submitted on 3 Aug 2016 HAL is a multi-disciplinary open access L’archive ouverte pluridisciplinaire HAL, est archive for the deposit and dissemination of sci- destinée au dépôt et à la diffusion de documents entific research documents, whether they are pub- scientifiques de niveau recherche, publiés ou non, lished or not. The documents may come from émanant des établissements d’enseignement et de teaching and research institutions in France or recherche français ou étrangers, des laboratoires abroad, or from public or private research centers. publics ou privés. Computational Trust Model for Repeated Trust Games Quang-Vinh Dang Claudia-Lavinia Ignat Universite´ de Lorraine, LORIA, F-54506 Inria, F-54600 Inria, F-54600 Universite´ de Lorraine, LORIA, F-54506 CNRS, LORIA, F-54506 CNRS, LORIA, F-54506 [email protected] [email protected] Abstract—Trust game is a money exchange game that has The trust game could be one-shot, i.e. the game ends after been widely used in behavioral economics for studying trust and one round of money exchange, or repeated, i.e. it lasts several collaboration between humans. In this game, exchange of money rounds [4], [5], [6]. The pairs of users could be fixed [7] or is entirely attributable to the existence of trust between users. The trust game could be one-shot, i.e. the game ends after one re-assigned before each round [8]. The total number of rounds round of money exchange, or repeated, i.e. it lasts several rounds. in the experiment can be reported to users [9], [10] or not [6]. Predicting user behavior in the repeated trust game is of critical These games provide different kinds of partner information to importance for the next movement of the partners. However, players, such as their gender, age and income [11], or their existing behavior prediction approaches uniquely rely on players past interaction history [12], [8]. Some of these games allow personal information such as their age, gender and income and do not consider their past behavior in the game. communication between players [9], [13] or specification of In this paper, we propose a computational trust metric that user contracts [14], [15]. is uniquely based on users past behavior and can predict the Predicting partner behavior in the repeated trust game is future behavior in repeated trust game. Our trust metric can of critical importance. In e-commerce, for instance, it is distinguish between users having different behavioral profiles very important to predict the behavior of the partners before and is resistant to fluctuating user behavior. We validate our model by using an empirical approach against data sets collected transaction completion. Game theory predicts that, in trust from several trust game experiments. We show that our model game, sender will send 0 and receiver will send back 0 [16]. is consistent with rating opinions of users, and our model can However, in experimental game theory we usually do not provide higher accuracy on predicting users’ behavior compared observe this user behavior. In fact, the sending behavior of with other naive models. users in large-scale settings follows the normal distribution Index Terms—trust management, computational trust, non- cooperative games, behavioral game theory [3]. Existing approaches on predicting behavior of participants in one-shot trust game were proposed using additional infor- I. INTRODUCTION mation related to players, such as the personal information Studying cooperation between users is a main topic of re- (age, gender, income, etc.) or the data collected through some search in economy, psychology and computer science. Invest- tests / questionnaires. However, the only reliable available in- ment game or trust game proposed by Berg [1] is an important formation for predicting users behavior is their behavior during tool to study trust and collaboration [2]. The experiment has previous transactions. No other work proposed prediction of been repeated thousands of times all over the world1. In this user behavior based on analysis of previous interactions. game two users exchange money in anonymous mode. The In this paper, we present a computational trust metric, which sender sends an amount to the receiver. This amount is tripled serves as a model for user behavior, to reflect and predict when it arrives to the receiver. The receiver then selects an user’s behavior in repeated trust games. We claim that, based amount that should not exceed the sum received to send back on users past behavior in the repeated trust game, we can to the sender. This time, the amount is not tripled when it model and predict their next behavior. We only focus on arrives to the sender. The money exchange between users is repeated version of original anonymous trust game, i.e. the entirely attributable to the existence of trust between them. players have no information about their partners. We show that The model of trust game can be applied for the cooperation our model is robust to several types of attacks. We validate between users that are located in different places that do not our model by using an empirical approach against data sets know each other. For instance, in e-commerce, buyers usually collected from several trust game experiments. We prove that need to pay before they can receive the goods, so they need our trust metric is consistent with users’ opinion about partners to trust the sellers. trustworthiness. The paper is structured as follows. We present related works 1According to Google Scholar, the paper has been cited 3,810 times, as on 27-Jun-2016. An incomplete study counted that there are more than 23,000 in Section II. In Section III we present our trust metric and in participants involved in trust game experiments [3]. section IV we describe its evaluation. Section V presents some limitations of our metric. Concluding remarks are presented in 2) The trust value can distinguish between different types Section VI. of users. 3) The trust value considers user behavior over time. II. RELATED WORK 4) The trust value encourages a stable behavior rather than Several approaches on predicting behavior of participants a fluctuating one. in one-shot trust game were proposed using additional infor- 5) The trust value is robust against attacks. mation related to players, such as their personal information A. Parameters initial values (e.g. age, gender, income) or evaluation collected through The values of the parameters used for the trust metric some tests or questionnaires users needed to fill in before the computation are displayed in Table I. The left side of the table experiment. contains the initial values of the corresponding parameters, Gunnthorsdottir et al. [17] used the Mach test on users while the right side of the table contains the constant values personality to predict their behavior in the game. Evans et of the corresponding parameters. al. [18] predicted users behavior based on the results of Propensity to Trust Survey. Using a similar idea, Yamagishi et TABLE I al. [19] defined a new model called attitudinal trust to predict PARAMETER INITIAL VALUES. users’ behavior in one-shot trust game. The attitudinal trust is calculated based on data collected through a questionnaire. α0 0. 0.3 Yen [20] claimed that users with higher income send more β0 0. φ 0.1 to their partners than users with lower income. Falk et al. [21] atf0 0. MAX ATF 2. confirmed this suggestion by showing that, students tend to expect trust0 0. threshold 0.25 send less than other social groups. trend factor0 0. c 0.9 current trust 0. Ashraf et al. [22] used the data collected in the dictator 0 aggregate trust0 0. game to predict the behavior of the same users in the trust change rate0 0. game. trust value0 0.5 In real world applications, sociometric information of users is usually not available [23]. Even if this information is available, such as a user login to a website by using Facebook B. Current trust account, it is not always reliable as users can declare false As we described in Section I, in each round, two users personal information. Our method to model and predict users’ interact by sending a non-negative amount. For senders, the behavior does not require any additional personal data. maximum amount they can send is set to 10, and for receivers, We are not aware of any prior work that predicted users’ the maximum amount they can send is the amount they behavior in repeated trust game. The theoretical prediction is received from the sender (i.e. three times of what the sender that in the repeated trust game participants keep exchanging sent). For both roles, we normalize the send proportiont as an amount of 0 [16]. the sending proportion of a user at round t, with t ≥ 1: Glaeser et al. [24] used average value of previous sending amount as the trust measurement of users in trust game. This sending amountt send proportiont = (1) trust metric as average of previous sending amounts was used maximum sending amountt later on by other research works [8], [3], [25]. However, It is obvious that 8t; 0 ≤ send proportiont ≤ 1. this average trust metric can not deal with user fluctuating We define next the trust metric for a single interaction behavior, as we discuss in Section IV.
Recommended publications
  • A Understanding Graph-Based Trust Evaluation in Online Social Networks: Methodologies and Challenges
    A Understanding Graph-based Trust Evaluation in Online Social Networks: Methodologies and Challenges Wenjun Jiang, Hunan University Guojun Wang, Central South University Md Zakirul Alam Bhuiyan, Temple University Jie Wu, Temple University Online social networks (OSNs) are becoming a popular method of meeting people and keeping in touch with friends. OSNs resort to trust evaluation models and algorithms, as to improve service qualities and enhance user experiences. Much research has been done to evaluate trust and predict the trustworthiness of a target, usually from the view of a source. Graph-based approaches make up a major portion of the existing works, in which the trust value is calculated through a trusted graph (or trusted network, web of trust, multiple trust chains). In this paper, we focus on graph-based trust evaluation models in OSNs, particularly in computer science literature. We first summarize the features of OSNs and the properties of trust. Then, we comparatively review two categories of graph-simplification based and graph-analogy based approaches, and discuss their individual problems and challenges. We also analyze the common challenges of all graph-based models. To provide an integrated view of trust evaluation, we conduct a brief review of its pre-and-post processes, i.e., the preparation and the validation of trust models, including information collection, performance evaluation, and related applications. Finally, we identify some open challenges that all trust models are facing. Categories and Subject Descriptors: A.1 [Introductory and Survey]; C.2.4 [Computer-Communication Networks]: Distributed Systems General Terms: Design, Reliability, Management Additional Key Words and Phrases: trusted graph, trust evaluation, simplification, analogy, online social networks (OSNs), trust models.
    [Show full text]
  • Decentralized Reputation Model and Trust Framework Blockchain and Smart Contracts
    IT 18 062 Examensarbete 30 hp December 2018 Decentralized Reputation Model and Trust Framework Blockchain and Smart contracts Sujata Tamang Institutionen för informationsteknologi Department of Information Technology Abstract Decentralized Reputation Model and Trust Framework: Blockchain and Smart contracts Sujata Tamang Teknisk- naturvetenskaplig fakultet UTH-enheten Blockchain technology is being researched in diverse domains for its ability to provide distributed, decentralized and time-stamped Besöksadress: transactions. It is attributed to by its fault-tolerant and zero- Ångströmlaboratoriet Lägerhyddsvägen 1 downtime characteristics with methods to ensure records of immutable Hus 4, Plan 0 data such that its modification is computationally infeasible. Trust frameworks and reputation models of an online interaction system are Postadress: responsible for providing enough information (e.g., in the form of Box 536 751 21 Uppsala trust score) to infer the trustworthiness of interacting entities. The risk of failure or probability of success when interacting with an Telefon: entity relies on the information provided by the reputation system. 018 – 471 30 03 Thus, it is crucial to have an accurate, reliable and immutable trust Telefax: score assigned by the reputation system. The centralized nature of 018 – 471 30 00 current trust systems, however, leaves the valuable information as such prone to both external and internal attacks. This master's thesis Hemsida: project, therefore, studies the use of blockchain technology as an http://www.teknat.uu.se/student infrastructure for an online interaction system that can guarantee a reliable and immutable trust score. It proposes a system of smart contracts that specify the logic for interactions and models trust among pseudonymous identities of the system.
    [Show full text]
  • Maintaining Security and Trust in Large Scale Public Key Infrastructures
    Maintaining Security and Trust in Large Scale Public Key Infrastructures Vom Fachbereich Informatik der Technischen Universit¨atDarmstadt genehmigte Dissertation zur Erlangung des Grades Doktor-Ingenieur (Dr.-Ing.) von Dipl. Wirtsch.-Inform. Johannes Braun geboren in Herrenberg. Referenten: Prof. Dr. Johannes Buchmann Prof. Dr. Max M¨uhlh¨auser Tag der Einreichung: 16.03.2015 Tag der m¨undlichen Pr¨ufung: 30.04.2015 Hochschulkennziffer: D 17 Darmstadt 2015 List of Publications [B1] Johannes Braun. Ubiquitous support of multi path probing: Preventing man in the middle attacks on Internet communication. IEEE Conference on Com- munications and Network Security (CNS 2014) - Poster Session, pages 510{ 511, IEEE Computer Society, 2014. Cited on pages 52 and 173. [B2] Johannes Braun, Florian Volk, Jiska Classen, Johannes Buchmann, and Max M¨uhlh¨auser.CA trust management for the Web PKI. Journal of Computer Security, 22: 913{959, IOS Press, 2014. Cited on pages 9, 66, 89, and 104. [B3] Johannes Braun, Johannes Buchmann, Ciaran Mullan, and Alex Wiesmaier. Long term confidentiality: a survey. Designs, Codes and Cryptography, 71(3): 459{478, Springer, 2014. Cited on page 161. [B4] Johannes Braun and Gregor Rynkowski. The potential of an individ- ualized set of trusted CAs: Defending against CA failures in the Web PKI. International Conference on Social Computing (SocialCom) - PAS- SAT 2013, pages 600{605, IEEE Computer Society, 2013. Extended version: http://eprint.iacr.org/2013/275. Cited on pages 9, 32, and 57. [B5] Johannes Braun, Florian Volk, Johannes Buchmann, and Max M¨uhlh¨auser. Trust views for the Web PKI. Public Key Infrastructures, Services and Appli- cations - EuroPKI 2013, vol.
    [Show full text]
  • Scalable Blockchains for Computational Trust
    A Scalable Blockchain Approach for Trusted Computation and Verifiable Simulation in Multi-Party Collaborations Ravi Kiran Raman,∗y Roman Vaculin,y Michael Hind,y Sekou L. Remy,y Eleftheria K. Pissadaki,y Nelson Kibichii Bore,y Roozbeh Daneshvar,y Biplav Srivastava,y and Kush R. Varshneyy ∗University of Illinois at Urbana-Champaign yIBM Research Abstract—In high-stakes multi-party policy making based on ology in this paper. machine learning and simulation models involving independent In AI, increasingly complex machine learning models with computing agents, a notion of trust in results is critical in facilitat- ever increasing complexity are being learned on vast datasets. ing transparency, accountability, and collaboration. Using a novel combination of distributed validation of atomic computation Lack of access to the training process and the hyperparameters blocks and a blockchain-based immutable audit mechanism, this used therein do not only hinder scientific reproducibility, work proposes a framework for distributed trust in computations. but also remove any basis to trust the reported results. In In particular we address the scalability problem by reducing particular, applications interested in the trained model might the storage and communication costs using a lossy compression not trust the agent performing the training. Other than sanity scheme. This framework guarantees not only verifiability of final results, but also the validity of local computations, and its cost- checks such as validation datasets, there exists no simpler benefit tradeoffs are studied using a synthetic example of training way to guarantee correctness than complete retraining. This a neural network. is impractical considering the extensive amounts of time and hardware required.
    [Show full text]
  • A Survey of Trust in Computer Science and the Semantic Web
    To appear in Journal of Web Semantics: Science, Services and Agents on the World Wide Web, 2007. A Survey of Trust in Computer Science and the Semantic Web Donovan Artz and Yolanda Gil Information Sciences Institute University of Southern California 4676 Admiralty Way, Marina del Rey CA 90292 +1 310-822-1511 [email protected] March 15, 2007 Abstract Trust is an integral component in many kinds of human interaction, allowing people to act under uncertainty and with the risk of negative consequences. For example, exchanging money for a service, giving access to your property, and choosing between conflicting sources of information all may utilize some form of trust. In computer science, trust is a widely- used term whose definition differs among researchers and application areas. Trust is an essential component of the vision for the Semantic Web, where both new problems and new applications of trust are being studied. This paper gives an overview of existing trust research in computer science and the Semantic Web. Keywords: Trust, Web of Trust, Policies, Reputation 1 Introduction Trust is a central component of the Semantic Web vision (Berners-Lee 1999; Berners-Lee et al 2001; Berners-Lee et al 2006). The Semantic Web stack (Berners-Lee 2000; Berners-Lee et al 2006) has included all along a trust layer to assimilate the ontology, rules, logic, and proof layers. Trust often refers to mechanisms to verify that the source of information is really who the source claims to be. Signatures and encryption mechanisms should allow any consumer of information to check the sources of that information.
    [Show full text]
  • Decentralized Trust Management: Risk Analysis and Trust Aggregation 1:3
    1 Decentralized Trust Management: Risk Analysis and Trust Aggregation XINXIN FAN, Institute of Computing Technology, Chinese Academy of Sciences, China LING LIU, School of Computer Science, Georgia Institute of Technology, USA RUI ZHANG, Institute of Information Engineering, Chinese Academy of Sciences, China QUANLIANG JING and JINGPING BI, Institute of Computing Technology, Chinese Academy of Sciences, China Decentralized trust management is used as a referral benchmark for assisting decision making by human or intelligence machines in open collaborative systems. During any given period of time, each participant may only interact with a few of other participants. Simply relying on direct trust may frequently resort to random team formation. Thus, trust aggregation becomes critical. It can leverage decentralized trust management to learn about indirect trust of every participant based on past transaction experiences. This paper presents alternative designs of decentralized trust management and their efficiency and robustness from three per- spectives. First, we study the risk factors and adverse effects of six common threat models. Second, we review the representative trust aggregation models and trust metrics. Third, we present an in-depth analysis and comparison of these reference trust aggregation methods with respect to effectiveness and robustness. We show our comparative study results through formal analysis and experimental evaluation. This comprehen- sive study advances the understanding of adverse effects of present and future threats and the robustness of different trust metrics. It may also serve as a guideline for research and development of next generation trust aggregation algorithms and services in the anticipation of risk factors and mischievous threats. CCS Concepts: • General and reference → Surveys and overviews; • Security and privacy → Trust frameworks; • Information systems → Collaborative and social computing systems and tools.
    [Show full text]
  • Social Computational Trust Model (SCTM): a Framework to Facilitate the Selection of Partners
    Social Computational Trust Model (SCTM): A Framework to Facilitate the Selection of Partners 1st Ameneh Deljoo 2nd Tom van Engers 3th Leon Gommans Informatics Institute, Faculty of Science Leibniz Center for Law AirFrance-KLM University of Amsterdam University of Amsterdam Amsterdam, the Netherlands Amsterdam, the Netherlands Amsterdam, the Netherlands [email protected] [email protected] [email protected] 4th Cees de Laat Informatics Institute, Faculty of Science University of Amsterdam Amsterdam, the Netherlands [email protected] Abstract—Creating a cyber security alliance among network and private organizations is required to arrange technical domain owners, as a means to minimize security incidents, counter measures. Sharing cyber intelligence among dif- has gained the interest of practitioners and academics in the ferent parties, such as internet & cloud service providers last few years. A cyber security alliance, like any membership and enterprise networks, becomes increasingly important. organization, requires the creation and maintenance of trust Additionally, networks have evolved over the time and among its members, in this case the network domain owners. became more complex and less connected, therefore the To promote the disclosure and sharing of cyber security infor- protection of such a complex network can often only be mation among the network domain owners, a trust framework guaranteed and financed as a shared effort. One of the is needed. benefits of information sharing among the members in an This paper discusses a social computational trust model alliance is improving the decision and policy making in the (SCTM), that helps alliance members to select the right different levels of organization and facilitating the selection partner to collaborate with and perform collective tasks, and of optimal cyber defense tactics during the attack period.
    [Show full text]
  • 101 Survey on Computational Trust and Reputation Models
    Survey on Computational Trust and Reputation Models DIEGO DE SIQUEIRA BRAGA, MARCO NIEMANN, and BERND HELLINGRATH, University of Muenster—ERCIS, Germany FERNANDO BUARQUE DE LIMA NETO, University of Pernambuco, Brazil Over the recent years, computational trust and reputation models have become an invaluable method to im- prove computer-computer and human-computer interaction. As a result, a considerable amount of research has been published trying to solve open problems and improving existing models. This survey will bring addi- tional structure into the already conducted research on both topics. After recapitulating the major underlying concepts, a new integrated review and analysis scheme for reputation and trust models is put forward. Using highly recognized review papers in this domain as a basis, this article will also introduce additional eval- uation metrics to account for characteristics so far unstudied. A subsequent application of the new review schema on 40 top recent publications in this scientific field revealed interesting insights. While the area of computational trust and reputation models is still a very active research branch, the analysis carried out here was able to show that some aspects have already started to converge, whereas others are still subject to vivid discussions. CCS Concepts: • General and reference → Surveys and overviews;•Information systems → Reputation systems;•Security and privacy → Trust frameworks;•Computing methodologies → Knowledge representation and reasoning; Intelligent agents; Additional Key Words and Phrases: Computational trust, reputation management systems ACM Reference format: Diego de Siqueira Braga, Marco Niemann, Bernd Hellingrath, and Fernando Buarque de Lima Neto. 2018. Survey on Computational Trust and Reputation Models. ACM Comput. Surv.
    [Show full text]
  • Asymmetric Trust and Causal Reasoning in Blockchain-Based Ais Percy Venegas1
    Asymmetric Trust and Causal Reasoning in Blockchain-based AIs Percy Venegas1 1Economy Monitor ABSTRACT We use genetic programming evolved networks, vector fields, and signal processing to study time varying-exposures where trust is implied (e.g. a conversion event from attention flow to financial commitment). The datasets are behavioral finance time series (from on-chain data, such as fees, and off-chain data, such as clickstreams), which we use to elaborate on various complexity metrics of causality, through the creation parametric network graphs. We discuss the related methods and applications and conclude with the notion of social memory irreversibility and value by memory as useful constructs that take advantage of the natural fact of the existence of trust asymmetries, that can be operationalized by embedded AIs that use distributed ledgers both as the substrate of their intelligence and as social computers. By being context-aware, those intelligent agents are able to intervene in problematic stressors and contribute to minimizing network fragility. Keywords: systemic risk, behavioral finance, economic complexity, evolutionary computation, computational trust, the blockchain, cryptocurrencies, market microstructure, reality mining. JEL Classification: G02, F63, B17, C53, C58 There is none deceived but he that trusts –Benjamin Franklin Introduction Given the current state of knowledge, it is relatively easy to have artificial intelligent agents to find patterns and to formulate predictions following some objective criteria. But to get become useful in comparison with human intelligence, it is crucial that those agents are able to ask: Why? Posing the question is an exercise on causal reasoning, a realization of the awareness of cause and effect.
    [Show full text]
  • State of the Art of Trust and Reputation Systems in E-Commerce Context
    State of the art of Trust and Reputation Systems in E-Commerce Context H. Rahimi* and H. El Bakkali, Information Security Research Team, University Mohamed V Rabat ENSIAS, Rabat, Morocco. [email protected] *[email protected] Abstract-This article proposes in depth comparative study of the most popular, used and analyzed Trust and Reputation System (TRS) according to the trust and reputation literature and in terms of specific trustworthiness criteria. This survey is realized relying on a selection of trustworthiness criteria that analyze and evaluate the maturity and effectiveness of TRS. These criteria describe the utility, the usability, the performance and the effectiveness of the TRS. We also provide a summary table of the compared TRS within a detailed and granular selection of trust and reputation aspects. I. INTRODUCTION Open electronic markets, online collaboration systems, distributed peer-to-peer applications, online social media require the establishment of mutual trust between service providers and service consumers. In fact, the major concerns of web-based services especially e-commerce applications is to overcome the inherent uncertainties and untrustworthiness risks and enhance the system's robustness and resistance against fraudulent users and distrustful ones. Besides, e-commerce platforms aim at adopting the most efficient approach that helps detect and analyze users' intentions in order to reveal and understand deceitful ones. Otherwise, the underlying purpose of e-commerce services which is to maximize the profit and the rate of purchase, would be threatened and deteriorated by fraudulent and ill-intentioned users. For this reason, Recommender Systems such as Trust and Reputation Systems (TRS), provide essential input for computational trust so as to predict future behaviors of peers basing on the past actions of a peer [1].
    [Show full text]
  • Towards a Theory of Trust in Networks of Humans and Computers
    Presented at the 19th International Workshop on Security Protocols, Cambridge, UK, March 28-30, 2011 (to appear in LNCS, Springer Verlag) Towards a Theory of Trust in Networks of Humans and Computers Virgil Gligor and Jeannette M. Wing Carnegie Mellon University Pittsburgh, Pennsylvania 15213 [email protected],[email protected] Abstract. We argue that a general theory of trust in networks of hu- mans and computers must be build on both a theory of behavioral trust and a theory of computational trust. This argument is motivated by increased participation of people in social networking, crowdsourcing, human computation, and socio-economic protocols, e.g., protocols mod- eled by trust and gift-exchange games [3, 10, 11], norms-establishing con- tracts [1], and scams [6, 35, 33]. User participation in these protocols re- lies primarily on trust, since on-line veri¯cation of protocol compliance is often impractical; e.g., veri¯cation can lead to undecidable problems, co- NP complete test procedures, and user inconvenience. Trust is captured by participant preferences (i.e., risk and betrayal aversion) and beliefs in the trustworthiness of other protocol participants [11, 10]. Both pref- erences and beliefs can be enhanced whenever protocol non-compliance leads to punishment of untrustworthy participants [11, 23]; i.e., it seems natural that betrayal aversion can be decreased and belief in trustwor- thiness increased by properly de¯ned punishment [1]. We argue that a general theory of trust should focus on the establishment of new trust relations where none were possible before. This focus would help create new economic opportunities by increasing the pool of usable services, removing cooperation barriers among users, and at the very least, taking advantage of \network e®ects." Hence a new theory of trust would also help focus security research in areas that promote trust-enhancement in- frastructures in human and computer networks.
    [Show full text]
  • The Importance of Trust in Computer Security Christian Jensen
    The Importance of Trust in Computer Security Christian Jensen To cite this version: Christian Jensen. The Importance of Trust in Computer Security. 8th IFIP International Conference on Trust Management (IFIPTM), Jul 2014, Singapore, Singapore. pp.1-12, 10.1007/978-3-662-43813- 8_1. hal-01381672 HAL Id: hal-01381672 https://hal.inria.fr/hal-01381672 Submitted on 14 Oct 2016 HAL is a multi-disciplinary open access L’archive ouverte pluridisciplinaire HAL, est archive for the deposit and dissemination of sci- destinée au dépôt et à la diffusion de documents entific research documents, whether they are pub- scientifiques de niveau recherche, publiés ou non, lished or not. The documents may come from émanant des établissements d’enseignement et de teaching and research institutions in France or recherche français ou étrangers, des laboratoires abroad, or from public or private research centers. publics ou privés. Distributed under a Creative Commons Attribution| 4.0 International License The Importance of Trust in Computer Security Christian D. Jensen Department of Applied Mathematics & Computer Science Technical University of Denmark DK-2800 Kgs. Lyngby, Denmark [email protected] Abstract. The computer security community has traditionally regarded security as a “hard” property that can be modelled and formally proven under certain simplifying assumptions. Traditional security technologies assume that computer users are either malicious, e.g. hackers or spies, or benevolent, competent and well informed about the security policies. Over the past two decades, however, computing has proliferated into all aspects of modern society and the spread of malicious software (malware) like worms, viruses and botnets have become an increasing threat.
    [Show full text]