The Anonymity Engine, Minimizing Quasi-Identifiers to Strengthen K-Anonymity

Total Page:16

File Type:pdf, Size:1020Kb

The Anonymity Engine, Minimizing Quasi-Identifiers to Strengthen K-Anonymity The Anonymity Engine, Minimizing Quasi-Identifiers to Strengthen k-Anonymity by ERIC MATTHEW LOBATO B.S., University of Colorado, 2015 A thesis submitted to the Faculty of the Graduate School of the University of Colorado in partial fulfillment of the requirement for the degree of Master of Science Interdisciplinary Telecommunications Department 2017 This thesis entitled: The Anonymity Engine, Minimizing Quasi-Identifiers to Strengthen k-Anonymity written by Eric Matthew Lobato has been approved for the Interdisciplinary Telecommunications Department (Joe McManus) (David Reed) (Levi Perigo) Date The final copy of this thesis has been examined by the signatories, and we find that both the content and the form meet acceptable presentation standards of scholarly work in the above mentioned discipline. IRB protocol # ____________________ IACUC protocol # __________________ ii Lobato, Eric Matthew (M.S, ITP) The Anonymity Engine, Minimizing Quasi-Identifiers to Strengthen k-Anonymity Thesis directed by Scholar in Residence Joe McManus The k-anonymity model has become a standard for anonymizing data. However, almost all applications of k-anonymity are used to anonymize large data sets of personally identifiable information owned by a trusted third party before being given to analysists To further research in this this area, this study created a tool called the Anonymity Engine. This tool was built as a web browser plugin that analyzes headers on all web traffic exiting the system and builds a database of relevant quasi-identifier. Users are notified in real time if a data packet would compromise their identity and give the option to not send the data. This tool has also been used to generate data that shows that modifying data before implementing k-anonymity can impact the results. These modified results show that it can make some users more anonymous while reducing the level of privacy for other users depending on the traffic. iii CONTENTS CHAPTER I. INTRODUCTION AND RESEARCH QUESTION ................................1 II. REVIEW OF THE LITERATURE ..........................................................5 III. RESEARCH DESIGN AND METHODOLOGY……………………....... 9 The Anonymity Engine .......................................................................9 Design ...............................................................................................13 Testing Plans .....................................................................................15 IV. RESULTS AND CONCLUSIONS……………………………………………. 18 Results ...............................................................................................18 Conclusions .......................................................................................24 BIBLIOGRAPHY…………………….………………………………………………26 APPENDIX A. Code Submission ....................................................................................27 B. RAW DATA ...........................................................................................43 iv Table 1. Hospital Records, without anonymization …………………………………………..2 2. Hospital records with k value of 3...…………………………………………………3 3. Database Schema..……...…………………………………………………..………17 4. Application Structure..…...………………………………………………………...18 5. Experiment Users, URLs and Traits……………………………………………......21 6. Experiment 1 Data with Raw Request Counts……………………………………..24 7. Experiment 1 Data with k values of 3……………………………………………...26 8. Experiment 2 Data with k values of 3……………………………………………...27 v Figure 1. Anonymity Engine GET Blocking Example ................................................15 2. Anonymity Engine POST Blocking Example(form data)………………….16 3. Anonymity Engine POST Blocking Example(binary data)………………….16 vi I. Introduction and Research Question Imagine a user who logs onto a computer and visits the following websites: www.denverpost.com to check the news www.apple.com/support/products to look at warranty plans for their computer forecast.weather.gov to check the weather www.ebay.com to bid on a tie. This may appear to be a normal browsing session, but to an advanced user with access to traffic sniffers could use this session to ascertain several identifying traits about the other user. First, that the person is from Colorado, identified from the zip code entered on the weather site and by the news site of choice. Second, that the person has recently broken their apple product, and finally that the person is shopping for neckwear. Perhaps the advanced user knows a friend in real life who matches these descriptions. If this was the case, they could easily learn more facts about their friend by following the traffic of this specific user and potentially learning things that their friend would wish to keep private. As this example shows, maintaining personal privacy when using the internet has become one of the most important issues of the modern day. Unfortunately, nontechnical users struggle to understand how their browsing patterns could link information about themselves to the traffic they generate. Meanwhile, technical users might not always be convinced that they are truly anonymous when using privacy software such as VPNs, proxies or the Tor browser. This research aims to create a new tool that will give users a clear idea of what their web traffic says about them and whether or not they are anonymous over time. This will fill a niche in privacy tools that currently remains open as few privacy tools describe what actual network packets say about a user. The tool that this research will produce is based on the existing model for privacy known as k- anonymity [1]. The k-anonymity model determines anonymity based on the idea that knowing a minimal amount of identifiers can link an individual in a data set. For a data set to meet k-anonymity it must meet the definition: 퐿푒푡 푅푇(퐴푖, … 퐴푛) 푏푒 푎 푇푎푏푙푒 푎푛푑 푄퐼푅푇 푏푒 푡ℎ푒 푞푢푎푠 − 푑푒푛푡푓푒푟 푎푠푠표푐푎푡푒푑 푤푡ℎ 푡. 푅푇 푠 푠푎푑 푡표 푠푎푡푠푓푦 푘 − 푎푛표푛푦푚푡푦 푓 푎푛푑 표푛푙푦 푓 푒푎푐ℎ 푠푒푞푢푒푛푐푒 표푓 푣푎푙푢푒푠 푛 푅푇[푄퐼푅푇]푎푝푝푒푎푟푠 푤푡ℎ 푎푡 푙푒푎푠푡 푘 표푐푐푢푟푎푛푐푒푠 푛 푅푇[푄퐼푅푇] Definition (1) This definition leaves much to be desired for those who are not mathematicians. Thankfully, there are many examples which clearly show the intention of what k-anonymity is attempting to illustrate. 1 The classic example, which was first published with the definition of k-anonymity, was to think of a scenario where you wake up one morning to find an ambulance taking your neighbor, Bob, to a hospital. Let’s say, for example, that you are the nosey type, and you investigate what happened by entering Bob’s house. Conveniently, you find a sheet of hospital records on his counter that looks like Table I. Table 1 Hospital Records While the hospital data may look like it is anonymized due to the fact that it contains no names or obvious identifiers, you as Bob’s neighbor can leverage some knowledge that you have about him to infer which patient is him. Let’s say that you know that Bob is 30 years old because you attended his party last week, and you know his zip code because it is the same as yours. This pieces of information are Bob’s quasi-identifiers; individually they mean nothing, but taken together they identify Bob. With just this background knowledge you can find that there is only one match on the records that fits this description and you realize that a heart attack caused by chest pains is what caused your dear neighbor to be sent to the hospital. Now let’s apply the theory of k-anonymity to the data set. We want to make sure that there are at least a k amount of rows that are the same. TABLE II does this for us by instating a k value of 3. 2 Table 2 Hospital Records with k of 3 This means that, at a minimum, data has been anonymized so that there are 3 rows with matching data. Now, if you had found this sheet in Bob’s house, you would be able to narrow down your choices but you would not be able to say for certain whether Bob was sent to the hospital due to a heart disease or cancer. This implementation of k-anonymity successfully prevented Bob’s privacy from being breached. When data such as this is used in the real world a trusted third party, such as the hospital record keepers, sanitize the data before it is shared with any relevant parties such as researchers. This system works well as long as the third party who is scrubbing the data is trustworthy themselves. One of the big differences with digital data is that unless the user is specifically trying to anonymize their data, unwanted third parties could potentially perform these linking attacks based on a user’s browsing history. What’s more is that as more and more technical users take steps to anonymize their traffic, it also becomes easier to link the data from users who are not taking steps to protect their data because they stand out in the crowd. Returning to the original example of a hacker observing a person’s browsing history we find that a similar tactic can be employed. Had the user chosen to view a weather website that was encrypted, and if they had instead gone to a more general news site as opposed to the Denver Post, the attacker would not have learned that the person was from Colorado and could not have guessed that he knew this user in real life. This research is intending to find out what happens if the user chooses to take up the job of that trusted third party and only send data that would not violate a k-anonymity table. This would not be possible in the real world, after all, you wouldn’t tell your doctor that age is greater than 20 if he asked you how old you were. Yet, in the computer world it
Recommended publications
  • A Tool for Protecting Children's Privacy Online☆
    Decision Support Systems 54 (2013) 1161–1173 Contents lists available at SciVerse ScienceDirect Decision Support Systems journal homepage: www.elsevier.com/locate/dss POCKET: A tool for protecting children's privacy online☆ France Bélanger a, Robert E. Crossler b,⁎, Janine S. Hiller c, Jung-Min Park d, Michael S. Hsiao d a 850 Drillfield Drive, Suite 3007, Blacksburg, VA 24061–0101, USA b PO Box 9581, Mississippi State University, Mississippi State, MS 39762, USA c 850 Drillfield Drive, Suite 2120, Blacksburg, VA 24061–0221, USA d 1185 Perry St. Room 302, Blacksburg, VA 24061–0111, USA article info abstract Article history: Children's privacy in the online environment has become critical. Use of the Internet is increasing for com- Received 3 August 2011 mercial purposes, in requests for information, and in the number of children who use the Internet for casual Received in revised form 12 September 2012 web surfing, chatting, games, schoolwork, e-mail, interactive learning, and other applications. Often, websites Accepted 11 November 2012 hosting these activities ask for personal information such as name, e-mail, street address, and phone number. Available online 19 November 2012 In the United States, the children's online privacy protection act (COPPA) of 1998 was enacted in reaction to widespread collection of information from children and subsequent abuses identified by the Federal Trade Keywords: Information privacy Commission (FTC). COPPA is aimed at protecting a child's privacy by requiring parental consent before Privacy collecting information from children under the age of 13. To date, however, the business practices used COPPA and the technical approaches employed to comply with COPPA fail to protect children's online privacy effec- Children tively.
    [Show full text]
  • Privacy-Enhancing Technologies for the Internet
    Privacy-enhancing technologies for the Internet Ian Goldberg David Wagner Eric Brewer University of California, Berkeley iang,daw,brewer ¡ @cs.berkeley.edu Abstract ing privacy issues on the Internet, and Section 3 provides some relevant background. We then discuss Internet pri- The increased use of the Internet for everyday activi- vacy technology chronologically, in three parts: Section 4 ties is bringing new threats to personal privacy. This pa- describes the technology of yesterday, Section 5 explains per gives an overview of existing and potential privacy- today’s technology, and Section 6 explores the technology enhancing technologies for the Internet, as well as moti- of tomorrow. Finally, we conclude in Section 7. vation and challenges for future work in this field. 2. Motivation 1. Introduction The threats to one’s privacy on the Internet are two-fold: your online actions could be (1) monitored by unauthorized Recently the Internet has seen tremendous growth, with parties and (2) logged and preserved for future access many the ranks of new users swelling at ever-increasing rates. years later. You might not realize that your personal infor- This expansion has catapulted it from the realm of academic mation has been monitored, logged, and subsequently dis- research towards new-found mainstream acceptance and in- closed; those who would compromise your privacy have no creased social relevance for the everyday individual. Yet incentive to warn you. this suddenly increased reliance on the Internet has the po- The threat of long-term storage and eventual disclosure tential to erode personal privacies we once took for granted. of personal information is especially acute on the Internet.
    [Show full text]
  • Counter-Forensic Privacy Tools a Forensic Evaluation
    Counter-Forensic Privacy Tools A Forensic Evaluation Matthew Geiger, Lorrie Faith Cranor June 2005 CMU-ISRI-05-119 Institute for Software Research, International, Carnegie Mellon University School of Computer Science Carnegie Mellon University Pittsburgh, PA 15213-3890 Abstract Modern operating systems and the applications that run on them generate copious amounts of data about their users’ activity. Users are increasingly aware of their privacy exposure from these records and from digital artifacts that linger after files are “deleted” on computers they use. Efforts to redress this privacy exposure have spawned a range of counter-forensic privacy tools – software designed to irretrievably eliminate records of computer system usage and other sensitive data. In this paper, we use forensic tools and techniques to evaluate the effectiveness of six counter-forensic software packages. The results highlight some significant shortfalls in the implementation and approach of these tools, leading to privacy concerns about the exposure of sensitive data. The findings also raise questions about the level of privacy protection that is realistic to expect from these tools, and others that take a similar approach. TABLE OF CONTENTS Introduction ........................................................................................3 Background.........................................................................................5 Testing Methodology.............................................................................7 Privacy tool testing...............................................................................9
    [Show full text]
  • Data Mining and Internet Profiling: Emerging Regulatory and Technological Approaches Ira S
    Data Mining and Internet Profiling: Emerging Regulatory and Technological Approaches Ira S. Rubinstein, Ronald D. Lee, & Paul M. Schwartzt INTRODUCTION The 9/11 terrorists, before their deadly attacks, sought invisibility through integration into the society they hoped to destroy. In a similar fashion, the terrorists who carried out subsequent attacks in Madrid and London attempted to blend into their host lands. This strategy has forced governments, including the United States, to rethink counter- terrorism strategies and tools. One of the current favored strategies involves data mining. In its pattern-based variant, data mining searches select individuals for scru- tiny by analyzing large data sets for suspicious data linkages and pat- terns. Because terrorists do not "stand out," intelligence and law en- forcement agents want to do more than rely exclusively on investiga- tions of known suspects. The new goal is to search "based on the premise that the planning of terrorist activity creates a pattern or 'sig- nature' that can be found in the ocean of transaction data created in the course of everyday life."' Accordingly, to identify and preempt terrorist activity, intelligence agencies have begun collecting, retaining, and analyzing voluminous and largely banal transactional information about the daily activities of hundreds of millions of people. Private organizations have their own reasons for gathering wide- spread information about individuals. With the expansion of internet- based services, companies can track and document a broad range of people's online activities and can develop comprehensive profiles of these people. Advertisers and marketing firms likewise have strong incentives to identify and reach internet users whose profiles have certain demographic, purchasing behavior, or other characteristics.
    [Show full text]
  • Paper Title (Use Style: Paper Title)
    Asian Journal of Convergence in Technology Volume II Issue III Issn No.:2350-1146, I.F-2.71 Ransomware: A Cyber Extortion Miss. Harshada U. Salvi Mr. Ravindra V. Kerkar Department of MCA, Department of MCA, Finolex Academy of Management & Technology Finolex Academy of Management & Technology Ratnagiri, Maharashtra Ratnagiri, Maharashtra [email protected] [email protected] Abstract— feres with the ordinary boot procedure. It displays ran- som demand on screen [1]. Life as we know it today would be hampered without comput- ers. They controls just about everything from basic communi- II. BACKGROUND cation, finances and even medical science. As internet technol- ogies are advancing more rapidly, more businesses and indi- The first ever recognized ransomware was the viduals are storing sensitive data electronically. Internet has "AIDS (Aids Info Desk) Trojan” released in 1989. It is also become the hunting ground for criminals to make profit, cause known as "PC Cyborg Trojan". It was written by Dr. Joseph disturbance and bring down organizations and governments. Popp. It replaced the AUTOEXEC.BAT file and it would Ransomware is the latest trend that criminals are using for then count the number of times the machine had booted. extorting cash from the victims. It is malware that denies you Once this boot count reaches 90, it would then hide directo- access to your system until you pay ransom. ries and encrypt the names of all the files on the C: drive and make the system unusable. To regain access, the user would Keywords— Ransomware, Extortion, Malware, have to send $189 to PC Cyborg Corp.
    [Show full text]
  • The Blinding Effect of Security Hubris on Data Privacy
    The Blinding Effect of Security Hubris on Data Privacy Provided by THE BLINDING EFFECT OF SECURITY HUBRIS ON DATA PRIVACY Executive summary In the humble early days of the Internet, anonymity Which had us begging the question: was a comforting given. Most people used the world Why not? wide web to look up information or communicate with complete strangers under pseudonyms. They’d then An easy answer to that is, of course, that these return to their “real life” and conduct their business at practices are more difficult and cumbersome, so people the office, pay their taxes with an accountant, and buy avoid having to do them. However, if data privacy is clothes at the mall. so important to such a large number of respondents, and trust is so low in other companies to do it, why are Today, real life and Internet life are blended into one. people shirking the responsibility? More often than not, users must include their full names, addresses, payment details, and vital financial After analyzing responses from participants in data when they interact online. It’s no surprise, then, Generation Z up to baby boomers, our findings show that with each begrudging entry of sensitive personal that perceived confidence in privacy practices is higher information, not to mention each news story about than reality. We determine this gap between perception companies such as Facebook and Google abusing that and reality to be a result of security hubris. Because personal information, users are having an emotional users follow many of the perceived-as-easier security reaction to data privacy.
    [Show full text]
  • Privacy-By-Design Framework for Assessing Internet of Things Applications and Platforms
    Privacy-by-Design Framework for Assessing Internet of Things Applications and Platforms Charith Perera 1, Ciaran McCormick 1, Arosha K. Bandara 1, Blaine A. Price 1 Bashar Nuseibeh 1,2 1 The Open University, Milton Keynes, UK 2 Lero - The Irish Software Research Centre, Limerick, Ireland fi[email protected] ABSTRACT services are called ‘middleware’ as they sit ‘in the mid- The Internet of Things (IoT) systems are designed and dle’ , in a layer above the operating system and network- developed either as standalone applications from the ing software and below domain-specific applications [1]. ground-up or with the help of IoT middleware platforms. Our proposed privacy-by-design (PbD) framework can They are designed to support different kinds of scenarios, be used to assess both IoT applications and middleware such as smart homes and smart cities. Thus far, privacy platforms without any changes and agnostic to their dif- concerns have not been explicitly considered by IoT ap- ferences. Therefore, in this paper, we use the terms ‘ap- plications and middleware platforms. This is partly due plication’ and ‘middleware platform’ interchangeably. to the lack of systematic methods for designing privacy Our research is motivated by a lack of privacy protec- that can guide the software development process in IoT. tion features in both IoT applications and middleware In this paper, we propose a set of guidelines, a privacy- platforms. We also recognise that existing privacy-by- by-design framework, that can be used to assess privacy design frameworks do not provide specific guidance that capabilities and gaps of existing IoT applications as well can be used by software engineers to design IoT applica- as middleware platforms.
    [Show full text]
  • On Realistically Attacking Tor with Website Fingerprinting
    Proceedings on Privacy Enhancing Technologies ; 2016 (4):21–36 Tao Wang* and Ian Goldberg On Realistically Attacking Tor with Website Fingerprinting Abstract: Website fingerprinting allows a local, passive ob- rate (TPR) [3, 25] and cutting down the false positive rate server monitoring a web-browsing client’s encrypted channel (FPR) [24] to practical levels (far below 1%). Researchers to determine her web activity. Previous attacks have shown have also applied WF techniques to circuit fingerprinting, al- that website fingerprinting could be a threat to anonymity net- lowing adversaries to discover and identify Tor hidden ser- works such as Tor under laboratory conditions. However, there vices [14]. They have shown that these attacks are compu- are significant differences between laboratory conditions and tationally cheap and effective in the open-world setting [24]. realistic conditions. First, in laboratory tests we collect the However, some researchers remain unconvinced that these at- training data set together with the testing data set, so the train- tacks are effective in the wild [13, 21]. ing data set is fresh, but an attacker may not be able to maintain Indeed, the attacks have not been demonstrated to be ef- a fresh data set. Second, laboratory packet sequences corre- fective in the wild; they were proven only under laboratory spond to a single page each, but for realistic packet sequences conditions. Recently, Juarez et al. [13] identified significant the split between pages is not obvious. Third, packet sequences differences between attacks in the wild and attacks proven un- may include background noise from other types of web traffic.
    [Show full text]
  • THE DILEMMA for FUTURE COMMUNICATION TECHNOLOGIES: How to CONSTITUTIONALLY DRESS the CRYPTO-GENIE1
    THE DILEMMA FOR FUTURE COMMUNICATION TECHNOLOGIES: How TO CONSTITUTIONALLY DRESS THE CRYPTO-GENIE1 Jason Kerben "The proliferation of encryption of technology threat- munication.4 This system of communication has ens the ability of law enforcement and national security officials to protect the nation's citizens against ter- been used throughout history. One of the earliest rorists, as well as organized criminals, drug traffickers known examples of cryptography was used by Ju- 2 and other violent criminals." lius Caesar when he sent military messages to his "If the freedom of the press . [or freedom of speech] armies.5 Most cryptographic system have two perishes, it will not be by sudden death . It will be a 6 long time dying from a debilitating disease caused by a basic functions: encoding and decoding. The en- series of erosive measures, each of which, if examined coding function converts the normal data com- singly, would have a great deal to be said for it."3 monly known as "plaintext" into incompre- The preceding two statements epitomize the hensible data commonly known as "ciphertext."7 enduring struggle that has pitted the law enforce- The decoding function reverses the process, by ment community against those who are con- changing the "ciphertext" back into "plaintext." cerned with protecting their privacy interests. In order to perform these functions, a sequence The expanded use of advanced technologies in of bits, or "keys" must be obtained by the sender communications has propelled the cryptography and receiver of each message.9 The strength of debate into the spotlight. the coded communication is greatly dependent Cryptography uses codes to create secret com- upon the length of the key.' 0 This system is an I The term "crypto-genie" was apparently first used by au- metric cryptography is for an individual to choose two secret thor Steven Levy in 1994.
    [Show full text]
  • The Tor Network: a Global Inquiry Into the Legal Status of Anonymity Networks
    Washington University Global Studies Law Review Volume 11 Issue 3 2012 The Tor Network: A Global Inquiry into the Legal Status of Anonymity Networks Keith D. Watson Washington University School of Law Follow this and additional works at: https://openscholarship.wustl.edu/law_globalstudies Part of the Comparative and Foreign Law Commons, and the Internet Law Commons Recommended Citation Keith D. Watson, The Tor Network: A Global Inquiry into the Legal Status of Anonymity Networks, 11 WASH. U. GLOBAL STUD. L. REV. 715 (2012), https://openscholarship.wustl.edu/law_globalstudies/vol11/iss3/6 This Note is brought to you for free and open access by the Law School at Washington University Open Scholarship. It has been accepted for inclusion in Washington University Global Studies Law Review by an authorized administrator of Washington University Open Scholarship. For more information, please contact [email protected]. THE TOR NETWORK: A GLOBAL INQUIRY INTO THE LEGAL STATUS OF ANONYMITY NETWORKS I. INTRODUCTION On July 29, 2010, Jacob Appelbaum, the only known American member of the whistle-blowing organization WikiLeaks, was detained by U.S. agents as he attempted to reenter the country.1 The agents frisked him, searched his bag, photocopied his receipts, and inspected his laptop.2 Appelbaum was questioned about WikiLeaks, which leaked numerous classified government documents relating to the war in Afghanistan only days earlier.3 The agents also questioned him about his views on the United States‘ involvement in Afghanistan and Iraq and his knowledge of the whereabouts of WikiLeaks founder Julian Assange.4 Since then, 1. Nathaniel Rich, The Most Dangerous Man in Cyberspace, ROLLING STONE, Sept.
    [Show full text]
  • Common Security Misconceptions
    I just don’t get it: Common Security Misconceptions A THESIS SUBMITTED TO THE FACULTY OF THE GRADUATE SCHOOL OF THE UNIVERSITY OF MINNESOTA BY Maz Jindeel IN PARTIAL FULFILLMENT OF THE REQUIREMENTS FOR THE DEGREE OF MASTER OF SCIENCE Peter Peterson June 2019 © Maz Jindeel 2019 Acknowledgements Thanks to Aleksandar Straumann, Jennie Smith, Brandon Geraci, and Peter Peterson for coding and all your contributions to this project. Thanks also to Abigail Pederson and Jennie for your work on the open-ended questions. i Dedication To my family, roommates, and friends without whose support this wouldn’t have been possible. Also to the Bogle family and R’hllor, who helped me through the long nights. ii Abstract Many security mistakes are made because of some underlying misconception about computer security. These misconceptions can be remedied by developing curricula targeting them, but they must first be identified. This paper outlines our process for identifying common security misconceptions by surveying experts and coding their responses and the results of that process. We also present open-ended questions which are preliminary version of a computer security concept inventory based on these misconceptions. iii Contents Contents iv List of Figures vii 1 Introduction 1 2 Background 2 2.1 Concept Inventories ............................ 2 2.2 Survey ................................... 4 2.3 Coding ................................... 5 3 Related Work 7 3.1 Cybersecurity Assessment Tools (CATS) ................ 7 3.2 Introductory Programming Concept Inventory ............ 8 3.3 Digital Logic Concept Inventory ..................... 9 3.4 Operating Systems Concept Inventory ................. 9 3.5 Expert vs Non-Expert Security Advice and Practices ......... 10 4 Methods 12 4.1 Survey Design ..............................
    [Show full text]
  • Americans and Online Privacy: the System Is Broken
    CORE Metadata, citation and similar papers at core.ac.uk Provided by ScholarlyCommons@Penn University of Pennsylvania ScholarlyCommons Departmental Papers (ASC) Annenberg School for Communication 6-2003 Americans and Online Privacy: The ysS tem is Broken Joseph Turow University of Pennsylvania, [email protected] Follow this and additional works at: https://repository.upenn.edu/asc_papers Part of the Communication Commons Recommended Citation Turow, J. (2003). Americans and Online Privacy: The ysS tem is Broken. Annenberg Public Policy Center of the University of Pennsylvania, Retrieved from https://repository.upenn.edu/asc_papers/526 This paper is posted at ScholarlyCommons. https://repository.upenn.edu/asc_papers/526 For more information, please contact [email protected]. Americans and Online Privacy: The ysS tem is Broken Disciplines Communication | Social and Behavioral Sciences This report is available at ScholarlyCommons: https://repository.upenn.edu/asc_papers/526 Americans and Online Privacy The System is Broken By Joseph Turow June 2003 Americans and Online Privacy The System is Broken Overview ............................................................................................... 3 Background .......................................................................................... 5 The Study and the Population ........................................................... 12 Enduring Concerns about Web Privacy ............................................. 16 Not Understanding Data Flow ..........................................................
    [Show full text]