Randomized Algorithms RAJEEV MOTWANI Department of Computer Science, Stanford University, Stanford, California

Total Page:16

File Type:pdf, Size:1020Kb

Randomized Algorithms RAJEEV MOTWANI Department of Computer Science, Stanford University, Stanford, California Randomized Algorithms RAJEEV MOTWANI Department of Computer Science, Stanford University, Stanford, California PRABHAKAR RAGHAVAN IBM Almaden Research Center, San Jose, California Randomized algorithms, once viewed as HISTORY AND SOURCES a tool in computational number theory, The roots of randomized algorithms can have by now found widespread applica- be traced back to Monte Carlo methods tion. Growth has been fueled by the two used in numerical analysis, statistical major benefits of randomization: sim- physics, and simulation. In complexity plicity and speed. For many applica- theory, the notion of a probabilistic Tur- tions a randomized algorithm is the ing machine was proposed by de Leeuw fastest algorithm available, or the sim- et al. [1955] and was further explored in plest, or both. the pioneering work of Rabin [1963] and A randomized algorithm is an algo- Gill [1977]. The earliest examples of rithm that uses random numbers to in- concrete randomized algorithms ap- fluence the choices it makes in the peared in the work of Berlekamp [1970], course of its computation. Thus its be- Rabin [1976], and Solovay and Strassen havior (typically quantified as running [1977]. Rabin [1976] explicitly proposed time or quality of output) varies from randomization as an algorithmic tool one execution to another even with a using as examples problems in compu- fixed input. In the analysis of a random- tational geometry and in number the- ized algorithm we establish bounds on ory. At about the same time, Solovay the expected value of a performance and Strassen [1977] presented a ran- measure (e.g., the running time of the domized primality-testing algorithm; algorithm) that are valid for every in- these were predated by the randomized put; the distribution of the performance polynomial-factoring algorithm pre- measure is on the random choices made sented by Berlekamp [1970]. by the algorithm based on the random Since then, an impressive array of bits provided to it. techniques for devising and analyzing randomized algorithms have been devel- oped. The reader may refer to Karp Caveat: The analysis of randomized [1991], Maffioli et al. [1985], and Welsh algorithms should be distinguished [1983] for recent surveys of the research from the probabilistic or average-case into randomized algorithms. The proba- analysis of an algorithm, in which it is bilistic (or “average-case”) analysis of assumed that the input is chosen from a algorithms (sometimes also called “dis- probability distribution. In the latter tributional complexity”) is surveyed by case, the analysis would typically imply Johnson [1984a], and this is contrasted only that the algorithm is good for most with randomized algorithms in his fol- inputs but not for all. lowing bulletin [1984b]. The work of R. Motwani was supported by an Alfred P. Sloan Research Fellowship, an IBM Faculty Partnership Award, and NSF Young Investigator Award CCR-9357849, with matching funds from IBM, Schlumberger Foundation, Shell Foundation, and Xerox Corporation. Copyright © 1996, CRC Press. ACM Computing Surveys, Vol. 28, No. 1, March 1996 34 • Rajeev Motwani and Prabhakar Raghavan The recent book by the authors [Mot- population is representative of the pop- wani and Raghavan 1995] gives a com- ulation as a whole. Because computa- prehensive introduction to randomized tions involving small samples are inex- algorithms. pensive, their properties can be used to guide the computations of an algorithm PARADIGMS FOR RANDOMIZED attempting to determine some feature of ALGORITHMS the entire population. For instance, a simple randomized algorithm [Floyd In spite of the multitude of areas in and Rivest 1975] based on sampling which randomized algorithms find ap- finds the kth largest of n elements in plication, a handful of general princi- 1.5n 1 o(n) comparison steps, with ples underlie almost all of them. Using high probability. In contrast, it is the summary in Karp [1991], we known that any deterministic algorithm present these principles in the follow- must make at least 2n comparisons in ing. the worst case. Foiling an Adversary. In the classical worst-case analysis of deterministic al- Abundance of Witnesses. Often a gorithms, a lower bound is established computational problem can be reduced on the running time of algorithms by to finding a witness or a certificate that postulating an “adversary” that con- would efficiently verify an hypothesis. structs an input on which the algorithm (For example, to show that a number is fares poorly. The input thus constructed not prime, it suffices to exhibit any non- may be different for each deterministic trivial factor of that number.) For many algorithm. With a game-theoretic inter- problems, the witness lies in a search pretation of the relationship between an space that is too large to be searched algorithm and an adversary, we can exhaustively. However, if the search view a randomized algorithm as a prob- space were to contain a relatively large ability distribution on a set of determin- number of witnesses, a randomly chosen istic algorithms. (This observation un- element is likely to be a witness. Fur- derlies Yao’s [1977] adaptation of von ther, independent repetitions of the Neumann’s Mini-Max Theorem in game sampling reduce the probability that a theory into a technique for establishing witness is not found on any of the repe- limits on the performance improve- titions. The most striking examples of ments possible via the use of a random- this phenomenon occur in number the- ized algorithm.) Although the adversary ory. Indeed, the problem of testing a may be able to construct an input that given integer for primality has no foils one (or a small fraction) of the known deterministic polynomial-time deterministic algorithms in the set, it algorithm. There are, however, several may be impossible to devise a single randomized polynomial-time algorithms input that is likely to defeat a randomly [Solovay and Strassen 1978; Rabin chosen algorithm. For example, consider 1980] that will, on any input, correctly a uniform binary AND-OR tree with n perform this test with high probability. leaves. Any deterministic algorithm that evaluates such a tree can be forced Fingerprinting and Hashing. A fin- to read the Boolean values at every one gerprint is the image of an element from of the n leaves. However, there is a a (large) universe under a mapping into simple randomized algorithm [Snir another (smaller) universe. Finger- 1985] for which the expected number of prints obtained via random mappings leaves read on any input is O(n0.794). have many useful properties. For exam- ple, in pattern-matching applications Random Sampling. A pervasive theme [Karp and Rabin 1987] it can be shown in randomized algorithms is the idea that two strings are likely to be identi- that a small random sample from a cal if their fingerprints are identical; ACM Computing Surveys, Vol. 28, No. 1, March 1996 Randomized Algorithms •35 comparing the short fingerprints is con- the resources. This paradigm has found siderably faster than comparing the many interesting applications in paral- strings themselves. Another example is lel and distributed computing where re- hashing [Carter and Wegman 1979], source utilization decisions have to be where the elements of a set S (drawn made locally, without global knowledge. from a universe U) are stored in a table Consider packet routing in an n-node of size linear in uSu (even though uUu .. butterfly network. It is known that uSu) with the guarantee that the ex- V(=n) steps are required by any deter- pected number of elements in S mapped ministic oblivious algorithm (a class of to a given location in the table is O(1). simple routing algorithms in which the This leads to efficient schemes for decid- route followed by a packet is indepen- ing membership in S. Random finger- dent of the routes of other packets). In prints have found a variety of applica- contrast, based on ideas of Valiant tions in generating pseudorandom [1982], it has been shown [Aleliunas numbers and complexity theory (for in- 1982; Upfal 1984] that there is a ran- stance, the verification of algebraic domized algorithm that terminates in identities [Freivalds 1977]). O(log n) steps with high probability. Random Reordering. A large class of Rapidly Mixing Markov Chains. In problems has the property that a rela- counting problems, the goal is to deter- tively naive algorithm A can be shown mine the number of combinatorial ob- to perform extremely well provided the jects with a specified property. When input data is presented in a random the space of objects is large, an appeal- order. Although A may have poor worst- ing solution is the use of the Monte case performance, randomly reordering Carlo approach of determining the num- the input data ensures that the input is ber of desired objects in a random sam- unlikely to be in one of the orderings ple of the entire space. In a number of that is pathological for A. The earliest cases, it can be shown that picking a instance of this phenomenon can be uniform random sample is as difficult as found in the behavior of the Quicksort the counting problem itself. A particu- algorithm [Hoare 1962]. The random re- larly successful technique for dealing ordering approach has been particularly with such problems is to generate near- successful in tackling problems in data uniform random samples by defining a structures and computational geometry. Markov chain on the elements of the For instance, there are simple algo- population, and showing that a short rithms for computing the convex hull of random walk using this Markov chain is n points in the plane that process the likely to sample the population uni- input one point at a time. Such algo- formly. This method is at the core of a rithms are doomed to take V(n2) steps if number of algorithms used in statistical the input points are presented in an physics [Sinclair 1992].
Recommended publications
  • Challenges in Web Search Engines
    Challenges in Web Search Engines Monika R. Henzinger Rajeev Motwani* Craig Silverstein Google Inc. Department of Computer Science Google Inc. 2400 Bayshore Parkway Stanford University 2400 Bayshore Parkway Mountain View, CA 94043 Stanford, CA 94305 Mountain View, CA 94043 [email protected] [email protected] [email protected] Abstract or a combination thereof. There are web ranking optimiza• tion services which, for a fee, claim to place a given web site This article presents a high-level discussion of highly on a given search engine. some problems that are unique to web search en• Unfortunately, spamming has become so prevalent that ev• gines. The goal is to raise awareness and stimulate ery commercial search engine has had to take measures to research in these areas. identify and remove spam. Without such measures, the qual• ity of the rankings suffers severely. Traditional research in information retrieval has not had to 1 Introduction deal with this problem of "malicious" content in the corpora. Quite certainly, this problem is not present in the benchmark Web search engines are faced with a number of difficult prob• document collections used by researchers in the past; indeed, lems in maintaining or enhancing the quality of their perfor• those collections consist exclusively of high-quality content mance. These problems are either unique to this domain, or such as newspaper or scientific articles. Similarly, the spam novel variants of problems that have been studied in the liter• problem is not present in the context of intranets, the web that ature. Our goal in writing this article is to raise awareness of exists within a corporation.
    [Show full text]
  • Stanford University's Economic Impact Via Innovation and Entrepreneurship
    Full text available at: http://dx.doi.org/10.1561/0300000074 Impact: Stanford University’s Economic Impact via Innovation and Entrepreneurship Charles E. Eesley Associate Professor in Management Science & Engineering W.M. Keck Foundation Faculty Scholar, School of Engineering Stanford University William F. Miller† Herbert Hoover Professor of Public and Private Management Emeritus Professor of Computer Science Emeritus and former Provost Stanford University and Faculty Co-director, SPRIE Boston — Delft Full text available at: http://dx.doi.org/10.1561/0300000074 Contents 1 Executive Summary2 1.1 Regional and Local Impact................. 3 1.2 Stanford’s Approach ..................... 4 1.3 Nonprofits and Social Innovation .............. 8 1.4 Alumni Founders and Leaders ................ 9 2 Creating an Entrepreneurial Ecosystem 12 2.1 History of Stanford and Silicon Valley ........... 12 3 Analyzing Stanford’s Entrepreneurial Footprint 17 3.1 Case Study: Google Inc., the Global Reach of One Stanford Startup ............................ 20 3.2 The Types of Companies Stanford Graduates Create .... 22 3.3 The BASES Study ...................... 30 4 Funding Startup Businesses 33 4.1 Study of Investors ...................... 38 4.2 Alumni Initiatives: Stanford Angels & Entrepreneurs Alumni Group ............................ 44 4.3 Case Example: Clint Korver ................. 44 5 How Stanford’s Academic Experience Creates Entrepreneurs 46 Full text available at: http://dx.doi.org/10.1561/0300000074 6 Changing Patterns in Entrepreneurial Career Paths 52 7 Social Innovation, Non-Profits, and Social Entrepreneurs 57 7.1 Case Example: Eric Krock .................. 58 7.2 Stanford Centers and Programs for Social Entrepreneurs . 59 7.3 Case Example: Miriam Rivera ................ 61 7.4 Creating Non-Profit Organizations ............. 63 8 The Lean Startup 68 9 How Stanford Supports Entrepreneurship—Programs, Cen- ters, Projects 77 9.1 Stanford Technology Ventures Program .........
    [Show full text]
  • SIGMOD Flyer
    DATES: Research paper SIGMOD 2006 abstracts Nov. 15, 2005 Research papers, 25th ACM SIGMOD International Conference on demonstrations, Management of Data industrial talks, tutorials, panels Nov. 29, 2005 June 26- June 29, 2006 Author Notification Feb. 24, 2006 Chicago, USA The annual ACM SIGMOD conference is a leading international forum for database researchers, developers, and users to explore cutting-edge ideas and results, and to exchange techniques, tools, and experiences. We invite the submission of original research contributions as well as proposals for demonstrations, tutorials, industrial presentations, and panels. We encourage submissions relating to all aspects of data management defined broadly and particularly ORGANIZERS: encourage work that represent deep technical insights or present new abstractions and novel approaches to problems of significance. We especially welcome submissions that help identify and solve data management systems issues by General Chair leveraging knowledge of applications and related areas, such as information retrieval and search, operating systems & Clement Yu, U. of Illinois storage technologies, and web services. Areas of interest include but are not limited to: at Chicago • Benchmarking and performance evaluation Vice Gen. Chair • Data cleaning and integration Peter Scheuermann, Northwestern Univ. • Database monitoring and tuning PC Chair • Data privacy and security Surajit Chaudhuri, • Data warehousing and decision-support systems Microsoft Research • Embedded, sensor, mobile databases and applications Demo. Chair Anastassia Ailamaki, CMU • Managing uncertain and imprecise information Industrial PC Chair • Peer-to-peer data management Alon Halevy, U. of • Personalized information systems Washington, Seattle • Query processing and optimization Panels Chair Christian S. Jensen, • Replication, caching, and publish-subscribe systems Aalborg University • Text search and database querying Tutorials Chair • Semi-structured data David DeWitt, U.
    [Show full text]
  • The Best Nurturers in Computer Science Research
    The Best Nurturers in Computer Science Research Bharath Kumar M. Y. N. Srikant IISc-CSA-TR-2004-10 http://archive.csa.iisc.ernet.in/TR/2004/10/ Computer Science and Automation Indian Institute of Science, India October 2004 The Best Nurturers in Computer Science Research Bharath Kumar M.∗ Y. N. Srikant† Abstract The paper presents a heuristic for mining nurturers in temporally organized collaboration networks: people who facilitate the growth and success of the young ones. Specifically, this heuristic is applied to the computer science bibliographic data to find the best nurturers in computer science research. The measure of success is parameterized, and the paper demonstrates experiments and results with publication count and citations as success metrics. Rather than just the nurturer’s success, the heuristic captures the influence he has had in the indepen- dent success of the relatively young in the network. These results can hence be a useful resource to graduate students and post-doctoral can- didates. The heuristic is extended to accurately yield ranked nurturers inside a particular time period. Interestingly, there is a recognizable deviation between the rankings of the most successful researchers and the best nurturers, which although is obvious from a social perspective has not been statistically demonstrated. Keywords: Social Network Analysis, Bibliometrics, Temporal Data Mining. 1 Introduction Consider a student Arjun, who has finished his under-graduate degree in Computer Science, and is seeking a PhD degree followed by a successful career in Computer Science research. How does he choose his research advisor? He has the following options with him: 1. Look up the rankings of various universities [1], and apply to any “rea- sonably good” professor in any of the top universities.
    [Show full text]
  • Practical Considerations in Benchmarking Digital Testing Systems
    Foreword The papers in these proceedings were presented at the 39th Annual Symposium on Foundations of Computer Science (FOCS ‘98), sponsored by the IEEE Computer Society Technical Committee on Mathematical Foundations of Computing. The conference was held in Palo Alto, California, November g-11,1998. The program committee consisted of Miklos Ajtai (IBM Almaden), Mihir Bellare (UC San Diego), Allan Borodin (Toronto), Edith Cohen (AT&T Labs), Sally Goldman (Washington), David Karger (MIT), Jon Kleinberg (Cornell), Rajeev Motwani (chair, Stanford)), Seffi Naor (Technion), Christos Papadimitriou (Berkeley), Toni Pitassi (Arizona), Dan Spielman (MIT), Eli Upfal (Brown), Emo Welzl (ETH Zurich), David Williamson (IBM TJ Watson), and Frances Yao (Xerox PARC). The program committee met on June 26-28, 1998, and selected 76 papers from the 204 detailed abstracts submitted. The submissions were not refereed, and many of them represent reports of continuing research. It is expected that most of these papers will appear in a more complete and polished form in scientific journals in the future. The program committee selected two papers to jointly receive the Machtey Award for the best student-authored paper. These two papers were: “A Factor 2 Approximation Algorithm for the Generalized Steiner Network Problem” by Kamal Jain, and “The Shortest Vector in a Lattice is Hard to Approximate to within Some Constant” by Daniele Micciancio. There were many excellent candidates for this award, each one deserving. At this conference we organized, for the first time, a set of three tutorials: “Geometric Computation and the Art of Sampling” by Jiri Matousek; “Theoretical Issues in Probabilistic Artificial Intelligence” by Michael Kearns; and, “Information Retrieval on the Web” by Andrei Broder and Monika Henzinger.
    [Show full text]
  • The Pagerank Citation Ranking: Bringing Order to the Web
    The PageRank Citation Ranking: Bringing Order to the Web Marlon Dias [email protected] Information Retrieval DCC/UFMG - 2017 Introduction Paper: The PageRank Citation Ranking: Bringing Order to the Web, 1999 Authors: Lawrence Page, Sergey Brin, Rajeev Motwani, Terry Winograd Page and Brin were MS students at Stanford They founded Google in September, 98. Most of this presentation is based on the original paper (link) The Initiative's focus is to dramatically “ advance the means to collect, store, and organize information in digital forms, and make it available for searching, retrieval, and processing via communication networks -- all in user-friendly ways. Stanford WebBase project Pagerank Motivation ▪ Web is vastly large and heterogeneous ▫ Original paper's estimation were over 150 M pages and 1.7 billion of links ▪ Pages are extremely diverse ▫ Ranging from "What does the fox say? " to journals about IR ▪ Web Page present some "structure" ▫ Pagerank takes advantage of links structure Pagerank Motivation ▪ Inspiration: Academic citation ▪ Papers ▫ are well defined units of work ▫ are roughly similar in quality ▫ are used to extend the body of knowledge ▫ can have their "quality" measured in number of citations Pagerank Motivation ▪ Web pages, on the other hand ▫ proliferate free of quality control or publishing costs ▫ huge numbers of pages can be created easily ▸ artificially inflating citation counts ▫ They vary on much wider scale than academic papers in quality, usage, citations and length Pagerank Motivation A research article A random archived about the effects of message posting cellphone use on asking an obscure driver attention is very question about an IBM different from an computer is very advertisement for a different from the IBM particular cellular home page provider The average web page quality experienced “ by a user is higher than the quality of the average web page.
    [Show full text]
  • The Pagerank Citation Ranking: Bringing Order to the Web Paper Review
    The PageRank Citation Ranking: Bringing Order to the Web Paper Review Anand Singh Kunwar 1 Introduction The PageRank [1] algorithm decribes a way to compute ranks or ratings of web pages. This comes from a simple intuition. Consider the web to be a directed graph with each node being a web page. Each web page has certain links to it and certain links from it. Let these links be edges of our directed graph. Something like what we can see in the following graphical representation of 3 pages. a c b The pages a, b and c have some outgoing and incoming edges. We compute the total PageRank of these pages by iterating over the following algorithm. X Rank(v) Rank(u) = c + cE(u) Nv v2Lu Here Nv are number of pages being linked by v, Lu is the set of pages which have links to u, E(u) is some vector that correspond to the source of rank and c is the normalising constant. The reason why the E(u) was introduced was for cases when a loop has no out-edges. In this case, our ranks of nodes inside this loop would blow up. The paper also provides us with a real life intuition for this. Authors provide us with a random surfer model. This model states that a web surfer will click consecutive pages like a random walk on graphs. However, in case this surfer encounters a loop of webpages with no other outgoing webpage than the ones in the loop, it is highly unlikely that this surfer will continue.
    [Show full text]
  • David Karger Rajeev Motwani Y Madhu Sudan Z
    Approximate Graph Coloring by Semidenite Programming y z David Karger Rajeev Motwani Madhu Sudan Abstract We consider the problem of coloring k colorable graphs with the fewest p ossible colors We present a randomized p olynomial time algorithm which colors a colorable graph on n vertices 12 12 13 14 with minfO log log n O n log ng colors where is the maximum degree of any vertex Besides giving the b est known approximation ratio in terms of n this marks the rst nontrivial approximation result as a function of the maximum degree This result can 12 12k b e generalized to k colorable graphs to obtain a coloring using minfO log log n 12 13(k +1) O n log ng colors Our results are inspired by the recent work of Go emans and Williamson who used an algorithm for semidenite optimization problems which generalize lin ear programs to obtain improved approximations for the MAX CUT and MAX SAT problems An intriguing outcome of our work is a duality relationship established b etween the value of the optimum solution to our semidenite program and the Lovasz function We show lower b ounds on the gap b etween the optimum solution of our semidenite program and the actual chromatic numb er by duality this also demonstrates interesting new facts ab out the function MIT Lab oratory for Computer Science Cambridge MA Email kargermitedu URL httptheorylcsmitedukarger Supp orted by a Hertz Foundation Graduate Fellowship by NSF Young In vestigator Award CCR with matching funds from IBM Schlumb erger Foundation Shell Foundation and Xerox Corp oration
    [Show full text]
  • Rajeev Motwani
    RAJEEV MOTWANI A Memorial Celebration Memorial Church, Stanford University ~ September 25, 2009 Memorial Church, Stanford University ~ September 25, 2009 Pre-Service Music Professor Donald Knuth, Organist Pastorale in F Major ~ J.S. Bach UÊ Adagio and Allegro for Mechanical Organ ~ Mozart Blue Rondo a la Turk ~ Dave Brubeck U Sonata I ~ Paul Hindemith U Toccata in D Minor ~ J.S. Bach Take Five ~ Paul Desmond Rabbi Patricia Karlin-Neumann, Senior Associate Dean for Religious Life Stanford University Welcome John L. Hennessy, President, Stanford University Opening Blessings Robert P. Goldman, Professor of Sanskrit, University of California Berkeley Music Song of the Soul ................................................................................................... Naitri Jadeja Family Introductions Ram Shriram Reflections Sergey Brin UÊ Jennifer Widom UÊ Sep Kamvar UÊ Gautam Bhargava UÊ Lakshmi Pratury Closing Prayer Robert P. Goldman, Professor of Sanskrit, University of California Berkeley Closing Remarks Ron Conway Music Grand Chœur Dialogué ~ E. Gigout Cover photo ©2009 Nicole Scarborough The Song of the Soul UÊ Ê>ÊÌÊÌ iÊ`]ÊÀÊÌiiVÌ]ÊÀÊi}]Ê UÊ Ê`ÊÌÊ >ÛiÊvi>ÀÊvÊ`i>Ì ]Ê>ÃÊÊ`ÊÌÊ >ÛiÊ`i>Ì ° nor the reflections of inner self; I am not the five senses I have no separation from my true self, I am beyond that. no doubt about my existence, I am not the earth, nor the fire, nor the wind nor have I discrimination on the basis of birth. I am, indeed, that eternal knowing and bliss, that I have no father or mother, nor did I have a birth. love and pure consciousness. I am not the relative, nor the friend, nor the guru, nor the disciple.
    [Show full text]
  • Vcs Aim to Out-Angel the Angels
    NEWS ANALYSIS April 2, 2007, 12:00AM EST VCs Aim to Out-Angel the Angels Responding to the emergence of a new breed of wealthy investor, venture capitalists are boosting their early-stage investments in startups by Aaron Ricadela In October, as startup Jaxtr hit up venture capital firms for its first round of funding, it landed an unusual arrangement. Instead of taking a few million in cash from a firm that would hope to one day book a fat return, Jaxtr took a loan—just $1.5 million, from no less than four VC firms and three angel investors. None got the usual perk of a seat on Jaxtr's board. The result is plenty of independence for Jaxtr, a maker of software that routes calls from blogs and MySpace profiles to cell phones. "You're still basically on your own," says Jaxtr Chief Executive Konstantin Guericke, one of the co-founders of networking site LinkedIn. Jaxtr's tale illustrates the new calculus governing high-tech venture capital. For years, angel investors and traditional venture firms existed in a sort of symbiosis. Wealthy tech-industry veterans willing to open their checkbooks for $100,000 or so—the angels, as they're known—could bootstrap promising young companies before serious money, to the tune of six or more figures, from venture firms arrived. Angling for a slice of Jaxtr, however, were both groups: Mayfield Fund, Draper Fisher Jurvetson, Draper Richards, and the Founders Fund on one hand; angel investors Ron Conway, Reid Hoffman, and Rajeev Motwani on the other. "The company was a bit in the driver's seat," one investor says.
    [Show full text]
  • Distributed Clustering Algorithm for Large Scale Clustering Problems
    IT 13 079 Examensarbete 30 hp November 2013 Distributed clustering algorithm for large scale clustering problems Daniele Bacarella Institutionen för informationsteknologi Department of Information Technology Abstract Distributed clustering algorithm for large scale clustering problems Daniele Bacarella Teknisk- naturvetenskaplig fakultet UTH-enheten Clustering is a task which has got much attention in data mining. The task of finding subsets of objects sharing some sort of common attributes is applied in various fields Besöksadress: such as biology, medicine, business and computer science. A document search engine Ångströmlaboratoriet Lägerhyddsvägen 1 for instance, takes advantage of the information obtained clustering the document Hus 4, Plan 0 database to return a result with relevant information to the query. Two main factors that make clustering a challenging task are the size of the dataset and the Postadress: dimensionality of the objects to cluster. Sometimes the character of the object makes Box 536 751 21 Uppsala it difficult identify its attributes. This is the case of the image clustering. A common approach is comparing two images using their visual features like the colors or shapes Telefon: they contain. However, sometimes they come along with textual information claiming 018 – 471 30 03 to be sufficiently descriptive of the content (e.g. tags on web images). Telefax: The purpose of this thesis work is to propose a text-based image clustering algorithm 018 – 471 30 00 through the combined application of two techniques namely Minhash Locality Sensitive Hashing (MinHash LSH) and Frequent itemset Mining. Hemsida: http://www.teknat.uu.se/student Handledare: Björn Lyttkens Lindén Ämnesgranskare: Kjell Orsborn Examinator: Ivan Christoff IT 13 079 Tryckt av: Reprocentralen ITC ”L’arte rinnova i popoli e ne rivela la vita.
    [Show full text]
  • Arxiv:1907.02513V2 [Cs.LG] 28 Jun 2021
    Locally Private k-Means Clustering Uri Stemmer [email protected] Ben-Gurion University, Beer-Sheva, Israel Google Research, Tel Aviv, Israel Abstract We design a new algorithm for the Euclidean k-means problem that operates in the local model of differential privacy. Unlike in the non-private literature, differentially private algorithms for the k-means objective incur both additive and multiplicative errors. Our algorithm significantly reduces the additive error while keeping the multiplicative error the same as in previous state-of-the-art results. Specifically, on a database of size n, our algorithm guarantees O(1) multiplicative error and n1/2+a additive error for an arbitrarily small constant a > 0. All previous algorithms≈ in the local model had additive error n2/3+a. Our techniques extend to k-median clustering. ≈ We show that the additive error we obtain is almost optimal in terms of its dependency on the database size n. Specifically, we give a simple lower bound showing that every locally-private algorithm for the k-means objective must have additive error at least √n. ≈ Keywords: Differential privacy, local model, clustering, k-means, k-median. 1. Introduction In center-based clustering, we aim to find a “best” set of centers (w.r.t. some cost function), and then partition the data points into clusters by assigning each data point to its nearest center. With over 60 years of research, center-based clustering is an intensively-studied key- problem in unsupervised learning (see Hartigan (1975) for a textbook). One of the most well-studied problems in this context is the Euclidean k-means problem.
    [Show full text]