Deterministic and Efficient Minimal Perfect Hashing Schemes

Total Page:16

File Type:pdf, Size:1020Kb

Deterministic and Efficient Minimal Perfect Hashing Schemes Deterministic and efficient minimal perfect hashing schemes Leandro M. Zatesko 1 Jair Donadelli 2 Abstract: This paper presents deterministic versions to the hashing schemes of Botelho, Kohayakawa and Ziviani (2005) and Botelho, Pagh and Ziviani (2007), also proves a statement left as open problem in the former work, related to the correct- ness proof and to the complexity analysis of their scheme. Our deterministic variants have been implemented and executed over datasets with up to 25,000,000 keys and have brought equivalent performance results between the deterministic and the origi- nal randomized algorithms. Resumo: Neste trabalho apresentamos versões determinísticas para os esquemas de hashing de Botelho, Kohayakawa e Ziviani (2005) e de Botelho, Pagh e Ziviani (2007). Também respondemos a um problema deixado em aberto no primeiro dos trabalhos, relacionado à prova da corretude e à análise de complexidade do esquema por eles proposto. As versões determinísticas desenvolvidas foram implementadas e testadas sobre conjuntos de dados com até 25.000.000 de chaves, e os resultados verificados se mostraram equivalentes aos dos algoritmos aleatorizados originais. 1 Introduction A minimal perfect hashing scheme, as defined in [1, 2, 3], is an algorithm that, given a set S with n keys from an universe U, constructs a hash function h: U →{0,...,n−1} which maps without collision S to {0,...,n − 1}. We are interested only in hashing schemes whose outputs are hash functions with O(1) lookup time. For our purposes, every key x is assumed to be a chain of at most L symbols taken from a finite alphabet Σ, for a fixed constant L. As an example, keys can be URLs of length at most L which we are trying to map to memory addresses, and the alphabet would then contain decimal digits, Latin letters and some special characters like / and ?. Hashing is a widely studied topic in Computer Science. Mapping n objects bijectively to hash table addresses {0,...,n − 1} is a very often problem, and minimal perfect hash functions, in particular, are useful in situations related to “efficient storage and fast retrieval of items from static sets, such as words in natural languages, reserved words in programming 1Departamento de Informática, UFPR, Brazil, PO Box 19081 [email protected] 2Centro de Matemática, Computação e Cognição, UFABC, Brazil, CEP 09210-170 [email protected] Deterministic and efficient minimal perfect hashing schemes languages or interactive systems, universal resource locations (URLs) in Web search engines, or item sets in data mining techniques” [1]. Derandomization is an important subject of Computational Complexity and a way to understand whether randomness in algorithms is necessary. Formally, a central problem about complexity of randomized algorithms is “P = BPP?”. For example, the problem of polyno- mial identity testing is in BPP, which means that it is solvable by a polynomial-time Monte Carlo algorithm [4], an algorithm whose answer can be wrong with bounded probability. De- randomizing polynomial identity tests has deep consequences in Computational Complexity [5]. On the other hand, the celebrated polynomial-time deterministic algorithm for primality testing [6] is a successful derandomization of a Monte Carlo polynomial-time algorithm [7]. Differently, the hashing schemes we study are known as Las Vegas algorithms [4], which means that their answer is always right, but the time complexity is a random variable. Prob- lems solvable by Las Vegas algorithms with expected polynomial-time complexity form the class ZPP ⊆ BPP. We shall present derandomized versions to the hashing schemes of Botelho, Ko- hayakawa and Ziviani (2005) and of Botelho, Pagh and Ziviani (2007), from now on referred as BMZ and BDZ, respectively. These schemes are Las Vegas algorithms that, given a set with n keys, construct in expected time O(n) a hash function which in O(1)-evaluation time maps without collision the keys to the set {0,...,n − 1}. The problem of constructing a minimal perfect O(1)-evaluation time hash function given a set with n keys is, of course, in P [8], therefore our work just shows that these very practical algorithms didn’t need to be randomized to achieve a good average performance. Actually we derandomize the schemes in a very simple manner, and the resulting algorithms are schemes of O(n) average-case time complexity. Additionally, we also give a proof for a question left open in [1] (Equation 2 below), closing the complexity analysis and the correctness proof of the BMZ scheme. In what follows, unless stated otherwise, we use the term graph to refer to a simple graph, that is an unweighted, undirected graph containing no loops or multiple edges. The term critical subgraph of a graphrefers to the maximal subgraphwith minimumdegree δ > 2. The term parent of a vertex in a graph search is used according to the traditional meaning, as could be found in [9], in any graph search the term tree edge refers to an edge in which one endpoint is the parent of the other, and the term back edge refers to an edge which is not a tree edge. Also, we write a(n) ≈ b if a(n) → b as n → ∞. This paper is organized as follows. In Section 2 we shall give a brief review on related works, in Section 3 we shall present deterministic versions of the schemes BMZ and BDZ, in Section 4 we shall prove a graph theoretical result related to the complexity analysis and to the correctness proof of BMZ scheme, in Section 5 we shall give performance comparisons between the randomized and derandomized schemes and finally in Section 6 we shall close with some considerations about hashing and our results. RITA • Volume 20 • Número 2 • 2013 57 Deterministic and efficient minimal perfect hashing schemes 2 Related works It is known that finding a perfect hash function for sets with n keys cannot be done in o(n) time [3], although many O(n)-time perfect hashing schemes are known from litera- ture [2, 10, 11, 12]. For example, the randomized hashing scheme presented in [3], which maps the n keys to the edges of an acyclic graph on 2.09n vertices and then uses a depth-first search to label the edges with the values 1,...,n, constructs a minimal perfect hash function in O(n) expected time. The constructed hash function requires O(nlogn) bits to be stored, and this amount of space is proportional to the size of the graph. This important hashing scheme inspired the BMZ scheme [1], which, allowing the graph to be cyclic, reduces the number of vertices to 1.15n. In 1984 Fredman and Komlós [13] proved that nlge + lglgu + O(logn) is a lower bound for the space of O(1)-evaluation time hash functions built by a minimal perfect scheme on an universe with u objects. Remark that this means about lge =∼ 1.443 bit per key. In ad- dition, Melhorn [14] presented in 1984 a theoretical scheme with which proved the lower bound to be tight. His scheme however was an exponential-time algorithm. Both schemes of [3] and [1], although perfect, minimal, practical and efficient in time, construct hash func- tions represented by an undesirable amount of space, if we take into account that it is possi- ble to have minimal perfect hashing schemes whose output hash functions require only O(n) bits [14]. Even the space of the latter being smaller than that of the former, it does not escape from the asymptotic O(nlogn). The practical, minimal, perfect and O(n)-expected time BDZ scheme [2], presented in 2007, not only achieves the O(n) space to the representation of the constructed hash function but also gets this amount to be 2.62n, just a little greater than 1.443n. More recently, better practical hashing schemes were proposed [15, 16]. The one by Belazzougui, Botelho and Di- etzfelbinger [15], known as CHD, generates in O(n) time a minimal perfect O(1)-evaluation time hash function which requires just about 2.06 bits per key. The authors’ experiment show that CHD is more efficient than other schemes concerning to running and evaluation time too. Nevertheless, one can still set CHD parameters to obtain better results according to each application. For example, if one does not need the hash function to be minimal, CHD can map without collision the n keys to addresses 1,...,m, where m = 1.23n, in a way that the generated hash function requires about 1.4 bits per key. Evenmore, if one sets m = 2n, one gets 0.67 bits per key. CHD also shows up very efficient for k-perfect hashing, where at most k keys can be mapped to the same address. BMZ, BDZ, CHD and other hashing schemes were implemented in CMPH (C Minimal Perfect Hashing) library, developed and maintained at SourceForge.net by the authors themselves. Below, for sake of completeness, we give a short review of the BMZ and BDZ schemes, though we strongly recommend [1, 2] for more details. 58 RITA • Volume 20 • Número 2 • 2013 Deterministic and efficient minimal perfect hashing schemes 2.1 BMZ This is a hashing scheme proposed in [1] which constructs in O(n) expected time a minimal perfect hash function given a set S with n keys. It maps S to the set E(G) of the n edges of a graph G on 1.15n vertices and then tries to find a way to assign labels to the vertices so that the edges labels, defined to be the sum of endpoints labels, will be the whole set {0,...,n − 1}. Two properties about G are required: Property P1: The critical subgraph of G, denoted by Gcrit, must be connected.
Recommended publications
  • Skip B-Trees
    Skip B-Trees Ittai Abraham1, James Aspnes2?, and Jian Yuan3 1 The Institute of Computer Science, The Hebrew University of Jerusalem, [email protected] 2 Department of Computer Science, Yale University, [email protected] 3 Google, [email protected] Abstract. We describe a new data structure, the Skip B-Tree that combines the advantages of skip graphs with features of traditional B-trees. A skip B-Tree provides efficient search, insertion and deletion operations. The data structure is highly fault tolerant even to adversarial failures, and allows for particularly simple repair mechanisms. Related resource keys are kept in blocks near each other enabling efficient range queries. Using this data structure, we describe a new distributed peer-to-peer network, the Distributed Skip B-Tree. Given m data items stored in a system with n nodes, the network allows to perform a range search operation for r consecutive keys that costs only O(logb m + r/b) where b = Θ(m/n). In addition, our distributed Skip B-tree search network has provable polylogarithmic costs for all its other basic operations like insert, delete, and node join. To the best of our knowledge, all previous distributed search networks either provide a range search operation whose cost is worse than ours or may require a linear cost for some basic operation like insert, delete, and node join. ? Supported in part by NSF grants CCR-0098078, CNS-0305258, and CNS-0435201. 1 Introduction Peer-to-peer systems provide a decentralized way to share resources among machines. An ideal peer-to-peer network should have such properties as decentralization, scalability, fault-tolerance, self-stabilization, load- balancing, dynamic addition and deletion of nodes, efficient query searching and exploiting spatial as well as temporal locality in searches.
    [Show full text]
  • Grad Cohort 2015
    2015 GradCohort www.cra-w.org @CRAWomen SAN FRANCISCO APRIL 10-11, 2015 Hyatt Regency San Francisco Dear Grad Cohort Participant, We welcome you to the 2015 CRA-Women Graduate 2015 Cohort Workshop! The next two days are filled with sessions where 25 senior computing researchers and professionals will be sharing their strategies and experiences to help increase your graduate school and career success. There are also plenty of opportunities to meet and network with these successful senior women as well as graduate students from other Grad universities. We hope that you will take the utmost advantage of this unique experience by actively participating in discussions, developing peer networks, and building mentoring relationships. The 2015 CRA-Women Graduate Cohort Workshop is Cohort made possible through generous contributions by Microsoft Research, Computing Research Association, National Science Foundation, Google, a private foundation, Alfred P. Sloan Foundation, Two Sigma, SAN FRANCISCO Intel Corporation, ACM SIGARCH, ACM SIGMICRO, ACM SIGGRAPH, ACM SIGOPS, ACM SIGPLAN, ACM SIGSOFT, Yahoo! Labs, Facebook, IBM and in some cases department funds from participating universities/ institutions. Please join us in thanking them for their kind support. We hope that you take home many nuggets and connections from this workshop to help you in your journey to make an impact in the world through computing. Be ready to be inspired, learn, and meet many interesting technical women. Sincerely, — Lori Clarke, Sandhya Dwarkadas, and Ayanna Howard CO-CHAIRS,
    [Show full text]
  • The CNN Problem and Other K-Server Variants
    The CNN Problem and Other k-Server Variants Elias Koutsoupias 1 Department of Informatics, University of Athens, Athens, Greece and Computer Science Department, University of California, Los Angeles, USA David Scot Taylor Department of Computer Science, San Jos´eState University, San Jos´e, California, USA Abstract We study several interesting variants of the k-server problem. In the cnn problem, one server services requests in the Euclidean plane. The difference from the k- server problem is that the server does not have to move to a request, but it has only to move to a point that lies in the same horizontal or vertical line with the request. This, for example, models the problem faced by a crew of a certain news network trying to shoot scenes on the streets of Manhattan from a distance; for any event at an intersection, the crew has only to be on a matching street or avenue. The cnn problem contains as special cases two important problems: the bridge problem, also known as the cow-path problem, and the weighted 2-server problem in which the 2 servers may have different speeds. We√ show that any deterministic online algorithm has competitive ratio at least 6 + 17. We also show that some successful algorithms for the k-server problem fail to be competitive. In particular, no memoryless randomized algorithm can be competitive. We also consider another variant of the k-server problem, in which servers can move simultaneously, and we wish to minimize the time spent waiting for service. This is equivalent to the regular k-server problem under the L∞ norm for movement 1 costs.
    [Show full text]
  • The Price of Privacy for Low-Rank Factorization
    The Price of Privacy for Low-rank Factorization Jalaj Upadhyay Johns Hopkins University Baltimore, MD - 21201, USA. [email protected] Abstract In this paper, we study what price one has to pay to release differentially private low-rank factorization of a matrix. We consider various settings that are close to the real world applications of low-rank factorization: (i) the manner in which matrices are updated (row by row or in an arbitrary manner), (ii) whether matrices are distributed or not, and (iii) how the output is produced (once at the end of all updates, also known as one-shot algorithms or continually). Even though these settings are well studied without privacy, surprisingly, there are no private algorithm for these settings (except when a matrix is updated row by row). We present the first set of differentially private algorithms for all these settings. Our algorithms when private matrix is updated in an arbitrary manner promise differential privacy with respect to two stronger privacy guarantees than previously studied, use space and time comparable to the non-private algorithm, and achieve optimal accuracy. To complement our positive results, we also prove that the space required by our algorithms is optimal up to logarithmic factors. When data matrices are distributed over multiple servers, we give a non-interactive differentially private algorithm with communication cost independent of dimension. In concise, we give algorithms that incur optimal cost across all parameters of interest. We also perform experiments to verify that all our algorithms perform well in practice and outperform the best known algorithm until now for large range of parameters.
    [Show full text]
  • Cuckoo Hashing
    Cuckoo Hashing Rasmus Pagh* BRICSy, Department of Computer Science, Aarhus University Ny Munkegade Bldg. 540, DK{8000 A˚rhus C, Denmark. E-mail: [email protected] and Flemming Friche Rodlerz ON-AIR A/S, Digtervejen 9, 9200 Aalborg SV, Denmark. E-mail: ff[email protected] We present a simple dictionary with worst case constant lookup time, equal- ing the theoretical performance of the classic dynamic perfect hashing scheme of Dietzfelbinger et al. (Dynamic perfect hashing: Upper and lower bounds. SIAM J. Comput., 23(4):738{761, 1994). The space usage is similar to that of binary search trees, i.e., three words per key on average. Besides being conceptually much simpler than previous dynamic dictionaries with worst case constant lookup time, our data structure is interesting in that it does not use perfect hashing, but rather a variant of open addressing where keys can be moved back in their probe sequences. An implementation inspired by our algorithm, but using weaker hash func- tions, is found to be quite practical. It is competitive with the best known dictionaries having an average case (but no nontrivial worst case) guarantee. Key Words: data structures, dictionaries, information retrieval, searching, hash- ing, experiments * Partially supported by the Future and Emerging Technologies programme of the EU under contract number IST-1999-14186 (ALCOM-FT). This work was initiated while visiting Stanford University. y Basic Research in Computer Science (www.brics.dk), funded by the Danish National Research Foundation. z This work was done while at Aarhus University. 1 2 PAGH AND RODLER 1. INTRODUCTION The dictionary data structure is ubiquitous in computer science.
    [Show full text]
  • The Price of Privacy for Low-Rank Factorization
    The Price of Privacy for Low-rank Factorization Jalaj Upadhyay Johns Hopkins University Baltimore, MD - 21201, USA. [email protected] Abstract In this paper, we study what price one has to pay to release differentially private low-rank factorization of a matrix. We consider various settings that are close to the real world applications of low-rank factorization: (i) the manner in which matrices are updated (row by row or in an arbitrary manner), (ii) whether matrices are distributed or not, and (iii) how the output is produced (once at the end of all updates, also known as one-shot algorithms or continually). Even though these settings are well studied without privacy, surprisingly, there are no private algorithm for these settings (except when a matrix is updated row by row). We present the first set of differentially private algorithms for all these settings. Our algorithms when private matrix is updated in an arbitrary manner promise differential privacy with respect to two stronger privacy guarantees than previously studied, use space and time comparable to the non-private algorithm, and achieve optimal accuracy. To complement our positive results, we also prove that the space required by our algorithms is optimal up to logarithmic factors. When data matrices are distributed over multiple servers, we give a non-interactive differentially private algorithm with communication cost independent of dimension. In concise, we give algorithms that incur optimal cost across all parameters of interest. We also perform experiments to verify that all our algorithms perform well in practice and outperform the best known algorithm until now for large range of parameters.
    [Show full text]
  • Annual Report of the ACM Awards Committee for the Period July 1, 2005 - June 30, 2006
    Annual Report of the ACM Awards Committee for the Period July 1, 2005 - June 30, 2006 1. BASIC INFORMATION 1.1 List of Committee Members and Terms of Office Calvin C. Gotlieb, Co-Chair 4/1/98-6/30/06 James Jay Horning, Co-Chair 7/02-6/30/06 Martin Abadi 8/04-12/31/06 Hal Abelson 12/03-12/31/09 Alfred V. Aho 2/01-12/31/05 David H. Bailey 12/31/08 Ruzena Bajcsy 1/04-12/31/07 Mary Gray Baker 11/04-12/31/07 Victor Basili 7/02-12/31/07 Michel Beaudouin-Lafon 8/04-12/31/07 Alan Berenbaum 12/03-12/31/05 Reinaldo Bergamaschi 12/31/09 Nina Bhatti Judith Bishop 6/03-12/31/06 Dines Bjørner 8/04-12/31/07 Ronald Boisvert 10/04-12/31/08 Stephen R. Bourne 12/31/10 Eric Brewer 3/05- John Seely Brown 1/01-12/31/07 E.G. Coffman 6/03-12/31/06 Patrick Cousot 9/04-12/31/07 Carlos de Lucena 2/01-12/31/05 James W. Demmel 12/31/09 Adel S. Elmaghraby 6/02-12/31/05 Joan Feigenbaum 5/01-12/31/05 Christian Freksa 7/05- Bernard Galler 2/01-12/31/05 Nicholas Georganas 12/03-12/31/07 Carlo Ghezzi 12/31/08 Adele Goldberg 12/31/09 Sandra Graham 3/05-12/31/06 James N. Gray 6/02-12/31/06 William D. Gropp 12/31/06 David Harel 1/05-12/31/09 Laura Hill 12/31/11 Charles H.
    [Show full text]
  • Algorithms for Managing Data in Distributed Systems
    Algorithms for Managing Data in Distributed Systems Jared Saia A dissertation submitted in partial fulfillment of the requirements for the degree of Doctor of Philosophy University of Washington 2002 Program Authorized to Offer Degree: Computer Science and Engineering University of Washington Graduate School This is to certify that I have examined this copy of a doctoral dissertation by Jared Saia and have found that it is complete and satisfactory in all respects, and that any and all revisions required by the final examining committee have been made. Chair of Supervisory Committee: Anna R. Karlin Reading Committee: Anna Karlin Richard Ladner Steve Gribble Date: In presenting this dissertation in partial fulfillment of the requirements for the Doctoral degree at the University of Washington, I agree that the Library shall make its copies freely available for inspection. I further agree that extensive copying of this dissertation is allowable only for scholarly purposes, consistent with ”fair use” as prescribed in the U.S. Copyright Law. Requests for copying or reproduction of thisdissertation may be referred to Bell Howell Information and Learning, 300 North Zeeb Road, Ann Arbor, MI 48106-1346, to whom the author has granted ”the right to reproduce and sell (a) copies of the manuscript in microform and/or (b) printed copies of the manuscript made from microform.” Signature Date University of Washington Abstract Algorithms for Managing Data in Distributed Systems by Jared Saia Chair of Supervisory Committee: Professor Anna R. Karlin Computer Science and Engineering This dissertation describes provably good algorithms for two fundamental problems in the area of data management for distributed systems: attack-resistance and data migration.
    [Show full text]
  • Game Theory, Alive Anna R
    Game Theory, Alive Anna R. Karlin Yuval Peres 10.1090/mbk/101 Game Theory, Alive Game Theory, Alive Anna R. Karlin Yuval Peres AMERICAN MATHEMATICAL SOCIETY Providence, Rhode Island 2010 Mathematics Subject Classification. Primary 91A10, 91A12, 91A18, 91B12, 91A24, 91A43, 91A26, 91A28, 91A46, 91B26. For additional information and updates on this book, visit www.ams.org/bookpages/mbk-101 Library of Congress Cataloging-in-Publication Data Names: Karlin, Anna R. | Peres, Y. (Yuval) Title: Game theory, alive / Anna Karlin, Yuval Peres. Description: Providence, Rhode Island : American Mathematical Society, [2016] | Includes bibliographical references and index. Identifiers: LCCN 2016038151 | ISBN 9781470419820 (alk. paper) Subjects: LCSH: Game theory. | AMS: Game theory, economics, social and behavioral sciences – Game theory – Noncooperative games. msc | Game theory, economics, social and behavioral sciences – Game theory – Cooperative games. msc | Game theory, economics, social and behavioral sciences – Game theory – Games in extensive form. msc | Game theory, economics, social and behavioral sciences – Mathematical economics – Voting theory. msc | Game theory, economics, social and behavioral sciences – Game theory – Positional games (pursuit and evasion, etc.). msc | Game theory, economics, social and behavioral sciences – Game theory – Games involving graphs. msc | Game theory, economics, social and behavioral sciences – Game theory – Rationality, learning. msc | Game theory, economics, social and behavioral sciences – Game theory – Signaling,
    [Show full text]
  • The Women of UW CSE December 2009 the Women of CSE
    CSE ACM-W brown bag lunch, November 2009: From the left are Prof. Anna Karlin, Nancy Do, Nadine Tabing, Sylvia Tashev, Kate Moore, Sophia Wang, Kathleen Tuite, and Shulin Yang. The NY Times features UW CSE in “Computer Science Takes Steps to Bring Women to the Fold.” Cherie Cheung, Julie Letchner, Ed Lazowska, Sierra Michels-Slettvet, and Tanya Bragin watch Roxana Geambasu demonstrate a research prototype. The CSE contingent at the 2009 Grace Hopper Celebration of Women in Computing: Back row: Kim Todd, Allison Obourn, Rita Sodt, Alyssa Harding, Coral Peterson, Sylvia Tashev, Kathleen Tuite, Sara Su (alum), and Megan Reardon. Front row: Stephanie Smallman, Angie Zhu, Yuki Liang, Kate Moore, Lynn Yang, Kristi Morton, Julie Letchner, Miryung Kim (alum), Leilani Battle, and Nancy Do. The Women of UW CSE December 2009 The Women of CSE More than 100 women are members of the UW Computer Science & Engineering student and faculty community. Here are some of them ... Gail Alverson I have been with Cray Inc., the supercomputer company, since I graduated from the University of Washington Department of Computer Science with my PhD in 1991. I love the challenges of working with parallelism and the race to achieve top sustainable speed. At Cray I've grown from being a software engineer designing and developing parallel runtime systems to leading a set of technical projects for our VP of Research and Development (who is also a woman!). I'm honored to be an Affiliate Professor in the department of Computer Science and Engineering, where I have taught classes with a focus on software engineering.
    [Show full text]
  • Table of Contents
    Table of Contents Editors’ Introduction 1 HU FU and MATT WEINBERG Job Market Candidate Profiles 2019 2 VASILIS GKATZELIS and JASON HARTLINE Revenue Management and Market Analytics in Management Science 37 KALYAN TALLURI and GABRIEL WEINTRAUB Simple Mechanisms for Subadditive Buyers via Duality 39 YANG CAI and MINGFEI ZHAO Strategyproof Linear Regression in High Dimensions: An Overview 54 YILING CHEN, CHARA PODIMATA, ARIEL D. PROCACCIA, and NISARG SHAH Recent Developments in Prophet Inequalities 61 JOSE CORREA, PATRICIO FONCEA, RUBEN HOEKSMA, TIM OOSTERWIJK, and TJARK VREDEVELD An Improved Welfare Guarantee for First-Price Auctions 71 DARRELL HOY, SAM TAGGART, and ZIHE WANG ACM SIGecom Exchanges, Vol. 17, No. 1, November 2018 By submitting an article to the SIGecom Exchanges, and on condition of accep- tance by the editor, authors grant to ACM the following non-exclusive, perpetual, worldwide rights: —to publish the article in print; —to digitize and post the article in the electronic version of the publication; —to include the article in the ACM Digital Library and in any Digital Library related services; —to allow users to make a personal copy of the article for noncommercial, educa- tional or research purposes. Contributing authors retain copyright to the article, and ACM will refer requests for republication directly to them. Editors’ Introduction HU FU University of British Columbia and MATT WEINBERG Princeton University This issue of SIGEcom Exchanges contains one brief announcement followed by four research letters. The 2019 SIGEcom Job Candidates Profiles, compiled last year again by Jason Hartline and Vasilis Gkatzelis, were published separately and are now officially included in the issue.
    [Show full text]
  • The K-Server Problem
    The k-server problem Elias Koutsoupias∗ April 13, 2009 Abstract The k-server problem is perhaps the most influential online problem: natural, crisp, with a surprising technical depth that manifests the richness of competitive analysis. The k-server conjecture, which was posed more that two decades ago when the problem was first studied within the competitive analysis framework, is still open and has been a major driving force for the development of the area online algorithms. This article surveys some major results for the k-server. 1 Introduction The k-server problem in a metric space (or weighted graph) is the problem of moving around efficiently k servers to service requests that appear online at the points of the metric space. The k servers start at a fixed set of k points which we call the initial configuration. Then in each time step a request appears at some point of the metric space and some server has to move to that point to service the request. We assume that the requests appear sequentially and that a new request appears only after the current request has been serviced. The goal is to minimize the total distance traveled by the servers. More formally, the definition of the problem is as follows: Let M be a metric space and let d(a1,a2) denote the distance in M of points a1 and a2. The distance d is nonnegative, symmetric, and satisfies the triangle inequality. For simplicity, we allow distinct points to be at distance 0, or more formally, M is a pseudometric rather than metric.
    [Show full text]