ACM SIGACT News Distributed Computing Column 28
Total Page:16
File Type:pdf, Size:1020Kb
Load more
Recommended publications
-
Edsger Dijkstra: the Man Who Carried Computer Science on His Shoulders
INFERENCE / Vol. 5, No. 3 Edsger Dijkstra The Man Who Carried Computer Science on His Shoulders Krzysztof Apt s it turned out, the train I had taken from dsger dijkstra was born in Rotterdam in 1930. Nijmegen to Eindhoven arrived late. To make He described his father, at one time the president matters worse, I was then unable to find the right of the Dutch Chemical Society, as “an excellent Aoffice in the university building. When I eventually arrived Echemist,” and his mother as “a brilliant mathematician for my appointment, I was more than half an hour behind who had no job.”1 In 1948, Dijkstra achieved remarkable schedule. The professor completely ignored my profuse results when he completed secondary school at the famous apologies and proceeded to take a full hour for the meet- Erasmiaans Gymnasium in Rotterdam. His school diploma ing. It was the first time I met Edsger Wybe Dijkstra. shows that he earned the highest possible grade in no less At the time of our meeting in 1975, Dijkstra was 45 than six out of thirteen subjects. He then enrolled at the years old. The most prestigious award in computer sci- University of Leiden to study physics. ence, the ACM Turing Award, had been conferred on In September 1951, Dijkstra’s father suggested he attend him three years earlier. Almost twenty years his junior, I a three-week course on programming in Cambridge. It knew very little about the field—I had only learned what turned out to be an idea with far-reaching consequences. a flowchart was a couple of weeks earlier. -
A Study on Asynchronous Randomized Consensus Title Algorithms for Byzantine Fault Tolerant Replication
A Study on Asynchronous Randomized Consensus Title Algorithms for Byzantine Fault Tolerant Replication Author(s) 中村, 純哉 Citation Issue Date Text Version ETD URL https://doi.org/10.18910/34568 DOI 10.18910/34568 rights Note Osaka University Knowledge Archive : OUKA https://ir.library.osaka-u.ac.jp/ Osaka University A Study on Asynchronous Randomized Consensus Algorithms for Byzantine Fault Tolerant Replication Submitted to Graduate School of Information Science and Technology Osaka University January 2014 Junya NAKAMURA iii List of Major Publications Journal Papers 1. Junya Nakamura, Tadashi Araragi, Toshimitsu Masuzawa, and Shigeru Masuyama, \A method of parallelizing consensuses for accelerating byzantine fault tolerance," IEICE Trans- actions on Information and Systems, vol. E97-D, no. 1, 2014. (to appear). 2. Junya Nakamura, Tadashi Araragi, Shigeru Masuyama, and Toshimitsu Masuzawa, “Effi- cient randomized byzantine fault-tolerant replication based on special valued coin tossing," IEICE Transactions on Information and Systems, vol. E97-D, no. 2, 2014. (to appear). Conference Papers 3. Junya Nakamura, Tadashi Araragi, and Shigeru Masuyama, \Asynchronous byzantine request- set agreement algorithm for replication," in Proceedings of the 1st AAAC Annual Meeting, p. 35, 2008. 4. Junya Nakamura, Tadashi Araragi, and Shigeru Masuyama, \Acceleration of byzantine fault tolerance by parallelizing consensuses," in Proceedings of the 10th International Conference on Parallel and Distributed Computing, Applications and Technologies, PDCAT '09, pp. 80{ 87, Dec. 2009. Technical Reports 5. Junya Nakamura, Tadashi Araragi, and Shigeru Masuyama, \Byzantine agreement on the order of processing received requests is solvable deterministically in asynchronous systems," in IEICE Technical Report, vol. 106 of COMP2006-35, pp. 33{40, Oct. -
Auto-Stabilisation : Solution Elégante Pour Lutter Contre Les Fautes Sylvie
Auto-stabilisation : Solution Elégante pour Lutter contre Les Fautes Sylvie Delaët Rapport scientifique présenté en vue de l’obtention de l’Habilitation à Diriger les Recherches. Soutenu le 8 novembre 2013 devant le jury composé de : Alain Denise, Professeur, Université Paris Sud, France Shlomi Dolev, Professeur, Ben Gurion University of the Negev, Israël Pierre Fraigniaud, Directeur de recherche CNRS, France Ted Herman, Professeur, University of Iowa, Etats Unis d’Amérique Nicolas Thiéry, Professeur, l’Université Paris Sud, France Franck Petit, Professeur, Université Paris Pierre et Marie Curie, France Après avis des rapporteurs : Ted Herman, Professeur, University of Iowa, Etats Unis d’Amérique Rachid Guerraoui, Professeur, Ecole Polytechnique Fédérale de Lausanne, Suisse Franck Petit, Université Paris Pierre et Marie Curie, France ii Sommaire Remerciements iii 1 Ma vision de ce document 1 1.1 Organisation du document . .1 2 Un peu de cuisine(s) 3 2.1 L'auberge espagnole de l'algorithmique r´epartie. .4 2.2 Exemples de probl`emes`ar´esoudreet quelques solutions . .8 3 Comprendre le monde 15 3.1 Comprendre l'auto-stabilisation . 15 3.2 Comprendre l'algorithmique r´epartie . 30 3.3 Comprendre mes contributions . 34 4 Construire l'avenir 43 4.1 Perspectives d'am´eliorationdes techniques . 43 4.2 Perspective d'ouverture `ala diff´erence . 45 4.3 Perspective de diffusion plus large . 47 4.4 Perspective de l'ordinateur auto-stabilisant . 48 A Agrafage 51 A.1 Cerner les probl`emes . 51 A.2 D´emasquerles planqu´es . 59 A.3 Recenser les pr´esents . 67 A.4 Obtenir l'auto-stabilisation gratuitement . -
Arxiv:1901.04709V1 [Cs.GT]
Self-Stabilization Through the Lens of Game Theory Krzysztof R. Apt1,2 and Ehsan Shoja3 1 CWI, Amsterdam, The Netherlands 2 MIMUW, University of Warsaw, Warsaw, Poland 3 Sharif University of Technology, Tehran, Iran Abstract. In 1974 E.W. Dijkstra introduced the seminal concept of self- stabilization that turned out to be one of the main approaches to fault- tolerant computing. We show here how his three solutions can be for- malized and reasoned about using the concepts of game theory. We also determine the precise number of steps needed to reach self-stabilization in his first solution. 1 Introduction In 1974 Edsger W. Dijkstra introduced in a two-page article [10] the notion of self-stabilization. The paper was completely ignored until 1983, when Leslie Lamport stressed its importance in his invited talk at the ACM Symposium on Principles of Distributed Computing (PODC), published a year later as [21]. Things have changed since then. According to Google Scholar Dijkstra’s paper has been by now cited more than 2300 times. It became one of the main ap- proaches to fault tolerant computing. An early survey was published in 1993 as [26], while the research on the subject until 2000 was summarized in the book [13]. In 2002 Dijkstra’s paper won the PODC influential paper award (renamed in 2003 to Dijkstra Prize). The literature on the subject initiated by it continues to grow. There are annual Self-Stabilizing Systems Workshops, the 18th edition of which took part in 2016. The idea proposed by Dijkstra is very simple. Consider a distributed system arXiv:1901.04709v1 [cs.GT] 15 Jan 2019 viewed as a network of machines. -
Distributed Algorithms for Wireless Networks
A THEORETICAL VIEW OF DISTRIBUTED SYSTEMS Nancy Lynch, MIT EECS, CSAIL National Science Foundation, April 1, 2021 Theory for Distributed Systems • We have worked on theory for distributed systems, trying to understand (mathematically) their capabilities and limitations. • This has included: • Defining abstract, mathematical models for problems solved by distributed systems, and for the algorithms used to solve them. • Developing new algorithms. • Producing rigorous proofs, of correctness, performance, fault-tolerance. • Proving impossibility results and lower bounds, expressing inherent limitations of distributed systems for solving problems. • Developing foundations for modeling, analyzing distributed systems. • Kinds of systems: • Distributed data-management systems. • Wired, wireless communication systems. • Biological systems: Insect colonies, developmental biology, brains. This talk: 1. Algorithms for Traditional Distributed Systems 2. Impossibility Results 3. Foundations 4. Algorithms for New Distributed Systems 1. Algorithms for Traditional Distributed Systems • Mutual exclusion in shared-memory systems, resource allocation: Fischer, Burns,…late 70s and early 80s. • Dolev, Lynch, Pinter, Stark, Weihl. Reaching approximate agreement in the presence of faults. JACM,1986. • Lundelius, Lynch. A new fault-tolerant algorithm for clock synchronization. Information and Computation,1988. • Dwork, Lynch, Stockmeyer. Consensus in the presence of partial synchrony. JACM,1988. Dijkstra Prize, 2007. 1A. Dwork, Lynch, Stockmeyer [DLS] “This paper introduces a number of practically motivated partial synchrony models that lie between the completely synchronous and the completely asynchronous models and in which consensus is solvable. It gave practitioners the right tool for building fault-tolerant systems and contributed to the understanding that safety can be maintained at all times, despite the impossibility of consensus, and progress is facilitated during periods of stability. -
2020 SIGACT REPORT SIGACT EC – Eric Allender, Shuchi Chawla, Nicole Immorlica, Samir Khuller (Chair), Bobby Kleinberg September 14Th, 2020
2020 SIGACT REPORT SIGACT EC – Eric Allender, Shuchi Chawla, Nicole Immorlica, Samir Khuller (chair), Bobby Kleinberg September 14th, 2020 SIGACT Mission Statement: The primary mission of ACM SIGACT (Association for Computing Machinery Special Interest Group on Algorithms and Computation Theory) is to foster and promote the discovery and dissemination of high quality research in the domain of theoretical computer science. The field of theoretical computer science is the rigorous study of all computational phenomena - natural, artificial or man-made. This includes the diverse areas of algorithms, data structures, complexity theory, distributed computation, parallel computation, VLSI, machine learning, computational biology, computational geometry, information theory, cryptography, quantum computation, computational number theory and algebra, program semantics and verification, automata theory, and the study of randomness. Work in this field is often distinguished by its emphasis on mathematical technique and rigor. 1. Awards ▪ 2020 Gödel Prize: This was awarded to Robin A. Moser and Gábor Tardos for their paper “A constructive proof of the general Lovász Local Lemma”, Journal of the ACM, Vol 57 (2), 2010. The Lovász Local Lemma (LLL) is a fundamental tool of the probabilistic method. It enables one to show the existence of certain objects even though they occur with exponentially small probability. The original proof was not algorithmic, and subsequent algorithmic versions had significant losses in parameters. This paper provides a simple, powerful algorithmic paradigm that converts almost all known applications of the LLL into randomized algorithms matching the bounds of the existence proof. The paper further gives a derandomized algorithm, a parallel algorithm, and an extension to the “lopsided” LLL. -
LNCS 4308, Pp
On Distributed Verification Amos Korman and Shay Kutten Faculty of IE&M, Technion, Haifa 32000, Israel Abstract. This paper describes the invited talk given at the 8th Inter- national Conference on Distributed Computing and Networking (ICDCN 2006), at the Indian Institute of Technology Guwahati, India. This talk was intended to give a partial survey and to motivate further studies of distributed verification. To serve the purpose of motivating, we allow ourselves to speculate freely on the potential impact of such research. In the context of sequential computing, it is common to assume that the task of verifying a property of an object may be much easier than computing it (consider, for example, solving an NP-Complete problem versus verifying a witness). Extrapolating from the impact the separation of these two notions (computing and verifying) had in the context of se- quential computing, the separation may prove to have a profound impact on the field of distributed computing as well. In addition, in the context of distributed computing, the motivation for the separation seems even stronger than in the centralized sequential case. In this paper we explain some motivations for specific definitions, sur- vey some very related notions and their motivations in the literature, survey some examples for problems and solutions, and mention some ad- ditional general results such as general algorithmic methods and general lower bounds. Since this paper is mostly intended to “give a taste” rather than be a comprehensive survey, we apologize to authors of additional related papers that we did not mention or detailed. 1 Introduction This paper addresses the problem of locally verifying global properties. -
Stabilization
Chapter 13 Stabilization A large branch of research in distributed computing deals with fault-tolerance. Being able to tolerate a considerable fraction of failing or even maliciously be- having (\Byzantine") nodes while trying to reach consensus (on e.g. the output of a function) among the nodes that work properly is crucial for building reli- able systems. However, consensus protocols require that a majority of the nodes remains non-faulty all the time. Can we design a distributed system that survives transient (short-lived) failures, even if all nodes are temporarily failing? In other words, can we build a distributed system that repairs itself ? 13.1 Self-Stabilization Definition 13.1 (Self-Stabilization). A distributed system is self-stabilizing if, starting from an arbitrary state, it is guaranteed to converge to a legitimate state. If the system is in a legitimate state, it is guaranteed to remain there, provided that no further faults happen. A state is legitimate if the state satisfies the specifications of the distributed system. Remarks: • What kind of transient failures can we tolerate? An adversary can crash nodes, or make nodes behave Byzantine. Indeed, temporarily an adversary can do harm in even worse ways, e.g. by corrupting the volatile memory of a node (without the node noticing { not unlike the movie Memento), or by corrupting messages on the fly (without anybody noticing). However, as all failures are transient, eventually all nodes must work correctly again, that is, crashed nodes get res- urrected, Byzantine nodes stop being malicious, messages are being delivered reliably, and the memory of the nodes is secure. -
Population Protocols: Expressiveness, Succinctness and Automatic Verification
Fakultat¨ fur¨ Informatik Technische Universitat¨ Munchen¨ Population Protocols: Expressiveness, Succinctness and Automatic Verification. Stefan Jaax Vollst¨andigerAbdruck der von der Fakult¨atf¨urInformatik der Technischen Universit¨at M¨unchen zur Erlangung des akademischen Grades eines Doktor der Naturwissenschaften (Dr. rer. nat.) genehmigten Dissertation. Vorsitzender: Prof. Dr. Helmut Seidl Pr¨ufendeder Dissertation: 1. Prof. Dr. Francisco Javier Esparza Estaun 2. Prof. Rupak Majumdar, Ph.D., Technische Universit¨atKaiserslautern Die Dissertation wurde am 03.03.2020 bei der Technischen Universit¨atM¨unchen eingereicht und durch die Fakult¨atf¨urInformatik am 04.06.2020 angenommen. Abstract Population protocols (Angluin et al., PODC, 2004) are a model of distributed computa- tion in which identical, finite-state, passively mobile agents interact in pairs to achieve a common goal. In the basic model of population protocols, agents compute number predicates by reaching a stable consensus. It is well known that population protocols compute precisely the semilinear predicates, or, equivalently, the predicates definable in Presburger arithmetic, the first-order theory of the natural numbers equipped with addition and the standard linear order. This thesis investigates three fundamental questions of the theory of population pro- tocols: Space complexity, verification complexity, and expressiveness of reasonable ex- tensions. Space Complexity. We show that every quantifier-free Presburger predicate ' aug- mented with remainder predicates is computable by a population protocol with poly(j'j) states, where j'j denotes the size of ' in binary encoding. Further, the protocol can be constructed in polynomial time. This is a major improvement to the previously known construction, which requires 2poly(j'j) states. As a special case, we consider predicates of the form 'c(x) = x ≥ c, where c is a positive integer constant. -
Mastering Concurrent Computing Through Sequential Thinking
review articles DOI:10.1145/3363823 we do not have good tools to build ef- A 50-year history of concurrency. ficient, scalable, and reliable concur- rent systems. BY SERGIO RAJSBAUM AND MICHEL RAYNAL Concurrency was once a specialized discipline for experts, but today the chal- lenge is for the entire information tech- nology community because of two dis- ruptive phenomena: the development of Mastering networking communications, and the end of the ability to increase processors speed at an exponential rate. Increases in performance come through concur- Concurrent rency, as in multicore architectures. Concurrency is also critical to achieve fault-tolerant, distributed services, as in global databases, cloud computing, and Computing blockchain applications. Concurrent computing through sequen- tial thinking. Right from the start in the 1960s, the main way of dealing with con- through currency has been by reduction to se- quential reasoning. Transforming problems in the concurrent domain into simpler problems in the sequential Sequential domain, yields benefits for specifying, implementing, and verifying concur- rent programs. It is a two-sided strategy, together with a bridge connecting the Thinking two sides. First, a sequential specificationof an object (or service) that can be ac- key insights ˽ A main way of dealing with the enormous challenges of building concurrent systems is by reduction to sequential I must appeal to the patience of the wondering readers, thinking. Over more than 50 years, more sophisticated techniques have been suffering as I am from the sequential nature of human developed to build complex systems in communication. this way. 12 ˽ The strategy starts by designing —E.W. -
Population Protocols Are Fast
Population Protocols Are Fast Adrian Kosowski1 and Przemysław Uznanski´ 2 1Inria Paris, France 2 ETH Zürich, Switzerland Abstract A population protocol describes a set of state change rules for a population of n indistinguish- able finite-state agents (automata), undergoing random pairwise interactions. Within this very basic framework, it is possible to resolve a number of fundamental tasks in distributed computing, includ- ing: leader election, aggregate and threshold functions on the population, such as majority compu- tation, and plurality consensus. For the first time, we show that solutions to all of these problems can be obtained quickly using finite-state protocols. For any input, the designed finite-state proto- cols converge under a fair random scheduler to an output which is correct with high probability in expected O(poly log n) parallel time. In the same setting, we also show protocols which always reach a valid solution, in expected parallel time O(nε), where the number of states of the interacting automata depends only on the choice of ε > 0. The stated time bounds hold for any semi-linear predicate computable in the population protocol framework. The key ingredient of our result is the decentralized design of a hierarchy of phase-clocks, which tick at different rates, with the rates of adjacent clocks separated by a factor of Θ(log n). The construction of this clock hierarchy relies on a new protocol composition technique, combined with an adapted analysis of a self-organizing process of oscillatory dynamics. This clock hierarchy is used to provide nested synchronization primitives, which allow us to view the population in a global manner and design protocols using a high-level imperative programming language with a (limited) capacity for loops and branching instructions. -
Kein Folientitel
182.703: Problems in Distributed Computing (Part 3) WS 2019 Ulrich Schmid Institute of Computer Engineering, TU Vienna Embedded Computing Systems Group E191-02 [email protected] Content (Part 3) The Role of Synchrony Conditions Failure Detectors Real-Time Clocks Partially Synchronous Models Models supporting lock-step round simulations Weaker partially synchronous models Dynamic distributed systems U. Schmid 182.703 PRDC 2 The Role of Synchrony Conditions U. Schmid 182.703 PRDC 3 Recall Distributed Agreement (Consensus) Yes No Yes NoYes? None? meet All meet Yes No No U. Schmid 182.703 PRDC 4 Recall Consensus Impossibility (FLP) Fischer, Lynch und Paterson [FLP85]: “There is no deterministic algorithm for solving consensus in an asynchronous distributed system in the presence of a single crash failure.” Key problem: Distinguish slow from dead! U. Schmid 182.703 PRDC 5 Consensus Solvability in ParSync [DDS87] (I) Dolev, Dwork and Stockmeyer investigated consensus solvability in Partially Synchronous Systems (ParSync), varying 5 „synchrony handles“ : • Processors synchronous / asynchronous • Communication synchronous / asynchronous • Message order synchronous (system-wide consistent) / asynchronous (out-of-order) • Send steps broadcast / unicast • Computing steps atomic rec+send / separate rec, send U. Schmid 182.703 PRDC 6 Consensus Solvability in ParSync [DDS87] (II) Wait-free consensus possible Consensus impossible s/r Consensus possible s+r for f=1 ucast bcast async sync Global message order message Global async sync Communication U. Schmid 182.703 PRDC 7 The Role of Synchrony Conditions Enable failure detection Enforce event ordering • Distinguish „old“ from „new“ • Ruling out existence of stale (in-transit) information • Distinguish slow from dead • Creating non-overlapping „phases of operation“ (rounds) U.