Cook's Theory and Twentieth Century Mathematics
Total Page:16
File Type:pdf, Size:1020Kb
Load more
Recommended publications
-
Efficient Algorithms with Asymmetric Read and Write Costs
Efficient Algorithms with Asymmetric Read and Write Costs Guy E. Blelloch1, Jeremy T. Fineman2, Phillip B. Gibbons1, Yan Gu1, and Julian Shun3 1 Carnegie Mellon University 2 Georgetown University 3 University of California, Berkeley Abstract In several emerging technologies for computer memory (main memory), the cost of reading is significantly cheaper than the cost of writing. Such asymmetry in memory costs poses a fun- damentally different model from the RAM for algorithm design. In this paper we study lower and upper bounds for various problems under such asymmetric read and write costs. We con- sider both the case in which all but O(1) memory has asymmetric cost, and the case of a small cache of symmetric memory. We model both cases using the (M, ω)-ARAM, in which there is a small (symmetric) memory of size M and a large unbounded (asymmetric) memory, both random access, and where reading from the large memory has unit cost, but writing has cost ω 1. For FFT and sorting networks we show a lower bound cost of Ω(ωn logωM n), which indicates that it is not possible to achieve asymptotic improvements with cheaper reads when ω is bounded by a polynomial in M. Moreover, there is an asymptotic gap (of min(ω, log n)/ log(ωM)) between the cost of sorting networks and comparison sorting in the model. This contrasts with the RAM, and most other models, in which the asymptotic costs are the same. We also show a lower bound for computations on an n × n diamond DAG of Ω(ωn2/M) cost, which indicates no asymptotic improvement is achievable with fast reads. -
Great Ideas in Computing
Great Ideas in Computing University of Toronto CSC196 Winter/Spring 2019 Week 6: October 19-23 (2020) 1 / 17 Announcements I added one final question to Assignment 2. It concerns search engines. The question may need a little clarification. I have also clarified question 1 where I have defined (in a non standard way) the meaning of a strict binary search tree which is what I had in mind. Please answer the question for a strict binary search tree. If you answered the quiz question for a non strict binary search tree withn a proper explanation you will get full credit. Quick poll: how many students feel that Q1 was a fair quiz? A1 has now been graded by Marta. I will scan over the assignments and hope to release the grades later today. If you plan to make a regrading request, you have up to one week to submit your request. You must specify clearly why you feel that a question may not have been graded fairly. In general, students did well which is what I expected. 2 / 17 Agenda for the week We will continue to discuss search engines. We ended on what is slide 10 (in Week 5) on Friday and we will continue with where we left off. I was surprised that in our poll, most students felt that the people advocating the \AI view" of search \won the debate" whereas today I will try to argue that the people (e.g., Salton and others) advocating the \combinatorial, algebraic, statistical view" won the debate as to current search engines. -
Computational Complexity Computational Complexity
In 1965, the year Juris Hartmanis became Chair Computational of the new CS Department at Cornell, he and his KLEENE HIERARCHY colleague Richard Stearns published the paper On : complexity the computational complexity of algorithms in the Transactions of the American Mathematical Society. RE CO-RE RECURSIVE The paper introduced a new fi eld and gave it its name. Immediately recognized as a fundamental advance, it attracted the best talent to the fi eld. This diagram Theoretical computer science was immediately EXPSPACE shows how the broadened from automata theory, formal languages, NEXPTIME fi eld believes and algorithms to include computational complexity. EXPTIME complexity classes look. It As Richard Karp said in his Turing Award lecture, PSPACE = IP : is known that P “All of us who read their paper could not fail P-HIERARCHY to realize that we now had a satisfactory formal : is different from ExpTime, but framework for pursuing the questions that Edmonds NP CO-NP had raised earlier in an intuitive fashion —questions P there is no proof about whether, for instance, the traveling salesman NLOG SPACE that NP ≠ P and problem is solvable in polynomial time.” LOG SPACE PSPACE ≠ P. Hartmanis and Stearns showed that computational equivalences and separations among complexity problems have an inherent complexity, which can be classes, fundamental and hard open problems, quantifi ed in terms of the number of steps needed on and unexpected connections to distant fi elds of a simple model of a computer, the multi-tape Turing study. An early example of the surprises that lurk machine. In a subsequent paper with Philip Lewis, in the structure of complexity classes is the Gap they proved analogous results for the number of Theorem, proved by Hartmanis’s student Allan tape cells used. -
Efficient Algorithms with Asymmetric Read and Write Costs
Efficient Algorithms with Asymmetric Read and Write Costs Guy E. Blelloch1, Jeremy T. Fineman2, Phillip B. Gibbons1, Yan Gu1, and Julian Shun3 1 Carnegie Mellon University 2 Georgetown University 3 University of California, Berkeley Abstract In several emerging technologies for computer memory (main memory), the cost of reading is significantly cheaper than the cost of writing. Such asymmetry in memory costs poses a fun- damentally different model from the RAM for algorithm design. In this paper we study lower and upper bounds for various problems under such asymmetric read and write costs. We con- sider both the case in which all but O(1) memory has asymmetric cost, and the case of a small cache of symmetric memory. We model both cases using the (M, ω)-ARAM, in which there is a small (symmetric) memory of size M and a large unbounded (asymmetric) memory, both random access, and where reading from the large memory has unit cost, but writing has cost ω 1. For FFT and sorting networks we show a lower bound cost of Ω(ωn logωM n), which indicates that it is not possible to achieve asymptotic improvements with cheaper reads when ω is bounded by a polynomial in M. Moreover, there is an asymptotic gap (of min(ω, log n)/ log(ωM)) between the cost of sorting networks and comparison sorting in the model. This contrasts with the RAM, and most other models, in which the asymptotic costs are the same. We also show a lower bound for computations on an n × n diamond DAG of Ω(ωn2/M) cost, which indicates no asymptotic improvement is achievable with fast reads. -
You and Your Research & the Elements of Style
You and Your Research & The Elements of Style Philip Wadler University of Edinburgh Logic Mentoring Workshop Saarbrucken,¨ Monday 6 July 2020 Part I You and Your Research Richard W. Hamming, 1915–1998 • Los Alamos, 1945. • Bell Labs, 1946–1976. • Naval Postgraduate School, 1976–1998. • Turing Award, 1968. (Third time given.) • IEEE Hamming Medal, 1987. It’s not luck, it’s not brains, it’s courage Say to yourself, ‘Yes, I would like to do first-class work.’ Our society frowns on people who set out to do really good work. You’re not supposed to; luck is supposed to descend on you and you do great things by chance. Well, that’s a kind of dumb thing to say. ··· How about having lots of ‘brains?’ It sounds good. Most of you in this room probably have more than enough brains to do first-class work. But great work is something else than mere brains. ··· One of the characteristics of successful scientists is having courage. Once you get your courage up and believe that you can do important problems, then you can. If you think you can’t, almost surely you are not going to. — Richard Hamming, You and Your Research Develop reusable solutions How do I obey Newton’s rule? He said, ‘If I have seen further than others, it is because I’ve stood on the shoulders of giants.’ These days we stand on each other’s feet! Now if you are much of a mathematician you know that the effort to gen- eralize often means that the solution is simple. -
P Versus NP Richard M. Karp
P Versus NP Computational complexity theory is the branch of theoretical computer science concerned with the fundamental limits on the efficiency of automatic computation. It focuses on problems that appear to Richard M. Karp require a very large number of computation steps for their solution. The inputs and outputs to a problem are sequences of symbols drawn from a finite alphabet; there is no limit on the length of the input, and the fundamental question about a problem is the rate of growth of the number of required computation steps as a function of the length of the input. Some problems seem to require a very rapidly growing number of steps. One such problem is the inde- pendent set problem: given a graph, consisting of points called vertices and lines called edges connect- ing pairs of vertices, a set of vertices is called independent if no two vertices in the set are connected by a line. Given a graph and a positive integer n, the problem is to decide whether the graph contains an independent set of size n. Every known algorithm to solve the independent set problem encounters a combinatorial explosion, in which the number of required computation steps grows exponentially as a function of the size of the graph. On the other hand, the problem of deciding whether a given set of vertices is an independent set in a given graph is solvable by inspection. There are many such dichotomies, in which it is hard to decide whether a given type of structure exists within an input object (the existence problem), but it is easy to decide whether a given structure is of the required type (the verification problem). -
History of Modern Applied Mathematics in Mathematics Education
HISTORY OF MODERN APPLIED MATHEMATICS IN MATHEMATICS EDUCATION UFFE THOMAS JANKVISI [I] When conversations turn to using history of mathematics in in-issues, of mathematics. When using history as a tool to classrooms, the rnferent is typically the old, often antique, improve leaining or instruction, we may distinguish at least history of the discipline (e g, Calinger, 1996; Fauvel & van two different uses: history as a motivational or affective Maanen, 2000; Jahnke et al, 1996; Katz, 2000) [2] This tool, and histmy as a cognitive tool Together with history tendency might be expected, given that old mathematics is as a goal these two uses of histoty as a tool are used to struc often more closely related to school mathematics However, ture discussion of the educational benefits of choosing a there seem to be some clear advantages of including histo history of modern applied mathematics ries of more modern applied mathematics 01 histories of History as a goal 'in itself' does not refor to teaching his modem applications of mathematics [3] tory of mathematics per se, but using histo1y to surface One (justified) objection to integrating elements of the meta-aspects of the discipline Of course, in specific teach history of modetn applied mathematics is that it is often com ing situations, using histmy as a goal may have the positive plex and difficult While this may be so in most instances, it side effect of offering students insight into mathematical is worthwhile to search for cases where it isn't so I consider in-issues of a specific history But the impo1tant detail is three here. -
Program of the Sessions San Diego, California, January 9–12, 2013
Program of the Sessions San Diego, California, January 9–12, 2013 AMS Short Course on Random Matrices, Part Monday, January 7 I MAA Short Course on Conceptual Climate Models, Part I 9:00 AM –3:45PM Room 4, Upper Level, San Diego Convention Center 8:30 AM –5:30PM Room 5B, Upper Level, San Diego Convention Center Organizer: Van Vu,YaleUniversity Organizers: Esther Widiasih,University of Arizona 8:00AM Registration outside Room 5A, SDCC Mary Lou Zeeman,Bowdoin upper level. College 9:00AM Random Matrices: The Universality James Walsh, Oberlin (5) phenomenon for Wigner ensemble. College Preliminary report. 7:30AM Registration outside Room 5A, SDCC Terence Tao, University of California Los upper level. Angles 8:30AM Zero-dimensional energy balance models. 10:45AM Universality of random matrices and (1) Hans Kaper, Georgetown University (6) Dyson Brownian Motion. Preliminary 10:30AM Hands-on Session: Dynamics of energy report. (2) balance models, I. Laszlo Erdos, LMU, Munich Anna Barry*, Institute for Math and Its Applications, and Samantha 2:30PM Free probability and Random matrices. Oestreicher*, University of Minnesota (7) Preliminary report. Alice Guionnet, Massachusetts Institute 2:00PM One-dimensional energy balance models. of Technology (3) Hans Kaper, Georgetown University 4:00PM Hands-on Session: Dynamics of energy NSF-EHR Grant Proposal Writing Workshop (4) balance models, II. Anna Barry*, Institute for Math and Its Applications, and Samantha 3:00 PM –6:00PM Marina Ballroom Oestreicher*, University of Minnesota F, 3rd Floor, Marriott The time limit for each AMS contributed paper in the sessions meeting will be found in Volume 34, Issue 1 of Abstracts is ten minutes. -
Sinai Awarded 2014 Abel Prize
Sinai Awarded 2014 Abel Prize The Norwegian Academy of Sci- Sinai’s first remarkable contribution, inspired ence and Letters has awarded by Kolmogorov, was to develop an invariant of the Abel Prize for 2014 to dynamical systems. This invariant has become Yakov Sinai of Princeton Uni- known as the Kolmogorov-Sinai entropy, and it versity and the Landau Insti- has become a central notion for studying the com- tute for Theoretical Physics plexity of a system through a measure-theoretical of the Russian Academy of description of its trajectories. It has led to very Sciences “for his fundamen- important advances in the classification of dynami- tal contributions to dynamical cal systems. systems, ergodic theory, and Sinai has been at the forefront of ergodic theory. mathematical physics.” The He proved the first ergodicity theorems for scat- Photo courtesy of Princeton University Mathematics Department. Abel Prize recognizes contribu- tering billiards in the style of Boltzmann, work tions of extraordinary depth he continued with Bunimovich and Chernov. He Yakov Sinai and influence in the mathemat- constructed Markov partitions for systems defined ical sciences and has been awarded annually since by iterations of Anosov diffeomorphisms, which 2003. The prize carries a cash award of approxi- led to a series of outstanding works showing the mately US$1 million. Sinai received the Abel Prize power of symbolic dynamics to describe various at a ceremony in Oslo, Norway, on May 20, 2014. classes of mixing systems. With Ruelle and Bowen, Sinai discovered the Citation notion of SRB measures: a rather general and Ever since the time of Newton, differential equa- distinguished invariant measure for dissipative tions have been used by mathematicians, scientists, systems with chaotic behavior. -
A Short History of Computational Complexity
The Computational Complexity Column by Lance FORTNOW NEC Laboratories America 4 Independence Way, Princeton, NJ 08540, USA [email protected] http://www.neci.nj.nec.com/homepages/fortnow/beatcs Every third year the Conference on Computational Complexity is held in Europe and this summer the University of Aarhus (Denmark) will host the meeting July 7-10. More details at the conference web page http://www.computationalcomplexity.org This month we present a historical view of computational complexity written by Steve Homer and myself. This is a preliminary version of a chapter to be included in an upcoming North-Holland Handbook of the History of Mathematical Logic edited by Dirk van Dalen, John Dawson and Aki Kanamori. A Short History of Computational Complexity Lance Fortnow1 Steve Homer2 NEC Research Institute Computer Science Department 4 Independence Way Boston University Princeton, NJ 08540 111 Cummington Street Boston, MA 02215 1 Introduction It all started with a machine. In 1936, Turing developed his theoretical com- putational model. He based his model on how he perceived mathematicians think. As digital computers were developed in the 40's and 50's, the Turing machine proved itself as the right theoretical model for computation. Quickly though we discovered that the basic Turing machine model fails to account for the amount of time or memory needed by a computer, a critical issue today but even more so in those early days of computing. The key idea to measure time and space as a function of the length of the input came in the early 1960's by Hartmanis and Stearns. -
Homogenization 2001, Proceedings of the First HMS2000 International School and Conference on Ho- Mogenization
in \Homogenization 2001, Proceedings of the First HMS2000 International School and Conference on Ho- mogenization. Naples, Complesso Monte S. Angelo, June 18-22 and 23-27, 2001, Ed. L. Carbone and R. De Arcangelis, 191{211, Gakkotosho, Tokyo, 2003". On Homogenization and Γ-convergence Luc TARTAR1 In memory of Ennio DE GIORGI When in the Fall of 1976 I had chosen \Homog´en´eisationdans les ´equationsaux d´eriv´eespartielles" (Homogenization in partial differential equations) for the title of my Peccot lectures, which I gave in the beginning of 1977 at Coll`egede France in Paris, I did not know of the term Γ-convergence, which I first heard almost a year after, in a talk that Ennio DE GIORGI gave in the seminar that Jacques-Louis LIONS was organizing at Coll`egede France on Friday afternoons. I had not found the definition of Γ-convergence really new, as it was quite similar to questions that were already discussed in control theory under the name of relaxation (which was a more general question than what most people mean by that term now), and it was the convergence in the sense of Umberto MOSCO [Mos] but without the restriction to convex functionals, and it was the natural nonlinear analog of a result concerning G-convergence that Ennio DE GIORGI had obtained with Sergio SPAGNOLO [DG&Spa]; however, Ennio DE GIORGI's talk contained a quite interesting example, for which he referred to Luciano MODICA (and Stefano MORTOLA) [Mod&Mor], where functionals involving surface integrals appeared as Γ-limits of functionals involving volume integrals, and I thought that it was the interesting part of the concept, so I had found it similar to previous questions but I had felt that the point of view was slightly different. -
A Variation of Levin Search for All Well-Defined Problems
A Variation of Levin Search for All Well-Defined Problems Fouad B. Chedid A’Sharqiyah University, Ibra, Oman [email protected] July 10, 2021 Abstract In 1973, L.A. Levin published an algorithm that solves any inversion problem π as quickly as the fastest algorithm p∗ computing a solution for ∗ ∗ ∗ π in time bounded by 2l(p ).t , where l(p ) is the length of the binary ∗ ∗ ∗ encoding of p , and t is the runtime of p plus the time to verify its correctness. In 2002, M. Hutter published an algorithm that solves any ∗ well-defined problem π as quickly as the fastest algorithm p computing a solution for π in time bounded by 5.tp(x)+ dp.timetp (x)+ cp, where l(p)+l(tp) l(f)+1 2 dp = 40.2 and cp = 40.2 .O(l(f) ), where l(f) is the length of the binary encoding of a proof f that produces a pair (p,tp), where tp(x) is a provable time bound on the runtime of the fastest program p ∗ provably equivalent to p . In this paper, we rewrite Levin Search using the ideas of Hutter so that we have a new simple algorithm that solves any ∗ well-defined problem π as quickly as the fastest algorithm p computing 2 a solution for π in time bounded by O(l(f) ).tp(x). keywords: Computational Complexity; Algorithmic Information Theory; Levin Search. arXiv:1702.03152v1 [cs.CC] 10 Feb 2017 1 Introduction We recall that the class NP is the set of all decision problems that can be solved efficiently on a nondeterministic Turing Machine.