On the Work of Madhu Sudan Von Avi Wigderson

Total Page:16

File Type:pdf, Size:1020Kb

On the Work of Madhu Sudan Von Avi Wigderson Avi Wigderson On the Work of Madhu Sudan von Avi Wigderson Madhu Sudan is the recipient of the 2002 Nevanlinna Prize. Sudan has made fundamental contributions to two major areas of research, the connections between them, and their applications. The first area is coding theory. Established by Shannon and Hamming over fifty years ago, it is the mathematical study of the possibility of, and the limits on, reliable communication over noisy media. The second area is probabilistically checkable proofs (PCPs). By contrast, it is only ten years old. It studies the minimal resources required for probabilistic verification of standard mathematical proofs. My plan is to briefly introduce these areas, their mo- Probabilistic Checking of Proofs tivation, and foundational questions and then to ex- One informal variant of the celebrated P versus NP plain Sudan’s main contributions to each. Before we question asks,“Can mathematicians, past and future, get to the specific works of Madhu Sudan, let us start be replaced by an efficient computer program?” We with a couple of comments that will set up the con- first define these notions and then explain the PCP text of his work. theorem and its impact on this foundational question. Madhu Sudan works in computational complexity theory. This research discipline attempts to rigor- Efficient Computation ously define and study efficient versions of objects and notions arising in computational settings. This Throughout, by an efficient algorithm (or program, machine, or procedure) we mean an algorithm which focus on efficiency is of course natural when studying 1 computation itself, but it also has proved extremely runs at most some fixed polynomial time in the fruitful in studying other fundamental notions such length of its input. The input is always a finite string as proof, randomness, knowledge, and more. Here of symbols from a fixed, finite alphabet. Note that I will try to explain how the efficiency “eyeglasses” an algorithm is an object of fixed size which is sup- were key in this study of the notions proof (again) posed to solve a problem on all inputs of all (finite) and error correction. lengths. A problem is efficiently computable if it can be solved by an efficient algorithm. Theoretical computer science is an extremely inter- Definition 1. The class P is the class of all prob- active and collaborative community. Sudan’s work lems solvable by efficient algorithms. was not done in a vacuum, and much of the back- ground to it, conceptual and technical, was developed For example, the problems Integer Multiplication, Determinant, Linear Programming, Univariate Poly- by other people. The space I have does not allow me 2 to give proper credit to all these people. A much bet- nomial Factorization, and (recently established ) P ter job has been done by Sudan himself; his home- Testing Primality are in . page (http://theory.lcs.mit.edu/~madhu/)con- Let us restrict attention (for a while) to algorithms tains several surveys of these areas which give proper whose output is always “accept” or “reject”. Such an historical accounts and references. In particular see algorithm solves a decision problem. The set L of in- [13] for a survey on PCPs and [15] for a survey on puts which are accepted by A is called the language the work on error correction. recognized by A. Statements of the form “x ∈ L” 1 Time refers to the number of elementary steps taken by the algorithm. The choice of “polynomial” to represent efficiency is both small enough to often imply practicality and large enough to make the definition independent of particular aspects of the model, e.g., the choice of allowed “elementary operations”. 2[SeeMitteilungen 4–2002, S. 14–21. Anm. d. Red.] 64 DMV-Mitteilungen 1/2003 On the Work of Madhu Sudan are correctly classified as “true” or “false” by the ef- proof verification is defined by the well-known effi- ficient algorithm, deterministically (and without any cient (anonymous) algorithm referee.3 As humans “outside help”). we are simply not interested in theorems whose proofs take, say, longer than our lifetime (or the three-month Efficient Verification deadline given by editor)toreadandverify. In contrast, allowing an efficient algorithm to use But is this notion of efficient verification – reading “outside help” (a guess or an alleged proof) naturally through the statement and proof, and checking that defines a proof system. We say that a language L is every new lemma indeed follows from previous ones efficiently verifiable if there is an efficient algorithm (and known results) – the best we can hope for? Cer- V (for “Verifier”) and a fixed polynomial p for which tainly as referees we would love some shortcuts as the following completeness and soundness conditions long as they do not change our notion of mathemat- hold: ical truth too much. Are there such shortcuts? – For every x ∈ L there exists a string π of length |π|≤p(|x|) such that V accepts the joint input Efficient Probabilistic Verification (x, π). A major paradigm in computational complexity is al- ∈ – For every x L, for every string π of length lowing algorithms to flip coins. We postulate access | |≤ | | π p( x ), V rejects the joint input (x, π). to a supply of independent unbiased random vari- Naturally, we can view all strings x in L as theorems ables which the probabilistic (or randomized) algo- of the proof system V . Those strings π which cause rithm can use in its computation on a given input. V to accept x are legitimate proofs of the theorem We comment that the very rich theories (which we x ∈ L in this system. have no room to discuss) of pseudorandomness and Definition 2. The class NP is the class of all lan- of weak random sources attempt to bridge the gap guages that are efficiently verifiable. between this postulate and “real-life” generation of random bits in computer programs. It is clear that P⊆NP. Are they equal? This is the “P versus NP” question [5], one of the most impor- The notion of efficiency remains the same: Proba- tant open scientific problems today. Not only mathe- bilistic algorithms can make only a polynomial num- maticians but scientists and engineers as well daily ber of steps in the input length. However, the output attempt to perform tasks (create theories and de- becomes a random variable. We demand that the signs) whose success can hopefully be efficiently ver- probability of error, on every input, never exceed a ified. Reflect on the practical and philosophical im- given small bound. (Note that can be taken to be, pact of a positive answer to the question: If P = NP, e.g., 1/3, since repeating the algorithm with fresh in- then much of their (creative!) work can be performed dependent randomness and taking majority vote of efficiently by one computer program. the answers can decrease the error exponentially in Many important computational problems, like the the number of repetitions.) Travelling Salesman, Integer Programming, Map Returning to proof systems, we now allow the verifier Coloring, Systems of Quadratic Equations, and In- to be a probabilistic algorithm. As above, we allow it teger Factorization are (when properly coded as lan- to err (namely, accept false “proofs”) with extremely guages) in NP. small probability. The gain would be extreme effi- We stress two aspects of efficient verification. The ciency: The verifier will access only a constant num- purported “proof” π for the statement “x ∈ L”must ber of symbols in the alleged proof. Naturally, the be short, and the verification procedure must be effi- positions of the viewed symbols can be randomly cho- cient. It is important to note that all standard (log- sen. What kind of theorems can be proved in the ical) proof systems used in mathematics conform to resulting proof system? First, let us formalize it. the second restriction: Since only “local inferences” We say that a language L has a probabilistically are made from “easily recognizable” axioms, verifica- checkable proof if there is an efficient probabilis- tion is always efficient in the total length of statement tic algorithm V , a fixed polynomial p,andafixed and proof. The first restriction, on the length of the constant c for which the following completeness and proof, is natural, since we want the verification to be probabilistic soundness conditions hold. efficient in terms of the length of the statement. – For every x ∈ L there exists a string π,oflength An excellent, albeit informal, example is the lan- |π|≤p(|x|), such that V accepts the joint input guage math of all mathematical statements, whose (x, π) with probability 1. 3 This system can, of course, be formalized. However, it is better to have the social process of mathematics in mind before we plunge into the notions of the next subsection. DMV-Mitteilungen 1/2003 65 Avi Wigderson – For every x ∈ L, for every string π of length of proof systems to replace the referee for journal |π|≤p(|x|), V rejects the joint input (x, π) with papers). In this sense the PCP theorem is just a probability ≥ 1/2. statement of philosophical importance. However, the – On every input (x, π), V can access at most c bits PCP theorem does have significant implications of of π. immediate relevance to the theory of computation, Note again that executing the verifier independently and we explain this next. a constant number of times will reduce the “sound- ness” error to an arbitrarily small constant without Hardness of Approximation changing the definition. Also note that randomness The first and foremost contribution thus far to under- in the verifier is essential if it probes only a constant standing the mystery of the P versus NP question (or even logarithmic) number of symbols in the proof.
Recommended publications
  • The Multiplicative Weights Update Method: a Meta Algorithm and Applications
    The Multiplicative Weights Update Method: a Meta Algorithm and Applications Sanjeev Arora∗ Elad Hazan Satyen Kale Abstract Algorithms in varied fields use the idea of maintaining a distribution over a certain set and use the multiplicative update rule to iteratively change these weights. Their analysis are usually very similar and rely on an exponential potential function. We present a simple meta algorithm that unifies these disparate algorithms and drives them as simple instantiations of the meta algo- rithm. 1 Introduction Algorithms in varied fields work as follows: a distribution is maintained on a certain set, and at each step the probability assigned to i is multi- plied or divided by (1 + C(i)) where C(i) is some kind of “payoff” for element i. (Rescaling may be needed to ensure that the new values form a distribution.) Some examples include: the Ada Boost algorithm in ma- chine learning [FS97]; algorithms for game playing studied in economics (see below), the Plotkin-Shmoys-Tardos algorithm for packing and covering LPs [PST91], and its improvements in the case of flow problems by Young, Garg-Konneman, and Fleischer [You95, GK98, Fle00]; Impagliazzo’s proof of the Yao XOR lemma [Imp95], etc. The analysis of the running time uses a potential function argument and the final running time is proportional to 1/2. It has been clear to most researchers that these results are very similar, see for instance, Khandekar’s PhD thesis [Kha04]. Here we point out that these are all instances of the same (more general) algorithm. This meta ∗This project supported by David and Lucile Packard Fellowship and NSF grant CCR- 0205594.
    [Show full text]
  • Four Results of Jon Kleinberg a Talk for St.Petersburg Mathematical Society
    Four Results of Jon Kleinberg A Talk for St.Petersburg Mathematical Society Yury Lifshits Steklov Institute of Mathematics at St.Petersburg May 2007 1 / 43 2 Hubs and Authorities 3 Nearest Neighbors: Faster Than Brute Force 4 Navigation in a Small World 5 Bursty Structure in Streams Outline 1 Nevanlinna Prize for Jon Kleinberg History of Nevanlinna Prize Who is Jon Kleinberg 2 / 43 3 Nearest Neighbors: Faster Than Brute Force 4 Navigation in a Small World 5 Bursty Structure in Streams Outline 1 Nevanlinna Prize for Jon Kleinberg History of Nevanlinna Prize Who is Jon Kleinberg 2 Hubs and Authorities 2 / 43 4 Navigation in a Small World 5 Bursty Structure in Streams Outline 1 Nevanlinna Prize for Jon Kleinberg History of Nevanlinna Prize Who is Jon Kleinberg 2 Hubs and Authorities 3 Nearest Neighbors: Faster Than Brute Force 2 / 43 5 Bursty Structure in Streams Outline 1 Nevanlinna Prize for Jon Kleinberg History of Nevanlinna Prize Who is Jon Kleinberg 2 Hubs and Authorities 3 Nearest Neighbors: Faster Than Brute Force 4 Navigation in a Small World 2 / 43 Outline 1 Nevanlinna Prize for Jon Kleinberg History of Nevanlinna Prize Who is Jon Kleinberg 2 Hubs and Authorities 3 Nearest Neighbors: Faster Than Brute Force 4 Navigation in a Small World 5 Bursty Structure in Streams 2 / 43 Part I History of Nevanlinna Prize Career of Jon Kleinberg 3 / 43 Nevanlinna Prize The Rolf Nevanlinna Prize is awarded once every 4 years at the International Congress of Mathematicians, for outstanding contributions in Mathematical Aspects of Information Sciences including: 1 All mathematical aspects of computer science, including complexity theory, logic of programming languages, analysis of algorithms, cryptography, computer vision, pattern recognition, information processing and modelling of intelligence.
    [Show full text]
  • Prahladh Harsha
    Prahladh Harsha Toyota Technological Institute, Chicago phone : +1-773-834-2549 University Press Building fax : +1-773-834-9881 1427 East 60th Street, Second Floor Chicago, IL 60637 email : [email protected] http://ttic.uchicago.edu/∼prahladh Research Interests Computational Complexity, Probabilistically Checkable Proofs (PCPs), Information Theory, Prop- erty Testing, Proof Complexity, Communication Complexity. Education • Doctor of Philosophy (PhD) Computer Science, Massachusetts Institute of Technology, 2004 Research Advisor : Professor Madhu Sudan PhD Thesis: Robust PCPs of Proximity and Shorter PCPs • Master of Science (SM) Computer Science, Massachusetts Institute of Technology, 2000 • Bachelor of Technology (BTech) Computer Science and Engineering, Indian Institute of Technology, Madras, 1998 Work Experience • Toyota Technological Institute, Chicago September 2004 – Present Research Assistant Professor • Technion, Israel Institute of Technology, Haifa February 2007 – May 2007 Visiting Scientist • Microsoft Research, Silicon Valley January 2005 – September 2005 Postdoctoral Researcher Honours and Awards • Summer Research Fellow 1997, Jawaharlal Nehru Center for Advanced Scientific Research, Ban- galore. • Rajiv Gandhi Science Talent Research Fellow 1997, Jawaharlal Nehru Center for Advanced Scien- tific Research, Bangalore. • Award Winner in the Indian National Mathematical Olympiad (INMO) 1993, National Board of Higher Mathematics (NBHM). • National Board of Higher Mathematics (NBHM) Nurture Program award 1995-1998. The Nurture program 1995-1998, coordinated by Prof. Alladi Sitaram, Indian Statistical Institute, Bangalore involves various topics in higher mathematics. 1 • Ranked 7th in the All India Joint Entrance Examination (JEE) for admission into the Indian Institutes of Technology (among the 100,000 candidates who appeared for the examination). • Papers invited to special issues – “Robust PCPs of Proximity, Shorter PCPs, and Applications to Coding” (with Eli Ben- Sasson, Oded Goldreich, Madhu Sudan, and Salil Vadhan).
    [Show full text]
  • January 2011 Prizes and Awards
    January 2011 Prizes and Awards 4:25 P.M., Friday, January 7, 2011 PROGRAM SUMMARY OF AWARDS OPENING REMARKS FOR AMS George E. Andrews, President BÔCHER MEMORIAL PRIZE: ASAF NAOR, GUNTHER UHLMANN American Mathematical Society FRANK NELSON COLE PRIZE IN NUMBER THEORY: CHANDRASHEKHAR KHARE AND DEBORAH AND FRANKLIN TEPPER HAIMO AWARDS FOR DISTINGUISHED COLLEGE OR UNIVERSITY JEAN-PIERRE WINTENBERGER TEACHING OF MATHEMATICS LEVI L. CONANT PRIZE: DAVID VOGAN Mathematical Association of America JOSEPH L. DOOB PRIZE: PETER KRONHEIMER AND TOMASZ MROWKA EULER BOOK PRIZE LEONARD EISENBUD PRIZE FOR MATHEMATICS AND PHYSICS: HERBERT SPOHN Mathematical Association of America RUTH LYTTLE SATTER PRIZE IN MATHEMATICS: AMIE WILKINSON DAVID P. R OBBINS PRIZE LEROY P. S TEELE PRIZE FOR LIFETIME ACHIEVEMENT: JOHN WILLARD MILNOR Mathematical Association of America LEROY P. S TEELE PRIZE FOR MATHEMATICAL EXPOSITION: HENRYK IWANIEC BÔCHER MEMORIAL PRIZE LEROY P. S TEELE PRIZE FOR SEMINAL CONTRIBUTION TO RESEARCH: INGRID DAUBECHIES American Mathematical Society FOR AMS-MAA-SIAM LEVI L. CONANT PRIZE American Mathematical Society FRANK AND BRENNIE MORGAN PRIZE FOR OUTSTANDING RESEARCH IN MATHEMATICS BY AN UNDERGRADUATE STUDENT: MARIA MONKS LEONARD EISENBUD PRIZE FOR MATHEMATICS AND OR PHYSICS F AWM American Mathematical Society LOUISE HAY AWARD FOR CONTRIBUTIONS TO MATHEMATICS EDUCATION: PATRICIA CAMPBELL RUTH LYTTLE SATTER PRIZE IN MATHEMATICS M. GWENETH HUMPHREYS AWARD FOR MENTORSHIP OF UNDERGRADUATE WOMEN IN MATHEMATICS: American Mathematical Society RHONDA HUGHES ALICE T. S CHAFER PRIZE FOR EXCELLENCE IN MATHEMATICS BY AN UNDERGRADUATE WOMAN: LOUISE HAY AWARD FOR CONTRIBUTIONS TO MATHEMATICS EDUCATION SHERRY GONG Association for Women in Mathematics ALICE T. S CHAFER PRIZE FOR EXCELLENCE IN MATHEMATICS BY AN UNDERGRADUATE WOMAN FOR JPBM Association for Women in Mathematics COMMUNICATIONS AWARD: NICOLAS FALACCI AND CHERYL HEUTON M.
    [Show full text]
  • Some Hardness Escalation Results in Computational Complexity Theory Pritish Kamath
    Some Hardness Escalation Results in Computational Complexity Theory by Pritish Kamath B.Tech. Indian Institute of Technology Bombay (2012) S.M. Massachusetts Institute of Technology (2015) Submitted to Department of Electrical Engineering and Computer Science in partial fulfillment of the requirements for the degree of Doctor of Philosophy in Electrical Engineering & Computer Science at Massachusetts Institute of Technology February 2020 ⃝c Massachusetts Institute of Technology 2019. All rights reserved. Author: ............................................................. Department of Electrical Engineering and Computer Science September 16, 2019 Certified by: ............................................................. Ronitt Rubinfeld Professor of Electrical Engineering and Computer Science, MIT Thesis Supervisor Certified by: ............................................................. Madhu Sudan Gordon McKay Professor of Computer Science, Harvard University Thesis Supervisor Accepted by: ............................................................. Leslie A. Kolodziejski Professor of Electrical Engineering and Computer Science, MIT Chair, Department Committee on Graduate Students Some Hardness Escalation Results in Computational Complexity Theory by Pritish Kamath Submitted to Department of Electrical Engineering and Computer Science on September 16, 2019, in partial fulfillment of the requirements for the degree of Doctor of Philosophy in Computer Science & Engineering Abstract In this thesis, we prove new hardness escalation results in computational complexity theory; a phenomenon where hardness results against seemingly weak models of computation for any problem can be lifted, in a black box manner, to much stronger models of computation by considering a simple gadget composed version of the original problem. For any unsatisfiable CNF formula F that is hard to refute in the Resolution proof system, we show that a gadget-composed version of F is hard to refute in any proof system whose lines are computed by efficient communication protocols.
    [Show full text]
  • Pairwise Independence and Derandomization
    Pairwise Independence and Derandomization Pairwise Independence and Derandomization Michael Luby Digital Fountain Fremont, CA, USA Avi Wigderson Institute for Advanced Study Princeton, NJ, USA [email protected] Boston – Delft Foundations and TrendsR in Theoretical Computer Science Published, sold and distributed by: now Publishers Inc. PO Box 1024 Hanover, MA 02339 USA Tel. +1-781-985-4510 www.nowpublishers.com [email protected] Outside North America: now Publishers Inc. PO Box 179 2600 AD Delft The Netherlands Tel. +31-6-51115274 A Cataloging-in-Publication record is available from the Library of Congress The preferred citation for this publication is M. Luby and A. Wigderson, Pairwise R Independence and Derandomization, Foundation and Trends in Theoretical Com- puter Science, vol 1, no 4, pp 237–301, 2005 Printed on acid-free paper ISBN: 1-933019-22-0 c 2006 M. Luby and A. Wigderson All rights reserved. No part of this publication may be reproduced, stored in a retrieval system, or transmitted in any form or by any means, mechanical, photocopying, recording or otherwise, without prior written permission of the publishers. Photocopying. In the USA: This journal is registered at the Copyright Clearance Cen- ter, Inc., 222 Rosewood Drive, Danvers, MA 01923. Authorization to photocopy items for internal or personal use, or the internal or personal use of specific clients, is granted by now Publishers Inc for users registered with the Copyright Clearance Center (CCC). The ‘services’ for users can be found on the internet at: www.copyright.com For those organizations that have been granted a photocopy license, a separate system of payment has been arranged.
    [Show full text]
  • David Karger Rajeev Motwani Y Madhu Sudan Z
    Approximate Graph Coloring by Semidenite Programming y z David Karger Rajeev Motwani Madhu Sudan Abstract We consider the problem of coloring k colorable graphs with the fewest p ossible colors We present a randomized p olynomial time algorithm which colors a colorable graph on n vertices 12 12 13 14 with minfO log log n O n log ng colors where is the maximum degree of any vertex Besides giving the b est known approximation ratio in terms of n this marks the rst nontrivial approximation result as a function of the maximum degree This result can 12 12k b e generalized to k colorable graphs to obtain a coloring using minfO log log n 12 13(k +1) O n log ng colors Our results are inspired by the recent work of Go emans and Williamson who used an algorithm for semidenite optimization problems which generalize lin ear programs to obtain improved approximations for the MAX CUT and MAX SAT problems An intriguing outcome of our work is a duality relationship established b etween the value of the optimum solution to our semidenite program and the Lovasz function We show lower b ounds on the gap b etween the optimum solution of our semidenite program and the actual chromatic numb er by duality this also demonstrates interesting new facts ab out the function MIT Lab oratory for Computer Science Cambridge MA Email kargermitedu URL httptheorylcsmitedukarger Supp orted by a Hertz Foundation Graduate Fellowship by NSF Young In vestigator Award CCR with matching funds from IBM Schlumb erger Foundation Shell Foundation and Xerox Corp oration
    [Show full text]
  • On Differentially Private Graph Sparsification and Applications
    On Differentially Private Graph Sparsification and Applications Raman Arora Jalaj Upadhyay Johns Hopkins University Rutgers University [email protected] [email protected] Abstract In this paper, we study private sparsification of graphs. In particular, we give an algorithm that given an input graph, returns a sparse graph which approximates the spectrum of the input graph while ensuring differential privacy. This allows one to solve many graph problems privately yet efficiently and accurately. This is exemplified with application of the proposed meta-algorithm to graph algorithms for privately answering cut-queries, as well as practical algorithms for computing MAX-CUT and SPARSEST-CUT with better accuracy than previously known. We also give an efficient private algorithm to learn Laplacian eigenmap on a graph. 1 Introduction Data from social and communication networks have become a rich source to gain useful insights into the social, behavioral, and information sciences. Such data is naturally modeled as observations on a graph, and encodes rich, fine-grained, and structured information. At the same time, due to the seamless nature of data acquisition, often collected through personal devices, the individual information content in network data is often highly sensitive. This raises valid privacy concerns pertaining the analysis and release of such data. We address these concerns in this paper by presenting a novel algorithm that can be used to publish a succinct differentially private representation of network data with minimal degradation in accuracy for various graph related tasks. There are several notions of differential privacy one can consider in the setting described above. Depending on privacy requirements, one can consider edge level privacy that renders two graphs that differ in a single edge as in-distinguishable based on the algorithm’s output; this is the setting studied in many recent works [9, 19, 25, 54].
    [Show full text]
  • Avi Wigderson
    On the Work of Madhu Sudan Avi Wigderson Madhu Sudan is the recipient of the 2002 Nevan- “eyeglasses” were key in this study of the no- linna Prize. Sudan has made fundamental contri- tions proof (again) and error correction. butions to two major areas of research, the con- • Theoretical computer science is an extremely nections between them, and their applications. interactive and collaborative community. The first area is coding theory. Established by Sudan’s work was not done in a vacuum, and Shannon and Hamming over fifty years ago, it is much of the background to it, conceptual and technical, was developed by other people. The the mathematical study of the possibility of, and space I have does not allow me to give proper the limits on, reliable communication over noisy credit to all these people. A much better job media. The second area is probabilistically check- has been done by Sudan himself; his homepage able proofs (PCPs). By contrast, it is only ten years (http://theory.lcs.mit.edu/~madhu/) con- old. It studies the minimal resources required for tains several surveys of these areas which give probabilistic verification of standard mathemati- proper historical accounts and references. In cal proofs. particular see [13] for a survey on PCPs and [15] My plan is to briefly introduce these areas, their for a survey on the work on error correction. motivation, and foundational questions and then to explain Sudan’s main contributions to each. Be- Probabilistic Checking of Proofs fore we get to the specific works of Madhu Sudan, One informal variant of the celebrated P versus let us start with a couple of comments that will set NP question asks, Can mathematicians, past and up the context of his work.
    [Show full text]
  • Lecture Notes in Computer Science 4833 Commenced Publication in 1973 Founding and Former Series Editors: Gerhard Goos, Juris Hartmanis, and Jan Van Leeuwen
    Lecture Notes in Computer Science 4833 Commenced Publication in 1973 Founding and Former Series Editors: Gerhard Goos, Juris Hartmanis, and Jan van Leeuwen Editorial Board David Hutchison Lancaster University, UK Takeo Kanade Carnegie Mellon University, Pittsburgh, PA, USA Josef Kittler University of Surrey, Guildford, UK Jon M. Kleinberg Cornell University, Ithaca, NY, USA Friedemann Mattern ETH Zurich, Switzerland John C. Mitchell Stanford University, CA, USA Moni Naor Weizmann Institute of Science, Rehovot, Israel Oscar Nierstrasz University of Bern, Switzerland C. Pandu Rangan Indian Institute of Technology, Madras, India Bernhard Steffen University of Dortmund, Germany Madhu Sudan Massachusetts Institute of Technology, MA, USA Demetri Terzopoulos University of California, Los Angeles, CA, USA Doug Tygar University of California, Berkeley, CA, USA Moshe Y. Vardi Rice University, Houston, TX, USA Gerhard Weikum Max-Planck Institute of Computer Science, Saarbruecken, Germany Kaoru Kurosawa (Ed.) Advances in Cryptology – ASIACRYPT 2007 13th International Conference on the Theory and Application of Cryptology and Information Security Kuching, Malaysia, December 2-6, 2007 Proceedings 13 Volume Editor Kaoru Kurosawa Ibaraki University Department of Computer and Information Sciences 4-12-1 Nakanarusawa Hitachi, Ibaraki 316-8511, Japan E-mail: [email protected] Library of Congress Control Number: 2007939450 CR Subject Classification (1998): E.3, D.4.6, F.2.1-2, K.6.5, C.2, J.1, G.2 LNCS Sublibrary: SL 4 – Security and Cryptology ISSN 0302-9743 ISBN-10 3-540-76899-8 Springer Berlin Heidelberg New York ISBN-13 978-3-540-76899-9 Springer Berlin Heidelberg New York This work is subject to copyright.
    [Show full text]
  • Proof Verification and Hardness of Approximation Problems
    Proof Verification and Hardness of Approximation Problems Sanjeev Arora∗ Carsten Lundy Rajeev Motwaniz Madhu Sudanx Mario Szegedy{ Abstract that finding a solution within any constant factor of optimal is also NP-hard. Garey and Johnson The class PCP(f(n); g(n)) consists of all languages [GJ76] studied MAX-CLIQUE: the problem of find- L for which there exists a polynomial-time probabilistic ing the largest clique in a graph. They showed oracle machine that uses O(f(n)) random bits, queries that if a polynomial time algorithm computes MAX- O(g(n)) bits of its oracle and behaves as follows: If CLIQUE within a constant multiplicative factor then x 2 L then there exists an oracle y such that the ma- MAX-CLIQUE has a polynomial-time approximation chine accepts for all random choices but if x 62 L then scheme (PTAS) [GJ78], i.e., for any c > 1 there ex- for every oracle y the machine rejects with high proba- ists a polynomial-time algorithm that approximates bility. Arora and Safra very recently characterized NP MAX-CLIQUE within a factor of c. They used graph as PCP(log n; (log log n)O(1)). We improve on their products to construct \gap increasing reductions" that result by showing that NP = PCP(log n; 1). Our re- mapped an instance of the clique problem into an- sult has the following consequences: other instance in order to enlarge the relative gap of 1. MAXSNP-hard problems (e.g., metric TSP, the clique sizes. Berman and Schnitger [BS92] used MAX-SAT, MAX-CUT) do not have polynomial the ideas of increasing gaps to show that if the clique time approximation schemes unless P=NP.
    [Show full text]
  • Pseudorandom Generators Without the XOR Lemma£
    Pseudorandom generators without the XOR Lemma£ [Extended Abstract] Þ Ü Madhu SudanÝ Luca Trevisan Salil Vadhan ½ ÁÒØÖÓ Impagliazzo and Wigderson [IW97] have recently shown that if This paper continues the exploration of hardness versus random- Ç ´Òµ there exists a decision problem solvable in time ¾ and hav- ness trade-offs, that is, results showing that randomized algorithms Òµ ª´ can be efficiently simulated deterministically if certain complexity- Ò È ing circuit complexity ¾ (for all but finitely many ) then theoretic assumptions are true. We present two new approaches BÈÈ . This result is a culmination of a series of works showing con- nections between the existence of hard predicates and the existence to proving the recent result of Impagliazzo and Wigderson [IW97] Ç ´Òµ of good pseudorandom generators. that, if there is a decision problem computable in time ¾ and ª´Òµ Ò The construction of Impagliazzo and Wigderson goes through having circuit complexity ¾ for all but finitely many , then BÈÈ three phases of “hardness amplification” (a multivariate polyno- È . Impagliazzo and Wigderson prove their result by pre- mial encoding, a first derandomized XOR Lemma, and a second senting a “randomness-efficient amplification of hardness” based derandomized XOR Lemma) that are composed with the Nisan– on a derandomized version of Yao’s XOR Lemma. The hardness- Wigderson [NW94] generator. In this paper we present two dif- amplification procedure is then composed with the Nisan–Wigderson ferent approaches to proving the main result of Impagliazzo and (NW) generator [NW94] and this gives the result. The hardness Wigderson. In developing each approach, we introduce new tech- amplification goes through three steps: an encoding using multi- niques and prove new results that could be useful in future improve- variate polynomials (from [BFNW93]), a first derandomized XOR ments and/or applications of hardness-randomness trade-offs.
    [Show full text]