Bounded Independence Plus Noise

Total Page:16

File Type:pdf, Size:1020Kb

Bounded Independence Plus Noise Bounded Independence Plus Noise by Chin Ho Lee A Thesis Submitted in Partial Fulfilment of the Requirements for the Degree of Doctor of Philosophy in Computer Science Northeastern University August 2019 c 2019 Chin Ho Lee All rights reserved. To my parents Abstract Bounded Independence Plus Noise by Chin Ho Lee Doctor of Philosophy Khoury College of Computer Science Northeastern University Derandomization is a fundamental research area in theoretical computer science. In the past decade, researchers have been able to derandomize a number of computational problems, leading to breakthrough discoveries such as proving SL = L, explicit constructions of Ramsey graphs and optimal-rate error-correcting codes. Bounded independent and small-bias distributions are two pseudorandom primitives that are used extensively in derandomization. This thesis studies the power of these primitives under the perturbation of noise. We give positive and negative results on these perturbed distributions. In particular, we show that they are significantly more powerful than the unperturbed ones, and have the potential to solve long standing open problems such as proving RL = L and AC0[⊕] lower bounds. As applications, we give new lower bounds on the complexity of decoding error-correcting codes, nearly-optimal pseudorandom generators for old and new classes of tests, and limita- tions on the sum of small-bias distributions. i Acknowledgements Foremost, I would like to thank Manu, Daniel, Jon and Omer for taking the time to serve on my thesis committee. This thesis would have been impossible without the wise guidance of Manu. I am truly grateful for his patience during my six years of PhD. His unique perspectives on many matters, whether they are related to research or not, have made a great impact on my life. His striving for simplicity and his clarity of writing will always be examples for me to pursue in the future. I thank Ravi Boppana, Elad Haramaty, Johan H˚astadand Manu, with whom I have collaborated on several results related to this thesis, for sharing their ideas with me. I have learned a great deal of doing research through these collaborations. I am extremely grateful to Amnon Ta-Shma for hosting me at Tel-Aviv University, and to Dean Doron and Gil Cohen for many stimulating discussions during my visit. I thank Salil Vadhan for his excellent course on pseudorandomness at Harvard. My understanding of pseudorandomness would not have been the same without it. I thank Andrej Bogdanov for teaching me Fourier Analysis during my masters, a tool that is used extensively in this thesis. I also thank Andrej for being available for discussions whenever I went home for a visit, and for giving me the opportunities to give talks at the theory seminars in CUHK. My PhD life would have been miserable without my friends at Harvard and Khoury, in particular those at MD138 and WVH266. I thank them for keeping the office spaces full of positive energy, and organizing all kinds of activities to put my mind at ease when getting depressed from research. Finally, I thank my beloved Sabrina for her endless support, patience and love on the other side of the world. ii Contents Abstract.........................................i Acknowledgements................................... ii Table of Contents.................................... iii 1 Introduction1 1.1 Contribution of this thesis............................3 1.2 Organization of this thesis............................4 2 Bounded Independence Plus Noise Fools Products7 2.1 Our results....................................9 2.1.1 Application: The complexity of decoding................ 10 2.1.2 Application: Pseudorandomness..................... 14 2.1.3 Techniques................................. 16 2.2 Bounded independence plus noise fools products................ 17 2.2.1 Preliminaries............................... 20 2.2.2 Proof of Theorem 2.22.......................... 22 2.3 Proofs for Section 2.1.1.............................. 24 2.4 Pseudorandomness: I.............................. 26 2.5 Pseudorandomness, II.............................. 29 2.5.1 Proof of Theorem 2.38.......................... 29 2.5.2 A recursive generator........................... 32 2.6 Pseudorandomness, III.............................. 35 2.7 A lower bound on b and η ............................ 36 3 Pseudorandom Generators for Read-Once Polynomials 37 3.1 Our results.................................... 37 3.1.1 Techniques................................. 40 3.2 Bounded independence plus noise fools products................ 44 3.3 Pseudorandom generators............................ 48 3.4 On almost k-wise independent variables with small total-variance...... 52 3.4.1 Preliminaries............................... 53 3.4.2 Proof of Lemma 3.12........................... 54 3.5 Improved bound for bounded independence plus noise fools products.... 68 3.5.1 Noise reduces variance of bounded complex-valued functions..... 69 iii 3.5.2 XOR Lemma for bounded independence................ 71 3.5.3 Proof of Theorem 3.9........................... 73 3.5.4 Proof of Theorem 3.11.......................... 76 3.6 Small-bias plus noise fools degree-2 polynomials................ 77 3.7 Proof of Claim 3.8................................ 78 3.8 Moment bounds for sum of almost d-wise independent variables....... 80 4 Fourier Bounds and Pseudorandom Generators for Product Tests 85 4.1 Our results.................................... 85 4.1.1 Techniques................................. 89 4.2 Fourier spectrum of product tests........................ 92 4.2.1 Schur-concavity of g ........................... 97 4.2.2 Lower bound............................... 99 4.3 Pseudorandom generators............................ 99 4.3.1 Generator for product tests....................... 101 4.3.2 Almost-optimal generator for XOR of Boolean functions....... 102 4.4 Level-d inequalities................................ 105 5 Some Limitations of the Sum of Small-Bias Distributions 109 5.1 Our results.................................... 110 5.2 Our techniques.................................. 113 5.3 Our counterexamples............................... 116 5.3.1 General circuits.............................. 116 5.3.2 NC2 circuits............................... 117 5.3.3 One-way log-space computation..................... 117 5.3.4 Depth 3 circuits, DNF formulas and AC0 circuits........... 118 5.3.5 Mod 3 linear functions.......................... 119 5.3.6 Sum of k copies of small-bias distributions............... 123 5.4 Mod 3 rank of k-wise independence....................... 124 5.4.1 Lower bound for almost k-wise independence.............. 124 5.4.2 Pairwise independence.......................... 125 5.5 Complexity of decoding.............................. 127 Bibilography 129 A Fooling read-once DNF formulas 141 iv Chapter 1 Introduction The theory of pseudorandomness studies explicit constructions of objects that appear random to restricted classes of tests. It has numerous connections to computer science and math- ematics, including algorithms, computational complexity, cryptography and combinatorics. In particular, the study of pseudorandomness is indispensable in understanding the power of randomness in computation, a fundamental research area in theoretical computer science. While it is known that certain tasks in areas such as cryptography and distributed computing are impossible without randomness, researchers have been able to derandomize a number of computational problems, showing that randomness often does not give us significant savings in computational resources over determinism. One central open question in derandomization is the BPP vs. P question. It asks whether probabilitistic polynomial-time algorithms can be made deterministic without a drastic draw- back in its running time. This question is largely open, as resolving this question would imply circuit lower bounds that seem beyond reach given our current techniques [IKW02, KI04]. Because of this, research in derandomization can be divided into two directions. • Conditional results: One line of research, pioneered by Blum and Micali [BM84], and Yao [Yao82], shows that derandomization can be realized under the assumption that hard functions exist. This approach has found a lot of success in cryptography, where cryptographic primitives are constructed based on the intractability of several specific computational problems. Indeed, this approach was first proposed for cryptographic applications, and the idea of trading hardness for randomness was first suggested by Shamir [Sha81], who constructed pseudorandom sequences assuming the hardness of the RSA encryption function. The seminal work of Nisan and Wigderson [NW94] shows that derandomizing BPP is possible under weaker complexity assumptions. This is followed by the subsequent work of Impagliazzo and Wigderson [IW99], which showed that if any problem in the class E requires circuits of super-polynomial size to approximate, then BPP = P. Since then, researchers have tried to optimize the hardness and randomness trade-off. • Unconditional results: Another line of research turns to derandomizing restricted classes of computation models for which we can prove unconditional lower bounds. 1 Two major classes of tests that have received a lot of attention since the late 80s are constant-depth circuits [AW89] and space-bounded computation [AKS87]. A major open problem in derandomizing space-bounded computation is the RL vs. L question, which is a space-analogue of BPP vs.
Recommended publications
  • Garbled Protocols and Two-Round MPC from Bilinear Maps
    58th Annual IEEE Symposium on Foundations of Computer Science Garbled Protocols and Two-Round MPC from Bilinear Maps Sanjam Garg Akshayaram Srinivasan Dept. of Computer Science Dept. of Computer Science University of California, Berkeley University of California, Berkeley Berkeley, USA Berkeley, USA Email: [email protected] Email: [email protected] Abstract—In this paper, we initiate the study of garbled (OT) protocol [57], [2], [51], [40] gives an easy solution to protocols — a generalization of Yao’s garbled circuits construc- the problem of (semi-honest) two-round secure computation tion to distributed protocols. More specifically, in a garbled in the two-party setting. However, the same problem for protocol construction, each party can independently generate a garbled protocol component along with pairs of input the multiparty setting turns out to be much harder. Beaver, labels. Additionally, it generates an encoding of its input. The Micali and Rogaway [7] show that garbled circuits can be evaluation procedure takes as input the set of all garbled used to realize a constant round multi-party computation protocol components and the labels corresponding to the input protocol. However, unlike the two-party case, this protocol encodings of all parties and outputs the entire transcript of the is not two rounds. distributed protocol. We provide constructions for garbling arbitrary protocols A. Garbled Protocols based on standard computational assumptions on bilinear maps (in the common random string model). Next, using In this paper, we introduce a generalization of Yao’s garbled protocols we obtain a general compiler that compresses construction from circuits to distributed protocols. We next any arbitrary round multiparty secure computation protocol elaborate on (i) what it means to garble a protocol, (ii) why into a two-round UC secure protocol.
    [Show full text]
  • Onyx: New Encryption and Signature Schemes with Multivariate Public Key in Degree 3
    Onyx: New Encryption and Signature Schemes with Multivariate Public Key in Degree 3 Gilles Macario-Rat1 and Jacques Patarin2 1 Orange, Orange Gardens, 46 avenue de la R´epublique,F-92320 Ch^atillon,France [email protected] 2 Versailles Laboratory of Mathematics, UVSQ, CNRS, University of Paris-Saclay [email protected] Abstract. In this paper, we present a new secret trapdoor function for the design of multivariate schemes that we call \Onyx", suitable for en- cryption and signature. It has been inspired by the schemes presented in [19,20]. From this idea, we present some efficient encryption and signa- ture multivariate schemes with explicit parameters that resist all known attacks. In particular they resist the two main (and often very power- ful) attacks in this area: the Gr¨obner attacks (to compute a solution of the system derived from the public key) and the MinRank attacks (to recover the secret key). Specific attacks due to the properties of the function and its differential are also addressed in this paper. The \Onyx" schemes have public key equations of degree 3. Despite this, the size of the public key may still be reasonable since we can use larger fields and smaller extension degrees. Onyx signatures can be as short as the \birth- day paradox" allows, i.e. twice the security level, or even shorter thanks to the Feistel-Patarin construction, like many other signatures schemes based on multivariate equations. Keywords: public-key cryptography, post-quantum multivariate cryptography, UOV, HFE, Gr¨obnerbasis, MinRank problem, differential attacks. 1 Introduction Many schemes in Multivariate cryptography have been broken.
    [Show full text]
  • ACM SIGLOG News 1 October 2015, Vol
    Volume 2, Number 4 Published by the Association for Computing Machinery Special Interest Group on Logic and Computation October 2015 SIGLOG news TABLE OF CONTENTS General Information 1 From the Editor Andrzej Murawski 2 Chair's Letter Prakash Panangaden Technical Columns 3 Automata Mikołaj Bojańczyk 16 Verication Neha Rungta Announcements 26 Gödel Prize - Call for Nominations 28 SIGLOG Monthly 175 SIGLOG NEWS Published by the ACM Special Interest Group on Logic and Computation SIGLOG Executive Committee Chair Prakash Panangaden McGill University Vice-Chair Luke Ong University of Oxford Treasurer Natarajan Shankar SRI International Secretary Alexandra Silva Radboud University Nijmegen Catuscia Palamidessi INRIA and LIX, Ecole´ Polytechnique EACSL President Anuj Dawar University of Cambridge EATCS President Luca Aceto Reykjavik University ACM ToCL E-in-C Dale Miller INRIA and LIX, Ecole´ Polytechnique Andrzej Murawski University of Warwick Veronique´ Cortier CNRS and LORIA, Nancy ADVISORY BOARD Mart´ın Abadi Google and UC Santa Cruz Phokion Kolaitis University of California, Santa Cruz Dexter Kozen Cornell University Gordon Plotkin University of Edinburgh Moshe Vardi Rice University COLUMN EDITORS Automata Mikołaj Bojanczyk´ University of Warsaw Complexity Neil Immerman University of Massachusetts Amherst Security and Privacy Matteo Maffei CISPA, Saarland University Semantics Mike Mislove Tulane University Verification Neha Rungta SGT Inc. and NASA Ames Notice to Contributing Authors to SIG Newsletters By submitting your article for distribution
    [Show full text]
  • The Gödel Prize 2020 - Call for Nominatonn
    The Gödel Prize 2020 - Call for Nominatonn Deadline: February 15, 2020 The Gödel Prize for outntanding papern in the area of theoretial iomputer niienie in nponnored jointly by the European Annoiiaton for Theoretial Computer Siienie (EATCS) and the Annoiiaton for Computng Maihinery, Speiial Innterent Group on Algorithmn and Computaton Theory (AC M SInGACT) The award in prenented annually, with the prenentaton taaing plaie alternately at the Innternatonal Colloquium on Automata, Languagen, and Programming (InCALP) and the AC M Symponium on Theory of Computng (STOC) The 28th Gödel Prize will be awarded at the 47th Innternatonal Colloquium on Automata, Languagen, and Programming to be held during 8-12 July, 2020 in Beijing The Prize in named in honour of Kurt Gödel in reiogniton of hin major iontributonn to mathematial logii and of hin interent, diniovered in a leter he wrote to John von Neumann nhortly before von Neumann’n death, in what han beiome the famoun “P vernun NP” quenton The Prize iniluden an award of USD 5,000 Award Committee: The 2020 Award Commitee ionnintn of Samnon Abramnay (Univernity of Oxford), Anuj Dawar (Chair, Univernity of Cambridge), Joan Feigenbaum (Yale Univernity), Robert Krauthgamer (Weizmann Innnttute), Daniel Spielman (Yale Univernity) and David Zuiaerman (Univernity of Texan, Auntn) Eligibility: The 2020 Prize rulen are given below and they nupernede any diferent interpretaton of the generii rule to be found on webniten of both SInGACT and EATCS Any renearih paper or nerien of papern by a ningle author or by
    [Show full text]
  • SETH-Based Lower Bounds for Subset Sum and Bicriteria Path
    SETH-Based Lower Bounds for Subset Sum and Bicriteria Path Amir Abboud1, Karl Bringmann2, Danny Hermelin3, and Dvir Shabtay3 1 Department of Computer Science, Stanford University, CA, USA [email protected] 2 Max Planck Institute for Informatics, Saarland Informatics Campus, Germany [email protected] 3 Department of Industrial Engineering and Management, Ben-Gurion University, Israel [email protected], [email protected] Abstract. Subset Sum and k-SAT are two of the most extensively studied problems in computer science, and conjectures about their hardness are among the cornerstones of fine-grained complexity. An important open problem in this area is to base the hardness of one of these problems on the other. Our main result is a tight reduction from k-SAT to Subset Sum on dense instances, proving that Bellman’s 1962 pseudo-polynomial O∗(T )-time algorithm for Subset Sum on n numbers and target T − cannot be improved to time T 1 ε · 2o(n) for any ε> 0, unless the Strong Exponential Time Hypothesis (SETH) fails. As a corollary, we prove a “Direct-OR” theorem for Subset Sum under SETH, offering a new tool for proving conditional lower bounds: It is now possible to assume that deciding whether one out of N − given instances of Subset Sum is a YES instance requires time (NT )1 o(1). As an application of this corollary, we prove a tight SETH-based lower bound for the classical Bicriteria s,t-Path problem, which is extensively studied in Operations Research. We separate its complexity from that of Subset Sum: On graphs with m edges and edge lengths bounded by L, we show that the O(Lm) pseudo- polynomial time algorithm by Joksch from 1966 cannot be improved to O˜(L + m), in contrast to a recent improvement for Subset Sum (Bringmann, SODA 2017).
    [Show full text]
  • Some Notions of Entropy for Cryptography∗
    Some Notions of Entropy for Cryptography∗ Leonid Reyzin Boston University Computer Science http://www.cs.bu.edu/~reyzin June 2, 2011 Abstract This paper presents a brief and (necessarily) incomplete survey of some notions of entropy that have been recently used in the analysis of cryptographic constructions. It focuses on min- entropy and its extensions to the cases when the adversary has correlated information and/or is computationally bounded. It also presents results that can be used to bound such entropy and apply it to the analysis of cryptographic constructions. 1 Information-Theoretic Case In many contexts, particularly in security-related ones, the ability to guess the value of a random variable (in a single attempt) is an important measures of the variable’s quality. This ability is captured by the following notion. Definition 1. A random variable X has min-entropy k, denoted H∞(X) = k, if max Pr[X = x] = 2−k. x Randomness extractors were defined to work with any distribution that has min-entropy [NZ96]. Moreover, strong extractors (whose outputs are nearly uniform even the presence of the seed) produce outputs that have, with high probability over the choice of seed, almost maximal min- entropy. Lemma 1 ([CKOR10]). If Ext : N × I → {0, 1}` is a (k, ε)-strong extractor with inputs from a set N and seeds from a distribution I, and X is a random variable taking values in N with H∞(X) ≥ k, ` then H∞(Ext(X; i)) ≥ ` − 1 with probability at least 1 − 2 ε over the choice of the seed i.
    [Show full text]
  • The Gödel Prize 2019 - Call for Nominations Deadline: February 15, 2019
    The Gödel Prize 2019 - Call for Nominations Deadline: February 15, 2019 The Gödel Prize for outstanding papers in the area of theoretical computer science is sponsored jointly by the European Association for Theoretical Computer Science (EATCS) and the Association for Computing Machinery, Special Interest Group on Algorithms and Computation Theory (ACM SIGACT). The award is presented annually, with the presentation taking place alternately at the International Colloquium on Automata, Languages, and Programming (ICALP) and the ACM Symposium on Theory of Computing (STOC). The 27th Gödel Prize will be awarded at 51st Annual ACM Symposium on the Theory of Computing to be held during June 23-26, 2019 in Phoenix, AZ. The Prize is named in honor of Kurt Gödel in recognition of his major contributions to mathematical logic and of his interest, discovered in a letter he wrote to John von Neumann shortly before von Neumann’s death, in what has become the famous “P versus NP” question. The Prize includes an award of USD 5,000. Award Committee: The 2019 Award Committee consists of Anuj Dawar (Cambridge University), Robert Krauthgamer (Weizmann Institute), Joan Feigenbaum (Yale University), Giuseppe Persiano (Università di Salerno), Omer Reingold (Chair, Stanford University) and Daniel Spielman (Yale University). Eligibility: The 2019 Prize rules are given below and they supersede any different interpretation of the generic rule to be found on websites of both SIGACT and EATCS. Any research paper or series of papers by a single author or by a team of authors is deemed eligible if: - The main results were not published (in either preliminary or final form) in a journal or conference proceedings before January 1st, 2006.
    [Show full text]
  • Bounded Independence Plus Noise Fools Products
    Bounded independence plus noise fools products Elad Haramaty∗ Chin Ho Lee∗ Emanuele Viola∗ May 4, 2017 Abstract Let D be a b-wise independent distribution over f0; 1gm. Let E be the \noise" distribution over f0; 1gm where the bits are independent and each bit is 1 with prob- ability η=2. We study which tests f : f0; 1gm ! [−1; 1] are "-fooled by D + E, i.e., j E[f(D + E)] − E[f(U)]j ≤ " where U is the uniform distribution. We show that D + E"-fools product tests f :(f0; 1gn)k ! [−1; 1] given by the product of k bounded functions on disjoint n-bit inputs with error " = k(1 − η)Ω(b2=m), where m = nk and b ≥ n. This bound is tight when b = Ω(m) and η ≥ (log k)=m. For b ≥ m2=3 log m and any constant η the distribution D + E also 0:1-fools log-space algorithms. We develop two applications of this type of results. First, we prove communi- cation lower bounds for decoding noisy codewords of length m split among k par- ties. For Reed{Solomon codes of dimension m=k where k = O(1), communication Ω(ηm) − O(log m) is required to decode one message symbol from a codeword with ηm errors, and communication O(ηm log m) suffices. Second, we obtain pseudorandom generators. We can "-fool product tests f :(f0; 1gn)k ! [−1; 1] under any permutation ~ 2 ~ p of the bits with seed lengths 2n + O(k log(1=")) andp O(n) + O( nk log 1=").
    [Show full text]
  • Cryptographic Assumptions: a Position Paper
    Cryptographic Assumptions: A Position Paper Shafi Goldwasser ∗ Yael Tauman Kalai y Abstract The mission of theoretical cryptography is to define and construct provably secure cryptographic protocols and schemes. Without proofs of security, cryptographic con- structs offer no guarantees whatsoever and no basis for evaluation and comparison. As most security proofs necessarily come in the form of a reduction between the security claim and an intractability assumption, such proofs are ultimately only as good as the assumptions they are based on. Thus, the complexity implications of every assumption we utilize should be of significant substance, and serve as the yard stick for the value of our proposals. Lately, the field of cryptography has seen a sharp increase in the number of new assumptions that are often complex to define and difficult to interpret. At times, these assumptions are hard to untangle from the constructions which utilize them. We believe that the lack of standards of what is accepted as a reasonable crypto- graphic assumption can be harmful to the credibility of our field. Therefore, there is a great need for measures according to which we classify and compare assumptions, as to which are safe and which are not. In this paper, we propose such a classification and review recently suggested assumptions in this light. This follows the footsteps of Naor (Crypto 2003). Our governing principle is relying on hardness assumptions that are independent of the cryptographic constructions. ∗MIT and Weizmann Institute. Email: shafi@theory.csail.mit.edu. yMicrosoft Research. Email: [email protected]. 1 Introduction Conjectures and assumptions are instrumental for the advancement of science.
    [Show full text]
  • EATCS General Assembly 2013
    EATCS General Assembly 2013 Luca Aceto School of Computer Science, Reykjavik University [email protected] 10 July 2013 Luca Aceto EATCS General Assembly 2013 Agenda for the general assembly 1 Reports on ICALP 2013 2 Award of the best student papers 3 Report on the organization of ICALP 2014 (Philip Bille) 4 Proposal for ICALP 2015 for approval by the General Assembly (Kazuo Iwama) 5 Orna Kupferman. The Gender Challenge at TCS. 6 Report from the president 7 Any other business Luca Aceto EATCS General Assembly 2013 In memoriam Kohei Honda Philippe Darondeau (ICALP 2012 invited speaker) Luca Aceto EATCS General Assembly 2013 Reports on ICALP 2013 1 Organizing committee (Agnis Skuskovniks) 2 Marta Kwiatkowska on behalf of the PC chairs Festive occasion This is the 40th ICALP! Luca Aceto EATCS General Assembly 2013 Award of the best student papers Best student paper for Track A Radu Curticapean. Counting matchings of size k is #W [1]-hard Best student paper for Track B Nicolas Basset. A maximal entropy stochastic process for a timed automaton. There was no eligible accepted paper for Track C this year. Luca Aceto EATCS General Assembly 2013 Upcoming ICALP conferences Report on the organization of ICALP 2014 in Copenhagen (Philip Bille) Novelty: Four-day ICALP! PC chairs: Track A: Elias Koutsoupias (Oxford, UK); Track B: Javier Esparza (TU M¨unich,Germany); Track C: Pierre Fraigniaud (Paris Diderot, FR). Invited speakers: Sanjeev Arora (Princeton University, USA), Maurice Herlihy (Brown University, USA), Viktor Kuncak (EPFL, CH) and Claire Mathieu (ENS Paris, France). Proposal for ICALP 2015 in Kyoto for approval by the General Assembly (Kazuo Iwama) Co-located with LICS 2015 First ICALP ever outside Europe! Luca Aceto EATCS General Assembly 2013 Upcoming ICALP conferences Report on the organization of ICALP 2014 in Copenhagen (Philip Bille) Novelty: Four-day ICALP! PC chairs: Track A: Elias Koutsoupias (Oxford, UK); Track B: Javier Esparza (TU M¨unich,Germany); Track C: Pierre Fraigniaud (Paris Diderot, FR).
    [Show full text]
  • SIGACT Viability
    SIGACT Name and Mission: SIGACT: SIGACT Algorithms & Computation Theory The primary mission of ACM SIGACT (Association for Computing Machinery Special Interest Group on Algorithms and Computation Theory) is to foster and promote the discovery and dissemination of high quality research in the domain of theoretical computer science. The field of theoretical computer science is interpreted broadly so as to include algorithms, data structures, complexity theory, distributed computation, parallel computation, VLSI, machine learning, computational biology, computational geometry, information theory, cryptography, quantum computation, computational number theory and algebra, program semantics and verification, automata theory, and the study of randomness. Work in this field is often distinguished by its emphasis on mathematical technique and rigor. Newsletters: Volume Issue Issue Date Number of Pages Actually Mailed 2014 45 01 March 2014 N/A 2013 44 04 December 2013 104 27-Dec-13 44 03 September 2013 96 30-Sep-13 44 02 June 2013 148 13-Jun-13 44 01 March 2013 116 18-Mar-13 2012 43 04 December 2012 140 29-Jan-13 43 03 September 2012 120 06-Sep-12 43 02 June 2012 144 25-Jun-12 43 01 March 2012 100 20-Mar-12 2011 42 04 December 2011 104 29-Dec-11 42 03 September 2011 108 29-Sep-11 42 02 June 2011 104 N/A 42 01 March 2011 140 23-Mar-11 2010 41 04 December 2010 128 15-Dec-10 41 03 September 2010 128 13-Sep-10 41 02 June 2010 92 17-Jun-10 41 01 March 2010 132 17-Mar-10 Membership: based on fiscal year dates 1st Year 2 + Years Total Year Professional
    [Show full text]
  • New Generic Algorithms for Hard Knapsacks
    New generic algorithms for hard knapsacks Nick Howgrave-Graham1 and Antoine Joux2 1 35 Park St, Arlington, MA 02474 [email protected] 2 dga and Universit´ede Versailles Saint-Quentin-en-Yvelines uvsq prism, 45 avenue des Etats-Unis,´ f-78035, Versailles cedex, France [email protected] Abstract. In this paper3, we study the complexity of solving hard knap- sack problems, i.e., knapsacks with a density close to 1 where lattice- based low density attacks are not an option. For such knapsacks, the cur- rent state-of-the-art is a 31-year old algorithm by Schroeppel and Shamir which is based on birthday paradox techniques and yields a running time of O~(2n=2) for knapsacks of n elements and uses O~(2n=4) storage. We propose here two new algorithms which improve on this bound, finally lowering the running time down to O~(20:3113 n) for almost all knapsacks of density 1. We also demonstrate the practicality of these algorithms with an implementation. 1 Introduction The 0{1 knapsack problem or subset sum problem is a famous NP-hard problem which has often been used in the construction of cryptosystems. An instance of this problem consists of a list of n positive integers (a1; a2; ··· ; an) together with another positive integer S. Given an instance, there exist two forms of knapsack problems. The first form is the decision knapsack problem, where we need to decide whether S can be written as: n X S = iai; i=1 with values of i in f0; 1g. The second form is the computational knapsack prob- lem where we need to recover a solution = (1; ··· ; n) if at least one exists.
    [Show full text]