Research Statement Prahladh Harsha

Total Page:16

File Type:pdf, Size:1020Kb

Research Statement Prahladh Harsha Research Statement Prahladh Harsha 1 Introduction My research interests are in the area of theoretical computer science, with special emphasis on computational complexity. The primary focus of my research has been in the area of probabilistically checkable proofs, and related areas such as property testing and information theory. The fundamental question in computational complexity is “what are the limits of feasible computation?” In fact, one of the articulations of this question is the famous “P vs. NP” question, where P refers to the class of problems that can be solved in polynomial time and NP to the class of problems whose solutions can be verified in polynomial time. To understand the limitations of efficient computation, we first need to understand what we mean by “efficient computation.” This natural connection between feasibility and hardness has many a time led to surprising consequences in complexity theory. One prime example is that of probabilistically checkable proofs. The original emphasis in the study of probabilistically checkable proofs was in the area of program checking. Surprisingly, it was soon realized that the existence of efficient probabilistic program checkers actually implied that the approximation versions of several NP-complete optimization problems were as intractable as the original optimization problems. Probabilistically checkable proofs provide an extremely efficient means of proof verification. The classical complexity class NP refers to the class of languages whose membership can be verified with the aid of a polynomial sized proof. Probabilistically checkable proofs (PCPs) are a means of encoding these proofs (and more generally any mathematical proof) into a format such that the encoded proof can be checked very efficiently, although in a probabilistic manner, by looking at it at only a constant number of locations (in fact, 3 bits suffice!) The main question addressed by my research in this area is the following: “how much does this encoding blow up the original proof while retaining the constant number of queries into the proof, and how efficiently (with respect to running time) can the checking be performed?” An important contribution of my work is the notion of a proof of proximity (also called PCP of proximity). A PCP of proximity is a strengthening of a PCP in the sense that it helps to decide if a statement is true with the help of an additional proof in the form of a PCP, by merely probing the statement at a few locations. In other words, a PCP of proximity makes constant probes not only to the proof but also to the statement whose truth it is checking. With such a stringent requirement, a PCP of proximity cannot distinguish true statements from false; however it can distinguish true statements from ones that are far from being true (in the sense that the statement is far from any true statement in Hamming distance). Thus, a PCP of proximity checks if a given statement is close to being true, without even reading the statement in its entirety! Hence, the name, proof of proximity. PCPs of proximity play a vital role in the construction of short PCPs, both in my work and in subsequent developments in the area of probabilistically checkable proofs. PCPs of proximity are also used in coding theory. All known constructions of locally testable codes are via PCPs of proximity. PCPs of proximity have also come very handy in simplifying the original proof of the PCP Theorem, which is one of the most involved proofs in complexity theory. In fact, the recent fully combinatorial proof of the PCP Theorem (due to Dinur [Din07]) crucially relies on PCPs of proximity. As mentioned above, the main focus of my research has been in the area of probabilistically checkable proofs. I have also worked in other areas such as property testing, information theory, proof complexity, network routing etc. Below I elaborate my work in three of these areas – probabilistically checkable proofs, property testing, and information theory. 2 Probabilistically checkable proofs The PCP Theorem The PCP Theorem [AS98, ALM+98] is one of the crowning achievements of com- plexity theory in the last decade. Probabilistically checkable proofs [BFLS91, FGL+96, AS98], as mentioned 1 earlier, are proofs that allow efficient probabilistic verification based on probing just a few bits of the proof. Informally speaking, the PCP Theorem states that any mathematical proof can be rewritten into a polyno- mially longer probabilistically checkable proof (PCP) such that its veracity can be checked very efficiently, although in a probabilistic manner, by looking at the rewritten proof at only a constant number of locations (in fact, 3 bits suffice) and furthermore proofs of false assertions are rejected with probability at least 1/2. The PCP Theorem has, since its discovery, attracted a lot of attention, motivated by its connection to inapproximability of optimization problems [FGL+96, AS98]. This connection led to a long line of fruitful research yielding inapproximability results (many of them optimal) of several optimization problems (e.g., Set-Cover [Fei98], Max-Clique [H˚as99], MAX-3SAT [H˚as01]). However, the significance of PCPs extends far beyond their applicability to deriving inapproximability results. The mere fact that proofs can be transformed into a format that supports super-fast probabilistic verification is remarkable. One would have naturally expected PCPs, as the name suggests, to lead to vast improvement in automated proof-checkers, theorem-provers, etc. However, unfortunately, this has not been the case. The chief reason why PCPs are not being used today in practice for automated proof-checking is that the blowup of the proof-size involved in all present constructions of PCPs makes it infeasible to do so. Just to put things in perspective, the original proof of the PCP Theorem [ALM+98] constructed PCPs of nearly cubic length with a query complexity roughly of the order of a million (in order to reject proofs of false assertion with probability at least 1/2). On the other hand, the 3-query optimal PCPs of [H˚as01, GLST98] 6 have length nearly n10 , which is still a polynomial! Even with respect to inapproximability results, though the PCP Theorem has been extremely successful in proving tight hardness results, the quantitative nature of these results has been rather unsatisfactory, once again due to the blowup involved in PCP constructions. To understand this it is instructive to compare the inapproximability hardness results obtained from the PCP Theorem with the optimization hardness results obtained from the usual textbook NP-completeness reductions. For example, the NP-completeness reduction from Satisfiability (SAT) to Clique transforms a Boolean formula on n variables to a graph with at most 10n vertices. On the other hand, the PCP reductions which show optimal inapproximability of Max- 6 Clique transform a Boolean formula of size n to a graph of size at least n10 . What these results imply in quantitative terms is that if one assumes solving satisfiability on formulae with 1, 000 variables is intractable, then NP-completeness reductions imply that solving Clique is intractable on graphs with 10, 000 vertices; while, the PCP reductions would imply that the optimal inapproximability hardness results for Max-Clique 6 sets in on graphs of size at least 100010 . 2.1 My research Short PCPs: Most of my work in the area of PCPs has focused on constructing short PCPs. In work done as part of my master’s thesis [HS00], I examine the size and query complexity of PCPs jointly and obtain a construction with reasonable performance in both parameters (more precisely n3 sized proofs with a query complexity of 16 bits). In a more recent work with Ben-Sasson, Goldreich, Sudan and Vadhan [BGH+06], I take a closer look at the PCP Theorem, simplify several parameters and obtain shorter PCPs. In quantitative terms, we obtain PCPs that are at most n · exp(logε n) in the size of the original proof for any ε > 0. PCPs of proximity and composition: Besides constructing short PCPs, the chief contribution of our work [BGH+06] is the simplification of PCP constructions. Previous construction of PCPs were extremely involved and elaborate. One of the main reasons for this is that “proof composition,” a key ingredient in all known PCP constructions, is a very involved process. We introduce “PCPs of proximity,” (a variant of PCPs mentioned earlier in Section 1) which facilitate very smooth composition – in fact, composition becomes almost definitional and syntactic given this variant. This new variant of PCPs and the corresponding composition have played a critical role in subsequent improvements in short PCP constructions (due to Ben- Sasson and Sudan [BS05] and Dinur [Din07]). Furthermore, these simplifications in the original proof of the PCP Theorem, in the guise of PCPs of proximity and the new composition, led to an alternate purely combinatorial proof of the PCP Theorem, due to Dinur [Din07]. This work [BGH+06] was invited to the 2 special issue of SIAM Journal on Computing on Randomness and Computation as well as the special issue of SIAM Journal on Computing for STOC 2004. Efficient PCPs: In the context of efficient proof verifiers, the running time of the verification process is as important a parameter as the length of the PCP. In fact, the emphases of the initial work of Babai et al. [BFLS91] in the area of PCPs was the time taken by the verifier and the length of the proof in the new format. In contrast, most succeeding works on PCPs have focused on the query complexity of the verifier and derived many strong inapproximability results for a wide variety of optimization problems; however, no later work seems to have returned to the question of the extreme efficiency of the verifier.
Recommended publications
  • 2020 SIGACT REPORT SIGACT EC – Eric Allender, Shuchi Chawla, Nicole Immorlica, Samir Khuller (Chair), Bobby Kleinberg September 14Th, 2020
    2020 SIGACT REPORT SIGACT EC – Eric Allender, Shuchi Chawla, Nicole Immorlica, Samir Khuller (chair), Bobby Kleinberg September 14th, 2020 SIGACT Mission Statement: The primary mission of ACM SIGACT (Association for Computing Machinery Special Interest Group on Algorithms and Computation Theory) is to foster and promote the discovery and dissemination of high quality research in the domain of theoretical computer science. The field of theoretical computer science is the rigorous study of all computational phenomena - natural, artificial or man-made. This includes the diverse areas of algorithms, data structures, complexity theory, distributed computation, parallel computation, VLSI, machine learning, computational biology, computational geometry, information theory, cryptography, quantum computation, computational number theory and algebra, program semantics and verification, automata theory, and the study of randomness. Work in this field is often distinguished by its emphasis on mathematical technique and rigor. 1. Awards ▪ 2020 Gödel Prize: This was awarded to Robin A. Moser and Gábor Tardos for their paper “A constructive proof of the general Lovász Local Lemma”, Journal of the ACM, Vol 57 (2), 2010. The Lovász Local Lemma (LLL) is a fundamental tool of the probabilistic method. It enables one to show the existence of certain objects even though they occur with exponentially small probability. The original proof was not algorithmic, and subsequent algorithmic versions had significant losses in parameters. This paper provides a simple, powerful algorithmic paradigm that converts almost all known applications of the LLL into randomized algorithms matching the bounds of the existence proof. The paper further gives a derandomized algorithm, a parallel algorithm, and an extension to the “lopsided” LLL.
    [Show full text]
  • Program Sunday Evening: Welcome Recep- Tion from 7Pm to 9Pm at the Staff Lounge of the Department of Computer Science, Ny Munkegade, Building 540, 2Nd floor
    Computational ELECTRONIC REGISTRATION Complexity The registration for CCC’03 is web based. Please register at http://www.brics.dk/Complexity2003/. Registration Fees (In Danish Kroner) Eighteenth Annual IEEE Conference Advance† Late Members‡∗ 1800 DKK 2200 DKK ∗ Sponsored by Nonmembers 2200 DKK 2800 DKK Students+ 500 DKK 600 DKK The IEEE Computer Society ∗The registration fee includes a copy of the proceedings, Technical Committee on receptions Sunday and Monday, the banquet Wednesday, and lunches Monday, Tuesday and Wednesday. Mathematical Foundations +The registration fee includes a copy of the proceedings, of Computing receptions Sunday and Monday, and lunches Monday, Tuesday and Wednesday. The banquet Wednesday is not included. †The advance registration deadline is June 15. ‡ACM, EATCS, IEEE, or SIGACT members. Extra proceedings/banquet tickets Extra proceedings are 350 DKK. Extra banquet tick- ets are 300 DKK. Both can be purchased when reg- istering and will also be available for sale on site. Alternative registration If electronic registration is not possible, please con- tact the organizers at one of the following: E-mail: [email protected] Mail: Complexity 2003 c/o Peter Bro Miltersen In cooperation with Department of Computer Science University of Aarhus ACM-SIGACT and EATCS Ny Munkegade, Building 540 DK 8000 Aarhus C, Denmark Fax: (+45) 8942 3255 July 7–10, 2003 Arhus,˚ Denmark Conference homepage Conference Information Information about this year’s conference is available Location All sessions of the conference and the on the Web at Kolmogorov workshop will be held in Auditorium http://www.brics.dk/Complexity2003/ F of the Department of Mathematical Sciences at Information about the Computational Complexity Aarhus University, Ny Munkegade, building 530, 1st conference is available at floor.
    [Show full text]
  • SIGACT News Complexity Theory Column 93
    SIGACT News Complexity Theory Column 93 Lane A. Hemaspaandra Dept. of Computer Science University of Rochester Rochester, NY 14627, USA Introduction to Complexity Theory Column 93 This issue This issue features Swastik Kopparty and Shubhangi Saraf's column, \Local Testing and Decoding of High-Rate Error-Correcting Codes." Warmest thanks to the authors for their fascinating, exciting column! Future issues And please stay tuned to future issues' Complexity Theory Columns, for articles by Neeraj Kayal and Chandan Saha (tentative topic: arithmetic circuit lower bounds); Stephen Fenner and Thomas Thierauf (tentative topic: related to the STOC 2016 quasi-NC bipartite perfect matching algorithm of Fenner, Gurjar, and Thierauf); Srinivasan Arunachalam and Ronald de Wolf (tentative topic: quantum machine learning); and Ryan O'Donnell and John Wright (tentative topic: learning and testing quantum states). 2016{2017 As this issue reaches you, the 2016{2017 academic year will be starting. For those of you who are faculty members, I wish you a theorem-filled year, including the treat of working on research with bright graduate and undergraduate students. For those you who are students (bright ones, of course|you're reading this column!), if you are not already working with a faculty member on research and would like to, don't be shy! Even if you're a young undergraduate, it makes a lot of sense to pop by your favorite theory professor's office, and let him or her know what has most interested you so far within theory, if something has (perhaps a nifty article you read by Kopparty and Saraf?), and that you find theory thrilling, if you do (admit it, you do!), and that you'd love to be doing theory research as soon as possible and would deeply appreciate any advice as to what you can do to best make that happen.
    [Show full text]
  • Toward Probabilistic Checking Against Non-Signaling Strategies with Constant Locality
    Electronic Colloquium on Computational Complexity, Report No. 144 (2020) Toward Probabilistic Checking against Non-Signaling Strategies with Constant Locality Mohammad Mahdi Jahanara Sajin Koroth Igor Shinkar [email protected] sajin [email protected] [email protected] Simon Fraser University Simon Fraser University Simon Fraser University September 7, 2020 Abstract Non-signaling strategies are a generalization of quantum strategies that have been studied in physics over the past three decades. Recently, they have found applications in theoretical computer science, including to proving inapproximability results for linear programming and to constructing protocols for delegating computation. A central tool for these applications is probabilistically checkable proof (PCPs) systems that are sound against non-signaling strategies. In this paper we show, assuming a certain geometrical hypothesis about noise robustness of non-signaling proofs (or, equivalently, about robustness to noise of solutions to the Sherali- Adams linear program), that a slight variant of the parallel repetition of the exponential-length constant-query PCP construction due to Arora et al. (JACM 1998) is sound against non-signaling strategies with constant locality. Our proof relies on the analysis of the linearity test and agreement test (also known as the direct product test) in the non-signaling setting. Keywords: direct product testing; linearity testing; non-signaling strategies; parallel repetition; proba- bilistically checkable proofs 1 ISSN 1433-8092 Contents 1 Introduction 3 1.1 Informal statement of the result . .5 1.2 Roadmap . .6 2 Preliminaries 6 2.1 Probabilistically Checkable Proofs . .6 2.2 Parallel repetition . .7 2.3 Non-signaling functions . .7 2.4 Permutation folded repeated non-signaling functions .
    [Show full text]
  • UC Berkeley UC Berkeley Electronic Theses and Dissertations
    UC Berkeley UC Berkeley Electronic Theses and Dissertations Title Hardness of Maximum Constraint Satisfaction Permalink https://escholarship.org/uc/item/5x33g1k7 Author Chan, Siu On Publication Date 2013 Peer reviewed|Thesis/dissertation eScholarship.org Powered by the California Digital Library University of California Hardness of Maximum Constraint Satisfaction by Siu On Chan A dissertation submitted in partial satisfaction of the requirements for the degree of Doctor of Philosophy in Computer Science in the Graduate Division of the University of California, Berkeley Committee in charge: Professor Elchanan Mossel, Chair Professor Luca Trevisan Professor Satish Rao Professor Michael Christ Spring 2013 Hardness of Maximum Constraint Satisfaction Creative Commons 3.0 BY: C 2013 by Siu On Chan 1 Abstract Hardness of Maximum Constraint Satisfaction by Siu On Chan Doctor of Philosophy in Computer Science University of California, Berkeley Professor Elchanan Mossel, Chair Maximum constraint satisfaction problem (Max-CSP) is a rich class of combinatorial op- timization problems. In this dissertation, we show optimal (up to a constant factor) NP- hardness for maximum constraint satisfaction problem with k variables per constraint (Max- k-CSP), whenever k is larger than the domain size. This follows from our main result con- cerning CSPs given by a predicate: a CSP is approximation resistant if its predicate contains a subgroup that is balanced pairwise independent. Our main result is related to previous works conditioned on the Unique-Games Conjecture and integrality gaps in sum-of-squares semidefinite programming hierarchies. Our main ingredient is a new gap-amplification technique inspired by XOR-lemmas. Using this technique, we also improve the NP-hardness of approximating Independent-Set on bounded-degree graphs, Almost-Coloring, Two-Prover-One-Round-Game, and various other problems.
    [Show full text]
  • László Lovász Avi Wigderson of Eötvös Loránd University of the Institute for Advanced Study, in Budapest, Hungary and Princeton, USA
    2021 The Norwegian Academy of Science and Letters has decided to award the Abel Prize for 2021 to László Lovász Avi Wigderson of Eötvös Loránd University of the Institute for Advanced Study, in Budapest, Hungary and Princeton, USA, “for their foundational contributions to theoretical computer science and discrete mathematics, and their leading role in shaping them into central fields of modern mathematics.” Theoretical Computer Science (TCS) is the study of computational lens”. Discrete structures such as the power and limitations of computing. Its roots go graphs, strings, permutations are central to TCS, back to the foundational works of Kurt Gödel, Alonzo and naturally discrete mathematics and TCS Church, Alan Turing, and John von Neumann, leading have been closely allied fields. While both these to the development of real physical computers. fields have benefited immensely from more TCS contains two complementary sub-disciplines: traditional areas of mathematics, there has been algorithm design which develops efficient methods growing influence in the reverse direction as well. for a multitude of computational problems; and Applications, concepts, and techniques from TCS computational complexity, which shows inherent have motivated new challenges, opened new limitations on the efficiency of algorithms. The notion directions of research, and solved important open of polynomial-time algorithms put forward in the problems in pure and applied mathematics. 1960s by Alan Cobham, Jack Edmonds, and others, and the famous P≠NP conjecture of Stephen Cook, László Lovász and Avi Wigderson have been leading Leonid Levin, and Richard Karp had strong impact on forces in these developments over the last decades. the field and on the work of Lovász and Wigderson.
    [Show full text]
  • Beating Brute Force for Polynomial Identity Testing of General Depth-3 Circuits
    Electronic Colloquium on Computational Complexity, Report No. 111 (2018) Beating Brute Force for Polynomial Identity Testing of General Depth-3 Circuits V. Arvind∗ Abhranil Chatterjeey Rajit Dattaz Partha Mukhopadhyayx June 4, 2018 Abstract Let C be a depth-3 ΣΠΣ arithmetic circuit of size s, computing a polynomial f 2 F[x1; : : : ; xn] (where F = Q or C) with fan-in of product gates bounded by d. We give a deterministic time 2d poly(n; s) polynomial identity testing algorithm to check whether f ≡ 0 or not. In the case of finite fields, for Char(F) > d we obtain a deterministic algorithm of running time 2γ·d poly(n; s), whereas for Char(F) ≤ d, we obtain a deterministic algorithm of running time 2(γ+2)·d log d poly(n; s) where γ ≤ 5. 1 Introduction Polynomial Identity Testing (PIT ) is the following well-studied algorith- mic problem: Given an arithmetic circuit C computing a polynomial in F[x1; : : : ; xn], determine whether C computes an identically zero polynomial or not. The problem can be presented either in the white-box model or in the black-box model. In the white-box model, the arithmetic circuit is given ex- plicitly as the input. In the black-box model, the arithmetic circuit is given n n black-box access. I.e., the circuit can be evaluated at any point in F (or in F , for a suitable extension field F ). In the last three decades, PIT has played a pivotal role in many important results in complexity theory and algorithms: Pri- mality Testing [AKS04], the PCP Theorem [ALM+98], IP = PSPACE [Sha90], graph matching algorithms [Lov79, MVV87].
    [Show full text]
  • Exploring Differential Obliviousness∗
    Exploring Differential Obliviousness∗ Amos Beimely Kobbi Nissimz Mohammad Zaherix October 4, 2019 Abstract In a recent paper, Chan et al. [SODA ’19] proposed a relaxation of the notion of (full) memory obliv- iousness, which was introduced by Goldreich and Ostrovsky [J. ACM ’96] and extensively researched by cryptographers. The new notion, differential obliviousness, requires that any two neighboring inputs exhibit similar memory access patterns, where the similarity requirement is that of differential privacy. Chan et al. demonstrated that differential obliviousness allows achieving improved efficiency for several algorithmic tasks, including sorting, merging of sorted lists, and range query data structures. In this work, we continue the exploration of differential obliviousness, focusing on algorithms that do not necessarily examine all their input. This choice is motivated by the fact that the existence of loga- rithmic overhead ORAM protocols implies that differential obliviousness can yield at most a logarithmic improvement in efficiency for computations that need to examine all their input. In particular, we explore property testing, where we show that differential obliviousness yields an almost linear improvement in overhead in the dense graph model, and at most quadratic improvement in the bounded degree model. We also explore tasks where a non-oblivious algorithm would need to explore different portions of the input, where the latter would depend on the input itself, and where we show that such a behavior can be maintained under differential obliviousness, but not under full obliviousness. Our examples suggest that there would be benefits in further exploring which class of computational tasks are amenable to differential obliviousness. arXiv:1905.01373v2 [cs.CR] 2 Oct 2019 ∗Work supported by NSF grant No.
    [Show full text]
  • Lipics-ISAAC-2019-6.Pdf (0.5
    Sliding Window Property Testing for Regular Languages Moses Ganardi Universität Siegen, Germany [email protected] Danny Hucke Universität Siegen, Germany [email protected] Markus Lohrey Universität Siegen, Germany [email protected] Tatiana Starikovskaya DI/ENS, PSL Research University, Paris, France [email protected] Abstract We study the problem of recognizing regular languages in a variant of the streaming model of computation, called the sliding window model. In this model, we are given a size of the sliding window n and a stream of symbols. At each time instant, we must decide whether the suffix of length n of the current stream (“the active window”) belongs to a given regular language. Recent works [14, 15] showed that the space complexity of an optimal deterministic sliding window algorithm for this problem is either constant, logarithmic or linear in the window size n and provided natural language theoretic characterizations of the space complexity classes. Subsequently, [16] extended this result to randomized algorithms to show that any such algorithm admits either constant, double logarithmic, logarithmic or linear space complexity. In this work, we make an important step forward and combine the sliding window model with the property testing setting, which results in ultra-efficient algorithms for all regular languages. Informally, a sliding window property tester must accept the active window if it belongs to the language and reject it if it is far from the language. We show that for every regular language, there is a deterministic sliding window property tester that uses logarithmic space and a randomized sliding window property tester with two-sided error that uses constant space.
    [Show full text]
  • On Adaptive Distance Estimation
    On Adaptive Distance Estimation Yeshwanth Cherapanamjeri Jelani Nelson Electrical Engineering and Computer Science Electrical Engineering and Computer Science University of California at Berkeley University of California at Berkeley Berkeley, CA 94720 Berkeley, CA 94720 [email protected] [email protected] Abstract We provide a static data structure for distance estimation which supports adaptive n d queries. Concretely, given a dataset X = xi i=1 of n points in R and 0 <p 2, we construct a randomized data structure{ with} low memory consumption and query time which, when later given any query point q Rd, outputs a (1 + ")- 2 approximation of q xi p with high probability for all i [n]. The main novelty is our data structure’sk correctness− k guarantee holds even when2 the sequence of queries can be chosen adaptively: an adversary is allowed to choose the jth query point qj in a way that depends on the answers reported by the data structure for q1,...,qj 1. Previous randomized Monte Carlo methods do not provide error guarantees− in the setting of adaptively chosen queries [JL84, Ind06, TZ12, IW18]. Our memory consumption is O˜((n + d)d/"2), slightly more than the O(nd) required to store X in memory explicitly, but with the benefit that our time to answer queries is only 2 O˜("− (n + d)), much faster than the naive ⇥(nd) time obtained from a linear scan in the case of n and d very large. Here O˜ hides log(nd/") factors. We discuss applications to nearest neighbor search and nonparametric estimation. Our method is simple and likely to be applicable to other domains: we describe a generic approach for transforming randomized Monte Carlo data structures which do not support adaptive queries to ones that do, and show that for the problem at hand, it can be applied to standard nonadaptive solutions to `p norm estimation with negligible overhead in query time and a factor d overhead in memory.
    [Show full text]
  • Open Main.Pdf
    The Pennsylvania State University The Graduate School EFFICIENT COMBINATORIAL METHODS IN SPARSIFICATION, SUMMARIZATION AND TESTING OF LARGE DATASETS A Dissertation in Computer Science and Engineering by Grigory Yaroslavtsev c 2014 Grigory Yaroslavtsev Submitted in Partial Fulfillment of the Requirements for the Degree of Doctor of Philosophy May 2014 The dissertation of Grigory Yaroslavtsev was reviewed and approved∗ by the following: Sofya Raskhodnikova Associate Professor of Computer Science and Engineering Chair of Committee and Dissertation Advisor Piotr Berman Associate Professor of Computer Science and Engineering Jason Morton Assistant Professor of Mathematics and Statistics Adam D. Smith Associate Professor of Computer Science and Engineering Raj Acharya Head of Department of Computer Science and Engineering ∗Signatures are on file in the Graduate School. Abstract Increasingly large amounts of structured data are being collected by personal computers, mobile devices, personal gadgets, sensors, etc., and stored in data centers operated by the government and private companies. Processing of such data to extract key information is one of the main challenges faced by computer scientists. Developing methods for construct- ing compact representations of large data is a natural way to approach this challenge. This thesis is focused on rigorous mathematical and algorithmic solutions for sparsi- fication and summarization of large amounts of information using discrete combinatorial methods. Areas of mathematics most closely related to it are graph theory, information theory and analysis of real-valued functions over discrete domains. These areas, somewhat surprisingly, turn out to be related when viewed through the computational lens. In this thesis we illustrate the power and limitations of methods for constructing small represen- tations of large data sets, such as graphs and databases, using a variety methods drawn from these areas.
    [Show full text]
  • Conference in Honor of Laci Babai's 60Th Birthday
    Combinatorics, Groups, Algorithms, and Complexity: Conference in honor of Laci Babai’s 60th birthday Akos Seress, Mario Szegedy To cite this version: Akos Seress, Mario Szegedy. Combinatorics, Groups, Algorithms, and Complexity: Conference in honor of Laci Babai’s 60th birthday. Discrete Mathematics and Theoretical Computer Science, DMTCS, 2010, special issue in honor of Laci Babai’s 60th birthday: Combinatorics, Groups, Al- gorithms, and Complexity, Vol. 13 no. 4 (4), pp.1-4. hal-00993123 HAL Id: hal-00993123 https://hal.inria.fr/hal-00993123 Submitted on 19 May 2014 HAL is a multi-disciplinary open access L’archive ouverte pluridisciplinaire HAL, est archive for the deposit and dissemination of sci- destinée au dépôt et à la diffusion de documents entific research documents, whether they are pub- scientifiques de niveau recherche, publiés ou non, lished or not. The documents may come from émanant des établissements d’enseignement et de teaching and research institutions in France or recherche français ou étrangers, des laboratoires abroad, or from public or private research centers. publics ou privés. Discrete Mathematics and Theoretical Computer Science DMTCS vol. 13:4, 2011, 1–4 Combinatorics, Groups, Algorithms, and Complexity: Conference in honor of Laci Babai’s 60th birthday Akos´ Seress12† Mario Szegedy3‡ (Guest Editors) 1School of Mathematics and Statistics, University of Western Australia, Crawley, Australia 2The Ohio State Univer- sity, Dept. of Mathematics 3Rutgers University, Dept. of Computer Science Keywords: Combinatorics, Groups, Algorithms, and Complexity This special issue of Discrete Mathematics & Theoretical Computer Science is dedicated to the con- ference Combinatorics, Groups, Algorithms, and Complexity held at The Ohio State University between March 21-25, 2010.
    [Show full text]