Applications of Matroid Theory and Combinatorial Optimization to Information and Coding Theory
Total Page:16
File Type:pdf, Size:1020Kb
Applications of Matroid Theory and Combinatorial Optimization to Information and Coding Theory Navin Kashyap (Queen’s University), Emina Soljanin (Alcatel-Lucent Bell Labs) Pascal Vontobel (Hewlett-Packard Laboratories) August 2–7, 2009 The aim of this workshop was to bring together experts and students from pure and applied mathematics, computer science, and engineering, who are working on related problems in the areas of matroid theory, combinatorial optimization, coding theory, secret sharing, network coding, and information inequalities. The goal was to foster exchange of mathematical ideas and tools that can help tackle some of the open problems of central importance in coding theory, secret sharing, and network coding, and at the same time, to get pure mathematicians and computer scientists to be interested in the kind of problems that arise in these applied fields. 1 Introduction Matroids are structures that abstract certain fundamental properties of dependence common to graphs and vector spaces. The theory of matroids has its origins in graph theory and linear algebra, and its most successful applications in the past have been in the areas of combinatorial optimization and network theory. Recently, however, there has been a flurry of new applications of this theory in the fields of information and coding theory. It is only natural to expect matroid theory to have an influence on the theory of error-correcting codes, as matrices over finite fields are objects of fundamental importance in both these areas of mathematics. Indeed, as far back as 1976, Greene [7] (re-)derived the MacWilliams identities — which relate the Hamming weight enumerators of a linear code and its dual — as special cases of an identity for the Tutte polynomial of a matroid. However, aside from such use of tools from matroid theory to re-derive results in coding theory that had already been proved by other means, each field has had surprisingly little impact on the other, until very recently. Matroid-theoretic methods are now starting to play an important role in the understanding of decoding algorithms for error-correcting codes. In a parallel and largely unrelated development, ideas from matroid theory are also finding other novel applications within the broader realm of information theory. Specifically, they are being applied to explore the fundamental limits of secret sharing schemes and network coding, and also to gain an understanding of information inequalities. We outline some of these recent developments next. 2 Background and Recent Developments Our workshop covered four major areas within the realm of information theory — coding theory, secret sharing, network coding, and information inequalities — which have seen a recent influx of ideas from 1 2 matroid theory and combinatorial optimization. We briefly discuss the applications of such ideas in each of these areas in turn. 2.1 Coding Theory The serious study of (channel) coding theory started with Shannon’s monumental 1948 paper [9]. Shannon stated the result that reliable communication is possible at rates up to channel capacity, meaning that for any desired symbol or block error probability there exists a channel code and a decoding algorithm that can achieve this symbol or block error probability as long as the rate of the channel code is smaller than the channel capacity. On the other hand, Shannon showed that if the rate is larger than capacity, the symbol and the block error probability must be bounded away from zero. Unfortunately, the proof of the above achievability result is nonconstructive, meaning that it shows only the existence of such channel codes and decoding algorithms. Therefore, since the appearance of Shannon’s theorem, the quest has been on to find codes with practical encoding and decoding algorithms that fulfill Shannon’s promise. The codes and decoding schemes that people have come up with can broadly be classified into two classes: “traditional schemes” and “modern schemes.” In “traditional schemes,” codes were proposed that have some desirable properties like large minimum Hamming distance (a typical example of such codes being the Reed- Solomon codes). However, given a code, it was usually unclear how to decode it efficiently. Often it took quite some time until such a decoding algorithm was found (e.g., the Berlekamp-Massey decoding algorithm for Reed-Solomon codes), if at all. In “modern schemes,” the situation is reversed: given an iterative decoding algorithm like the sum-product algorithm, the question is what codes work well together with such an iterative decoding algorithm. “Modern schemes” took off with the seminal paper by Berrou, Glavieu, and Thitimajshima in 1993 on turbo coding schemes [2]. Actually, codes and decoding algorithm in the spirit of “modern schemes” were already described in the early 1960s by Gallager in his Ph.D. thesis [5]. However, these schemes were, besides the work by Zyablov, Pinkser, and Tanner in the 1970s and 1980s, largely forgotten until the mid-1990s. Only then people started to appreciate Gallager’s revolutionary approach to coding theory. Gallager proposed to define codes in terms of graphs. Such graphs are now known as Tanner graphs: they are bipartite graphs where one class of vertices corresponds to codeword symbols and where the other class of vertices corresponds to parity-checks that are imposed on the adjacent codeword symbols. Decoding is then based on repeatedly sending messages with estimates about the value of the codeword symbols along edges, and to locally process these messages at vertices in order to produce new messages that are again sent along the edges. Especially for sparse Tanner graphs the resulting decoding algorithms have very low implementation complexity. In the last fifteen years, Tanner graphs and iterative decoding algorithms have been generalized to fac- tor graphs and algorithms operating on them, and many connections to techniques in statistical mechanics, graphical models, artificial intelligence, and combinatorial optimization were uncovered. The workshop talk by Kashyap (see Section 3.1) surveyed the connection between complexity measures for graphical models for a code and the treewidth (and other width parameters) of the associated matroid. On the other hand, the workshop talks by Wainwright and Vontobel (see Section 3.1) emphasized the connections between message- passing iterative decoding of codes and certain techniques from combinatorial optimization. In particular, they discussed the linear programming decoder by Feldman, Wainwright, and Karger [4], which is a low- complexity relaxation of an integer linear programming formulation of the maximum likelihood decoder. This linear programming decoder (and its variations) has paved the way for the use of tools from combinato- rial optimization and matroids in the design and analysis of decoding algorithms. 2.2 Secret Sharing The second major application of matroid-theoretic ideas that we mention here is with respect to secret-sharing schemes. A secret-sharing scheme is a method to distribute shares of a secret value among a certain number of participants such that qualified subsets of participants (e.g., subsets of a certain size) can recover the secret from their joint shares, but unqualified subsets of participants can obtain no information whatsoever about the secret by pooling together their shares. Secret-sharing schemes were originally motivated by the problem 3 of secure storage of cryptographic keys, but have since found numerous other applications in cryptography and distributed computing. It is not difficult to show that in a secret-sharing scheme, the size of each of the shares cannot be smaller than the size (information content) of the secret value. An ideal secret-sharing scheme is one in which all shares have the same size as the secret value. More generally, the information rate of a secret-sharing scheme is the ratio of the size of the secret to the maximum share size. In a secret-sharing scheme, the collection of qualified subsets of participants is called the access structure of the scheme. It is known that for any monotone increasing collection, Γ, of subsets of a finite set, one can define a secret-sharing scheme with access structure Γ. The information rate ρ(Γ) is defined to be the supremum of information rates among all secret-sharing schemes having access structure Γ. Γ is said to be an ideal access structure if it admits an ideal secret-sharing scheme. Brickell and Davenport [3] began a line of work relating ideal secret-sharing schemes to matroids. They showed that any ideal access structure is induced by a matroid in a very specific sense. However, it is also known that not every matroid gives rise to an ideal access structure; for example, the access structures induced by the Vamos´ matroid are not ideal. Characterizing the matroids that give rise to ideal access structures has remained an open problem. There has been some very recent work on computing the information rates of non-ideal access structures using polymatroid techniques, linear programming, and non-Shannon information inequalities. For example, it has been shown that for any access structure Γ induced by the Vamos´ matroid, ρ(Γ) ≤ 19/21, which shows that such access structures are far from being ideal. This, and other related results, were surveyed in the workshop talks by Padro´ and Beimel (see Section 3.3). Secret-sharing schemes have also been received some recent attention in the quantum domain, a topic covered in the workshop talk by Sarvepalli (see Section 3.3). 2.3 Network Coding Another novel application of matroid theory and combinatorial optimization within the realm of information theory is in the area of network coding [10]. Network coding is an elegant technique introduced at the turn of the millennium to improve network throughput and performance. Since then, it has attracted significant interest from diverse scientific communities of engineers, computer scientists, and mathematicians in both academia and industry. This workshop explored connections between network coding and combinatorial optimization, matroids, and non-Shannon inequalities.