On Unimodular Graphs

Total Page:16

File Type:pdf, Size:1020Kb

On Unimodular Graphs View metadata, citation and similar papers at core.ac.uk brought to you by CORE provided by Elsevier - Publisher Connector Linear Algebra and its Applications 421 (2007) 3–15 www.elsevier.com/locate/laa On unimodular graphs S. Akbari a, S.J. Kirkland b,∗ a Department of Mathematical Sciences, Sharif University of Technology, P.O. Box 11365-9415, Tehran, Iran b Department of Mathematics and Statistics, University of Regina, Regina, Saskatchewan, Canada S4S 0A2 Received 21 October 2005; accepted 18 October 2006 Submitted by R. Bhatia Abstract We study graphs whose adjacency matrices have determinant equal to 1 or −1, and characterize certain subclasses of these graphs. Graphs whose adjacency matrices are totally unimodular are also character- ized. For bipartite graphs having a unique perfect matching, we provide a formula for the inverse of the corresponding adjacency matrix, and address the problem of when that inverse is diagonally similar to a non-negative matrix. Special attention is paid to the case that such a graph is unicyclic. © 2006 Elsevier Inc. All rights reserved. AMS classification: 05C05; 05C50; 15A09; 15A15 Keywords: Determinant; Minor; Forest; Unicyclic graph; Unimodular graph 1. Introduction and preliminaries Let G be a graph with adjacency matrix A. Evidently if A is invertible, then its inverse is an integral matrix if and only if det(A) =±1. In a similar vein, if A is singular, there is a full rank submatrix, say B, so that the rows and columns of A can be permuted and partitioned as BC A = , CT D where A and B have equal rank (see [9, p. 179]). If det(B) =±1, then the reduced row echelon form for A is the integral matrix ∗ Corresponding author. Tel.: +1 306 585 4352; fax: +1 306 585 4020. E-mail addresses: [email protected] (S. Akbari), [email protected] (S.J. Kirkland). 0024-3795/$ - see front matter ( 2006 Elsevier Inc. All rights reserved. doi:10.1016/j.laa.2006.10.017 4 S. Akbari, S.J. Kirkland / Linear Algebra and its Applications 421 (2007) 3–15 IB−1C 00 (see [1] for more on the reduced row echelon form for certain classes of graphs). These observations prompt our interest in graphs whose adjacency matrices have determinant 1 or −1, and those graphs are the subject of this paper. Throughout this paper, we consider simple graphs with no loops or multiple edges, and we will use standard terminology and ideas from both graph theory and matrix theory. We refer the reader to [3] for background on the former and to [11] for background on the latter. For convenience, we recall a few key notions that will be used in the paper. We denote a path and a cycle on k vertices, by Pk and Ck, respectively. A forest is a graph with no cycles, and a connected forest is known as a tree. A graph is called unicyclic if it is connected and contains exactly one cycle. A matching of a graph G is a subgraph of G which is 1-regular. The number of edges in a matching is called its size.Amaximum matching is a matching with largest size. A perfect matching of G is a spanning matching of G. Clearly, if G has a perfect matching, then the number of vertices of G is even. A vertex is a pendant vertex if it has degree 1; a pendant edge is an edge that is incident with a pendant vertex. Let G be a graph with vertex set {v1,...,vn}. The adjacency matrix of G is the n × n matrix A such that Aij is1ifvi and vj are adjacent and is 0 otherwise. We will occasionally use det(G) to denote the determinant of the adjacency matrix for G. Remark 1. Let G be a bipartite graph with two parts of sizes m and n. Let B A = 0 BT 0 be the adjacency matrix of G.Ifm/= n, then B does not have full rank and so det(A) = 0. If m = n, then we have det(A) = (−1)n(det(B))2. Clearly, every perfect matching of G corresponds to a transversal in B with all cells 1, and conversely. Consequently, if G has no perfect matching, then det(G) = 0. If det(G) = 1, then the number of perfect matchings is odd. Also if G has exactly one perfect matching, then det(G) = (−1)n. Recall that a unimodular matrix is a square matrix with determinant 1 or −1, while a rectangular matrix is called totally unimodular if each of its minors is −1, 0 or 1. Similarly, a square matrix is called principally unimodular if each of its principal minors is −1, 0, or 1. Consequently, we will say that a graph is unimodular, totally unimodular or principally unimodular, according as its adjacency matrix is unimodular, totally unimodular or principally unimodular, respectively. Similarly, we will say that a graph is unital if its adjacency matrix has determinant 0, 1or−1. Evidently, the class of unital graphs properly includes the class of unimodular graphs. A signed bipartite graph is a bipartite graph in which each edge is given a weight of −1or +1. The weight of a cycle in a signed bipartite graph is the sum of the weights of its edges. A signed bipartite graph is called restricted unimodular if the weight of any cycle in G is divisible by 4. There is a natural correspondence between a signed bipartite graph G and its {−1, 0, 1} adjacency matrix A, where the rows and columns of A are indexed by the vertices of G, with Aij =−1, 0, 1, according as the edge between vertices vi and vj has weight −1, is absent, or has weight +1, respectively. The following result is due to Commoner [5]. Theorem A. A signed bipartite graph that is restricted unimodular has a totally unimodular adjacency matrix. S. Akbari, S.J. Kirkland / Linear Algebra and its Applications 421 (2007) 3–15 5 Fig. 1. G is totally unimodular but not restricted unimodular. Consider the graph G shown in Fig. 1. Evidently, G is bipartite, but if each edge weight is 1, then it is not restricted unimodular. A straightforward check shows that G is totally unimodular, so that the converse of Theorem A fails. An elementary graph is a graph G, each component of which is either a single edge or a cycle; let s(G) denote the number of components that are cycles. The following theorem of Harary (see [2, p. 44, 50]) uses elementary graphs to compute the determinant. Theorem B. For any graph G, we have |E()|−s() s() det(G) = (−1) 2 , where the summation is over all spanning elementary subgraphs of G. Equivalently, we have r s det(G) = f(r,s)(−1) 2 , where f(r,s)is the number of spanning elementary subgraphs with s cycles and r + s edges. Corollary 1. Let F be a forest on n vertices. If F does not have a perfect matching, then det(F ) = 0. If F has a perfect matching, then det(F ) = (−1)n/2. n− Corollary 2. If n is odd, then det(Cn) = (−1) 12. If n ≡ 0 mod 4, then det(Cn) = 0, while if n ≡ 2 mod 4, then det(Cn) =−4. Remark 2. If a graph G has no perfect matching, then its determinant is an even number. Thus unimodular graphs have an even number of vertices. 2. Unital and unimodular graphs In this section, we explore constructions of and characterizations for unital and unimodular graphs. Theorem 1. Let A be an n by n integral symmetric matrix with even entries on the main diagonal. Then (i) if n ≡ 0, mod 4, then det(A) ≡ 0 or 1, mod 4; (ii) if n ≡ 2 mod 4; then det(A) ≡ 0 or −1 mod 4. Proof. We proceed by induction on n, and note that if n = 2, then the assertion holds. Now suppose that n>2 is even. If all entries of A are even, then det(A) is 0 mod 4. Thus we may assume that x1 y there exists a principal submatrix of A, say P = yx, where x1 and x2 are even and y is odd. 2 −x y PBT We have P −1 = 2 mod 4. Writing, without loss of generality, A = , we have, y −x1 BC from the Schur complement formula for the determinant, det(A) = det(P ) det(C − BP−1BT) = −det(C − BP−1BT) mod 4. It is easily checked that C − BP−1BT is a symmetric integral matrix 6 S. Akbari, S.J. Kirkland / Linear Algebra and its Applications 421 (2007) 3–15 each of whose entries on the diagonal are even; applying the induction hypothesis now yields the desired conclusion. The following is immediate from Remark 2 and Theorem 1. Corollary 3. Let G be a graph of order n. If det(G) = 1, then n is divisible by 4, while if det(G) =−1, then n is congruent to 2 mod 4. The proof of the next lemma is simple and we omit it. Lemma 1. Let G be a graph with pendant vertex v which is adjacent to w. Then det(G) = −det(G \{v,w}). In particular,Gis unimodular (respectively, unital) if and only if G \{v,w} is unimodular (respectively, unital). Theorem 2. Let G be a unicyclic graph. Then det(G) ∈{0, ±1, ±2, ±4}. Proof. By Lemma 1 there is a subgraph of G, say H , whose connected components are either the cycle or are trees and such that |det(G)|=|det(H )|.
Recommended publications
  • Quiz7 Problem 1
    Quiz 7 Quiz7 Problem 1. Independence The Problem. In the parts below, cite which tests apply to decide on independence or dependence. Choose one test and show complete details. 1 ! 1 ! (a) Vectors ~v = ;~v = 1 −1 2 1 0 1 1 0 1 1 0 1 1 B C B C B C (b) Vectors ~v1 = @ −1 A ;~v2 = @ 1 A ;~v3 = @ 0 A 0 −1 −1 2 5 (c) Vectors ~v1;~v2;~v3 are data packages constructed from the equations y = x ; y = x ; y = x10 on (−∞; 1). (d) Vectors ~v1;~v2;~v3 are data packages constructed from the equations y = 1 + x; y = 1 − x; y = x3 on (−∞; 1). Basic Test. To show three vectors are independent, form the system of equations c1~v1 + c2~v2 + c3~v3 = ~0; then solve for c1; c2; c3. If the only solution possible is c1 = c2 = c3 = 0, then the vectors are independent. Linear Combination Test. A list of vectors is independent if and only if each vector in the list is not a linear combination of the remaining vectors, and each vector in the list is not zero. Subset Test. Any nonvoid subset of an independent set is independent. The basic test leads to three quick independence tests for column vectors. All tests use the augmented matrix A of the vectors, which for 3 vectors is A =< ~v1j~v2j~v3 >. Rank Test. The vectors are independent if and only if rank(A) equals the number of vectors. Determinant Test. Assume A is square. The vectors are independent if and only if the deter- minant of A is nonzero.
    [Show full text]
  • Theory of Angular Momentum Rotations About Different Axes Fail to Commute
    153 CHAPTER 3 3.1. Rotations and Angular Momentum Commutation Relations followed by a 90° rotation about the z-axis. The net results are different, as we can see from Figure 3.1. Our first basic task is to work out quantitatively the manner in which Theory of Angular Momentum rotations about different axes fail to commute. To this end, we first recall how to represent rotations in three dimensions by 3 X 3 real, orthogonal matrices. Consider a vector V with components Vx ' VY ' and When we rotate, the three components become some other set of numbers, V;, V;, and Vz'. The old and new components are related via a 3 X 3 orthogonal matrix R: V'X Vx V'y R Vy I, (3.1.1a) V'z RRT = RTR =1, (3.1.1b) where the superscript T stands for a transpose of a matrix. It is a property of orthogonal matrices that 2 2 /2 /2 /2 vVx + V2y + Vz =IVVx + Vy + Vz (3.1.2) is automatically satisfied. This chapter is concerned with a systematic treatment of angular momen- tum and related topics. The importance of angular momentum in modern z physics can hardly be overemphasized. A thorough understanding of angu- Z z lar momentum is essential in molecular, atomic, and nuclear spectroscopy; I I angular-momentum considerations play an important role in scattering and I I collision problems as well as in bound-state problems. Furthermore, angu- I lar-momentum concepts have important generalizations-isospin in nuclear physics, SU(3), SU(2)® U(l) in particle physics, and so forth.
    [Show full text]
  • Adjacency and Incidence Matrices
    Adjacency and Incidence Matrices 1 / 10 The Incidence Matrix of a Graph Definition Let G = (V ; E) be a graph where V = f1; 2;:::; ng and E = fe1; e2;:::; emg. The incidence matrix of G is an n × m matrix B = (bik ), where each row corresponds to a vertex and each column corresponds to an edge such that if ek is an edge between i and j, then all elements of column k are 0 except bik = bjk = 1. 1 2 e 21 1 13 f 61 0 07 3 B = 6 7 g 40 1 05 4 0 0 1 2 / 10 The First Theorem of Graph Theory Theorem If G is a multigraph with no loops and m edges, the sum of the degrees of all the vertices of G is 2m. Corollary The number of odd vertices in a loopless multigraph is even. 3 / 10 Linear Algebra and Incidence Matrices of Graphs Recall that the rank of a matrix is the dimension of its row space. Proposition Let G be a connected graph with n vertices and let B be the incidence matrix of G. Then the rank of B is n − 1 if G is bipartite and n otherwise. Example 1 2 e 21 1 13 f 61 0 07 3 B = 6 7 g 40 1 05 4 0 0 1 4 / 10 Linear Algebra and Incidence Matrices of Graphs Recall that the rank of a matrix is the dimension of its row space. Proposition Let G be a connected graph with n vertices and let B be the incidence matrix of G.
    [Show full text]
  • A Brief Introduction to Spectral Graph Theory
    A BRIEF INTRODUCTION TO SPECTRAL GRAPH THEORY CATHERINE BABECKI, KEVIN LIU, AND OMID SADEGHI MATH 563, SPRING 2020 Abstract. There are several matrices that can be associated to a graph. Spectral graph theory is the study of the spectrum, or set of eigenvalues, of these matrices and its relation to properties of the graph. We introduce the primary matrices associated with graphs, and discuss some interesting questions that spectral graph theory can answer. We also discuss a few applications. 1. Introduction and Definitions This work is primarily based on [1]. We expect the reader is familiar with some basic graph theory and linear algebra. We begin with some preliminary definitions. Definition 1. Let Γ be a graph without multiple edges. The adjacency matrix of Γ is the matrix A indexed by V (Γ), where Axy = 1 when there is an edge from x to y, and Axy = 0 otherwise. This can be generalized to multigraphs, where Axy becomes the number of edges from x to y. Definition 2. Let Γ be an undirected graph without loops. The incidence matrix of Γ is the matrix M, with rows indexed by vertices and columns indexed by edges, where Mxe = 1 whenever vertex x is an endpoint of edge e. For a directed graph without loss, the directed incidence matrix N is defined by Nxe = −1; 1; 0 corresponding to when x is the head of e, tail of e, or not on e. Definition 3. Let Γ be an undirected graph without loops. The Laplace matrix of Γ is the matrix L indexed by V (G) with zero row sums, where Lxy = −Axy for x 6= y.
    [Show full text]
  • Introduction
    Introduction This dissertation is a reading of chapters 16 (Introduction to Integer Liner Programming) and 19 (Totally Unimodular Matrices: fundamental properties and examples) in the book : Theory of Linear and Integer Programming, Alexander Schrijver, John Wiley & Sons © 1986. The chapter one is a collection of basic definitions (polyhedron, polyhedral cone, polytope etc.) and the statement of the decomposition theorem of polyhedra. Chapter two is “Introduction to Integer Linear Programming”. A finite set of vectors a1 at is a Hilbert basis if each integral vector b in cone { a1 at} is a non- ,… , ,… , negative integral combination of a1 at. We shall prove Hilbert basis theorem: Each ,… , rational polyhedral cone C is generated by an integral Hilbert basis. Further, an analogue of Caratheodory’s theorem is proved: If a system n a1x β1 amx βm has no integral solution then there are 2 or less constraints ƙ ,…, ƙ among above inequalities which already have no integral solution. Chapter three contains some basic result on totally unimodular matrices. The main theorem is due to Hoffman and Kruskal: Let A be an integral matrix then A is totally unimodular if and only if for each integral vector b the polyhedron x x 0 Ax b is integral. ʜ | ƚ ; ƙ ʝ Next, seven equivalent characterization of total unimodularity are proved. These characterizations are due to Hoffman & Kruskal, Ghouila-Houri, Camion and R.E.Gomory. Basic examples of totally unimodular matrices are incidence matrices of bipartite graphs & Directed graphs and Network matrices. We prove Konig-Egervary theorem for bipartite graph. 1 Chapter 1 Preliminaries Definition 1.1: (Polyhedron) A polyhedron P is the set of points that satisfies Μ finite number of linear inequalities i.e., P = {xɛ ő | ≤ b} where (A, b) is an Μ m (n + 1) matrix.
    [Show full text]
  • Incidence Matrices and Interval Graphs
    Pacific Journal of Mathematics INCIDENCE MATRICES AND INTERVAL GRAPHS DELBERT RAY FULKERSON AND OLIVER GROSS Vol. 15, No. 3 November 1965 PACIFIC JOURNAL OF MATHEMATICS Vol. 15, No. 3, 1965 INCIDENCE MATRICES AND INTERVAL GRAPHS D. R. FULKERSON AND 0. A. GROSS According to present genetic theory, the fine structure of genes consists of linearly ordered elements. A mutant gene is obtained by alteration of some connected portion of this structure. By examining data obtained from suitable experi- ments, it can be determined whether or not the blemished portions of two mutant genes intersect or not, and thus inter- section data for a large number of mutants can be represented as an undirected graph. If this graph is an "interval graph," then the observed data is consistent with a linear model of the gene. The problem of determining when a graph is an interval graph is a special case of the following problem concerning (0, l)-matrices: When can the rows of such a matrix be per- muted so as to make the l's in each column appear consecu- tively? A complete theory is obtained for this latter problem, culminating in a decomposition theorem which leads to a rapid algorithm for deciding the question, and for constructing the desired permutation when one exists. Let A — (dij) be an m by n matrix whose entries ai3 are all either 0 or 1. The matrix A may be regarded as the incidence matrix of elements el9 e2, , em vs. sets Sl9 S2, , Sn; that is, ai3 = 0 or 1 ac- cording as et is not or is a member of S3 .
    [Show full text]
  • Contents 5 Eigenvalues and Diagonalization
    Linear Algebra (part 5): Eigenvalues and Diagonalization (by Evan Dummit, 2017, v. 1.50) Contents 5 Eigenvalues and Diagonalization 1 5.1 Eigenvalues, Eigenvectors, and The Characteristic Polynomial . 1 5.1.1 Eigenvalues and Eigenvectors . 2 5.1.2 Eigenvalues and Eigenvectors of Matrices . 3 5.1.3 Eigenspaces . 6 5.2 Diagonalization . 9 5.3 Applications of Diagonalization . 14 5.3.1 Transition Matrices and Incidence Matrices . 14 5.3.2 Systems of Linear Dierential Equations . 16 5.3.3 Non-Diagonalizable Matrices and the Jordan Canonical Form . 19 5 Eigenvalues and Diagonalization In this chapter, we will discuss eigenvalues and eigenvectors: these are characteristic values (and characteristic vectors) associated to a linear operator T : V ! V that will allow us to study T in a particularly convenient way. Our ultimate goal is to describe methods for nding a basis for V such that the associated matrix for T has an especially simple form. We will rst describe diagonalization, the procedure for (trying to) nd a basis such that the associated matrix for T is a diagonal matrix, and characterize the linear operators that are diagonalizable. Then we will discuss a few applications of diagonalization, including the Cayley-Hamilton theorem that any matrix satises its characteristic polynomial, and close with a brief discussion of non-diagonalizable matrices. 5.1 Eigenvalues, Eigenvectors, and The Characteristic Polynomial • Suppose that we have a linear transformation T : V ! V from a (nite-dimensional) vector space V to itself. We would like to determine whether there exists a basis of such that the associated matrix β is a β V [T ]β diagonal matrix.
    [Show full text]
  • Eigenvalues of Graphs
    Eigenvalues of graphs L¶aszl¶oLov¶asz November 2007 Contents 1 Background from linear algebra 1 1.1 Basic facts about eigenvalues ............................. 1 1.2 Semide¯nite matrices .................................. 2 1.3 Cross product ...................................... 4 2 Eigenvalues of graphs 5 2.1 Matrices associated with graphs ............................ 5 2.2 The largest eigenvalue ................................. 6 2.2.1 Adjacency matrix ............................... 6 2.2.2 Laplacian .................................... 7 2.2.3 Transition matrix ................................ 7 2.3 The smallest eigenvalue ................................ 7 2.4 The eigenvalue gap ................................... 9 2.4.1 Expanders .................................... 10 2.4.2 Edge expansion (conductance) ........................ 10 2.4.3 Random walks ................................. 14 2.5 The number of di®erent eigenvalues .......................... 16 2.6 Spectra of graphs and optimization .......................... 18 Bibliography 19 1 Background from linear algebra 1.1 Basic facts about eigenvalues Let A be an n £ n real matrix. An eigenvector of A is a vector such that Ax is parallel to x; in other words, Ax = ¸x for some real or complex number ¸. This number ¸ is called the eigenvalue of A belonging to eigenvector v. Clearly ¸ is an eigenvalue i® the matrix A ¡ ¸I is singular, equivalently, i® det(A ¡ ¸I) = 0. This is an algebraic equation of degree n for ¸, and hence has n roots (with multiplicity). The trace of the square matrix A = (Aij ) is de¯ned as Xn tr(A) = Aii: i=1 1 The trace of A is the sum of the eigenvalues of A, each taken with the same multiplicity as it occurs among the roots of the equation det(A ¡ ¸I) = 0. If the matrix A is symmetric, then its eigenvalues and eigenvectors are particularly well behaved.
    [Show full text]
  • Notes on the Proof of the Van Der Waerden Permanent Conjecture
    Iowa State University Capstones, Theses and Creative Components Dissertations Spring 2018 Notes on the proof of the van der Waerden permanent conjecture Vicente Valle Martinez Iowa State University Follow this and additional works at: https://lib.dr.iastate.edu/creativecomponents Part of the Discrete Mathematics and Combinatorics Commons Recommended Citation Valle Martinez, Vicente, "Notes on the proof of the van der Waerden permanent conjecture" (2018). Creative Components. 12. https://lib.dr.iastate.edu/creativecomponents/12 This Creative Component is brought to you for free and open access by the Iowa State University Capstones, Theses and Dissertations at Iowa State University Digital Repository. It has been accepted for inclusion in Creative Components by an authorized administrator of Iowa State University Digital Repository. For more information, please contact [email protected]. Notes on the proof of the van der Waerden permanent conjecture by Vicente Valle Martinez A Creative Component submitted to the graduate faculty in partial fulfillment of the requirements for the degree of MASTER OF SCIENCE Major: Mathematics Program of Study Committee: Sung Yell Song, Major Professor Steve Butler Jonas Hartwig Leslie Hogben Iowa State University Ames, Iowa 2018 Copyright c Vicente Valle Martinez, 2018. All rights reserved. ii TABLE OF CONTENTS LIST OF FIGURES . iii ACKNOWLEDGEMENTS . iv ABSTRACT . .v CHAPTER 1. Introduction and Preliminaries . .1 1.1 Combinatorial interpretations . .1 1.2 Computational properties of permanents . .4 1.3 Computational complexity in computing permanents . .8 1.4 The organization of the rest of this component . .9 CHAPTER 2. Applications: Permanents of Special Types of Matrices . 10 2.1 Zeros-and-ones matrices .
    [Show full text]
  • The Assignment Problem and Totally Unimodular Matrices
    Methods and Models for Combinatorial Optimization The assignment problem and totally unimodular matrices L.DeGiovanni M.DiSumma G.Zambelli 1 The assignment problem Let G = (V,E) be a bipartite graph, where V is the vertex set and E is the edge set. Recall that “bipartite” means that V can be partitioned into two disjoint subsets V1, V2 such that, for every edge uv ∈ E, one of u and v is in V1 and the other is in V2. In the assignment problem we have |V1| = |V2| and there is a cost cuv for every edge uv ∈ E. We want to select a subset of edges such that every node is the end of exactly one selected edge, and the sum of the costs of the selected edges is minimized. This problem is called “assignment problem” because selecting edges with the above property can be interpreted as assigning each node in V1 to exactly one adjacent node in V2 and vice versa. To model this problem, we define a binary variable xuv for every u ∈ V1 and v ∈ V2 such that uv ∈ E, where 1 if u is assigned to v (i.e., edge uv is selected), x = uv 0 otherwise. The total cost of the assignment is given by cuvxuv. uvX∈E The condition that for every node u in V1 exactly one adjacent node v ∈ V2 is assigned to v1 can be modeled with the constraint xuv =1, u ∈ V1, v∈VX2:uv∈E while the condition that for every node v in V2 exactly one adjacent node u ∈ V1 is assigned to v2 can be modeled with the constraint xuv =1, v ∈ V2.
    [Show full text]
  • Perfect, Ideal and Balanced Matrices: a Survey
    Perfect Ideal and Balanced Matrices Michele Conforti y Gerard Cornuejols Ajai Kap o or z and Kristina Vuskovic December Abstract In this pap er we survey results and op en problems on p erfect ideal and balanced matrices These matrices arise in connection with the set packing and set covering problems two integer programming mo dels with a wide range of applications We concentrate on some of the b eautiful p olyhedral results that have b een obtained in this area in the last thirty years This survey rst app eared in Ricerca Operativa Dipartimento di Matematica Pura ed Applicata Universitadi Padova Via Belzoni Padova Italy y Graduate School of Industrial administration Carnegie Mellon University Schenley Park Pittsburgh PA USA z Department of Mathematics University of Kentucky Lexington Ky USA This work was supp orted in part by NSF grant DMI and ONR grant N Introduction The integer programming mo dels known as set packing and set covering have a wide range of applications As examples the set packing mo del o ccurs in the phasing of trac lights Stoer in pattern recognition Lee Shan and Yang and the set covering mo del in scheduling crews for railways airlines buses Caprara Fischetti and Toth lo cation theory and vehicle routing Sometimes due to the sp ecial structure of the constraint matrix the natural linear programming relaxation yields an optimal solution that is integer thus solving the problem We investigate conditions under which this integrality prop erty holds Let A b e a matrix This matrix is perfect if the fractional
    [Show full text]
  • Order Number 9807786 Effectiveness in Representations of Positive Definite Quadratic Forms Icaza Perez, Marfa In6s, Ph.D
    Order Number 9807786 Effectiveness in representations of positive definite quadratic form s Icaza Perez, Marfa In6s, Ph.D. The Ohio State University, 1992 UMI 300 N. ZeebRd. Ann Arbor, MI 48106 E ffectiveness in R epresentations o f P o s it iv e D e f i n it e Q u a d r a t ic F o r m s dissertation Presented in Partial Fulfillment of the Requirements for the Degree Doctor of Philosophy in the Graduate School of The Ohio State University By Maria Ines Icaza Perez, The Ohio State University 1992 Dissertation Committee: Approved by John S. Hsia Paul Ponomarev fJ Adviser Daniel B. Shapiro lepartment of Mathematics A cknowledgements I begin these acknowledgements by first expressing my gratitude to Professor John S. Hsia. Beyond accepting me as his student, he provided guidance and encourage­ ment which has proven invaluable throughout the course of this work. I thank Professors Paul Ponomarev and Daniel B. Shapiro for reading my thesis and for their helpful suggestions. I also express my appreciation to Professor Shapiro for giving me the opportunity to come to The Ohio State University and for helping me during the first years of my studies. My recognition goes to my undergraduate teacher Professor Ricardo Baeza for teaching me mathematics and the discipline of hard work. I thank my classmates from the Quadratic Forms Seminar especially, Y. Y. Shao for all the cheerful mathematical conversations we had during these months. These years were more enjoyable thanks to my friends from here and there. I thank them all.
    [Show full text]