Additional Details and Fortification for Chapter 1

Total Page:16

File Type:pdf, Size:1020Kb

Additional Details and Fortification for Chapter 1 APPENDIX A Additional Details and Fortification for Chapter 1 A.1 Matrix Classes and Special Matrices The matrices can be grouped into several classes based on their operational prop- erties. A short list of various classes of matrices is given in Tables A.1 and A.2. Some of these have already been described earlier, for example, elementary, sym- metric, hermitian, othogonal, unitary, positive definite/semidefinite, negative defi- nite/semidefinite, real, imaginary, and reducible/irreducible. Some of the matrix classes are defined based on the existence of associated matri- ces. For instance, A is a diagonalizable matrix if there exists nonsingular matrices T such that TAT −1 = D results in a diagonal matrix D. Connected with diagonalizable matrices are normal matrices. A matrix B is a normal matrix if BB∗ = B∗B.Normal matrices are guaranteed to be diagonalizable matrices. However, defective matri- ces are not diagonalizable. Once a matrix has been identified to be diagonalizable, then the following fact can be used for easier computation of integral powers of the matrix: A = T −1DT → Ak = T −1DT T −1DT ··· T −1DT = T −1DkT and then take advantage of the fact that ⎛ ⎞ dk 0 ⎜ 1 ⎟ K . D = ⎝ .. ⎠ K 0 dN Another set of related classes of matrices are the idempotent, projection, invo- lutory, nilpotent, and convergent matrices. These classes are based on the results of integral powers. Matrix A is idempotent if A2 = A, and if, in addition, A is hermitian, then A is known as a projection matrix. Projection matrices are used to partition an N-dimensional space into two subspaces that are orthogonal to each other. A matrix B is involutory if it is its own inverse, that is, if B2 = I. For example, a reflection matrix such as the Householder matrix is given by 2 H = I − vv∗ v∗v where v is a nonzero vector, and then H = H−1. A convergent matrix (also known as k stable matrix) C is a matrix for which limk→∞ C = 0. These matrices are important 561 562 Appendix A: Additional Details and Fortification for Chapter 1 Table A.1. Matrix classes (based on operational properties) Class Definition Remarks Convergent (Stable) lim Ak = 0 k→∞ r α Ak = 0 Defective (Deficient) k k=0 αr = 0; r < N T −1AT is diagonal Diagonalizable for some nonsingular T Any matrix that scales, interchanges, or adds Elementary multiples of rows or • Used in Gaussian elimination columns of another matrix B Gram A = B∗B for some B • Are Hermitian • (B + B∗) /2 is the hermitian Hermitian A∗ = A part of B. • B∗B and BB∗ are hermitian Idempotent A2 = A • det(A) = 1ordet(A) = 0 • Examples: identity matrix Involutory A2 = I,i.e.A = A−1 reverse unit matrices, symmetric orthogonal matrices Negative definite x∗Ax < 0 x = 0 Negative semidefinite x∗Ax ≤ 0 x = 0 k Nilpotent (of degree k) A = 0; k > 0 • det(A) = 0 Normal AA∗ = A∗A • Are diagonalizable Nonsingular (Invertible) |A| = 0 for procedures that implement iterative computations. If, in addition, k < ∞ for Ck = 0, then the stable matrix will belong to the subclass of nilpotent matrices. Aside from the classifications given in Tables A.1 and A.2, we also list some spe- cial matrices based on the structure and composition of the matrices. These are given in Table A.3. Some of the items in this table serve as a glossary of terms for the special matrices already described in this chapter. Some of the matrices refer to matrix struc- tures based on the positions of zero and nonzero elements such as banded, sparse, triangular, tridiagonal, diagonal, bidiagonal, anti-diagonal, and Hessenberg. Some involve additional specifications on the elements themselves. These include iden- tity, reverse identity, shift, real, complex, polynomial, rational, positive/negative, or nonpositive/nonnegative matrices. For instance, positive (or nonnegative) matrices Appendix A: Additional Details and Fortification for Chapter 1 563 Table A.2. Matrix classes (based on operations) Class Definition Remarks Orthogonal AT = A−1 Positive definite x∗Ax > 0;x = 0 Positive semidefinite x∗Ax ≥ 0;x = 0 Projection Idempotent and Hermitian There exists permutation P Reducible such that A = PAPT is block triangular • det(A) = 0ifN is odd • a = 0, thus trace(A) = 0 Skew-symmetric AT =−A ii • (B − BT )/2isthe skew-symmetric part of B • aii = 0 or pure imaginary Skew-hermitian A∗ =−A • (B − B∗)/2isthe skew-hermitian part of B • BT B and BBT are both symmetric but generally not equal Symmetric A = AT • (B + BT )/2isthe symmetric part of B Unitary A∗ = A−1 are matrices having only positive (or nonnegative) elements.1 Some special matrices depend on specifications on the pattern of the nonzero elements. For instance, we have Jordan, Toeplitz, Shift, Hankel, and circulant matrices, as well as their block matrix versions, that is, block-Jordan, block-Toeplitz, and so forth. There are also special matrices that depend on collective properties of the rows or columns. For instance, stochastic matrices are positive matrices in which the sum of the elements within each row should sum up to unity. Another example are diagonally dominant matrices, where for the elements of any fixed row, the sum of the magnitudes of off-diagonal elements should be less than the magnitude of the diagonal element in that row. Finally, there are matrices whose entries depend on their row and column indices, such as Fourier, Haddamard, Hilbert, and Cauchy matrices. Fourier and Haddamard matrices are used in signal-processing applications. As can be expected, these tables are not exhaustive. Instead, the collection shows that there are several classes and special matrices found in the literature. They often contain interesting patterns and properties such as analytical formulas for determinants, trace, inverses, and so forth, that could be taken advantage of during analysis and computations. 1 Note that positive matrices are not the same as positive definite matrices. For instance, with 15 1 −2 A = B = 51 02 A is positive but not positive definite, whereas B is positive definite but not positive. 564 Appendix A: Additional Details and Fortification for Chapter 1 Table A.3. Matrices classes (based on structure and composition) Name Definition Remarks • AB (or BA) will reverse ⎛ ⎞ sequence of rows (columns) α α 0 1 of B,scaledby i ⎝ ⎠ N Antidiagonal A = ... • det(A) = (−1) αi αN 0 • MATLAB: A=flipud(diag(alpha)) where alpha=(α1,...,αN) ⎧ ⎨ i > j + p • p is the right-bandwidth Band (or banded) a = 0if or ij ⎩ • q is the left-bandwidth j > i + q • = N α det(A) i=1 i • Let B = A−1 then if j > i, bij = 0 ⎛ ⎞ α 1 1 0 if j = i, b = ⎜ β α ⎟ ii α ⎜ 1 2 ⎟ i Bidiagonal = ⎜ ⎟ i−1 A ⎜ . ⎟ 1 βk (Stieltjes) ⎝ .. .. ⎠ if i > j, bij = − αi αk k=j 0 βN−1 αN • MATLAB: A=diag(v)+diag(w,-1) where v= (α1,...,αN) w= (β1,...,βN−1) • Often used to indicate Binary aij = 0or1 incidence relationship between i and j • Are nonsingular (but often ill-conditioned for large N) For given x and y • = det(A) 1 N i−1 a = ; x + y = 0 = = fij Cauchy ij x + y i j i2 j 1 i j N N + and elements of x and y i=1 j=1(xi yj ) = − − are distinct where fij (xi xj )(yi yj ) • MATLAB: A=gallery(‘cauchy’,x,y) ⎛ ⎞ • α α ··· α Are normal matrices 1 2 N • ⎜ α α ··· α ⎟ Are special case of Toeplitz = ⎜ N 1 N−1 ⎟ • Circulant A ⎝ ··· ⎠ MATLAB: A=gallery(‘circul’,alpha) α2 α3 ··· α1 where alpha= (α1, ···,αN) ⎛ ⎞ − ··· − − pn−1 p1 p0 • ⎜ ⎟ pk are coefficients of a ⎜ ⎟ ⎜ ⎟ polynomial: ⎜ 100 ⎟ N n−1 Companion A = ⎜ ⎟ s + pN−1s + p1s + p0 ⎜ ⎟ . • MATLAB: A=compan(p) ⎝ .. ⎠ where p= (1, p − ,...,p , p ) 010 n 1 1 0 Complex aij are complex-valued Appendix A: Additional Details and Fortification for Chapter 1 565 Name Definition Remarks ⎛ ⎞ α 1 0 • = α ⎜ ⎟ det(A) i i = ⎜ . ⎟ • Diagonal A ⎝ .. ⎠ MATLAB: A=diag(alpha) where alpha= (α1,...,αN) 0 αN |a | > a • Nonsingular (based on Diagonally dominant ii ij i=j Gersgorin’s theorem) i = 1, 2,...,N • Are orthogonal √ (i−1)(j−1) • Used in Fourier transforms aij = (1/ N)W • MATLAB: Fourier h=ones(N,1)*[0:N-1]; √ 2π W = exp − −1 W=exp(-2*pi/N*1i); N A=W.ˆ(h.*h’)/sqrt(N) Identity matrix with 4 • Used to rotate points elements replaced based on in hyperplane Givens (Rotation) given p and q: • Useful in matrix reduction app = aqq = cos(θ) to Hessenberg form apq =−aqp = sin(θ) • Are orthogonal H [=]2k × 2k k • Elements either 1 or −1 11 Hadamard H = ⊗ H − • Are orthogonal k 1 −1 k 1 • MATLAB: A=hadamard(2ˆk) H0 = 1 ⎛ ⎞ • ··· βα Each anti-diagonal has the ⎜ ··· ··· γ ⎟ same value = ⎜ ⎟ • Hankel A ⎝ β ··· ··· ⎠ MATLAB: A=hankel([v,w]) = ...,β,α αγ ··· where v ( ) w= (α,γ,...) • Useful in finding eigenvalues • For square B, there is unitary Q such that A = Q∗BQ is a + , = 0 Hessenberg j k j upper hessenberg 2 ≤ k ≤ (N − j) • MATLAB: [Q,A]=hess(B); where A=(Q’)(B)(Q) • Symmetric and positive definite 1 • MATLAB: Hilbert a = ij i + j − 1 h=[1:N]; A=gallery(‘cauchy’,h,h-1) ⎛ ⎞ • 10 Often denoted by IN ⎜ ⎟ • det(A) = 1 Identity A = ⎜ .. ⎟ ⎝ . ⎠ • AB = BA = B 01 • MATLAB: A=eye(N) (continued) 566 Appendix A: Additional Details and Fortification for Chapter 1 Table A.3 (continued) Name Definition Remarks A = iB Imaginary where B√is real and i = −1 ⎛ ⎞ s 10 ⎜ ⎟ • Are bidiagonal ⎜ . ⎟ ⎜ .. .. ⎟ • det(A) = sN Jordan block A = ⎜ ⎟ ⎜ .
Recommended publications
  • Perron–Frobenius Theory of Nonnegative Matrices
    Buy from AMAZON.com http://www.amazon.com/exec/obidos/ASIN/0898714540 CHAPTER 8 Perron–Frobenius Theory of Nonnegative Matrices 8.1 INTRODUCTION m×n A ∈ is said to be a nonnegative matrix whenever each aij ≥ 0, and this is denoted by writing A ≥ 0. In general, A ≥ B means that each aij ≥ bij. Similarly, A is a positive matrix when each aij > 0, and this is denoted by writing A > 0. More generally, A > B means that each aij >bij. Applications abound with nonnegative and positive matrices. In fact, many of the applications considered in this text involve nonnegative matrices. For example, the connectivity matrix C in Example 3.5.2 (p. 100) is nonnegative. The discrete Laplacian L from Example 7.6.2 (p. 563) leads to a nonnegative matrix because (4I − L) ≥ 0. The matrix eAt that defines the solution of the system of differential equations in the mixing problem of Example 7.9.7 (p. 610) is nonnegative for all t ≥ 0. And the system of difference equations p(k)=Ap(k − 1) resulting from the shell game of Example 7.10.8 (p. 635) has Please report violations to [email protected] a nonnegativeCOPYRIGHTED coefficient matrix A. Since nonnegative matrices are pervasive, it’s natural to investigate their properties, and that’s the purpose of this chapter. A primary issue concerns It is illegal to print, duplicate, or distribute this material the extent to which the properties A > 0 or A ≥ 0 translate to spectral properties—e.g., to what extent does A have positive (or nonnegative) eigen- values and eigenvectors? The topic is called the “Perron–Frobenius theory” because it evolved from the contributions of the German mathematicians Oskar (or Oscar) Perron 89 and 89 Oskar Perron (1880–1975) originally set out to fulfill his father’s wishes to be in the family busi- Buy online from SIAM Copyright c 2000 SIAM http://www.ec-securehost.com/SIAM/ot71.html Buy from AMAZON.com 662 Chapter 8 Perron–Frobenius Theory of Nonnegative Matrices http://www.amazon.com/exec/obidos/ASIN/0898714540 Ferdinand Georg Frobenius.
    [Show full text]
  • Lecture Notes in Advanced Matrix Computations
    Lecture Notes in Advanced Matrix Computations Lectures by Dr. Michael Tsatsomeros Throughout these notes, signifies end proof, and N signifies end of example. Table of Contents Table of Contents i Lecture 0 Technicalities 1 0.1 Importance . 1 0.2 Background . 1 0.3 Material . 1 0.4 Matrix Multiplication . 1 Lecture 1 The Cost of Matrix Algebra 2 1.1 Block Matrices . 2 1.2 Systems of Linear Equations . 3 1.3 Triangular Systems . 4 Lecture 2 LU Decomposition 5 2.1 Gaussian Elimination and LU Decomposition . 5 Lecture 3 Partial Pivoting 6 3.1 LU Decomposition continued . 6 3.2 Gaussian Elimination with (Partial) Pivoting . 7 Lecture 4 Complete Pivoting 8 4.1 LU Decomposition continued . 8 4.2 Positive Definite Systems . 9 Lecture 5 Cholesky Decomposition 10 5.1 Positive Definiteness and Cholesky Decomposition . 10 Lecture 6 Vector and Matrix Norms 11 6.1 Vector Norms . 11 6.2 Matrix Norms . 13 Lecture 7 Vector and Matrix Norms continued 13 7.1 Matrix Norms . 13 7.2 Condition Numbers . 15 Notes by Jakob Streipel. Last updated April 27, 2018. i TABLE OF CONTENTS ii Lecture 8 Condition Numbers 16 8.1 Solving Perturbed Systems . 16 Lecture 9 Estimating the Condition Number 18 9.1 More on Condition Numbers . 18 9.2 Ill-conditioning by Scaling . 19 9.3 Estimating the Condition Number . 20 Lecture 10 Perturbing Not Only the Vector 21 10.1 Perturbing the Coefficient Matrix . 21 10.2 Perturbing Everything . 22 Lecture 11 Error Analysis After The Fact 23 11.1 A Posteriori Error Analysis Using the Residue .
    [Show full text]
  • 4.6 Matrices Dimensions: Number of Rows and Columns a Is a 2 X 3 Matrix
    Matrices are rectangular arrangements of data that are used to represent information in tabular form. 1 0 4 A = 3 - 6 8 a23 4.6 Matrices Dimensions: number of rows and columns A is a 2 x 3 matrix Elements of a matrix A are denoted by aij. a23 = 8 1 2 Data about many kinds of problems can often be represented by Matrix of coefficients matrix. Solutions to many problems can be obtained by solving e.g Average temperatures in 3 different cities for each month: systems of linear equations. For example, the constraints of a problem are represented by the system of linear equations 23 26 38 47 58 71 78 77 69 55 39 33 x + y = 70 A = 14 21 33 38 44 57 61 59 49 38 25 21 3 cities 24x + 14y = 1180 35 46 54 67 78 86 91 94 89 75 62 51 12 months Jan - Dec 1 1 rd The matrix A = Average temp. in the 3 24 14 city in July, a37, is 91. is the matrix of coefficient for this system of linear equations. 3 4 In a matrix, the arrangement of the entries is significant. Square Matrix is a matrix in which the number of Therefore, for two matrices to be equal they must have rows equals the number of columns. the same dimensions and the same entries in each location. •Main Diagonal: in a n x n square matrix, the elements a , a , a , …, a form the main Example: Let 11 22 33 nn diagonal of the matrix.
    [Show full text]
  • Recent Developments in Boolean Matrix Factorization
    Proceedings of the Twenty-Ninth International Joint Conference on Artificial Intelligence (IJCAI-20) Survey Track Recent Developments in Boolean Matrix Factorization Pauli Miettinen1∗ and Stefan Neumann2y 1University of Eastern Finland, Kuopio, Finland 2University of Vienna, Vienna, Austria pauli.miettinen@uef.fi, [email protected] Abstract analysts, and van den Broeck and Darwiche [2013] to lifted inference community, to name but a few examples. Even more The goal of Boolean Matrix Factorization (BMF) is recently, Ravanbakhsh et al. [2016] studied the problem in to approximate a given binary matrix as the product the framework of machine learning, increasing the interest of of two low-rank binary factor matrices, where the that community, and Chandran et al. [2016]; Ban et al. [2019]; product of the factor matrices is computed under Fomin et al. [2019] (re-)launched the interest in the problem the Boolean algebra. While the problem is computa- in the theory community. tionally hard, it is also attractive because the binary The various fields in which BMF is used are connected by nature of the factor matrices makes them highly in- the desire to “keep the data binary”. This can be because terpretable. In the last decade, BMF has received a the factorization is used as a pre-processing step, and the considerable amount of attention in the data mining subsequent methods require binary input, or because binary and formal concept analysis communities and, more matrices are more interpretable in the application domain. In recently, the machine learning and the theory com- the latter case, Boolean algebra is often preferred over the field munities also started studying BMF.
    [Show full text]
  • Chapter 7 Powers of Matrices
    Chapter 7 Powers of Matrices 7.1 Introduction In chapter 5 we had seen that many natural (repetitive) processes such as the rabbit examples can be described by a discrete linear system (1) ~un+1 = A~un, n = 0, 1, 2,..., and that the solution of such a system has the form n (2) ~un = A ~u0. n In addition, we had learned how to find an explicit expression for A and hence for ~un by using the constituent matrices of A. In this chapter, however, we will be primarily interested in “long range predictions”, i.e. we want to know the behaviour of ~un for n large (or as n → ∞). In view of (2), this is implied by an understanding of the limit matrix n A∞ = lim A , n→∞ provided that this limit exists. Thus we shall: 1) Establish a criterion to guarantee the existence of this limit (cf. Theorem 7.6); 2) Find a quick method to compute this limit when it exists (cf. Theorems 7.7 and 7.8). We then apply these methods to analyze two application problems: 1) The shipping of commodities (cf. section 7.7) 2) Rat mazes (cf. section 7.7) As we had seen in section 5.9, the latter give rise to Markov chains, which we shall study here in more detail (cf. section 7.7). 312 Chapter 7: Powers of Matrices 7.2 Powers of Numbers As was mentioned in the introduction, we want to study here the behaviour of the powers An of a matrix A as n → ∞. However, before considering the general case, it is useful to first investigate the situation for 1 × 1 matrices, i.e.
    [Show full text]
  • Topological Recursion and Random Finite Noncommutative Geometries
    Western University Scholarship@Western Electronic Thesis and Dissertation Repository 8-21-2018 2:30 PM Topological Recursion and Random Finite Noncommutative Geometries Shahab Azarfar The University of Western Ontario Supervisor Khalkhali, Masoud The University of Western Ontario Graduate Program in Mathematics A thesis submitted in partial fulfillment of the equirr ements for the degree in Doctor of Philosophy © Shahab Azarfar 2018 Follow this and additional works at: https://ir.lib.uwo.ca/etd Part of the Discrete Mathematics and Combinatorics Commons, Geometry and Topology Commons, and the Quantum Physics Commons Recommended Citation Azarfar, Shahab, "Topological Recursion and Random Finite Noncommutative Geometries" (2018). Electronic Thesis and Dissertation Repository. 5546. https://ir.lib.uwo.ca/etd/5546 This Dissertation/Thesis is brought to you for free and open access by Scholarship@Western. It has been accepted for inclusion in Electronic Thesis and Dissertation Repository by an authorized administrator of Scholarship@Western. For more information, please contact [email protected]. Abstract In this thesis, we investigate a model for quantum gravity on finite noncommutative spaces using the topological recursion method originated from random matrix theory. More precisely, we consider a particular type of finite noncommutative geometries, in the sense of Connes, called spectral triples of type (1; 0) , introduced by Barrett. A random spectral triple of type (1; 0) has a fixed fermion space, and the moduli space of its Dirac operator D = fH; ·} ;H 2 HN , encoding all the possible geometries over the fermion −S(D) space, is the space of Hermitian matrices HN . A distribution of the form e dD is considered over the moduli space of Dirac operators.
    [Show full text]
  • Students' Solutions Manual Applied Linear Algebra
    Students’ Solutions Manual for Applied Linear Algebra by Peter J.Olver and Chehrzad Shakiban Second Edition Undergraduate Texts in Mathematics Springer, New York, 2018. ISBN 978–3–319–91040–6 To the Student These solutions are a resource for students studying the second edition of our text Applied Linear Algebra, published by Springer in 2018. An expanded solutions manual is available for registered instructors of courses adopting it as the textbook. Using the Manual The material taught in this book requires an active engagement with the exercises, and we urge you not to read the solutions in advance. Rather, you should use the ones in this manual as a means of verifying that your solution is correct. (It is our hope that all solutions appearing here are correct; errors should be reported to the authors.) If you get stuck on an exercise, try skimming the solution to get a hint for how to proceed, but then work out the exercise yourself. The more you can do on your own, the more you will learn. Please note: for students taking a course based on Applied Linear Algebra, copying solutions from this Manual can place you in violation of academic honesty. In particular, many solutions here just provide the final answer, and for full credit one must also supply an explanation of how this is found. Acknowledgements We thank a number of people, who are named in the text, for corrections to the solutions manuals that accompanied the first edition. Of course, as authors, we take full respon- sibility for all errors that may yet appear.
    [Show full text]
  • A Complete Bibliography of Publications in Linear Algebra and Its Applications: 2010–2019
    A Complete Bibliography of Publications in Linear Algebra and its Applications: 2010{2019 Nelson H. F. Beebe University of Utah Department of Mathematics, 110 LCB 155 S 1400 E RM 233 Salt Lake City, UT 84112-0090 USA Tel: +1 801 581 5254 FAX: +1 801 581 4148 E-mail: [email protected], [email protected], [email protected] (Internet) WWW URL: http://www.math.utah.edu/~beebe/ 12 March 2021 Version 1.74 Title word cross-reference KY14, Rim12, Rud12, YHH12, vdH14]. 24 [KAAK11]. 2n − 3[BCS10,ˇ Hil13]. 2 × 2 [CGRVC13, CGSCZ10, CM14, DW11, DMS10, JK11, KJK13, MSvW12, Yan14]. (−1; 1) [AAFG12].´ (0; 1) 2 × 2 × 2 [Ber13b]. 3 [BBS12b, NP10, Ghe14a]. (2; 2; 0) [BZWL13, Bre14, CILL12, CKAC14, Fri12, [CI13, PH12]. (A; B) [PP13b]. (α, β) GOvdD14, GX12a, Kal13b, KK14, YHH12]. [HW11, HZM10]. (C; λ; µ)[dMR12].(`; m) p 3n2 − 2 2n3=2 − 3n [MR13]. [DFG10]. (H; m)[BOZ10].(κ, τ) p 3n2 − 2 2n3=23n [MR14a]. 3 × 3 [CSZ10, CR10c]. (λ, 2) [BBS12b]. (m; s; 0) [Dru14, GLZ14, Sev14]. 3 × 3 × 2 [Ber13b]. [GH13b]. (n − 3) [CGO10]. (n − 3; 2; 1) 3 × 3 × 3 [BH13b]. 4 [Ban13a, BDK11, [CCGR13]. (!) [CL12a]. (P; R)[KNS14]. BZ12b, CK13a, FP14, NSW13, Nor14]. 4 × 4 (R; S )[Tre12].−1 [LZG14]. 0 [AKZ13, σ [CJR11]. 5 Ano12-30, CGGS13, DLMZ14, Wu10a]. 1 [BH13b, CHY12, KRH14, Kol13, MW14a]. [Ano12-30, AHL+14, CGGS13, GM14, 5 × 5 [BAD09, DA10, Hil12a, Spe11]. 5 × n Kal13b, LM12, Wu10a]. 1=n [CNPP12]. [CJR11]. 6 1 <t<2 [Seo14]. 2 [AIS14, AM14, AKA13, [DK13c, DK11, DK12a, DK13b, Kar11a].
    [Show full text]
  • Volume 3 (2010), Number 2
    APLIMAT - JOURNAL OF APPLIED MATHEMATICS VOLUME 3 (2010), NUMBER 2 APLIMAT - JOURNAL OF APPLIED MATHEMATICS VOLUME 3 (2010), NUMBER 2 Edited by: Slovak University of Technology in Bratislava Editor - in - Chief: KOVÁČOVÁ Monika (Slovak Republic) Editorial Board: CARKOVS Jevgenijs (Latvia ) CZANNER Gabriela (USA) CZANNER Silvester (Great Britain) DE LA VILLA Augustin (Spain) DOLEŽALOVÁ Jarmila (Czech Republic) FEČKAN Michal (Slovak Republic) FERREIRA M. A. Martins (Portugal) FRANCAVIGLIA Mauro (Italy) KARPÍŠEK Zdeněk (Czech Republic) KOROTOV Sergey (Finland) LORENZI Marcella Giulia (Italy) MESIAR Radko (Slovak Republic) TALAŠOVÁ Jana (Czech Republic) VELICHOVÁ Daniela (Slovak Republic) Editorial Office: Institute of natural sciences, humanities and social sciences Faculty of Mechanical Engineering Slovak University of Technology in Bratislava Námestie slobody 17 812 31 Bratislava Correspodence concerning subscriptions, claims and distribution: F.X. spol s.r.o Azalková 21 821 00 Bratislava [email protected] Frequency: One volume per year consisting of three issues at price of 120 EUR, per volume, including surface mail shipment abroad. Registration number EV 2540/08 Information and instructions for authors are available on the address: http://www.journal.aplimat.com/ Printed by: FX spol s.r.o, Azalková 21, 821 00 Bratislava Copyright © STU 2007-2010, Bratislava All rights reserved. No part may be reproduced, stored in a retrieval system, or transmitted in any form or by any means, electronic, mechanical, photocopying, recording, or otherwise,
    [Show full text]
  • COMPUTING RELATIVELY LARGE ALGEBRAIC STRUCTURES by AUTOMATED THEORY EXPLORATION By
    COMPUTING RELATIVELY LARGE ALGEBRAIC STRUCTURES BY AUTOMATED THEORY EXPLORATION by QURATUL-AIN MAHESAR A thesis submitted to The University of Birmingham for the degree of DOCTOR OF PHILOSOPHY School of Computer Science College of Engineering and Physical Sciences The University of Birmingham March 2014 University of Birmingham Research Archive e-theses repository This unpublished thesis/dissertation is copyright of the author and/or third parties. The intellectual property rights of the author or third parties in respect of this work are as defined by The Copyright Designs and Patents Act 1988 or as modified by any successor legislation. Any use made of information contained in this thesis/dissertation must be in accordance with that legislation and must be properly acknowledged. Further distribution or reproduction in any format is prohibited without the permission of the copyright holder. Abstract Automated reasoning technology provides means for inference in a formal context via a multitude of disparate reasoning techniques. Combining different techniques not only increases the effectiveness of single systems but also provides a more powerful approach to solving hard problems. Consequently combined reasoning systems have been successfully employed to solve non-trivial mathematical problems in combinatorially rich domains that are intractable by traditional mathematical means. Nevertheless, the lack of domain specific knowledge often limits the effectiveness of these systems. In this thesis we investigate how the combination of diverse reasoning techniques can be employed to pre-compute additional knowledge to enable mathematical discovery in finite and potentially infinite domains that is otherwise not feasible. In particular, we demonstrate how we can exploit bespoke symbolic computations and automated theorem proving to automatically compute and evolve the structural knowledge of small size finite structures in the algebraic theory of quasigroups.
    [Show full text]
  • Formal Matrix Integrals and Combinatorics of Maps
    SPhT-T05/241 Formal matrix integrals and combinatorics of maps B. Eynard 1 Service de Physique Th´eorique de Saclay, F-91191 Gif-sur-Yvette Cedex, France. Abstract: This article is a short review on the relationship between convergent matrix inte- grals, formal matrix integrals, and combinatorics of maps. 1 Introduction This article is a short review on the relationship between convergent matrix inte- grals, formal matrix integrals, and combinatorics of maps. We briefly summa- rize results developed over the last 30 years, as well as more recent discoveries. We recall that formal matrix integrals are identical to combinatorial generating functions for maps, and that formal matrix integrals are in general very different from arXiv:math-ph/0611087v3 19 Dec 2006 convergent matrix integrals. Both may coincide perturbatively (i.e. up to terms smaller than any negative power of N), only for some potentials which correspond to negative weights for the maps, and therefore not very interesting from the combinatorics point of view. We also recall that both convergent and formal matrix integrals are solutions of the same set of loop equations, and that loop equations do not have a unique solution in general. Finally, we give a list of the classical matrix models which have played an important role in physics in the past decades. Some of them are now well understood, some are still difficult challenges. Matrix integrals were first introduced by physicists [55], mostly in two ways: - in nuclear physics, solid state physics, quantum chaos, convergent matrix integrals are 1 E-mail: [email protected] 1 studied for the eigenvalues statistical properties [48, 33, 5, 52].
    [Show full text]
  • Memory Capacity of Neural Turing Machines with Matrix Representation
    Memory Capacity of Neural Turing Machines with Matrix Representation Animesh Renanse * [email protected] Department of Electronics & Electrical Engineering Indian Institute of Technology Guwahati Assam, India Rohitash Chandra * [email protected] School of Mathematics and Statistics University of New South Wales Sydney, NSW 2052, Australia Alok Sharma [email protected] Laboratory for Medical Science Mathematics RIKEN Center for Integrative Medical Sciences Yokohama, Japan *Equal contributions Editor: Abstract It is well known that recurrent neural networks (RNNs) faced limitations in learning long- term dependencies that have been addressed by memory structures in long short-term memory (LSTM) networks. Matrix neural networks feature matrix representation which inherently preserves the spatial structure of data and has the potential to provide better memory structures when compared to canonical neural networks that use vector repre- sentation. Neural Turing machines (NTMs) are novel RNNs that implement notion of programmable computers with neural network controllers to feature algorithms that have copying, sorting, and associative recall tasks. In this paper, we study augmentation of mem- arXiv:2104.07454v1 [cs.LG] 11 Apr 2021 ory capacity with matrix representation of RNNs and NTMs (MatNTMs). We investigate if matrix representation has a better memory capacity than the vector representations in conventional neural networks. We use a probabilistic model of the memory capacity us- ing Fisher information and investigate how the memory capacity for matrix representation networks are limited under various constraints, and in general, without any constraints. In the case of memory capacity without any constraints, we found that the upper bound on memory capacity to be N 2 for an N ×N state matrix.
    [Show full text]