Multilinear Algebra for Visual Geometry

Total Page:16

File Type:pdf, Size:1020Kb

Multilinear Algebra for Visual Geometry multilinear algebra for visual geometry Alberto Ruiz http://dis.um.es/~alberto DIS, University of Murcia MVIGRO workshop Berlin, june 28th 2013 contents 1. linear spaces 2. tensor diagrams 3. Grassmann algebra 4. multiview geometry 5. tensor equations intro diagrams multiview equations linear algebra the most important computational tool: linear = solvable nonlinear problems: linearization + iteration multilinear problems are easy (' linear) visual geometry is multilinear (' solved) intro diagrams multiview equations linear spaces I objects are represented by arrays of coordinates (depending on chosen basis, with proper rules of transformation) I linear transformations are extremely simple (just multiply and add), and invertible in closed form (in least squares sense) I duality: linear transformations are also a linear space intro diagrams multiview equations linear spaces objects are linear combinations of other elements: i vector x has coordinates x in basis B = feig X i x = x ei i Einstein's convention: i x = x ei intro diagrams multiview equations change of basis 1 1 0 0 c1 c2 3 1 0 i [e1; e2] = [e1; e2]C; C = 2 2 = ej = cj ei c1 c2 1 2 x1 x1 x01 x = x1e + x2e = [e ; e ] = [e ; e ]C C−1 = [e0 ; e0 ] 1 2 1 2 x2 1 2 x2 1 2 x02 | 0{z 0 } [e1;e2] | {z } 2x013 4x025 intro diagrams multiview equations covariance and contravariance if the basis transforms according to C: 0 0 [e1; e2] = [e1; e2]C vector coordinates transform according to C−1: x01 x1 = C−1 x02 x2 in the previous example: 7 3 1 2 2 0;4 −0;2 7 = ; = 4 1 2 1 1 −0;2 0;6 4 intro diagrams multiview equations covariance and contravariance I things that transform like the basis are covariant (subindex, `row', `function-like') I things that transform `inversely' are contravariant (superindex, `column', `vector-like') vector coordinates are contravariant intro diagrams multiview equations scalar linear functions n the coordinates of linear functions R ! R (`covectors') are covariant: T 1 n k f(x) = a x = a1x + : : : anx = akx 0 a Tx0 = aTC C−1x |{z} | {z } a0T x0 0 r ak = ckar intro diagrams multiview equations linear transformations j in bases feig and fgig the transformation A has coordinates fai g j j i y = ai x in bases fe0 = Ceg and fg0 = Dgg: A0 = D−1AC 0t j i t t r r as = csajqi ; qi dt = δi the `input' is covariant and the `output' is contravariant intro diagrams multiview equations bilinear functions bilinear functions have two covariant (function-like) inputs. C(u; v) = uTCv conics (quadratic forms) are similar to bilinear functions: C(x) = xTCx = 0 line (dual) conics have two contravariant (vector-like) inputs, and so on. intro diagrams multiview equations multilinear functions several vector and covector arguments, linear in all of them I application and composition: contraction I partial application: produces a multilinear function of the remaining arguments I duality: no fixed input/output role a transformation of vectors transforms covectors in the opposite direction intro diagrams multiview equations matrix notation too many things are described by raw 2D arrays: I data sets (e.g. a collection of vectors) I linear transformations I bilinear functions, quadratic forms I homographies, Fundamental Matrix, etc.. matrix notation and manipulation tricks are not suitable for multilinear algebra alternative to algebraic indexed expressions: tensor diagrams intro diagrams multiview equations scalar, invariant same value s in all bases square edges intro diagrams multiview equations vector k coordinates x depend on chosen basis fekg k x = x ek contravariant slot: convex, `pointed' edge intro diagrams multiview equations linear function T 1 n k f(x) = a x = a1x + ::: + anx = akx 0 a Tx0 = aTCC−1x covariant slot: `concave' edge intro diagrams multiview equations function application T fa(x) = a x produces an invariant: index contraction = connecting complementary slots intro diagrams multiview equations transformation i j i y = FA(x) = Ax = ajx = y covariant and contravariant slots: intro diagrams multiview equations transformation works both ways intro diagrams multiview equations dual transformations intro diagrams multiview equations composition intro diagrams multiview equations identity intro diagrams multiview equations bilinear function e.g. an inner product given two vectors produces an invariant: a linear function is the inner product with some vector: intro diagrams multiview equations quadratic forms exercises: I conics I dual conics I tangent to a conic / polar lines intro diagrams multiview equations self-contraction? intro diagrams multiview equations tensor product creates a multilinear function from the product of the results of several multilinear functions: multiply function coordinates before the arguments are applied intro diagrams multiview equations trace any transformation can be decomposed as (sum of) there is an invariant: intro diagrams multiview equations contraction multilinear function application and/or composition tensor product followed by sum of `diagonal slice' of repeated indexes. implemented by `matrization' and normal matrix product intro diagrams multiview equations homogeneous coordinates projective geometry: geometric objects represented by subspaces n in P points and hyperplanes are dual points are contravariant, hyperplanes are covariant 2 in P points and lines are easy, e.g. incidence line-point: lT x = 0 3 in P points and planes are easy, lines are awkward we want a uniform representation of points, lines, planes, incidence, etc. for spaces of any dimension intro diagrams multiview equations symmetric tensors a multilinear function with undistinguished slots F (:::; x;:::; y;:::) = F (:::; y;:::; x;:::) commutative function symmetric coordinates intro diagrams multiview equations antisymmetric tensors F (:::; x;:::; y;:::) = −F (:::; y;:::; x;:::) F (:::; x;:::; x;:::) = 0 responds only to the linear independent component of the input F (:::; x + y;:::; x + z;:::) = F (:::; y;:::; z;:::) intro diagrams multiview equations full antisymmetric tensor Levi-Civita symbol 3 e.g. in R : when fully applied computes the `determinant' hypervolume of parallelepiped defined by the inputs undistinguished inputs (except sign) extremely useful: universal linear solver intro diagrams multiview equations exterior product antisymmetrized tensor product obtains an antisymmetric tensor represents the subspace spanned by the inputs ‘flat’, `blade', `r-vector', `multivector', `extensor' intro diagrams multiview equations exterior product representation of the subspace, independent of the `generators' (up to scale) intro diagrams multiview equations exterior product An r-vector can be represented as a dense multilinear array, or as coordinates in suitable basis: ei ^ ej = eij ei ^ ei = 0 eij = −eji Grassmann Algebra cf. geometric algebra, Pl¨ucker coordinates, etc. intro diagrams multiview equations example a 3D line T a = [1; 2; 0; 0] = e1 + 2e2 T b = [0; 3; 0; 4] = 3e2 + 4e3 a ^ b = 3e12 + 4e14 + 8e24 2 0 3 0 43 6−3 0 0 87 a ^ b = 6 7 4 0 0 0 05 −4 −8 0 0 intro diagrams multiview equations the whole space represented by the full contravariant antisymmetric tensor: intro diagrams multiview equations transformation of r-vectors intro diagrams multiview equations metric tensor vectors depicted as arrows in space are actually not very meaninful without an inner product, giving distances and angles. intro diagrams multiview equations metric tensor dual coordinates (raise/lower indexes): intro diagrams multiview equations inner product of r-vectors induced by the standard inner product for vectors: what does it mean? intro diagrams multiview equations inner product of r-vectors obtains the orthogonal complement of the projection x a a ^ b (a ^ b) · x b intro diagrams multiview equations Hodge dual there is an isomorphism between r-vectors and (n − r)-vectors e.g. dual of a line in 3D space: inner product with whole space gives `complementary' subspace (x ^ y is automatically computed by contraction with ) intro diagrams multiview equations cross product 3 in R : intro diagrams multiview equations [x]× k [x]× = ijkx = aij = 2 3 a 2 0 a b3 b 4 5 4−a 0 c5 c −b −c 0 intro diagrams multiview equations dualization universal tool for geometric constructions: I ijk::: transforms `span of points' into `generic intersection of hyperplanes' ijk::: I intersects (dual) hyperplanes to get `span of points' intro diagrams multiview equations normal vs dual representation P2 p ∗p l= p1 ^ p2 *l 3 × 3 3 × 3 3 3 point line P3 π= p1 ^ p2 ^ p3 4 × 4 × 4 ∗p p 4 × 4 × 4 l= p1 ^ p2 ∗l ∗π 4 × 4 4 × 4 4 4 point line plane intro diagrams multiview equations join (^) and meet (_) join meet l l p π _ l p π _ l l l is analogous to a NOR logic gate intro diagrams multiview equations cross product trick in P2 p1 p ∗l p 2 1 ∗l2 ∗l p1 ∗l1 ∗l2 ∗l p p 2 intro diagrams multiview equations incidence geometric objects Q = a ^ b ^ c ^ ::: and Q = x ^ y ^ ::: P is in Q: I Q ^ P = 0 I ∗Q · P = PQ = 0 example: incidence line (a ^ b) - point (x) 3 P2 P P4 = 0 ijk 03 03×3 intro diagrams multiview equations rule 1 dual of dual = identity (up to scale) intro diagrams multiview equations rule 2 transformation of whole space = whole space intro diagrams multiview equations inversion intro diagrams multiview equations preimage point reprojection line reprojection intro diagrams multiview equations nullspace the preimage of nothing intro diagrams multiview equations rank deficient nullspace intro diagrams multiview equations rank deficient generalization of rule 2 intro diagrams multiview equations
Recommended publications
  • Multivector Differentiation and Linear Algebra 0.5Cm 17Th Santaló
    Multivector differentiation and Linear Algebra 17th Santalo´ Summer School 2016, Santander Joan Lasenby Signal Processing Group, Engineering Department, Cambridge, UK and Trinity College Cambridge [email protected], www-sigproc.eng.cam.ac.uk/ s jl 23 August 2016 1 / 78 Examples of differentiation wrt multivectors. Linear Algebra: matrices and tensors as linear functions mapping between elements of the algebra. Functional Differentiation: very briefly... Summary Overview The Multivector Derivative. 2 / 78 Linear Algebra: matrices and tensors as linear functions mapping between elements of the algebra. Functional Differentiation: very briefly... Summary Overview The Multivector Derivative. Examples of differentiation wrt multivectors. 3 / 78 Functional Differentiation: very briefly... Summary Overview The Multivector Derivative. Examples of differentiation wrt multivectors. Linear Algebra: matrices and tensors as linear functions mapping between elements of the algebra. 4 / 78 Summary Overview The Multivector Derivative. Examples of differentiation wrt multivectors. Linear Algebra: matrices and tensors as linear functions mapping between elements of the algebra. Functional Differentiation: very briefly... 5 / 78 Overview The Multivector Derivative. Examples of differentiation wrt multivectors. Linear Algebra: matrices and tensors as linear functions mapping between elements of the algebra. Functional Differentiation: very briefly... Summary 6 / 78 We now want to generalise this idea to enable us to find the derivative of F(X), in the A ‘direction’ – where X is a general mixed grade multivector (so F(X) is a general multivector valued function of X). Let us use ∗ to denote taking the scalar part, ie P ∗ Q ≡ hPQi. Then, provided A has same grades as X, it makes sense to define: F(X + tA) − F(X) A ∗ ¶XF(X) = lim t!0 t The Multivector Derivative Recall our definition of the directional derivative in the a direction F(x + ea) − F(x) a·r F(x) = lim e!0 e 7 / 78 Let us use ∗ to denote taking the scalar part, ie P ∗ Q ≡ hPQi.
    [Show full text]
  • Multilinear Algebra and Applications July 15, 2014
    Multilinear Algebra and Applications July 15, 2014. Contents Chapter 1. Introduction 1 Chapter 2. Review of Linear Algebra 5 2.1. Vector Spaces and Subspaces 5 2.2. Bases 7 2.3. The Einstein convention 10 2.3.1. Change of bases, revisited 12 2.3.2. The Kronecker delta symbol 13 2.4. Linear Transformations 14 2.4.1. Similar matrices 18 2.5. Eigenbases 19 Chapter 3. Multilinear Forms 23 3.1. Linear Forms 23 3.1.1. Definition, Examples, Dual and Dual Basis 23 3.1.2. Transformation of Linear Forms under a Change of Basis 26 3.2. Bilinear Forms 30 3.2.1. Definition, Examples and Basis 30 3.2.2. Tensor product of two linear forms on V 32 3.2.3. Transformation of Bilinear Forms under a Change of Basis 33 3.3. Multilinear forms 34 3.4. Examples 35 3.4.1. A Bilinear Form 35 3.4.2. A Trilinear Form 36 3.5. Basic Operation on Multilinear Forms 37 Chapter 4. Inner Products 39 4.1. Definitions and First Properties 39 4.1.1. Correspondence Between Inner Products and Symmetric Positive Definite Matrices 40 4.1.1.1. From Inner Products to Symmetric Positive Definite Matrices 42 4.1.1.2. From Symmetric Positive Definite Matrices to Inner Products 42 4.1.2. Orthonormal Basis 42 4.2. Reciprocal Basis 46 4.2.1. Properties of Reciprocal Bases 48 4.2.2. Change of basis from a basis to its reciprocal basis g 50 B B III IV CONTENTS 4.2.3.
    [Show full text]
  • A Clifford Dyadic Superfield from Bilateral Interactions of Geometric Multispin Dirac Theory
    A CLIFFORD DYADIC SUPERFIELD FROM BILATERAL INTERACTIONS OF GEOMETRIC MULTISPIN DIRAC THEORY WILLIAM M. PEZZAGLIA JR. Department of Physia, Santa Clam University Santa Clam, CA 95053, U.S.A., [email protected] and ALFRED W. DIFFER Department of Phyaia, American River College Sacramento, CA 958i1, U.S.A. (Received: November 5, 1993) Abstract. Multivector quantum mechanics utilizes wavefunctions which a.re Clifford ag­ gregates (e.g. sum of scalar, vector, bivector). This is equivalent to multispinors con­ structed of Dirac matrices, with the representation independent form of the generators geometrically interpreted as the basis vectors of spacetime. Multiple generations of par­ ticles appear as left ideals of the algebra, coupled only by now-allowed right-side applied (dextral) operations. A generalized bilateral (two-sided operation) coupling is propoeed which includes the above mentioned dextrad field, and the spin-gauge interaction as partic­ ular cases. This leads to a new principle of poly-dimensional covariance, in which physical laws are invariant under the reshuffling of coordinate geometry. Such a multigeometric su­ perfield equation is proposed, whi~h is sourced by a bilateral current. In order to express the superfield in representation and coordinate free form, we introduce Eddington E-F double-frame numbers. Symmetric tensors can now be represented as 4D "dyads", which actually are elements of a global SD Clifford algebra.. As a restricted example, the dyadic field created by the Greider-Ross multivector current (of a Dirac electron) describes both electromagnetic and Morris-Greider gravitational interactions. Key words: spin-gauge, multivector, clifford, dyadic 1. Introduction Multi vector physics is a grand scheme in which we attempt to describe all ba­ sic physical structure and phenomena by a single geometrically interpretable Algebra.
    [Show full text]
  • Causalx: Causal Explanations and Block Multilinear Factor Analysis
    To appear: Proc. of the 2020 25th International Conference on Pattern Recognition (ICPR 2020) Milan, Italy, Jan. 10-15, 2021. CausalX: Causal eXplanations and Block Multilinear Factor Analysis M. Alex O. Vasilescu1;2 Eric Kim2;1 Xiao S. Zeng2 [email protected] [email protected] [email protected] 1Tensor Vision Technologies, Los Angeles, California 2Department of Computer Science,University of California, Los Angeles Abstract—By adhering to the dictum, “No causation without I. INTRODUCTION:PROBLEM DEFINITION manipulation (treatment, intervention)”, cause and effect data Developing causal explanations for correct results or for failures analysis represents changes in observed data in terms of changes from mathematical equations and data is important in developing in the causal factors. When causal factors are not amenable for a trustworthy artificial intelligence, and retaining public trust. active manipulation in the real world due to current technological limitations or ethical considerations, a counterfactual approach Causal explanations are germane to the “right to an explanation” performs an intervention on the model of data formation. In the statute [15], [13] i.e., to data driven decisions, such as those case of object representation or activity (temporal object) rep- that rely on images. Computer graphics and computer vision resentation, varying object parts is generally unfeasible whether problems, also known as forward and inverse imaging problems, they be spatial and/or temporal. Multilinear algebra, the algebra have been cast as causal inference questions [40], [42] consistent of higher order tensors, is a suitable and transparent framework for disentangling the causal factors of data formation. Learning a with Donald Rubin’s quantitative definition of causality, where part-based intrinsic causal factor representations in a multilinear “A causes B” means “the effect of A is B”, a measurable framework requires applying a set of interventions on a part- and experimentally repeatable quantity [14], [17].
    [Show full text]
  • Geometric-Algebra Adaptive Filters Wilder B
    1 Geometric-Algebra Adaptive Filters Wilder B. Lopes∗, Member, IEEE, Cassio G. Lopesy, Senior Member, IEEE Abstract—This paper presents a new class of adaptive filters, namely Geometric-Algebra Adaptive Filters (GAAFs). They are Faces generated by formulating the underlying minimization problem (a deterministic cost function) from the perspective of Geometric Algebra (GA), a comprehensive mathematical language well- Edges suited for the description of geometric transformations. Also, (directed lines) differently from standard adaptive-filtering theory, Geometric Calculus (the extension of GA to differential calculus) allows Fig. 1. A polyhedron (3-dimensional polytope) can be completely described for applying the same derivation techniques regardless of the by the geometric multiplication of its edges (oriented lines, vectors), which type (subalgebra) of the data, i.e., real, complex numbers, generate the faces and hypersurfaces (in the case of a general n-dimensional quaternions, etc. Relying on those characteristics (among others), polytope). a deterministic quadratic cost function is posed, from which the GAAFs are devised, providing a generalization of regular adaptive filters to subalgebras of GA. From the obtained update rule, it is shown how to recover the following least-mean squares perform calculus with hypercomplex quantities, i.e., elements (LMS) adaptive filter variants: real-entries LMS, complex LMS, that generalize complex numbers for higher dimensions [2]– and quaternions LMS. Mean-square analysis and simulations in [10]. a system identification scenario are provided, showing very good agreement for different levels of measurement noise. GA-based AFs were first introduced in [11], [12], where they were successfully employed to estimate the geometric Index Terms—Adaptive filtering, geometric algebra, quater- transformation (rotation and translation) that aligns a pair of nions.
    [Show full text]
  • 28. Exterior Powers
    28. Exterior powers 28.1 Desiderata 28.2 Definitions, uniqueness, existence 28.3 Some elementary facts 28.4 Exterior powers Vif of maps 28.5 Exterior powers of free modules 28.6 Determinants revisited 28.7 Minors of matrices 28.8 Uniqueness in the structure theorem 28.9 Cartan's lemma 28.10 Cayley-Hamilton Theorem 28.11 Worked examples While many of the arguments here have analogues for tensor products, it is worthwhile to repeat these arguments with the relevant variations, both for practice, and to be sensitive to the differences. 1. Desiderata Again, we review missing items in our development of linear algebra. We are missing a development of determinants of matrices whose entries may be in commutative rings, rather than fields. We would like an intrinsic definition of determinants of endomorphisms, rather than one that depends upon a choice of coordinates, even if we eventually prove that the determinant is independent of the coordinates. We anticipate that Artin's axiomatization of determinants of matrices should be mirrored in much of what we do here. We want a direct and natural proof of the Cayley-Hamilton theorem. Linear algebra over fields is insufficient, since the introduction of the indeterminate x in the definition of the characteristic polynomial takes us outside the class of vector spaces over fields. We want to give a conceptual proof for the uniqueness part of the structure theorem for finitely-generated modules over principal ideal domains. Multi-linear algebra over fields is surely insufficient for this. 417 418 Exterior powers 2. Definitions, uniqueness, existence Let R be a commutative ring with 1.
    [Show full text]
  • Multilinear Algebra in Data Analysis: Tensors, Symmetric Tensors, Nonnegative Tensors
    Multilinear Algebra in Data Analysis: tensors, symmetric tensors, nonnegative tensors Lek-Heng Lim Stanford University Workshop on Algorithms for Modern Massive Datasets Stanford, CA June 21–24, 2006 Thanks: G. Carlsson, L. De Lathauwer, J.M. Landsberg, M. Mahoney, L. Qi, B. Sturmfels; Collaborators: P. Comon, V. de Silva, P. Drineas, G. Golub References http://www-sccm.stanford.edu/nf-publications-tech.html [CGLM2] P. Comon, G. Golub, L.-H. Lim, and B. Mourrain, “Symmetric tensors and symmetric tensor rank,” SCCM Tech. Rep., 06-02, 2006. [CGLM1] P. Comon, B. Mourrain, L.-H. Lim, and G.H. Golub, “Genericity and rank deficiency of high order symmetric tensors,” Proc. IEEE Int. Con- ference on Acoustics, Speech, and Signal Processing (ICASSP), 31 (2006), no. 3, pp. 125–128. [dSL] V. de Silva and L.-H. Lim, “Tensor rank and the ill-posedness of the best low-rank approximation problem,” SCCM Tech. Rep., 06-06 (2006). [GL] G. Golub and L.-H. Lim, “Nonnegative decomposition and approximation of nonnegative matrices and tensors,” SCCM Tech. Rep., 06-01 (2006), forthcoming. [L] L.-H. Lim, “Singular values and eigenvalues of tensors: a variational approach,” Proc. IEEE Int. Workshop on Computational Advances in Multi- Sensor Adaptive Processing (CAMSAP), 1 (2005), pp. 129–132. 2 What is not a tensor, I • What is a vector? – Mathematician: An element of a vector space. – Physicist: “What kind of physical quantities can be rep- resented by vectors?” Answer: Once a basis is chosen, an n-dimensional vector is something that is represented by n real numbers only if those real numbers transform themselves as expected (ie.
    [Show full text]
  • Multilinear Algebra
    Appendix A Multilinear Algebra This chapter presents concepts from multilinear algebra based on the basic properties of finite dimensional vector spaces and linear maps. The primary aim of the chapter is to give a concise introduction to alternating tensors which are necessary to define differential forms on manifolds. Many of the stated definitions and propositions can be found in Lee [1], Chaps. 11, 12 and 14. Some definitions and propositions are complemented by short and simple examples. First, in Sect. A.1 dual and bidual vector spaces are discussed. Subsequently, in Sects. A.2–A.4, tensors and alternating tensors together with operations such as the tensor and wedge product are introduced. Lastly, in Sect. A.5, the concepts which are necessary to introduce the wedge product are summarized in eight steps. A.1 The Dual Space Let V be a real vector space of finite dimension dim V = n.Let(e1,...,en) be a basis of V . Then every v ∈ V can be uniquely represented as a linear combination i v = v ei , (A.1) where summation convention over repeated indices is applied. The coefficients vi ∈ R arereferredtoascomponents of the vector v. Throughout the whole chapter, only finite dimensional real vector spaces, typically denoted by V , are treated. When not stated differently, summation convention is applied. Definition A.1 (Dual Space)Thedual space of V is the set of real-valued linear functionals ∗ V := {ω : V → R : ω linear} . (A.2) The elements of the dual space V ∗ are called linear forms on V . © Springer International Publishing Switzerland 2015 123 S.R.
    [Show full text]
  • The Construction of Spinors in Geometric Algebra
    The Construction of Spinors in Geometric Algebra Matthew R. Francis∗ and Arthur Kosowsky† Dept. of Physics and Astronomy, Rutgers University 136 Frelinghuysen Road, Piscataway, NJ 08854 (Dated: February 4, 2008) The relationship between spinors and Clifford (or geometric) algebra has long been studied, but little consistency may be found between the various approaches. However, when spinors are defined to be elements of the even subalgebra of some real geometric algebra, the gap between algebraic, geometric, and physical methods is closed. Spinors are developed in any number of dimensions from a discussion of spin groups, followed by the specific cases of U(1), SU(2), and SL(2, C) spinors. The physical observables in Schr¨odinger-Pauli theory and Dirac theory are found, and the relationship between Dirac, Lorentz, Weyl, and Majorana spinors is made explicit. The use of a real geometric algebra, as opposed to one defined over the complex numbers, provides a simpler construction and advantages of conceptual and theoretical clarity not available in other approaches. I. INTRODUCTION Spinors are used in a wide range of fields, from the quantum physics of fermions and general relativity, to fairly abstract areas of algebra and geometry. Independent of the particular application, the defining characteristic of spinors is their behavior under rotations: for a given angle θ that a vector or tensorial object rotates, a spinor rotates by θ/2, and hence takes two full rotations to return to its original configuration. The spin groups, which are universal coverings of the rotation groups, govern this behavior, and are frequently defined in the language of geometric (Clifford) algebras [1, 2].
    [Show full text]
  • Low-Level Image Processing with the Structure Multivector
    Low-Level Image Processing with the Structure Multivector Michael Felsberg Bericht Nr. 0202 Institut f¨ur Informatik und Praktische Mathematik der Christian-Albrechts-Universitat¨ zu Kiel Olshausenstr. 40 D – 24098 Kiel e-mail: [email protected] 12. Marz¨ 2002 Dieser Bericht enthalt¨ die Dissertation des Verfassers 1. Gutachter Prof. G. Sommer (Kiel) 2. Gutachter Prof. U. Heute (Kiel) 3. Gutachter Prof. J. J. Koenderink (Utrecht) Datum der mundlichen¨ Prufung:¨ 12.2.2002 To Regina ABSTRACT The present thesis deals with two-dimensional signal processing for computer vi- sion. The main topic is the development of a sophisticated generalization of the one-dimensional analytic signal to two dimensions. Motivated by the fundamental property of the latter, the invariance – equivariance constraint, and by its relation to complex analysis and potential theory, a two-dimensional approach is derived. This method is called the monogenic signal and it is based on the Riesz transform instead of the Hilbert transform. By means of this linear approach it is possible to estimate the local orientation and the local phase of signals which are projections of one-dimensional functions to two dimensions. For general two-dimensional signals, however, the monogenic signal has to be further extended, yielding the structure multivector. The latter approach combines the ideas of the structure tensor and the quaternionic analytic signal. A rich feature set can be extracted from the structure multivector, which contains measures for local amplitudes, the local anisotropy, the local orientation, and two local phases. Both, the monogenic signal and the struc- ture multivector are combined with an appropriate scale-space approach, resulting in generalized quadrature filters.
    [Show full text]
  • Lecture Notes on Linear and Multilinear Algebra 2301-610
    Lecture Notes on Linear and Multilinear Algebra 2301-610 Wicharn Lewkeeratiyutkul Department of Mathematics and Computer Science Faculty of Science Chulalongkorn University August 2014 Contents Preface iii 1 Vector Spaces 1 1.1 Vector Spaces and Subspaces . 2 1.2 Basis and Dimension . 10 1.3 Linear Maps . 18 1.4 Matrix Representation . 32 1.5 Change of Bases . 42 1.6 Sums and Direct Sums . 48 1.7 Quotient Spaces . 61 1.8 Dual Spaces . 65 2 Multilinear Algebra 73 2.1 Free Vector Spaces . 73 2.2 Multilinear Maps and Tensor Products . 78 2.3 Determinants . 95 2.4 Exterior Products . 101 3 Canonical Forms 107 3.1 Polynomials . 107 3.2 Diagonalization . 115 3.3 Minimal Polynomial . 128 3.4 Jordan Canonical Forms . 141 i ii CONTENTS 4 Inner Product Spaces 161 4.1 Bilinear and Sesquilinear Forms . 161 4.2 Inner Product Spaces . 167 4.3 Operators on Inner Product Spaces . 180 4.4 Spectral Theorem . 190 Bibliography 203 Preface This book grew out of the lecture notes for the course 2301-610 Linear and Multilinaer Algebra given at the Deparment of Mathematics, Faculty of Science, Chulalongkorn University that I have taught in the past 5 years. Linear Algebra is one of the most important subjects in Mathematics, with numerous applications in pure and applied sciences. A more theoretical linear algebra course will emphasize on linear maps between vector spaces, while an applied-oriented course will mainly work with matrices. Matrices have an ad- vantage of being easier to compute, while it is easier to establish the results by working with linear maps.
    [Show full text]
  • Numerical Multilinear Algebra I
    Numerical Multilinear Algebra I Lek-Heng Lim University of California, Berkeley January 5{7, 2009 L.-H. Lim (ICM Lecture) Numerical Multilinear Algebra I January 5{7, 2009 1 / 55 Hope Past 50 years, Numerical Linear Algebra played indispensable role in the statistical analysis of two-way data, the numerical solution of partial differential equations arising from vector fields, the numerical solution of second-order optimization methods. Next step | development of Numerical Multilinear Algebra for the statistical analysis of multi-way data, the numerical solution of partial differential equations arising from tensor fields, the numerical solution of higher-order optimization methods. L.-H. Lim (ICM Lecture) Numerical Multilinear Algebra I January 5{7, 2009 2 / 55 DARPA mathematical challenge eight One of the twenty three mathematical challenges announced at DARPA Tech 2007. Problem Beyond convex optimization: can linear algebra be replaced by algebraic geometry in a systematic way? Algebraic geometry in a slogan: polynomials are to algebraic geometry what matrices are to linear algebra. Polynomial f 2 R[x1;:::; xn] of degree d can be expressed as > > f (x) = a0 + a1 x + x A2x + A3(x; x; x) + ··· + Ad (x;:::; x): n n×n n×n×n n×···×n a0 2 R; a1 2 R ; A2 2 R ; A3 2 R ;:::; Ad 2 R . Numerical linear algebra: d = 2. Numerical multilinear algebra: d > 2. L.-H. Lim (ICM Lecture) Numerical Multilinear Algebra I January 5{7, 2009 3 / 55 Motivation Why multilinear: \Classification of mathematical problems as linear and nonlinear is like classification of the Universe as bananas and non-bananas." Nonlinear | too general.
    [Show full text]