Numerical Solution of Saddle Point Problems
Total Page:16
File Type:pdf, Size:1020Kb
Load more
Recommended publications
-
38. a Preconditioner for the Schur Complement Domain Decomposition Method J.-M
Fourteenth International Conference on Domain Decomposition Methods Editors: Ismael Herrera , David E. Keyes, Olof B. Widlund, Robert Yates c 2003 DDM.org 38. A preconditioner for the Schur complement domain decomposition method J.-M. Cros 1 1. Introduction. This paper presents a preconditioner for the Schur complement domain decomposition method inspired by the dual-primal FETI method [4]. Indeed the proposed method enforces the continuity of the preconditioned gradient at cross-points di- rectly by a reformulation of the classical Neumann-Neumann preconditioner. In the case of elasticity problems discretized by finite elements, the degrees of freedom corresponding to the cross-points coming from domain decomposition, in the stiffness matrix, are separated from the rest. Elimination of the remaining degrees of freedom results in a Schur complement ma- trix for the cross-points. This assembled matrix represents the coarse problem. The method is not mathematically optimal as shown by numerical results but its use is rather economical. The paper is organized as follows: in sections 2 and 3, the Schur complement method and the formulation of the Neumann-Neumann preconditioner are briefly recalled to introduce the notations. Section 4 is devoted to the reformulation of the Neumann-Neumann precon- ditioner. In section 5, the proposed method is compared with other domain decomposition methods such as generalized Neumann-Neumann algorithm [7][9], one-level FETI method [5] and dual-primal FETI method. Performances on a parallel machine are also given for structural analysis problems. 2. The Schur complement domain decomposition method. Let Ω denote the computational domain of an elasticity problem. -
Fast Solution of Sparse Linear Systems with Adaptive Choice of Preconditioners Zakariae Jorti
Fast solution of sparse linear systems with adaptive choice of preconditioners Zakariae Jorti To cite this version: Zakariae Jorti. Fast solution of sparse linear systems with adaptive choice of preconditioners. Gen- eral Mathematics [math.GM]. Sorbonne Université, 2019. English. NNT : 2019SORUS149. tel- 02425679v2 HAL Id: tel-02425679 https://tel.archives-ouvertes.fr/tel-02425679v2 Submitted on 15 Feb 2021 HAL is a multi-disciplinary open access L’archive ouverte pluridisciplinaire HAL, est archive for the deposit and dissemination of sci- destinée au dépôt et à la diffusion de documents entific research documents, whether they are pub- scientifiques de niveau recherche, publiés ou non, lished or not. The documents may come from émanant des établissements d’enseignement et de teaching and research institutions in France or recherche français ou étrangers, des laboratoires abroad, or from public or private research centers. publics ou privés. Sorbonne Université École doctorale de Sciences Mathématiques de Paris Centre Laboratoire Jacques-Louis Lions Résolution rapide de systèmes linéaires creux avec choix adaptatif de préconditionneurs Par Zakariae Jorti Thèse de doctorat de Mathématiques appliquées Dirigée par Laura Grigori Co-supervisée par Ani Anciaux-Sedrakian et Soleiman Yousef Présentée et soutenue publiquement le 03/10/2019 Devant un jury composé de: M. Tromeur-Dervout Damien, Professeur, Université de Lyon, Rapporteur M. Schenk Olaf, Professeur, Università della Svizzera italiana, Rapporteur M. Hecht Frédéric, Professeur, Sorbonne Université, Président du jury Mme. Emad Nahid, Professeur, Université Paris-Saclay, Examinateur M. Vasseur Xavier, Ingénieur de recherche, ISAE-SUPAERO, Examinateur Mme. Grigori Laura, Directrice de recherche, Inria Paris, Directrice de thèse Mme. Anciaux-Sedrakian Ani, Ingénieur de recherche, IFPEN, Co-encadrante de thèse M. -
Efficient “Black-Box” Multigrid Solvers for Convection-Dominated Problems
EFFICIENT \BLACK-BOX" MULTIGRID SOLVERS FOR CONVECTION-DOMINATED PROBLEMS A thesis submitted to the University of Manchester for the degree of Doctor of Philosophy in the Faculty of Engineering and Physical Science 2011 Glyn Owen Rees School of Computer Science 2 Contents Declaration 12 Copyright 13 Acknowledgement 14 1 Introduction 15 2 The Convection-Di®usion Problem 18 2.1 The continuous problem . 18 2.2 The Galerkin approximation method . 22 2.2.1 The ¯nite element method . 26 2.3 The Petrov-Galerkin (SUPG) approximation . 29 2.4 Properties of the discrete operator . 34 2.5 The Navier-Stokes problem . 37 3 Methods for Solving Linear Algebraic Systems 42 3.1 Basic iterative methods . 43 3.1.1 The Jacobi method . 45 3.1.2 The Gauss-Seidel method . 46 3.1.3 Convergence of splitting iterations . 48 3.1.4 Incomplete LU (ILU) factorisation . 49 3.2 Krylov methods . 56 3.2.1 The GMRES method . 59 3.2.2 Preconditioned GMRES method . 62 3.3 Multigrid method . 67 3.3.1 Geometric multigrid method . 67 3.3.2 Algebraic multigrid method . 75 3.3.3 Parallel AMG . 80 3.3.4 Literature summary of multigrid . 81 3.3.5 Multigrid preconditioning of Krylov solvers . 84 3 3.4 A new tILU smoother . 85 4 Two-Dimensional Case Studies 89 4.1 The di®usion problem . 90 4.2 Geometric multigrid preconditioning . 95 4.2.1 Constant uni-directional wind . 95 4.2.2 Double glazing problem - recirculating wind . 106 4.2.3 Combined uni-directional and recirculating wind . -
PENALTY METHODS for the SOLUTION of GENERALIZED NASH EQUILIBRIUM PROBLEMS (WITH COMPLETE TEST PROBLEMS) Francisco Facchinei1
PENALTY METHODS FOR THE SOLUTION OF GENERALIZED NASH EQUILIBRIUM PROBLEMS (WITH COMPLETE TEST PROBLEMS) Francisco Facchinei1 and Christian Kanzow2 1 Sapienza University of Rome Department of Computer and System Sciences “A. Ruberti” Via Ariosto 25, 00185 Roma, Italy e-mail: [email protected] 2 University of W¨urzburg Institute of Mathematics Am Hubland, 97074 W¨urzburg, Germany e-mail: [email protected] February 11, 2009 Abstract The generalized Nash equilibrium problem (GNEP) is an extension of the classical Nash equilibrium problem where both the objective functions and the constraints of each player may depend on the rivals’ strategies. This class of problems has a multitude of important engineering applications and yet solution algorithms are extremely scarce. In this paper, we analyze in detail a globally convergent penalty method that has favorable theoretical properties. We also consider strengthened results for a particular subclass of problems very often considered in the literature. Basically our method reduces the GNEP to a single penalized (and nonsmooth) Nash equilibrium problem. We suggest a suitable method for the solution of the latter penalized problem and present extensive numerical results. Key Words: Nash equilibrium problem, Generalized Nash equilibrium problem, Jointly convex problem, Exact penalty function, Global convergence. 1 Introduction In this paper, we consider the Generalized Nash Equilibrium Problem (GNEP) and describe new globally convergent methods for its solution. The GNEP is an extension of the classical Nash Equilibrium Problem (NEP) where the feasible sets of each player may depend on ν n the rivals’ strategies. We assume there are N players and denote by x ∈ R ν the vector representing the ν-th player’s strategy. -
Chapter 7 Iterative Methods for Large Sparse Linear Systems
Chapter 7 Iterative methods for large sparse linear systems In this chapter we revisit the problem of solving linear systems of equations, but now in the context of large sparse systems. The price to pay for the direct methods based on matrix factorization is that the factors of a sparse matrix may not be sparse, so that for large sparse systems the memory cost make direct methods too expensive, in memory and in execution time. Instead we introduce iterative methods, for which matrix sparsity is exploited to develop fast algorithms with a low memory footprint. 7.1 Sparse matrix algebra Large sparse matrices We say that the matrix A Rn is large if n is large, and that A is sparse if most of the elements are2 zero. If a matrix is not sparse, we say that the matrix is dense. Whereas for a dense matrix the number of nonzero elements is (n2), for a sparse matrix it is only (n), which has obvious implicationsO for the memory footprint and efficiencyO for algorithms that exploit the sparsity of a matrix. AdiagonalmatrixisasparsematrixA =(aij), for which aij =0for all i = j,andadiagonalmatrixcanbegeneralizedtoabanded matrix, 6 for which there exists a number p,thebandwidth,suchthataij =0forall i<j p or i>j+ p.Forexample,atridiagonal matrix A is a banded − 59 CHAPTER 7. ITERATIVE METHODS FOR LARGE SPARSE 60 LINEAR SYSTEMS matrix with p =1, xx0000 xxx000 20 xxx003 A = , (7.1) 600xxx07 6 7 6000xxx7 6 7 60000xx7 6 7 where x represents a nonzero4 element. 5 Compressed row storage The compressed row storage (CRS) format is a data structure for efficient represention of a sparse matrix by three arrays, containing the nonzero values, the respective column indices, and the extents of the rows. -
Chebyshev and Fourier Spectral Methods 2000
Chebyshev and Fourier Spectral Methods Second Edition John P. Boyd University of Michigan Ann Arbor, Michigan 48109-2143 email: [email protected] http://www-personal.engin.umich.edu/jpboyd/ 2000 DOVER Publications, Inc. 31 East 2nd Street Mineola, New York 11501 1 Dedication To Marilyn, Ian, and Emma “A computation is a temptation that should be resisted as long as possible.” — J. P. Boyd, paraphrasing T. S. Eliot i Contents PREFACE x Acknowledgments xiv Errata and Extended-Bibliography xvi 1 Introduction 1 1.1 Series expansions .................................. 1 1.2 First Example .................................... 2 1.3 Comparison with finite element methods .................... 4 1.4 Comparisons with Finite Differences ....................... 6 1.5 Parallel Computers ................................. 9 1.6 Choice of basis functions .............................. 9 1.7 Boundary conditions ................................ 10 1.8 Non-Interpolating and Pseudospectral ...................... 12 1.9 Nonlinearity ..................................... 13 1.10 Time-dependent problems ............................. 15 1.11 FAQ: Frequently Asked Questions ........................ 16 1.12 The Chrysalis .................................... 17 2 Chebyshev & Fourier Series 19 2.1 Introduction ..................................... 19 2.2 Fourier series .................................... 20 2.3 Orders of Convergence ............................... 25 2.4 Convergence Order ................................. 27 2.5 Assumption of Equal Errors ........................... -
First-Order Methods for Semidefinite Programming
FIRST-ORDER METHODS FOR SEMIDEFINITE PROGRAMMING Zaiwen Wen Advisor: Professor Donald Goldfarb Submitted in partial fulfillment of the requirements for the degree of Doctor of Philosophy in the Graduate School of Arts and Sciences COLUMBIA UNIVERSITY 2009 c 2009 Zaiwen Wen All Rights Reserved ABSTRACT First-order Methods For Semidefinite Programming Zaiwen Wen Semidefinite programming (SDP) problems are concerned with minimizing a linear function of a symmetric positive definite matrix subject to linear equality constraints. These convex problems are solvable in polynomial time by interior point methods. However, if the number of constraints m in an SDP is of order O(n2) when the unknown positive semidefinite matrix is n n, interior point methods become impractical both in terms of the × time (O(n6)) and the amount of memory (O(m2)) required at each iteration to form the m × m positive definite Schur complement matrix M and compute the search direction by finding the Cholesky factorization of M. This significantly limits the application of interior-point methods. In comparison, the computational cost of each iteration of first-order optimization approaches is much cheaper, particularly, if any sparsity in the SDP constraints or other special structure is exploited. This dissertation is devoted to the development, analysis and evaluation of two first-order approaches that are able to solve large SDP problems which have been challenging for interior point methods. In chapter 2, we present a row-by-row (RBR) method based on solving a sequence of problems obtained by restricting the n-dimensional positive semidefinite constraint on the matrix X. By fixing any (n 1)-dimensional principal submatrix of X and using its − (generalized) Schur complement, the positive semidefinite constraint is reduced to a sim- ple second-order cone constraint. -
Constraint-Handling Techniques for Highly Constrained Optimization
Andreas Petrow Constraint-Handling Techniques for highly Constrained Optimization Intelligent Cooperative Systems Computational Intelligence Constraint-Handling Techniques for highly Constrained Optimization Master Thesis Andreas Petrow March 22, 2019 Supervisor: Prof. Dr.-Ing. habil. Sanaz Mostaghim Advisor: M.Sc. Heiner Zille Advisor: Dr.-Ing. Markus Schori Andreas Petrow: Constraint-Handling Techniques for highly Con- strained Optimization Otto-von-Guericke Universität Intelligent Cooperative Systems Computational Intelligence Magdeburg, 2019. Abstract This thesis relates to constraint handling techniques for highly constrained opti- mization. Many real world optimization problems are non-linear and restricted to constraints. To solve non-linear optimization problems population-based meta- heuristics are used commonly. To enable such optimization methods to fulfill the optimization with respect to the constraints modifications are required. Therefore a research field considering such constraint handling techniques establishes. How- ever the combination of large-scale optimization problems with constraint handling techniques is rare. Therefore this thesis involves with common constraint handling techniques and evaluates these on two benchmark problems. I Contents List of Figures VII List of TablesXI List of Acronyms XIII 1. Introduction and Motivation1 1.1. Aim of this thesis...........................2 1.2. Structure of this thesis........................3 2. Basics5 2.1. Constrained Optimization Problem..................5 2.2. Optimization Method.........................6 2.2.1. Classical optimization....................7 2.2.2. Population-based Optimization Methods...........7 3. Related Work 11 3.1. Penalty Methods........................... 12 3.1.1. Static Penalty Functions................... 12 3.1.2. Co-Evolution Penalty Methods................ 13 3.1.3. Dynamic Penalty Functions................. 13 3.1.4. Adaptive Penalty Functions................. 13 3.2. -
Domain Decomposition Solvers (FETI) Divide Et Impera
Domain Decomposition solvers (FETI) a random walk in history and some current trends Daniel J. Rixen Technische Universität München Institute of Applied Mechanics www.amm.mw.tum.de [email protected] 8-10 October 2014 39th Woudschoten Conference, organised by the Werkgemeenschap Scientific Computing (WSC) 1 Divide et impera Center for Aerospace Structures CU, Boulder wikipedia When splitting the problem in parts and asking different cpu‘s (or threads) to take care of subproblems, will the problem be solved faster ? FETI, Primal Schur (Balancing) method around 1990 …….. basic methods, mesh decomposer technology 1990-2001 .…….. improvements •! preconditioners, coarse grids •! application to Helmholtz, dynamics, non-linear ... Here the concepts are outlined using some mechanical interpretation. For mathematical details, see lecture of Axel Klawonn. !"!! !"#$%&'()$&'(*+*',-%-.&$(*&$-%'#$%-'(*-/$0%1*()%-).1*,2%()'(%()$%,.,3$4*-($,+$% .5%'%-.6"(*.,%*&76*$-%'%6.2*+'6%+.,(8'0*+(*.,9%1)*6$%$,2*,$$#-%&*2)(%+.,-*0$#%'% ,"&$#*+'6%#$-"6(%'-%()$%.,6:%#$'-.,';6$%2.'6<%% ="+)%.,$%-*0$0%>*$1-%-$$&%(.%#$?$+(%)"&',%6*&*('(*.,-%#'()$#%()',%.;@$+(*>$%>'6"$-<%% A,%*(-$65%&'()$&'(*+-%*-%',%*,0*>*-*;6$%.#2',*-&%",*(*,2%()$.#$(*+'6%+.,($&76'(*.,% ',0%'+(*>$%'776*+'(*.,<%%%%% %%%%%%% R. Courant %B % in Variational! Methods for the solution of problems of equilibrium and vibrations Bulletin of American Mathematical Society, 49, pp.1-23, 1943 Here the concepts are outlined using some mechanical interpretation. For mathematical details, see lecture of Axel Klawonn. Content -
Relaxed Modulus-Based Matrix Splitting Methods for the Linear Complementarity Problem †
S S symmetry Article Relaxed Modulus-Based Matrix Splitting Methods for the Linear Complementarity Problem † Shiliang Wu 1, *, Cuixia Li 1 and Praveen Agarwal 2,3 1 School of Mathematics, Yunnan Normal University, Kunming 650500, China; [email protected] 2 Department of Mathrematics, Anand International College of Engineering, Jaipur 303012, India; [email protected] 3 Nonlinear Dynamics Research Center (NDRC), Ajman University, Ajman, United Arab Emirates * Correspondence: [email protected] † This research was supported by National Natural Science Foundation of China (No.11961082). Abstract: In this paper, we obtain a new equivalent fixed-point form of the linear complementarity problem by introducing a relaxed matrix and establish a class of relaxed modulus-based matrix split- ting iteration methods for solving the linear complementarity problem. Some sufficient conditions for guaranteeing the convergence of relaxed modulus-based matrix splitting iteration methods are presented. Numerical examples are offered to show the efficacy of the proposed methods. Keywords: linear complementarity problem; matrix splitting; iteration method; convergence MSC: 90C33; 65F10; 65F50; 65G40 1. Introduction Citation: Wu, S.; Li, C.; Agarwal, P. Relaxed Modulus-Based Matrix In this paper, we focus on the iterative solution of the linear complementarity problem, Splitting Methods for the Linear abbreviated as ‘LCP(q, A)’, whose form is Complementarity Problem. Symmetry T 2021, 13, 503. https://doi.org/ w = Az + q ≥ 0, z ≥ 0 and z w = 0, (1) 10.3390/sym13030503 × where A 2 Rn n and q 2 Rn are given, and z 2 Rn is unknown, and for two s × t matrices Academic Editor: Jan Awrejcewicz G = (gij) and H = (hij) the order G ≥ (>)H means gij ≥ (>)hij for any i and j. -
Trigonometric Transform Splitting Methods for Real Symmetric Toeplitz Systems
View metadata, citation and similar papers at core.ac.uk brought to you by CORE provided by Universidade do Minho: RepositoriUM TRIGONOMETRIC TRANSFORM SPLITTING METHODS FOR REAL SYMMETRIC TOEPLITZ SYSTEMS ZHONGYUN LIU∗, NIANCI WU∗, XIAORONG QIN∗, AND YULIN ZHANGy Abstract. In this paper we study efficient iterative methods for real symmetric Toeplitz systems based on the trigonometric transformation splitting (TTS) of the real symmetric Toeplitz matrix A. Theoretical analyses show that if the generating function f of the n × n Toeplitz matrix A is a real positive even function, then the TTS iterative methods converge to the unique solution of the linear system of equations for sufficient large n. Moreover, we derive an upper bound of the contraction factor of the TTS iteration which is dependent solely on the spectra of the two TTS matrices involved. Different from the CSCS iterative method in [19] in which all operations counts concern complex op- erations when the DFTs are employed, even if the Toeplitz matrix A is real and symmetric, our method only involves real arithmetics when the DCTs and DSTs are used. The numerical experiments show that our method works better than CSCS iterative method and much better than the positive definite and skew- symmetric splitting (PSS) iterative method in [3] and the symmetric Gauss-Seidel (SGS) iterative method. Key words. Sine transform, Cosine transform, Matrix splitting, Iterative methods, Real Toeplitz matrices. AMS subject classifications. 15A23, 65F10, 65F15. 1. Introduction. Consider the iterative solution to the following linear system of equa- tions Ax = b (1.1) by the two-step splitting iteration with alternation, where b 2 Rn and A 2 Rn×n is a symmetric positive definite Toeplitz matrix. -
Numerical Linear Algebra
Numerical Linear Algebra Comparison of Several Numerical Methods used to solve Poisson’s Equation Christiana Mavroyiakoumou University of Oxford A case study report submitted for the degree of M.Sc. in Mathematical Modelling and Scientific Computing Hilary 2017 1 Introduction In this report, we apply the finite difference scheme to the Poisson equation with homogeneous Dirichlet boundary conditions. This yields a system of linear equations with a large sparse system matrix that is a classical test problem for comparing direct and iterative linear solvers. The solution of sparse linear systems by iterative methods has become one of the core applications and research areas of scientific computing. The size of systems that are solved routinely has increased tremendously over time. This is because the discretisation of partial differential equations can lead to systems that are arbitrarily large. We introduce the reader to the general theory of regular splitting methods and present some of the classical methods such as Jacobi, Gauss-Seidel, Relaxed Jacobi, SOR and SSOR. Iterative methods yield the solution U of a linear system after an infinite number of steps. At each step, iterative methods require the computation of the residual of the system. In this report, we use iterative solution methods in order to solve n n n AU “ f;A P R ˆ ; f P R : (1.1) Linear systems can be solved by fixed point iteration. To do so, one must transform the system of linear equations into a fixed point form and this is in general achieved by splitting the matrix A into two parts, A “ M ´ N.