Fall 2012 FDI Short Courses

Total Page:16

File Type:pdf, Size:1020Kb

Fall 2012 FDI Short Courses Fall 2012 FDI Short Courses Overview This fall, ARC will offer six two-session short courses through the Virginia Tech Faculty Development Institute (FDI). These courses will generally be held each on Monday afternoon from 3:00-4:45 in Torgersen 3060. Course descriptions are below. Parallel MATLAB (with ICAM) 1] Parallel Matlab at VT and PARFOR (Monday, September 10, 3-4:45pm, Torgersen 3060) This short course is the first in a two-part series on parallel programming in Matlab. We present an overview of Matlab Parallel Computing resources at VT and discuss several approaches to parallelism including: 1) PARallel FOR loops, 2) Single Program Multiple Data constructs, and 3) Task computing. We then focus on parfor loops and discuss several examples that illustrate how to identify bottlenecks (candidates for parallelism) and several limitations on the use of parlor. A variety of example codes are available for student use. 2] Parallel Matlab and Single Program Multiple Data (Tuesday, September 11, 3-4:45pm, Torgersen 3060) This short course is the second in a two-part series on parallel programming in Matlab. This course focuses on Single Program Multiple Data (SPMD) constructs. We discuss SPMD workspace(s), the scope of variables and composite arrays. Construction and use of distributed (and codistributed) arrays is covered along with data exchange among the multiple workers. A variety of example codes are available for student use. Scientific Programming with Python 1] Scientific Programming with Python I (Monday, September 17, 3-4:45pm, Torgersen 3060) Python is a high level, object oriented, interpreted programming language that combines the power andflexibility of compiled languages with the ease-of-use of scripting, interpreted languages. Like C++ and Java, but unlike MATLAB, Python provides - in addition to mathematical libraries - a rich selection of libraries, from web and database programming to visualizaton. In the first part of this two-part course, we first discuss the essentials of Python programming, including lists, tuples, dictionaries, strings, sets, file objects, and control flow. Then we give an overview of the NumPy (fast, multidimensional arrays) and MatplotLib (2D and 3D plotting) packages. Throughout the course, we use iPython for interactive program development and MatPlotLib for plotting. • Course Slides 2] Scientific Programming with Python II (Tuesday, September 18, 3- 4:45pm, Torgersen 3060) In the second part of this two-part course, we delve into the NumPy, SciPy and PyMPI packages. First, we discuss NumPy's array objects, n-dimensional arrays, array slicing, indexing, iterators, arithmetic and comparison operations Then we describe ScipPy, including linear system solvers, LU factorization, matrix functions, and special matrices, sparse matrix storage, sparse direct and iterative linear system solvers, and statistical functions. Finally, we show how to use the Message Passing Interface (MPI) from Python. Throughout the course, we use MatPlotLib for plotting and iPython for interactive program development. • Course Slides Parallel Programming with OpenMP and MPI 1] Shared Memory Parallel Programming in OpenMP (Monday, September 24, 3-4:45pm, Torgersen 3060) This short course, the first in a two-part series on parallel programming in a HPC environment, will introduce basics of OpenMP. It will describe and provide examples of parallel programming using OpenMP in a shared memory environment and describe key considerations when choosing and designing an algorithm using each of these approaches. The course is designed for programmers with little or no experience with parallel computing. • Course Slides 2] Parallel Programming in MPI (Monday, October 1, 3-4:45pm, Torgersen 3060) This short course, the second in a two-part series on parallel programming in a HPC environment, will introduce basics of MPI programming for distributed memory systems, as well as hybrid approaches using MPI and OpenMP. The course is designed for programmers with little or no experience with parallel computing. The course will describe key considerations when choosing and designing an algorithm using each of these approaches. • Course Slides Visual Computing 1] Deep Media (Monday, October 8, 3-4:45pm, Torgersen 3060) [Enroll] This class, the first in a two-part series on visual computing, is a high-level exploration of the possibilities for visual communication using a variety of technologies and display venues; through demonstration and discussion, we will consider Deep Media's applicability to meet current research and pedagogical challenges. We will examine the various tools and data formats involved in the publishing pipelines for interactive 3D. Participants will critically examine the features and methods of different publishing approaches and will leave with an understanding of the ecology of tools available to translate and publish their content. We will also dive into the authoring process for 3D and 4D environments. From text editors to open-source/freeware tools to commercial packages, participants will examine how to build basic objects and environments with geometry and appearance types, lighting, animation, and interaction and publish them to the web. 2] High-Performance Visualization (Monday, October 15, 3-4:45pm, Torgersen 3060) This class, the second in a two-part series on visual computing, will present recent research in human perception and cognition that can serve to inform designers as they make choices regarding representations of various information types. This session will introduce two powerful, open-source visualization software packages in use at the cutting edge of science: Visit and Paraview. We will run through tutorials for each package including loading data, mapping multiple variables to visual form and the production of high-resolution imagery and movies. We will also examine how new graphics hardware at VT enables such scaling including the new GPU cluster Athena and the new Visionarium VisCube. Issues and techniques for interactive (and offline) remote rendering and immersive visualization will be demonstrated and discussed. GPU Programming 1] GPU Programming Using CUDA (Monday, October 22, 3-4:45pm, Torgersen 3060) The use of graphics processing units (GPUs) has become an increasingly accessible strategy for accelerating scientific calculations. CUDA is a programming language designed to target Nvidia GPU such as those on Hokiespeed. This session, the first in a two-part series on CUDA programming, provides a basic introduction to the CUDA programming language and the concepts needed to achieve high performance. • Course Slides 2] Hybrid Programming using CUDA, OpenMP and MPI (Monday, October 29, 3-4:45pm, Torgersen 3060) Supercomputing systems such as Hokiespeed utilize heterogeneous nodes, with multiple CPU cores and multiple GPU on each node. In order to fully harness the capabilities of this setup hybrid programming strategies are necessary. In this session, the second in a two-part series on CUDA programming, we address hybrid implementations of CUDA which utilize OpenMP to scale across cores within a node and MPI to scale across nodes. • Course Slides Parallel Programming with Intel Cilk Plus 1] Parallel Programming with Intel Cilk Plus 1 (Monday, November 5, 3- 4:45pm, Torgersen 3060) In the multi-core and vector mathematics era, parallel programming becomes ubiquitous. Intel Cilk Plus is an extension of C/C++ that supports task parallelism and data parallelism. In this course, we describe parallel programming patterns that model essential parallel programming strategies and show how to apply these patterns for writing effective and hardware-agnostic parallel programs using Cilk Plus. In the first part of this two-part course, we first describe structured parallel programming patterns and then discuss the constructs of Cilk Plus for SIMD parallelism (vectorization) and loop parallelization. 2] Parallel Programming with Intel Cilk Plus 2 (Wednesday, November 7, 3- 4:45pm, Torgersen 3060) In the multi-core and vector mathematics era, parallel programming becomes ubiquitous. Intel Cilk Plus is an extension of C/C++ that supports task parallelism and data parallelism. In this course we describe parallel programming patterns that model essential parallel programming strategies and show how to apply these patterns for writing effective and hardware-agnostic parallel programs using Cilk Plus. In the second part of this two-part course, we first describe and illustrate the task parallelism constructs of Cilk Plus, then we discuss how Cilk Plus achieves dynamic load balancing using work-stealing, and then we show how Cilk Plus avoids determinacy races using lock-free reducer hyperobjects. .
Recommended publications
  • Ada-95: a Guide for C and C++ Programmers
    Ada-95: A guide for C and C++ programmers by Simon Johnston 1995 Welcome ... to the Ada guide especially written for C and C++ programmers. Summary I have endeavered to present below a tutorial for C and C++ programmers to show them what Ada can provide and how to set about turning the knowledge and experience they have gained in C/C++ into good Ada programming. This really does expect the reader to be familiar with C/C++, although C only programmers should be able to read it OK if they skip section 3. My thanks to S. Tucker Taft for the mail that started me on this. 1 Contents 1 Ada Basics. 7 1.1 C/C++ types to Ada types. 8 1.1.1 Declaring new types and subtypes. 8 1.1.2 Simple types, Integers and Characters. 9 1.1.3 Strings. {3.6.3} ................................. 10 1.1.4 Floating {3.5.7} and Fixed {3.5.9} point. 10 1.1.5 Enumerations {3.5.1} and Ranges. 11 1.1.6 Arrays {3.6}................................... 13 1.1.7 Records {3.8}. ................................. 15 1.1.8 Access types (pointers) {3.10}.......................... 16 1.1.9 Ada advanced types and tricks. 18 1.1.10 C Unions in Ada, (food for thought). 22 1.2 C/C++ statements to Ada. 23 1.2.1 Compound Statement {5.6} ........................... 24 1.2.2 if Statement {5.3} ................................ 24 1.2.3 switch Statement {5.4} ............................. 25 1.2.4 Ada loops {5.5} ................................. 26 1.2.4.1 while Loop .
    [Show full text]
  • An Introduction to Numpy and Scipy
    An introduction to Numpy and Scipy Table of contents Table of contents ............................................................................................................................ 1 Overview ......................................................................................................................................... 2 Installation ...................................................................................................................................... 2 Other resources .............................................................................................................................. 2 Importing the NumPy module ........................................................................................................ 2 Arrays .............................................................................................................................................. 3 Other ways to create arrays............................................................................................................ 7 Array mathematics .......................................................................................................................... 8 Array iteration ............................................................................................................................... 10 Basic array operations .................................................................................................................. 11 Comparison operators and value testing ....................................................................................
    [Show full text]
  • Worksheet 4. Matrices in Matlab
    MS6021 Scientific Computation Worksheet 4 Worksheet 4. Matrices in Matlab Creating matrices in Matlab Matlab has a number of functions for generating elementary and common matri- ces. zeros Array of zeros ones Array of ones eye Identity matrix repmat Replicate and tile array blkdiag Creates block diagonal array rand Uniformly distributed randn Normally distributed random number linspace Linearly spaced vector logspace Logarithmically spaced vector meshgrid X and Y arrays for 3D plots : Regularly spaced vector : Array slicing If given a single argument they construct square matrices. octave:1> eye(4) ans = 1 0 0 0 0 1 0 0 0 0 1 0 0 0 0 1 If given two entries n and m they construct an n × m matrix. octave:2> rand(2,3) ans = 0.42647 0.81781 0.74878 0.69710 0.42857 0.24610 It is also possible to construct a matrix the same size as an existing matrix. octave:3> x=ones(4,5) x = 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 William Lee [email protected] 1 MS6021 Scientific Computation Worksheet 4 octave:4> y=zeros(size(x)) y = 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 • Construct a 4 by 4 matrix whose elements are random numbers evenly dis- tributed between 1 and 2. • Construct a 3 by 3 matrix whose off diagonal elements are 3 and whose diagonal elements are 2. 1001 • Construct the matrix 0101 0011 The size function returns the dimensions of a matrix, while length returns the largest of the dimensions (handy for vectors).
    [Show full text]
  • Slicing (Draft)
    Handling Parallelism in a Concurrency Model Mischael Schill, Sebastian Nanz, and Bertrand Meyer ETH Zurich, Switzerland [email protected] Abstract. Programming models for concurrency are optimized for deal- ing with nondeterminism, for example to handle asynchronously arriving events. To shield the developer from data race errors effectively, such models may prevent shared access to data altogether. However, this re- striction also makes them unsuitable for applications that require data parallelism. We present a library-based approach for permitting parallel access to arrays while preserving the safety guarantees of the original model. When applied to SCOOP, an object-oriented concurrency model, the approach exhibits a negligible performance overhead compared to or- dinary threaded implementations of two parallel benchmark programs. 1 Introduction Writing a multithreaded program can have a variety of very different motiva- tions [1]. Oftentimes, multithreading is a functional requirement: it enables ap- plications to remain responsive to input, for example when using a graphical user interface. Furthermore, it is also an effective program structuring technique that makes it possible to handle nondeterministic events in a modular way; develop- ers take advantage of this fact when designing reactive and event-based systems. In all these cases, multithreading is said to provide concurrency. In contrast to this, the multicore revolution has accentuated the use of multithreading for im- proving performance when executing programs on a multicore machine. In this case, multithreading is said to provide parallelism. Programming models for multithreaded programming generally support ei- ther concurrency or parallelism. For example, the Actor model [2] or Simple Con- current Object-Oriented Programming (SCOOP) [3,4] are typical concurrency models: they are optimized for coordination and event handling, and provide safety guarantees such as absence of data races.
    [Show full text]
  • Declaring Matrices in Python
    Declaring Matrices In Python Idiopathic Seamus regrinds her spoom so discreetly that Saul trauchled very ashamedly. Is Elvis cashesepigynous his whenpanel Husainyesternight. paper unwholesomely? Weber is slothfully terebinthine after disguisable Milo Return the array, writing about the term empty functions that can be exploring data in python matrices Returns an error message if a jitted function to declare an important advantages and! We declared within a python matrices as a line to declare an array! Transpose does this case there is. Today act this Python Array Tutorial we sure learn about arrays in Python Programming Here someone will get how Python array import module and how fly we. Matrices in Python programming Foundation Course and laid the basics to do this waterfall can initialize weights. Two-dimensional lists arrays Learn Python 3 Snakify. Asking for help, clarification, or responding to other answers. How arrogant I create 3x3 matrices Stack Overflow. What is declared a comparison operators. The same part back the array. Another Python module called array defines one-dimensional arrays so don't. By default, the elements of the bend may be leaving at all. It does not an annual step with use arrays because they there to be declared while lists don't because clothes are never of Python's syntax so lists are. How to wake a 3D NumPy array in Python Kite. Even if trigger already used Array slicing and indexing before, you may find something to evoke in this tutorial article. MATLAB Arrays as Python Variables MATLAB & Simulink. The easy way you declare array types is to subscript an elementary type according to the toil of dimensions.
    [Show full text]
  • Efficient Compilation of High Level Python Numerical Programs With
    Efficient Compilation of High Level Python Numerical Programs with Pythran Serge Guelton Pierrick Brunet Mehdi Amini Tel´ ecom´ Bretagne INRIA/MOAIS [email protected] [email protected] Abstract Note that due to dynamic typing, this function can take The Python language [5] has a rich ecosystem that now arrays of different shapes and types as input. provides a full toolkit to carry out scientific experiments, def r o s e n ( x ) : from core scientific routines with the Numpy package[3, 4], t 0 = 100 ∗ ( x [ 1 : ] − x [: −1] ∗∗ 2) ∗∗ 2 to scientific packages with Scipy, plotting facilities with the t 1 = (1 − x [ : − 1 ] ) ∗∗ 2 Matplotlib package, enhanced terminal and notebooks with return numpy.sum(t0 + t1) IPython. As a consequence, there has been a move from Listing 1: High-level implementation of the Rosenbrock historical languages like Fortran to Python, as showcased by function in Numpy. the success of the Scipy conference. As Python based scientific tools get widely used, the question of High performance Computing naturally arises, 1.3 Temporaries Elimination and it is the focus of many recent research. Indeed, although In Numpy, any point-to-point array operation allocates a new there is a great gain in productivity when using these tools, array that holds the computation result. This behavior is con- there is also a performance gap that needs to be filled. sistent with many Python standard module, but it is a very This extended abstract focuses on compilation techniques inefficient design choice, as it keeps on polluting the cache that are relevant for the optimization of high-level numerical with potentially large fresh storage and adds extra alloca- kernels written in Python using the Numpy package, illus- tion/deallocation operations, that have a very bad caching ef- trated on a simple kernel.
    [Show full text]
  • An Analysis of the D Programming Language Sumanth Yenduri University of Mississippi- Long Beach
    View metadata, citation and similar papers at core.ac.uk brought to you by CORE provided by CSUSB ScholarWorks Journal of International Technology and Information Management Volume 16 | Issue 3 Article 7 2007 An Analysis of the D Programming Language Sumanth Yenduri University of Mississippi- Long Beach Louise Perkins University of Southern Mississippi- Long Beach Md. Sarder University of Southern Mississippi- Long Beach Follow this and additional works at: http://scholarworks.lib.csusb.edu/jitim Part of the Business Intelligence Commons, E-Commerce Commons, Management Information Systems Commons, Management Sciences and Quantitative Methods Commons, Operational Research Commons, and the Technology and Innovation Commons Recommended Citation Yenduri, Sumanth; Perkins, Louise; and Sarder, Md. (2007) "An Analysis of the D Programming Language," Journal of International Technology and Information Management: Vol. 16: Iss. 3, Article 7. Available at: http://scholarworks.lib.csusb.edu/jitim/vol16/iss3/7 This Article is brought to you for free and open access by CSUSB ScholarWorks. It has been accepted for inclusion in Journal of International Technology and Information Management by an authorized administrator of CSUSB ScholarWorks. For more information, please contact [email protected]. Analysis of Programming Language D Journal of International Technology and Information Management An Analysis of the D Programming Language Sumanth Yenduri Louise Perkins Md. Sarder University of Southern Mississippi - Long Beach ABSTRACT The C language and its derivatives have been some of the dominant higher-level languages used, and the maturity has stemmed several newer languages that, while still relatively young, possess the strength of decades of trials and experimentation with programming concepts.
    [Show full text]
  • Performance Analyses and Code Transformations for MATLAB Applications Patryk Kiepas
    Performance analyses and code transformations for MATLAB applications Patryk Kiepas To cite this version: Patryk Kiepas. Performance analyses and code transformations for MATLAB applications. Computa- tion and Language [cs.CL]. Université Paris sciences et lettres, 2019. English. NNT : 2019PSLEM063. tel-02516727 HAL Id: tel-02516727 https://pastel.archives-ouvertes.fr/tel-02516727 Submitted on 24 Mar 2020 HAL is a multi-disciplinary open access L’archive ouverte pluridisciplinaire HAL, est archive for the deposit and dissemination of sci- destinée au dépôt et à la diffusion de documents entific research documents, whether they are pub- scientifiques de niveau recherche, publiés ou non, lished or not. The documents may come from émanant des établissements d’enseignement et de teaching and research institutions in France or recherche français ou étrangers, des laboratoires abroad, or from public or private research centers. publics ou privés. Préparée à MINES ParisTech Analyses de performances et transformations de code pour les applications MATLAB Performance analyses and code transformations for MATLAB applications Soutenue par Composition du jury : Patryk KIEPAS Christine EISENBEIS Le 19 decembre 2019 Directrice de recherche, Inria / Paris 11 Présidente du jury João Manuel Paiva CARDOSO Professeur, University of Porto Rapporteur Ecole doctorale n° 621 Erven ROHOU Ingénierie des Systèmes, Directeur de recherche, Inria Rennes Rapporteur Matériaux, Mécanique, Michel BARRETEAU Ingénieur de recherche, THALES Examinateur Énergétique Francois GIERSCH Ingénieur de recherche, THALES Invité Spécialité Claude TADONKI Informatique temps-réel, Chargé de recherche, MINES ParisTech Directeur de thèse robotique et automatique Corinne ANCOURT Maître de recherche, MINES ParisTech Co-directrice de thèse Jarosław KOŹLAK Professeur, AGH UST Co-directeur de thèse 2 Abstract MATLAB is an interactive computing environment with an easy programming language and a vast library of built-in functions.
    [Show full text]
  • Safe, Contract-Based Parallel Programming
    Presentation cover page EU Safe, Contract-Based Parallel Programming Ada-Europe 2017 June 2017 Vienna, Austria Tucker Taft AdaCore Inc www.adacore.com Outline • Vocabulary for Parallel Programming • Approaches to Supporting Parallel Programming – With examples from Ada 202X, Rust, ParaSail, Go, etc. • Fostering a Parallel Programming Mindset • Enforcing Safety in Parallel Programming • Additional Kinds of Contracts for Parallel Programming Parallel Lang Support 2 Vocabulary • Concurrency vs. Parallelism • Program, Processor, Process • Thread, Task, Job • Strand, Picothread, Tasklet, Lightweight Task, Work Item (cf. Workqueue Algorithms) • Server, Worker, Executor, Execution Agent, Kernel/OS Thread, Virtual CPU Parallel Lang Support 3 Vocabulary What is Concurrency vs. Parallelism? Concurrency Parallelism • “concurrent” • “parallel” programming programming constructs constructs allow allow programmer to … programmer to … Parallel Lang Support 4 Concurrency vs. Parallelism Concurrency Parallelism • “concurrent” • “parallel” programming programming constructs constructs allow allow programmer to programmer to divide and simplify the program by conquer a problem, using using multiple logical multiple threads to work in threads of control to reflect parallel on independent the natural concurrency in parts of the problem to the problem domain achieve a net speedup • heavier weight constructs • constructs need to be light OK weight both syntactically and at run-time Parallel Lang Support 5 Threads, Picothreads, Tasks, Tasklets, etc. • No uniformity
    [Show full text]
  • On the Performance of the Python Programming Language for Serial and Parallel Scientific Computations
    ISSN (Online) 2278-1021 ISSN (Print) 2319-5940 IJARCCE International Journal of Advanced Research in Computer and Communication Engineering ICACTRP 2017 International Conference on Advances in Computational Techniques and Research Practices Noida Institute of Engineering & Technology, Greater Noida Vol. 6, Special Issue 2, February 2017 On the Performance of the Python Programming Language for Serial and Parallel Scientific Computations Samriddha Prajapati1, Chitvan Gupta2 Student, CSE, NIET, Gr. NOIDA, India1 Assistant Professor, NIET, Gr. Noida, India 2 Abstract: In this paper the performance of scientific applications are discussed by using Python programming language. Firstly certain techniques and strategies are explained to improve the computational efficiency of serial Python codes. Then the basic programming techniques in Python for parallelizing scientific applications have been discussed. It is shown that an efficient implementation of array-related operations is essential for achieving better parallel [11] performance, as for the serial case. High performance can be achieved in serial and parallel computation by using a mixed language programming in array-related operations [11]. This has been proved by a collection of numerical experiments. Python [13] is also shown to be well suited for writing high-level parallel programs in less number of lines of codes. Keywords: Numpy, Pypar, Scipy, F2py. I BACKGROUND AND INTRODUCTION Earlier there was a strong tradition among computational vii. Documentation and support scientists to use compiled languages, in particular Fortran 77 and C, for numerical [2] simulation. As there is Many scientists normally feel more productive in Matlab increase in demand for software flexibility during the last than with compiled languages and separate visualization decade, it has also popularized more advanced compiled tools.
    [Show full text]
  • Making Legacy Fortran Code Type Safe Through Automated Program Transformation
    The Journal of Supercomputing https://doi.org/10.1007/s11227-021-03839-9 Making legacy Fortran code type safe through automated program transformation Wim Vanderbauwhede1 Accepted: 21 April 2021 © The Author(s) 2021 Abstract Fortran is still widely used in scientifc computing, and a very large corpus of legacy as well as new code is written in FORTRAN 77. In general this code is not type safe, so that incorrect programs can compile without errors. In this paper, we pre- sent a formal approach to ensure type safety of legacy Fortran code through auto- mated program transformation. The objective of this work is to reduce programming errors by guaranteeing type safety. We present the frst rigorous analysis of the type safety of FORTRAN 77 and the novel program transformation and type checking algorithms required to convert FORTRAN 77 subroutines and functions into pure, side-efect free subroutines and functions in Fortran 90. We have implemented these algorithms in a source-to-source compiler which type checks and automati- cally transforms the legacy code. We show that the resulting code is type safe and that the pure, side-efect free and referentially transparent subroutines can readily be ofoaded to accelerators. Keywords Fortran · Type safety · Type system · Program transformation · Acceleration 1 Introduction 1.1 The enduring appeal of Fortran The Fortran programming language has a long history. It was originally proposed by John Backus in 1957 for the purpose of facilitating scientifc programming, and has since become widely adopted amongst scientists, and been shown to be an efective language for use in supercomputing.
    [Show full text]
  • Parametric Fortran – User Manual
    Parametric Fortran – User Manual Ben Pflaum and Zhe Fu School of EECS Oregon State University September 14, 2006 1 Introduction Parametric Fortran adds elements of generic programming to Fortran through a concept of program templates. Such a template looks much like an ordinary Fortran program, except that it can be enriched by parameters to represent the varying aspects of data and control structures. Any part of a Fortran program can be parameterized, including statements, expressions, declarations, and subroutines. The Parametric Fortran compiler takes values for all the parameters used in a template and translates the program template into a plain Fortran program. Therefore, algorithms can be ex- pressed in Parametric Fortran generically by employing parameters to abstract from details of particular implementations. Different instances can then be created automatically by the Paramet- ric Fortran compiler when given the appropriate parameter values. Parametric Fortran is, in fact, not simply a language extension, but more like a framework to generate customized Fortran extensions for different situations, because the way parameter values affect the code generation is not fixed, but is part of the definition of a particular parameter definition. 1.1 Array addition The following example shows how to write a Parametric Fortran subroutine to add two arrays of arbitrary dimensions. An experienced Fortran programmer would see that this could easily be accomplished Fortran 90 array syntax. The purpose of this example is not to provide a better way to do array addition, but to show how to use parameter values to generate Fortran programs without the details of the program getting in the way.
    [Show full text]