Compiler Tools in APL 

Total Page:16

File Type:pdf, Size:1020Kb

Compiler Tools in APL  Compiler Tools in APL Robert Bernecky Gert Osterburg Snake Island Research Inc Technoma GmbH 18 Fifth Street, Ward’s Island Kurhessenstr. 14 Toronto, Ontario M5J 2B9 Frankfurt/M Canada Germany +1 416 203-0854 email: (IPSHARP) technoma@ipsaint [email protected] Keywords many common programming problems, especially those which are inherently or apparently iterative. Compilers, parsers, finite state automata, FSM, per- formance. 1 Introduction Abstract APL is a powerful interactive language which of- fers primitives to assist in creation of new pro- We present the design and implementation of APL grams. Among these are execute (â), which executes Intrinsic Functions for a Finite State Machine (also APL sentences, and fix (Ìfx), which creates APL known as a Finite State Automaton) which recog- functions from character tables. Given such primi- nizes regular languages, and a Parser which recog- tives, why would an APL programmer want compiler nizes a subset of context free languages, including tools? SLR(1), LALR(1), and LR(1). These are currently Several reasons for wanting additional tools are: being used on a commercial APL mainframe system Many applications allow users to specify infor- as part of a large real-time financial trading system, mation in the form of statements expressed not where they are part of a compiler which translates in APL syntax, but in a form closer to a nat- dealer specification statements into APL functions. ural language, or in a form closer to unnatural Use of these Intrinsic Functions more than doubled languages such as the more common program- the performance of the compiler. ming languages. Those statements must be an- In addition to making a significant performance alyzed for syntactic correctness and semantic improvement for a large production system, we show content according to their rules, which differ these functions to have value in effectively solving from the rules of APL. Such analysis typically This paper originally appeared in the APL92 Conference requires iteration, an area in which APL inter- Proceedings. [BO92] preters have traditionally performed poorly. 1 Many problems which are inherently iterative, PTgenerator A Parser Table generator, or which are not obviously non-iterative, are which takes definitions of a grammar and gram- amenable to easy and fast solutions using com- mar type, and generates a Parser Table. piler tool assists. FSM A Finite State Machine (or Finite State Au- Manually constructed syntax analyzers and tomaton) which takes an FSM table and source parsers are less likely to be fault-free than those text, and produces a list of machine states and constructed with the aid of already debugged partitions, which can be used to extract Tokens compiler tools. Hence, code reliability is im- from the text, and to perform certain classes of proved by the use of extant tools. validation on the text. Although a complete calculus of compiler tools is PARSER A Parsing function, which takes a beyond the scope of this paper, we have designed and Parser Table and Tokens, and produces a Parse implemented a subset of such tools, which have done Tree. an excellent job in practice on a number of applica- tions. Interpreter An interpreter or compiler function, which takes APL code fragments and a Parse Tree, and produces APL programs, or 2 A Calculus of Compiler Tools directly interprets the desired expressions. Syntax-directed program development is an impor- These tools provide a compiler or interpreter gen- tant area of computer science: Considerable re- erating device which is comparable to Lex and Yacc: search has resulted in the availability of a number of compiler-writing aids such as LEX and YACC FSMgenerator and FSM could be useful in [Joh86, LS86]. These tools have turned compiler regular pattern matching. SHARP APL’s LO- writing from an art into a science, making creation GOS, [AGDH86] for example, uses the termi- of simple compilers an undergraduate term project. nology, but has implemented only a very sim- The problems faced by APL application program- ple subset of regular expressions. With these mers are no different from those faced by program- functions, full regular expressions could be ex- mers working in other languages. Having efficient pressed quite efficiently. compiler tools available in APL expands the power of APL in quite unexpected ways, as will be shown FSMgenerator and PTgenerator would below. make up a parser generator. That is, they gen- What sort of tools are required for a calculus of erate Tables which, together with the FSM and compiler tools? Consider the following set of func- PARSER, constitute a proper parser. tions which deal with syntax-directed program de- velopment: Detailed discussion of the entire set of compiler tools is beyond the scope of this paper, so we re- FSMgenerator A Finite State Machine gen- strict discussion to the FSMgenerator, PTgen- erator which takes definitions expressed as reg- erator,andtheFSM. A later section deals with ular expressions, and generates an FSM table. the definitions of these tools. 2 Token Example Regular Expression Regular expressions are thus seen to be valuable nat 123 DIG * both in scanning for tokens in a traditional program- integer 123 -123 (!+ !-) nat ming language, or in the validation of statements in real 1.3 -1.3 (!+ !-) ( ( nat *) !. nat + a custom-designed end-user oriented command lan- nat !. (nat *) ) guage. expreal 1.3e3 real !e integer names abc a123 ALPH (ALPH + DIG) * 4 Compiler Tool Definitions 4.1 The FSM Generator Definition currency USD DM (!U !S !D) + (!D !M) The Finite State Machine (also known as a Finite Figure 1: Regular expressions for data validation State Automaton (FSA) or FSM) generator function is written in ISO standard APL, except for the use Regular Pattern recognized (p,q are of Ìsignal to signal errors during argument val- Expression instances of patterns) idation. Performance of the generator itself is usu- !x the character x ally not critical, because a generator is typically used ! the empty string only once, while its results are used many times. pq pfollowedbyq The duty of FSMgenerator is to take gram- p+q porq mar definitions expressed as regular expressions, and p* zero or more p produce a table which can be used as the left argu- Figure 2: Regular expression definition ment to the FSM Intrinsic Function. In working with typical grammars, the effort involved to create a fi- nite state machine definition by hand can be exces- 3 Regular Expressions sive. The FSMgenerator automates this process. Proving that a finite state machine definition is Regular expressions are commonly used in compiler correct is also difficult. The use of a function to cre- design as a way to compactly describe the legal con- ate these definitions ensures that they are correct if structs of a grammar. If we wish to construct a gram- the grammar is correct. mar which accepts numbers, characters, and curren- cies, we might construct regular expressions to de- 4.2 The PTgenerator Definition fine such a grammar. These allow precise definitions in a general way. See Figure 1 for an example of The PTgenerator function takes as argument a set of regular expressions for defining tokens. grammar rules as shown in the second part of Figure In fact, any pattern which can be derived from the 7. Rules are given by: recursive definition in Figure 2 can be recognized by Nonterminal ! a Finite State Machine. List-of-Terminals-and-Nonterminals As noted earlier, the reliability of grammars based on regular expressions and compiled by a generator Terminals are identified by a leading ’$’ and are are more likely to be correct than those constructed assumed to be defined via regular expressions. After manually in an ad hoc manner. performing some plausibility checks such as: 3 ! does each Nonterminal occuring right of ’ ’ which are not explicitly indicated in this row are as- also occur left of at least one rule? sumed to appear in the last column of the row. The remaining rows of the FSM table are next-state val- does each Nonterminal on the left occur also on ues, selected by the current state of the FSM (initially the right side (except for the first one, which is 1), and the column indicated by the next character in considered to be the starting Nonterminal). the argument. The rows are one-origin indices into the rows of the FSM table excluding the first row (or, The parser starts constructing the parsing tables zero-origin indices into the rows including the first according to the type (slr in Figure 7) of the gram- row). mar. For algorithmic details we refer to Aho, et al Legal values in these rows are ¢1 or legal one- [ASU86]. origin indices into the FSM table after the first row The result of PTgenerator is either the GOTO and has been dropped. Zeros and values greater than Actiontables(asshowninFigure8)oranerrormes- 1ÙÒFSMtable cause domain error. Other positive sage indicating the reason why the construction of values indicate the row of the FSMtable to be used the parsing table failed. as the next FSM state. In the SWAP application discussed later, an Negative one indicates an illegal transition. This SLR(1) grammar is used. A LALR(1) generator also causes the previously generated result element to be exists, but it runs fairly slowly. LALR(1) is also the negated, marking the end of a partition of legal char- technique used by Yacc. acters. The FSM resets to the initial state, and res- cans the offending character from that state. A fail- 4.3 The FSM Definition ure in this state causes negative one to be stored in the current result element, and the machine restarts A Finite State Machine derives its name from the fact on the next character, starting a new partition.
Recommended publications
  • Application and Interpretation
    Programming Languages: Application and Interpretation Shriram Krishnamurthi Brown University Copyright c 2003, Shriram Krishnamurthi This work is licensed under the Creative Commons Attribution-NonCommercial-ShareAlike 3.0 United States License. If you create a derivative work, please include the version information below in your attribution. This book is available free-of-cost from the author’s Web site. This version was generated on 2007-04-26. ii Preface The book is the textbook for the programming languages course at Brown University, which is taken pri- marily by third and fourth year undergraduates and beginning graduate (both MS and PhD) students. It seems very accessible to smart second year students too, and indeed those are some of my most successful students. The book has been used at over a dozen other universities as a primary or secondary text. The book’s material is worth one undergraduate course worth of credit. This book is the fruit of a vision for teaching programming languages by integrating the “two cultures” that have evolved in its pedagogy. One culture is based on interpreters, while the other emphasizes a survey of languages. Each approach has significant advantages but also huge drawbacks. The interpreter method writes programs to learn concepts, and has its heart the fundamental belief that by teaching the computer to execute a concept we more thoroughly learn it ourselves. While this reasoning is internally consistent, it fails to recognize that understanding definitions does not imply we understand consequences of those definitions. For instance, the difference between strict and lazy evaluation, or between static and dynamic scope, is only a few lines of interpreter code, but the consequences of these choices is enormous.
    [Show full text]
  • Formalizing the Expertise of the Assembly Language Programmer
    MASSACHUSETTS INSTITUTE OF TECHNOLOGY ARTIFICIAL INTELLIGENCE LABORATORY Working Paper 203 September 1980 Formalizing tte Expertise of the Assembly Language Programmer Roger DuWayne Duffey II Abstract: A novel compiler strategy for generating high quality code is described. The quality of the code results from reimplementing the program in the target language using knowledge of the program's behavior. The research is a first step towards formalizing the expertise of the assembly language programmer. The ultimate goal is to formalize code generation and implementation techniques it, the same way that parsing and code generation techniques have been formalized. An experimental code generator based on the reimplementation strategy will be constructed. The code generator will provide a framework for analyzing the costs, applicability, and effectiveness of various implementation techniques. Several common code generation problems will be studied. Code written by experienced programmers and code generated by a conventional optimizing compiler will provide standards of comparison. A.I. Laboratory Working Papers are produced for internal circulation, and may contain information that is, for 4.xample, too preliminary or too detailed for formal publication. It is not intended that they should be considered papers to which reference can be made in the literature. 0 MASSACHUSETTS INSTITUTE OF TECHNOLOGY 1980 I. Introduction This research proposes a novel compiler strategy for generating high quality code. The strategy evolved .from studying the differences between the code produced by experienced programmers and the code produced by conventional optimization techniques. The strategy divides compilation into four serial stages: analyzing the source implementation, undoing source implementation decisions, reimplementing the program in the target language, and assembling the object modules.
    [Show full text]
  • System Language: Understanding Systems
    SYSTEM LANGUAGE: UNDERSTANDING SYSTEMS George Mobus1 University of Washington Tacoma, Institute of Technology, 1900 Commerce St. Tacoma WA, 98402 Kevin Anderson University of Washington Tacoma, Institute of Technology, 1900 Commerce St. Tacoma WA, 98402 ABSTRACT Current languages for system modelling impose limitations on how a system is described. For example system dynamics languages (e.g. Stella) assume that the only concern in modelling a system is its dynamics which can be expressed in stocks, flows, and regulators only. A language for describing systems in a general framework provides guidance for the analysis of real systems as well as a way to construct models of those systems suitable for simulation. The language being developed, system language (SL) for lack of a catchier name, consists of: • A set of lexical elements, terms that represent abstractions of components and entities that are found in all dynamic, complex systems to one extent or another - e.g. regulator, process, flow, boundaries, interfaces, etc. • A syntax for constructing the structure of a system including: o describing the boundary and its conditions (including expansion of boundaries as needed) o describing the hierarchical network of connections and relations (e.g. system of systems) o describing interfaces and protocols for entities to exchange flows o describing the behaviour of elements in the system (e.g. functions) o providing specific identifiers naming the abstract lexical elements (e.g. electrical power flow) o providing a set of attributes appropriate to the nature of the element (e.g. voltage, amperage, etc.) • A semantics that establishes patterns of connectivity and behaviour including: o distinction of material, energy, and message (communications) flows o laws of nature to be observed, e.g.
    [Show full text]
  • The Java® Language Specification Java SE 11 Edition
    The Java® Language Specification Java SE 11 Edition James Gosling Bill Joy Guy Steele Gilad Bracha Alex Buckley Daniel Smith 2018-08-21 Specification: JSR-384 Java SE 11 (18.9) ("Specification") Version: 11 Status: Final Release Specification Lead: Oracle America, Inc. ("Specification Lead") Release: September 2018 Copyright © 1997, 2018, Oracle America, Inc. All rights reserved. The Specification provided herein is provided to you only under the Limited License Grant included herein as Appendix A. Please see Appendix A, Limited License Grant. Table of Contents 1 Introduction 1 1.1 Organization of the Specification 2 1.2 Example Programs 6 1.3 Notation 6 1.4 Relationship to Predefined Classes and Interfaces 7 1.5 Feedback 7 1.6 References 7 2 Grammars 9 2.1 Context-Free Grammars 9 2.2 The Lexical Grammar 9 2.3 The Syntactic Grammar 10 2.4 Grammar Notation 10 3 Lexical Structure 15 3.1 Unicode 15 3.2 Lexical Translations 16 3.3 Unicode Escapes 17 3.4 Line Terminators 19 3.5 Input Elements and Tokens 19 3.6 White Space 20 3.7 Comments 21 3.8 Identifiers 22 3.9 Keywords 24 3.10 Literals 25 3.10.1 Integer Literals 25 3.10.2 Floating-Point Literals 32 3.10.3 Boolean Literals 35 3.10.4 Character Literals 35 3.10.5 String Literals 36 3.10.6 Escape Sequences for Character and String Literals 38 3.10.7 The Null Literal 39 3.11 Separators 40 3.12 Operators 40 4 Types, Values, and Variables 41 4.1 The Kinds of Types and Values 41 4.2 Primitive Types and Values 42 4.2.1 Integral Types and Values 43 iii The Java® Language Specification 4.2.2 Integer
    [Show full text]
  • Programming Languages: Principles and Paradigms Prof
    Undergraduate Topics in Computer Science Undergraduate Topics in Computer Science (UTiCS) delivers high-quality instruc- tional content for undergraduates studying in all areas of computing and information science. From core foundational and theoretical material to final-year topics and ap- plications, UTiCS books take fresh, concise, and modern approach and are ideal for self-study or for a one- or two-semester course. The texts are all authored by established experts in their fields, reviewed by an international advisory board, and contain numerous examples and problems. Many include fully worked solutions. For further volumes: http://www.springer.com/series/7592 Maurizio Gabbrielli and Simone Martini Programming Languages: Principles and Paradigms Prof. Dr. Maurizio Gabbrielli Prof. Dr. Simone Martini Università di Bologna Università di Bologna Bologna Bologna Italy Italy Series editor Ian Mackie Advisory board Samson Abramsky, University of Oxford, UK Chris Hankin, Imperial College London, UK Dexter Kozen, Cornell University, USA Andrew Pitts, University of Cambridge, UK Hanne Riis Nielson, Technical University of Denmark, Denmark Steven Skiena, Stony Brook University, USA Iain Stewart, University of Durham, UK David Zhang, The Hong Kong Polytechnic University, Hong Kong ISSN 1863-7310 ISBN 978-1-84882-913-8 e-ISBN 978-1-84882-914-5 DOI 10.1007/978-1-84882-914-5 Springer London Dordrecht Heidelberg New York British Library Cataloguing in Publication Data A catalogue record for this book is available from the British Library Library of Congress Control Number: 2010922995 Translated from the original Italian edition: “Linguaggi di programmazione: principi e paradigmi”, pub- lished in 2006 by McGraw-Hill Companies, Publishing Group Italia.
    [Show full text]
  • Comparative Programming Languages
    Comparative Programming Languages hussein suleman uct csc304s 2003 Course Structure 15 lectures 2 assignments 1 x 2-week programming assignment 1 x 1-week “written” tutorial open-book final (1/2 paper) 1 Course Topics overview of paradigms evolution of languages assignment and expressions types, variables, binding and scope pointers and memory management control structures, subprograms runtime execution exceptions concurrency visual languages scripting languages Overview of Paradigms and PL Issues 2 Why study programming languages? Understand differences and similarities. Understand concepts and structures independently of languages. Learn reasons for design decisions in particular languages. Develop the ability to choose appropriate languages and paradigms for particular tasks. Develop the ability to design languages. Common paradigms paradigm n. pattern, model - Dorling Kindersley Dictionary Procedural a.k.a. Imperative imperative = peremptory, urgent, necessary peremptory = imperious, dictatorial, precluding opposition imperious = dominating, dictatorial Object-oriented Declarative a.k.a. Logic Functional 3 Examples Language Paradigm C C++ Java Clean Prolog Assembly Language Visual C++ HTML C# Javascript (Programming) Languages Languages are used to standardise communication. What makes it a programming language? Is HTML a programming language? Javascript? SQL? 3 traditional concepts: sequence selection iteration 4 Issues in comparing languages Simplicity Orthogonality Control structures Data types and type checking Syntax Abstractions
    [Show full text]
  • Demodularization of a SGLR Parser for Performance Gains
    Undoing Software Engineering: Demodularization of a SGLR Parser for Performance Gains Nik Kapitonenko TU Delft Abstract avoided by removing unnecessary architecture supporting the modularization. This ‘inlining’ may simplify the way com- JSGLR2 is a java implementation of the Scanner- ponents interact on a lower abstraction level, which may im- less Generalized LR-parsing (SGLR) algorithm. It prove parsing throughput. employs a modular architecture. This architec- This paper presents four inlined versions of one the JS- ture comes with a performance overhead for letting GLR2 variants, called “recovery”. A copy of the “recovery multiple components interact with each other. This variant” is created, and refactors it be be less modular. The paper looks into the size of the performance over- four inlined versions share a common base, but have slight head penalty for the recovery parser variant. It does differences in how they inline certain components. Next, us- so by creating an ‘inlined’ version of the recovery ing a dedicated evaluation suite, this paper presents the per- parser variant. The inlined recovery variant is a formance results of the original and the inlined variants. The JSGLR2 implementation that ignores the modular inlined variants are compared to the base case and each other. architecture, and hard-codes the components. The From the results, conclusions are drawn on which inlined performance of the inlined variant is measured with variants are the most successful. a pre-existing evaluation suite. The results show that there is a performance increase between the 1.3 Organization original, and the inlined variant. The paper is organized as follows into several sections.
    [Show full text]
  • The FFP Machine
    The FFP Machine Technical Report 87-014 1987 Gy ula A. Mag6 Donald F. Stanat The University of North Carolina at Chapel Hill Department of Computer Science New West Hall 035 A Chapel Hill . N.C . 27514 (This appe&l'l u a chapter in 'Topic:• ln High-Level Language Computer Arc:hitec:ture,' edited by V. Milutinovic, published by Computer Science Press.) THE FFP MACHINE Gyula Mag6* Donald F. Stanat Department of Computer Science University of North Carolina Chapel Hill, North Carolina ABSTRACT The FFP machine directl11 e:tecutes programs written in a general purpose high level functional programming language. It is a small-grain reduction machine which dy­ namicallJI creates independent submachines of varJ!ing size consisting of conglom­ erations of processor. for each available subcomputation. Buau..e the machine per­ forms problem decomposition and resource alloeation without e:tplicit programmer or software control, it is appropriate for d11namic and irregular computations. We descn"be an implementation that consists of a binarJI tree of onl11 two kinds of pro­ cessors; this implementation iB u:tensible and well-suited to VLSI implementation. I. INTRODUCTION A highly parallel machine architecture has been developed and extensively studied at the University of North Carolina at Chapel Hill [Mago79,MagMi84). The project has as its goal the construction of a machine with the following characteristics: 1. general purpose, 2. well-suited to VLSI technology, 3. extensible, 4. high performance. The FFP machine holds promise for satisfying all these criteria. Detailed simulations have shown the design to be sound, and recent advances in microelectronic fabrication technology and asynchronous design make a successful implementation feasible.
    [Show full text]
  • Twig: a Configurable Domain-Specific Language
    TWIG: A CONFIGURABLE DOMAIN-SPECIFIC LANGUAGE by GEOFFREY C. HULETTE A DISSERTATION Presented to the Department of Computer and Information Science and the Graduate School of the University of Oregon in partial fulfillment of the requirements for the degree of Doctor of Philosophy June 2012 DISSERTATION APPROVAL PAGE Student: Geoffrey C. Hulette Title: Twig: A Configurable Domain-Specific Language This dissertation has been accepted and approved in partial fulfillment of the requirements for the Doctor of Philosophy degree in the Department of Computer and Information Science by: Dr. Allen D. Malony Chair Dr. Michal Young Member Dr. Zena Ariola Member Dr. Shawn Lockery Outside Member and Kimberly Andrews Espy Vice President for Research & Innovation/ Dean of the Graduate School Original approval signatures are on file with the University of Oregon Graduate School. Degree awarded June 2012 ii c 2012 Geoffrey C. Hulette iii DISSERTATION ABSTRACT Geoffrey C. Hulette Doctor of Philosophy Department of Computer and Information Science June 2012 Title: Twig: A Configurable Domain-Specific Language Programmers design, write, and understand programs with a high-level structure in mind. Existing programming languages are not very good at capturing this structure because they must include low-level implementation details. To address this problem we introduce Twig, a programming language that allows for domain-specific logic to be encoded alongside low-level functionality. Twig's language is based on a simple, formal calculus that is amenable to both human and machine reasoning. Users may introduce rules that rewrite expressions, allowing for user-defined optimizations. Twig can also incorporate procedures written in a variety of low-level languages.
    [Show full text]
  • Chapter 16 Programming and Languages
    CCHHAAPPTTEERR 1166 PPRROOGGRRAAMMMMIINNGG AANNDD LLAANNGGUUAAGGEESS Programming and Programmers 3 Programming and Languages Notre Dame University Programming and Programmers Learning Module Objectives When you have • Understood what programmers do and do not do completed this • Knew how programmers define a problem, plan the solution, and the code, learning test, and document the program module you will have: Programming and Programmers 4 Programming and Languages Notre Dame University Understand what is programming, what programmers do and do not do Why You may already have used software, perhaps for word processing or Programming spreadsheets, to solve problems. Perhaps now you are curious to learn how programmers develop software. A program is a set of step-by-step instructions that directs the computer to do the tasks you want it to do and produce the results you want. A set of rules that provides a way of telling a computer what operations to perform is called a programming language. There is not, however, just one programming language; there are many. What In general, the programmer's job is to convert problem solutions into instructions programmers for the computer. That is, the programmer prepares the instructions of a computer do program and runs those instructions on the computer, tests the program to see if it is working properly, and makes corrections to the program. The programmer also writes a report on the program. These activities are all done for the purpose of helping a user fill a need, such as paying employees, billing customers, or admitting students to college. The programming activities just described could be done, perhaps, as solo activities, but a programmer typically interacts with a variety of people.
    [Show full text]
  • The Java® Language Specification Java SE 10 Edition
    The Java® Language Specification Java SE 10 Edition James Gosling Bill Joy Guy Steele Gilad Bracha Alex Buckley Daniel Smith 2018-02-20 Specification: JSR-383 Java SE 10 (18.3) ("Specification") Version: 10 Status: Final Release Release: March 2018 Copyright © 1997, 2018, Oracle America, Inc. 500 Oracle Parkway, Redwood City, California 94065, U.S.A. All rights reserved. The Specification provided herein is provided to you only under the Limited License Grant included herein as Appendix A. Please see Appendix A, Limited License Grant. Table of Contents 1 Introduction 1 1.1 Organization of the Specification 2 1.2 Example Programs 6 1.3 Notation 6 1.4 Relationship to Predefined Classes and Interfaces 7 1.5 Feedback 7 1.6 References 7 2 Grammars 9 2.1 Context-Free Grammars 9 2.2 The Lexical Grammar 9 2.3 The Syntactic Grammar 10 2.4 Grammar Notation 10 3 Lexical Structure 15 3.1 Unicode 15 3.2 Lexical Translations 16 3.3 Unicode Escapes 17 3.4 Line Terminators 19 3.5 Input Elements and Tokens 19 3.6 White Space 20 3.7 Comments 21 3.8 Identifiers 22 3.9 Keywords 24 3.10 Literals 25 3.10.1 Integer Literals 25 3.10.2 Floating-Point Literals 32 3.10.3 Boolean Literals 35 3.10.4 Character Literals 35 3.10.5 String Literals 36 3.10.6 Escape Sequences for Character and String Literals 38 3.10.7 The Null Literal 39 3.11 Separators 40 3.12 Operators 40 4 Types, Values, and Variables 41 4.1 The Kinds of Types and Values 41 4.2 Primitive Types and Values 42 4.2.1 Integral Types and Values 43 iii The Java® Language Specification 4.2.2 Integer Operations
    [Show full text]
  • Modeling of Motion Primitive Architectures Using Domain-Specific Languages
    Modeling of Motion Primitive Architectures using Domain-Specific Languages by Dipl.-Ing. Arne Nordmann Dissertation Faculty of Technology Bielefeld University Bielefeld, August 2015 to the most wonderful family one can wish for to the friends and colleages who joined me on this crazy journey they call doing a phd Printed on permanent paper according to ISO 9706. Abstract Many aspects in robotics, and their omnipresent ideal models, animals and humans, are still not understood or explored well enough, for example producing motions of animal- and human-like complexity. To explore the inner workings of systems studying this complexity, the essential concepts of interest need to be made explicit and raised from the code-level to a higher level of abstraction to be able to reason about them. This work introduces a model-driven engineering approach for complex movement control architectures based on motion primitives, which in recent years have been a central development towards adaptive and flexible control of complex and compliant robots. The goal is to realize rich motor skills through the composition of motion primitives. This thesis proposes a design process to analyze the control architectures of representative example systems to identify common abstractions. Identified and formalized concepts can then be used to automate software development of motion primitive architectures through model-driven engineering methods and domain-specific languages. It turns out that the introduced notion of motion primitives implemented as dynamical systems with machine learning capabilities, provide the computational building block for a large class of such control architectures. Building on the identified concepts, a set of modularized domain-specific languages allows the compact specifica- tion of motion primitive architectures.
    [Show full text]