How to Be Correct, Lazy and Efficient ?

Total Page:16

File Type:pdf, Size:1020Kb

How to Be Correct, Lazy and Efficient ? How to be correct, lazy and efficient ? C. Recanati Université Paris 13, av. JB. Clément, 1. Lambdix and the semantic problems 93430, Villetaneuse, of Lisp France This paper is an introduction to In this section the semantical defects of Lambdix, a lazy Lisp interpreter lisp1 are reviewed and shown not to exist implemented at the Research Laboratory in Lambdix. These defects are of the University of Paris XI (Laboratoire characteristic of early versions of lisp, but de Recherche en Informatique, Orsay). Lambdix was devised in the course of an they are still present in current versions investigation into the relationship between (although, of course, not all defects are the semantics of programming languages present in all versions); this is why we and their implementation; it was used to demonstrate that in the Lisp domain, think this little overview is not out of date semantic correctness is consistent with even if some of points we make are now efficiency, contrary to what has often been well-known. claimed. The first part of the paper is an overview of well-known semantic 1.1. The functional argument problem difficulties encountered by Lisp as well as an informal presentation of Lambdix; it is shown that the difficulties which Lisp Lisp was first thought of as being an encouters do not arise in Lambdix. The implementation of the lambda calculus. Its second part is about efficiency in syntax allows the definition of functions implementation models. It explains why Lambdix is better suited for lazy by means of lambda expressions, as for evaluation than previous models. The instance section ends by giving comparative execution time tables. (lambda (x y) (+ x y)) 1 By 'lisp' here we mean a family of languages rather than a particular language belonging to this family. page 1 for addition. This is a fundamental feature when the function returns the lambda of Lisp. Now serious problems arise when expression; consequently, when the these lambda expressions are used as anonymous lambda function is applied, x functional arguments - when a function recovers its previous value - which is 456 takes another function as argument, and or not defined. also when a function returns such a function as value. Example 2 $ (define apply (f x) (f x) ) Example 1 $ ( define Identity ( x ) $ ( define BuildConstFunc (x) ( apply (lambda (y) x) 2)) ( lambda(y) x )) The Identity function defined here The function BuildConstFunc is is constructed by applying a function supposed to return a function lambda of y, returning x. In Lambdix, a call to ( which returns x. This lambda function Identity 45) returns 45 but in lisp: ought to be a constant function since its argument y is not used. Then a call to ( $ ( Identity 45) (BuildConstFunc 0) 1) => 2 ought to return 0, as well as This is even more surprising since only y ( (BuildConsFunc 0) 2). This is of course seems bound to 2. The source of the the case in Lambdix, but not in Lisp trouble is the use of the name x in the where: definition of apply. Had apply been $ ((BuildConsFunc 0) 1) defined as => ** error - x not defined ** or $ (setq x 456) $ ( define apply (f z) $ ((BuildConsFunc 0) 1) (f z) ) => 456 the problem would have disappeared. Here The reason for this surprising it is not because the stack has been popped answer is the following. In most lisps too early that the binding fails, but environments (bindings between names because an intermediate binding has been and values) are represented in a stack inserted: x is first bound to 45; then the which is popped when the execution of the evaluation of apply inserts the binding of function terminates. The binding of x to 0 x to 2. The lambda is evaluated with y in the application of BuildConsFunc is lost page 2 bound to 2 and it returns x - which is now part of the benefits of functional bound to 2. programming is lost. Because of these functional argument problems, Lisp is not a true 1.2. Lexical scope vs dynamic scope second order functional language1, and The foregoing examples show that the names of formal parameters are of major 1 Some lisps proposed a function to solve importance in Lisp. This feature is due to the functional argument problem. This the type of variable binding used in Lisp, function, sometimes called closure, takes called 'most recent binding' (MRB, for two arguments - a list of formal short). MRB was perhaps, as Gowan has parameters and an expression - and returns wittily said, the 'most recent error' the application of a lambda expression. ([GOW72]). Originally motivated by the For instance, the value returned by the technical advantages of the closure function applied to the list '(x) implementation, dynamic scoping of the and an expression Expr - in an type illustrated by MRB has disastrous environment where x is bound to 2 - is consequences for the safety of the given by : language. How can you trust a language $ ( closure '(x) 'Expr ) when x =2 in which the names of formal parameters => ((lambda (x) Expr) '2 ) must be taken into account? The use of dynamic scoping in Although this can punctually solve the Lisp conflicts with the original model problem, many critics can be made to this introduced by Church, which was at the solution. First, it is not very efficient source of Lisp. In lambda calculus, a rule because it adds function calls and requires known as the ß-rule allows the reduction the evaluation of all the parameters to be of terms along the following pattern: saved. Moreover, it is an ad hoc solution. It requires the programmer to know when ( ( λ x . expr ) val ) the closure function is necessary since → expr [ x ← val ] the call to the closure function must be explicit. Furthermore, lisp distinguishing between an f-value and a c-value, some particular attention must be given on the way the interpreter pass the arguments. value interpretation and this additional call This requires the use of a special function must also be explicit. (called funcall ) to force the functional page 3 The ß-rule takes as input the application to allows the use of functional arguments and a value val of a lambda expression with is therefore a true second order language. formal parameter x and yields as output the same expression in which all tokens of 1.3. Evaluation order x have been replaced by val. The substitution is supposed to be determined 1.3.1. Call by need and call by value by the inner lambda binding in the (program) text, i.e. lexically. In lambda calculus, a term t defined by In most cases, dynamic scope and lexical scope yield the same result, but it (λ x y . x) A ((λ u . u u) (λ u . u u)) is very easy to construct cases with reduces to A because the first reduction diverging answers1 . This is why - like (the substitution of A to x) yields a term, λ many functional languages today y . A, which always reduces to A because (Common Lisp, Scheme, ML, etc.) - y does not occur in the core of this lambda Lambdix consistently uses lexical scope expression. But in Lisp the interpreter for formal parameters. Lambdix also computes the values of the arguments before the core of the function. This strategy of parameter evaluation is known 1 For instance the term t defined by as call by value. Since in the evaluation of t ≡ ( λ x . ( λ y . ( λ x . y) B ) x) A t all arguments are calculated first, the would reduced to A by ß-reduction calculation of the second argument - the t → ( λ y . ( λ x . y) B ) A) term (( λ u . u u) ( λ u . u u)) - generates → ( λ x . A) B ) an infinite loop: → A y ≡ (( λ u . u u) ( λ u . u u)) while it would reduced to B in the → (( λ u . u u) ( λ u . u u)) dynamic model: → (( λ u . u u) ( λ u . u u)) ( λ x . ( λ y . ( λ x . y) B ) x) A: → ... Stack ( λ y . ( λ x . y) B ) x) [ x – A ] In the framework of the lambda ( λ x . y) B ) [ x – A ][ y – x ] calculus, call by value can be understood y [ x – A ][ y – x ][ x – B ] as a strategy governing the order in which B the ß-reductions are performed. This A formal demonstration of the non strategy guarantees that if there is a unique equivalence of the two models can be solution, the calculus will converge on it. found in A. Eick and E. Fehr [EIC85]. Unfortunately it also guarantees that the page 4 calculus will never return if there is also arguments and if the latter were calculated an infinite derivation. Thus for terms only when necessary. Such a strategy of having both a finite and an infinite evaluation is known as call by need. For derivation, this strategy guarantees that the instance, the function f defined by finite derivation will not be found. Hence it is not a winning strategy. The winning ( de f (x y) strategy of the lambda-calculus requires (if (< x 0) that the leftmost inner term be reduced 1 first. If there is a finite solution, the (f (- x 1) (f x y)))) calculus will converge on it (the will never terminate from a call to (f 1 2) confluence property guarantees the unicity if the interpreter conforms to the call by of the finite solution). This strategy value strategy, but it returns 1 if the corresponds to call by need. interpreter conforms go the other strategy Functions in Lisp are strict1, (call by need ).
Recommended publications
  • Top-K Entity Resolution with Adaptive Locality-Sensitive Hashing
    Top-K Entity Resolution with Adaptive Locality-Sensitive Hashing Vasilis Verroios Hector Garcia-Molina Stanford University Stanford University [email protected] [email protected] ABSTRACT Given a set of records, entity resolution algorithms find all the records referring to each entity. In this paper, we study Entity 1 the problem of top-k entity resolution: finding all the records Entity 2 referring to the k largest (in terms of records) entities. Top-k Top-k Entity entity resolution is driven by many modern applications that Dataset Filtering Entities’ Records Resolution operate over just the few most popular entities in a dataset. We propose a novel approach, based on locality-sensitive hashing, that can very rapidly and accurately process mas- Entity k sive datasets. Our key insight is to adaptively decide how much processing each record requires to ascertain if it refers to a top-k entity or not: the less likely a record is to refer to Figure 1: Filtering stage in the overall workflow. a top-k entity, the less it is processed. The heavily reduced rest. As we will see, this fact makes it easier to find the top- amount of processing for the vast majority of records that k entities without computing all entities. We next illustrate do not refer to top-k entities, leads to significant speedups. two additional applications where one typically only needs Our experiments with images, web articles, and scientific top-k entities. Incidentally, two of the datasets we use for publications show a 2x to 25x speedup compared to the tra- our experiments are from these applications.
    [Show full text]
  • Solving the Funarg Problem with Static Types IFL ’21, September 01–03, 2021, Online
    Solving the Funarg Problem with Static Types Caleb Helbling Fırat Aksoy [email protected][email protected] Purdue University Istanbul University West Lafayette, Indiana, USA Istanbul, Turkey ABSTRACT downwards. The upwards funarg problem refers to the manage- The difficulty associated with storing closures in a stack-based en- ment of the stack when a function returns another function (that vironment is known as the funarg problem. The funarg problem may potentially have a non-empty closure), whereas the down- was first identified with the development of Lisp in the 1970s and wards funarg problem refers to the management of the stack when hasn’t received much attention since then. The modern solution a function is passed to another function (that may also have a non- taken by most languages is to allocate closures on the heap, or to empty closure). apply static analysis to determine when closures can be stack allo- Solving the upwards funarg problem is considered to be much cated. This is not a problem for most computing systems as there is more difficult than the downwards funarg problem. The downwards an abundance of memory. However, embedded systems often have funarg problem is easily solved by copying the closure’s lexical limited memory resources where heap allocation may cause mem- scope (captured variables) to the top of the program’s stack. For ory fragmentation. We present a simple extension to the prenex the upwards funarg problem, an issue arises with deallocation. A fragment of System F that allows closures to be stack-allocated. returning function may capture local variables, making the size of We demonstrate a concrete implementation of this system in the the closure dynamic.
    [Show full text]
  • Java: Exercises in Style
    Java: Exercises in Style Marco Faella June 1, 2018 i ii This work is licensed under the Creative Commons Attribution- NonCommercial 4.0 International License. To view a copy of this li- cense, visit http://creativecommons.org/licenses/by-nc/4.0/ or send a letter to Creative Commons, PO Box 1866, Mountain View, CA 94042, USA. Author: Marco Faella, [email protected] Preface to the Online Draft This draft is a partial preview of what may eventually become a book. The idea is to share it with the public and obtain feedback to improve its quality. So, if you read the whole draft, or maybe just a chapter, or you just glance over it long enough to spot a typo, let me know! I’m interested in all comments and suggestions for improvement, including (natural) language issues, programming, book structure, etc. I would like to thank all the students who have attended my Programming Languages class, during the past 11 years. They taught me that there is no limit to the number of ways you may think to have solved a Java assignment. Naples, Italy Marco Faella January 2018 iii Contents Contents iv I Preliminaries 1 1 Introduction 3 1.1 Software Qualities . 5 2 Problem Statement 9 2.1 Data Model and Representations . 10 3 Hello World! 13 4 Reference Implementation 17 4.1 Space Complexity . 21 4.2 Time Complexity . 23 II Software Qualities 29 5 Need for Speed 31 5.1 Optimizing “getAmount” and “addWater” . 32 5.2 Optimizing “connectTo” and “addWater” . 34 5.3 The Best Balance: Union-Find Algorithms .
    [Show full text]
  • Graph Reduction Without Pointers
    Graph Reduction Without Pointers TR89-045 December, 1989 William Daniel Partain The University of North Carolina at Chapel Hill Department of Computer Science ! I CB#3175, Sitterson Hall Chapel Hill, NC 27599-3175 UNC is an Equal Opportunity/Aflirmative Action Institution. Graph Reduction Without Pointers by William Daniel Partain A dissertation submitted to the faculty of the University of North Carolina at Chapel Hill in partial fulfillment of the requirements for the degree of Doctor of Philosophy in the Department of Computer Science. Chapel Hill, 1989 Approved by: Jfn F. Prins, reader ~ ~<---(­ CJ)~ ~ ;=tfJ\ Donald F. Stanat, reader @1989 William D. Partain ALL RIGHTS RESERVED II WILLIAM DANIEL PARTAIN. Graph Reduction Without Pointers (Under the direction of Gyula A. Mag6.) Abstract Graph reduction is one way to overcome the exponential space blow-ups that simple normal-order evaluation of the lambda-calculus is likely to suf­ fer. The lambda-calculus underlies lazy functional programming languages, which offer hope for improved programmer productivity based on stronger mathematical underpinnings. Because functional languages seem well-suited to highly-parallel machine implementations, graph reduction is often chosen as the basis for these machines' designs. Inherent to graph reduction is a commonly-accessible store holding nodes referenced through "pointers," unique global identifiers; graph operations cannot guarantee that nodes directly connected in the graph will be in nearby store locations. This absence of locality is inimical to parallel computers, which prefer isolated pieces of hardware working on self-contained parts of a program. In this dissertation, I develop an alternate reduction system using "sus­ pensions" (delayed substitutions), with terms represented as trees and vari­ ables by their binding indices (de Bruijn numbers).
    [Show full text]
  • Tamperproofing a Software Watermark by Encoding Constants
    Tamperproofing a Software Watermark by Encoding Constants by Yong He A Thesis Submitted in Partial Fulfilment of the Requirement for the Degree of MASTER OF SCIENCE in Computer Science University of Auckland Copyright © 2002 by Yong He i Abstract Software Watermarking is widely used for software ownership authentication, but it is susceptible to various de-watermarking attacks such as obfuscation. Dynamic Graph Watermarking is a relatively new technology for software watermarking, and is believed the most likely to withstand attacks, which are trying to distort watermark structure. In this thesis, we present a new technology for protecting Dynamic Graph Watermarks. This technology encodes some of the constants, which are found in a software program, into a tree structure that is similar to the watermark, and generates decoding functions to retrieve the value of the constants at program execution time. If the constant tree is modified, the value of some constants will be affected, destroying program correctness. An attacker cannot reliably distinguish the watermark tree from the constant tree, so they must preserve the watermark tree or risk introducing bugs into the program. Constant Encoding technology can be included in Dynamic Graph Watermarking systems as a plug-in module to improve the Dynamic Graph Watermark protection. In this thesis, we present a prototyping program for Constant Encoding technology, which we call the JSafeMark encoder. Besides addressing the issues about Constant Encoding technology, we also discuss the design and implementation of our JSafeMark encoder, and give a practical example to show how this technology can protect Dynamic Graph Watermarking. iii Acknowledgement I would like to thank Professor Clark Thomborson, my supervisor, for his guidance and enthusiasm.
    [Show full text]
  • N4397: a Low-Level API for Stackful Coroutines
    Document number: N4397 Date: 2015-04-09 Project: WG21, SG1 Author: Oliver Kowalke ([email protected]) A low-level API for stackful coroutines Abstract........................................................1 Background.....................................................2 Definitions......................................................2 Execution context.............................................2 Activation record.............................................2 Stack frame.................................................2 Heap-allocated activation record....................................2 Processor stack..............................................2 Application stack.............................................3 Thread stack................................................3 Side stack..................................................3 Linear stack................................................3 Linked stack................................................3 Non-contiguous stack..........................................3 Parent pointer tree............................................3 Coroutine.................................................3 Toplevel context function.........................................4 Asymmetric coroutine..........................................4 Symmetric coroutine...........................................4 Fiber/user-mode thread.........................................4 Resumable function............................................4 Suspend by return.............................................4 Suspend
    [Show full text]
  • Dynamic Graph-Based Software Watermarking
    Dynamic Graph-Based Software Watermarking † ‡ Christian Collberg∗, Clark Thomborson, Gregg M. Townsend Technical Report TR04-08 April 28, 2004 Abstract Watermarking embeds a secret message into a cover message. In media watermarking the secret is usually a copyright notice and the cover a digital image. Watermarking an object discourages intellectual property theft, or when such theft has occurred, allows us to prove ownership. The Software Watermarking problem can be described as follows. Embed a structure W into a program P such that: W can be reliably located and extracted from P even after P has been subjected to code transformations such as translation, optimization and obfuscation; W is stealthy; W has a high data rate; embedding W into P does not adversely affect the performance of P; and W has a mathematical property that allows us to argue that its presence in P is the result of deliberate actions. In this paper we describe a software watermarking technique in which a dynamic graph watermark is stored in the execution state of a program. Because of the hardness of pointer alias analysis such watermarks are difficult to attack automatically. 1 Introduction Steganography is the art of hiding a secret message inside a host (or cover) message. The purpose is to allow two parties to communicate surreptitiously, without raising suspicion from an eavesdropper. Thus, steganography and cryptog- raphy are complementary techniques: cryptography attempts to hide the contents of a message while steganography attempts to hide the existence of a message. History provides many examples of steganographic techniques used in the military and intelligence communities.
    [Show full text]
  • The Transmutation Chains Period Folding on the Use of the Parent Pointer Tree Data Structure
    The transmutation chains period folding on the use of the parent pointer tree data structure Przemysław Stanisz, Jerzy Cetnar Konferencję Użytkowników Komputerów Dużej Mocy – KU KDM'17 10 March 2017, Zakopane Software MCB = MCNP + TTA Two main methedologies required to describe and predict nuclear core behavior MCNP Monte Carlo N-Particle Transport Code is use to describe Boltzmann equation. Describe neutron distribution and spectrum TTA Transmutation Trajectory Analysis analyzes nuclide density by solve Bateman equation. Describe fuel evolution Bateman Equations Solution (general case) This particularly: The transmutation trajectory analysis Level: 1 1, A C TTA 2 2, B A B decomposition 3 3, A 4, C 5, D D 4 6, B 7, D The Bateman equations for transmutation trajectory 5 8, C dN i = production rate destruction rate (i 1,n) dt Cutoff Transition and Pasage functions The trajectory transition calculate the number density that goes from initial nuclide to the formed nuclide for a given time t Passage is defined as the total removal rate in the considered trajectory or a fraction of the nuclides in a chain that passed beyond n nuclide and is assigned (or not) to following nuclides in the chain for the considered period. Parent pointer tree data structure The proposed data structure is an N-ary tree in which each node has a pointer to its parent node, but no pointers to child nodes. the structure stores: -nuclide index, -transition passage values of each chain -information about generation number (i.e. length of the current / tree level) -special flags i.e nuclide is last in the chain.
    [Show full text]
  • H Function Based Tamper-Proofing Software Watermarking Scheme Jianqi Zhu1,2, Yanheng Liu1,2, Aimin Wang1,2* 1
    148 JOURNAL OF SOFTWARE, VOL. 6, NO. 1, JANUARY 2011 H Function based Tamper-proofing Software Watermarking Scheme Jianqi Zhu1,2, Yanheng Liu1,2, Aimin Wang1,2* 1. College of Computer Science and technology 2. Key Laboratory of Computation and Knowledge Engineering, Ministry of Education Jilin University, Changchun, China Emails: {zhujq, liuyh, wangam @jlu.edu.cn} Kexin Yin3 3. College of computer and science and engineering ChangChun University of Technology, Changchun, China Email: [email protected] Abstract—A novel constant tamper-proofing software execution state of a software object. More precisely, in watermark technique based on H encryption function is dynamic software watermarking, what has been presented. First we split the watermark into smaller pieces embedded is not the watermark itself but some codes before encoding them using CLOC scheme. With the which cause the watermark to be expressed, or extracted, watermark pieces, a many-to-one function (H function) as when the software is run. One of the most effective the decoding function is constructed in order to avoid the pattern-matching or reverse engineering attack. The results watermarking techniques proposed to date is the dynamic of the function are encoded into constants as the graph watermarking (DGW) scheme of Collberg et al [5] parameters of opaque predicates or appended to the [6] [7]. The algorithm starts by mapping the watermark condition branches of the program to make the pieces to a special data structure called Planted Plane Cubic relevant. The feature of interaction among the pieces Tree (PPCT), and when the program is executed the improves the tamper-proofing ability because there being PPCT will be constructed.
    [Show full text]
  • The Algorithms and Data Structures Course Multicomponent Complexity and Interdisciplinary Connections
    Technium Vol. 2, Issue 5 pp.161-171 (2020) ISSN: 2668-778X www.techniumscience.com The Algorithms and Data Structures course multicomponent complexity and interdisciplinary connections Gregory Korotenko Dnipro University of Technology, Ukraine [email protected] Leonid Korotenko Dnipro University of Technology, Ukraine [email protected] Abstract. The experience of modern education shows that the trajectory of preparing students is extremely important for their further adaptation in the structure of widespread digitalization, digital transformations and the accumulation of experience with digital platforms of business and production. At the same time, there continue to be reports of a constant (catastrophic) lack of personnel in this area. The basis of the competencies of specialists in the field of computing is the ability to solve algorithmic problems that they have not been encountered before. Therefore, this paper proposes to consider the features of teaching the course Algorithms and Data Structures and possible ways to reduce the level of complexity in its study. Keywords. learning, algorithms, data structures, data processing, programming, mathematics, interdisciplinarity. 1. Introduction The analysis of electronic resources and literary sources allows for the conclusion that the Algorithms and Data Structures course is one of the cornerstones for the study programs related to computing. Moreover, this can be considered not only from the point of view of entering the profession, but also from the point of view of preparation for an interview in hiring as a programmer, including at a large IT organization (Google, Microsoft, Apple, Amazon, etc.) or for a promising startup [1]. As a rule, this course is characterized by an abundance of complex data structures and various algorithms for their processing.
    [Show full text]
  • Strongly Typed Genetic Programming
    Strongly Typed Genetic Programming David J. Montana Bolt Beranek and Newman, Inc. 70 Fawcett Street Cambridge, MA 02 13 8 [email protected] Abstract Genetic programming is a powerful method for automatically generating computer pro- grams via the process of natural selection (Koza, 1992). However, in its standard form, there is no way to restrict the programs it generates to those where the functions operate on appropriate data types. In the case when the programs manipulate multiple data types and contain functions designed to operate on particular data types, this can lead to unneces- sarily large search times and/or unnecessarily poor generalization performance. Strongly typed genetic programming (STGP) is an enhanced version of genetic programming that enforces data-type constraints and whose use of generic functions and generic data types makes it more powerful than other approaches to type-constraint enforcement. After describing its operation, we illustrate its use on problems in two domains, matridvector manipulation and list manipulation, which require its generality. The examples are (1) the multidimensional least-squares regression problem, (2) the multidimensional Kalman fil- ter, (3) the list manipulation function NTH, and (4) the list manipulation function MAP- CAR. Keywords genetic programming, automatic programming, strong typing, generic functions 1. Introduction Genetic programming is a method of automatically generating computer programs to per- form specified tasks (Koza, 1992). It uses a genetic algorithm to search through a space of possible computer programs for one that is nearly optimal in its ability to perform a particular task. While it was not the first method of automatic programming to use genetic algorithms (one earlier approach is detailed in Cramer [1985]), it is so far the most successful.
    [Show full text]
  • Bottom-Up Β-Reduction: Uplinks and Λ-Dags∗ (Extended Version)
    Bottom-up β-reduction: uplinks and λ-DAGs∗ (extended version) Olin Shiversy Mitchell Wand Georgia Institute of Technology Northeastern University [email protected] [email protected] December 31, 2004 Abstract If we represent a λ-calculus term as a DAG rather than a tree, we can efficiently represent the sharing that arises from β-reduction, thus avoiding combinatorial ex- plosion in space. By adding uplinks from a child to its parents, we can efficiently implement β-reduction in a bottom-up manner, thus avoiding combinatorial ex- plosion in time required to search the term in a top-down fashion. We present an algorithm for performing β-reduction on λ-terms represented as uplinked DAGs; describe its proof of correctness; discuss its relation to alternate techniques such as Lamping graphs, explicit-substitution calculi and director strings; and present some timings of an implementation. Besides being both fast and parsimonious of space, the algorithm is particularly suited to applications such as compilers, theorem provers, and type-manipulation systems that may need to examine terms in-between reductions—i.e., the “readback” problem for our representation is triv- ial. Like Lamping graphs, and unlike director strings or the suspension λ calculus, the algorithm functions by side-effecting the term containing the redex; the rep- resentation is not a “persistent” one. The algorithm additionally has the charm of being quite simple; a complete implementation of the data structure and algorithm is 180 lines of SML. ∗This document is the text of BRICS technical report RS-04-38 reformatted for 8.5”×11” “letter size” paper, for the convenience of Americans who may not have A4 paper available for printing.
    [Show full text]