CRAMM: Virtual Memory Support for Garbage-Collected Applications

Total Page:16

File Type:pdf, Size:1020Kb

CRAMM: Virtual Memory Support for Garbage-Collected Applications CRAMM: Virtual Memory Support for Garbage-Collected Applications Ting Yang Emery D. Berger Scott F. Kaplan† J. Eliot B. Moss [email protected] [email protected] [email protected] [email protected] Dept. of Computer Science †Dept. of Mathematics and Computer Science University of Massachusetts Amherst Amherst College Amherst, MA 01003-9264 Amherst, MA 01002-5000 Abstract like Python and Ruby. Garbage collection’s popularity de- Existing virtual memory systems were designed to sup- rives from its many software engineering advantages over port applications written in C and C++, but do not pro- manual memory management, including eliminating dan- vide adequate support for garbage-collected applications. gling pointer errors and drastically reducing memory leaks. The performance of garbage-collected applications is ex- Garbage-collection application performance is highly tremely sensitive to heap size. Larger heaps reduce the sensitive to heap size. A smaller heap reduces the amount frequency of garbage collections, making them run several of memory referenced, but requires frequent garbage col- times faster. However, if the heap is too large to fit in avail- lections that hurt performance. A larger heap reduces the able RAM, garbage collection activity can trigger thrash- frequency of collections, thus improving performance by ing. Existing Java virtual machines attempt to adjust their up to 10x. However, if the heap cannot fit in available application heap sizes adaptively to fit in RAM, but suffer RAM, performance drops off suddenly and sharply. This performance degradations of up to 94% when subjected to is because garbage collection has a large working set (it bursts of memory pressure. touches the entire heap) and thus can trigger catastrophic We present CRAMM (Cooperative Robust Automatic page swapping that degrades performance and increases Memory Management), a system that solves these prob- collection pauses by orders of magnitude [18]. Hence, heap lems. CRAMM consists of two parts: (1) a new virtual size and main memory allocation need to be coordinated memory system that collects detailed reference informa- to achieve good performance. Unfortunately, current VM tion for (2) an analytical model tailored to the underly- systems do not provide sufficient support for this coordina- ing garbage collection algorithm. The CRAMM virtual tion, and thus do not support garbage-collected applications memory manager tracks recent reference behavior with low well. overhead. The CRAMM heap sizing model uses this infor- Choosing the appropriate heap size for a garbage- mation to compute a heap size that maximizes throughput while minimizing paging. We present extensive empirical results demonstrating CRAMM’s ability to maintain high JRockit -- pseudojbb performance in the face of changing application and sys- No Memory Pressure 1.2 tem load. Dynamic Memory Pressure 13.055 trans/ms 1 1 Introduction 0.8 The virtual memory (VM) systems in today’s operating systems were designed to support applications written in 0.6 the widely-used programming languages of the 80’s and 0.4 90’s, C and C++. To maximize the performance of these Normalized Throughput 0.2 applications, it is enough to fit their working sets in physi- 0.777 trans/ms 0.722 trans/ms 0 cal memory [16]. VM systems typically manage available 0 0.5 1 1.5 2 2.5 memory with an approximation of LRU [12, 13, 15, 16, 22], Normalized Elapsed Time which works reasonably well for legacy applications. However, garbage-collected languages are now increas- Figure 1: Impact of bursts of memory pressure on the per- ingly prevalent. These languages range from general- formance on the JRockit Java virtual machine, which de- purpose languages like Java and C# to scripting languages grades throughput by as much as 94%. collected application—one that is large enough to maxi- Java VirtualM achine (JVM ) mize throughput but small enough to avoid paging—is a G arbage Collector key performance challenge. The ideal heap size is one that makes the working set of garbage collection just fit Heap Size W orking Set within the process’s main memory allocation. However, an M anager Size M odel a priori best choice is impossible in multiprogrammed en- e y l e r b p g S o a a vironments where the amount of main memory allocated n l S M inorfault i Allowable M ajor m e a a e W h H overhead target v Faultoverhead c to each process constantly changes. Existing garbage- M A collected languages either ignore this problem, allowing M ajorFault W SS Estim ator e only static heap sizes, or adapt the heap size dynamically z i CostM onitor S l t using mechanisms that are only moderately effective. For o s i r M inorFault t L n e example, Figure 1 shows the effect of dynamic memory o CostM onitor v i C t pressure on an industrial-strength Java virtual machine, c a Page Fault n BEA’s JRockit [7], running a variant of the SPECjbb2000 I Handler Histogram benchmark. The solid lines depict program execution when the process fits in available RAM, while the dashed lines VirtualM em ory M anager(VM ) show execution under periodic bursts of memory pressure. This memory pressure dilates overall execution time by a Figure 2: The CRAMM system. The CRAMM VM system factor of 220%, and degrades performance by up to 94%. efficiently gathers detailed per-process reference informa- The problem with these adaptive approaches is not that tion, allowing the CRAMM heap size model to choose an their adaptivity mechanism is broken, but rather that they optimal heap size dynamically. are reactive. The only way these systems can detect whether the heap size is too large is to grow the heap un- til paging occurs, which leads to unacceptable performance CRAMM, including experimental measurements across 20 degradation. benchmarks and 5 garbage collectors, as well as compari- Contributions: This paper makes the following con- son to two industrial Java implementations. These results tributions. It presents CRAMM (Cooperative Robust Au- demonstrate CRAMM’s effectiveness in maintaining high tomatic Memory Management), a system that enables performance in the face of changes in application behavior garbage-collected applications to predict an appropriate and system load. heap size, allowing the system to maintain high perfor- In addition to serving the needs of garbage-collected mance while adjusting dynamically to changing memory applications, the CRAMM VM system is the first sys- pressure. tem to our knowledge to provide per-process and per-file CRAMM consists of two parts; Figure 2 presents an page management while efficiently gathering detailed ref- overview. The first part is the CRAMM VM system that erence histograms. This information can be used to im- dynamically gathers the working set size (WSS) of each pro- plement a wide range of recently-proposed memory man- cess, where we define the WSS as the main memory allo- agement systems, including compressed page caches [27], cation that yields a trivial amount of page swapping. To adaptive LRU mechanisms like EELRU [25], and informed accomplish this, the VM system maintains separate page prefetchers [20, 24]. lists for each process and computes an LRU reference his- The remainder of this paper is organized as follows. Sec- togram [25, 27] that captures detailed reference informa- tion 2 presents an overview of garbage collection algo- tion while incurring little overhead (around 1%). The sec- rithms and terminology used in this paper. Section 3 de- ond part of CRAMM is its heap sizing model, which con- rives the CRAMM heap sizing model, which relates appli- trols application heap size and is independent of any par- cation working set size to heap size. Section 4 describes ticular garbage collection algorithm. The CRAMM model the CRAMM VM system, which gathers detailed statistics correlates the WSS measured by the CRAMM VM to the allowing it to compute the precise current process work- current heap size. It then uses this correlation to select a ing set size. Section 5 presents empirical results, compar- new heap size that is as large as possible (thus maximizing ing static and previous adaptive approaches to CRAMM. throughput) while yielding little or no page faulting behav- Section 6 presents work most closely related to ours, and ior. We apply the CRAMM model to five different garbage Section 7 concludes. collection algorithms, demonstrating its generality. We have implemented the CRAMM VM system in 2 GC Behavior and Terminology the Linux kernel and the CRAMM heap sizing model in A garbage collector (GC) periodically and automatically the Jikes RVM research Java virtual machine [3]. We finds and reclaims heap-allocated objects that a program present the results of an extensive empirical evaluation of can no longer possibly use. We now sketch how, and 2 when, a GC may do this work, and along the way intro- MarkSweep -- SPEC _213_javac duce GC terminology and concepts critical to understand- 50 300 ing CRAMM. 45 250 Garbage collectors operate on the principle that if an 40 35 object is unreachable via any chain of pointers starting 200 from roots—pointers found in global/static variables and 30 on thread stacks—then the program cannot possibly use 25 150 the object in the future, and the collector can reclaim and 20 100 reuse the object’s space. Through a slight abuse of termi- 15 Working Set Size (MB) Execution Time (second) 10 nology, reachable objects are often called live and unreach- 50 5 Execution Time able ones dead. Reference counting collectors determine Working Set Size (95%) 0 0 (conservatively) that an object is unreachable when there 0 50 100 150 200 250 are no longer any pointers to it. Here, we focus primarily Heap Size (MB) on tracing collectors, which actually trace through pointer chains from roots, visiting reachable objects.
Recommended publications
  • Apache Harmony Project Tim Ellison Geir Magnusson Jr
    The Apache Harmony Project Tim Ellison Geir Magnusson Jr. Apache Harmony Project http://harmony.apache.org TS-7820 2007 JavaOneSM Conference | Session TS-7820 | Goal of This Talk In the next 45 minutes you will... Learn about the motivations, current status, and future plans of the Apache Harmony project 2007 JavaOneSM Conference | Session TS-7820 | 2 Agenda Project History Development Model Modularity VM Interface How Are We Doing? Relevance in the Age of OpenJDK Summary 2007 JavaOneSM Conference | Session TS-7820 | 3 Agenda Project History Development Model Modularity VM Interface How Are We Doing? Relevance in the Age of OpenJDK Summary 2007 JavaOneSM Conference | Session TS-7820 | 4 Apache Harmony In the Beginning May 2005—founded in the Apache Incubator Primary Goals 1. Compatible, independent implementation of Java™ Platform, Standard Edition (Java SE platform) under the Apache License 2. Community-developed, modular architecture allowing sharing and independent innovation 3. Protect IP rights of ecosystem 2007 JavaOneSM Conference | Session TS-7820 | 5 Apache Harmony Early history: 2005 Broad community discussion • Technical issues • Legal and IP issues • Project governance issues Goal: Consolidation and Consensus 2007 JavaOneSM Conference | Session TS-7820 | 6 Early History Early history: 2005/2006 Initial Code Contributions • Three Virtual machines ● JCHEVM, BootVM, DRLVM • Class Libraries ● Core classes, VM interface, test cases ● Security, beans, regex, Swing, AWT ● RMI and math 2007 JavaOneSM Conference | Session TS-7820 |
    [Show full text]
  • Oracle Lifetime Support Policy for Oracle Fusion Middleware Guide
    ORACLE INFORMATION-DRIVEN SUPPORT Oracle Lifetime Support Policy Oracle Fusion Middleware Oracle Fusion Middleware 8 Oracle Fusion Middleware Releases 8 Application Development Tools 9 Oracle’s Application Development Tools 9 Oracle’s GraalVM Enterprise Releases 9 Oracle Cloud Application Foundation Releases 10 Oracle’s Sun and Glassfish Application Server Releases 12 Oracle’s Java Releases 13 Oracle’s Sun JDK Releases 13 Oracle’s Blockchain Platform Releases 14 Business Intelligence 14 Oracle Business Intelligence EE Releases 14 Oracle’s HyperRoll Releases 21 Oracle’s Siebel Technology Releases 22 Oracle’s Siebel Applications Releases 22 Oracle Big Data Discovery Releases 23 Oracle Endeca Information Discovery Releases 23 Oracle’s Endeca Releases 25 Master Data Management and Data Integrator 26 Oracle’s GoldenGate Releases 26 Oracle Data Integrator Releases 28 Oracle Data Integrator (Formerly Sunopsis) Releases 28 Oracle’s Sun Master Data Management and Data Integrator Releases 28 Oracle’s Silver Creek and EDQP Releases 29 Oracle's Datanomic and EDQ Releases 30 Oracle WebCenter Portal Releases 31 Oracle’s Sun Portal Releases 32 Oracle WebCenter Content Releases 32 Oracle’s Stellent Releases (Enterprise Content Management) 34 Oracle’s Captovation Releases (Enterprise Content Management) 35 Oracle WebCenter Sites Releases 36 Oracle FatWire Releases (WebCenter Sites) 36 Oracle Identity and Access Management Releases 37 Oracle’s Bharosa Releases 40 Oracle’s Passlogix Releases 41 Oracle’s Bridgestream Releases 41 Oracle’s Oblix Releases 42
    [Show full text]
  • Thread Scheduling in Multi-Core Operating Systems Redha Gouicem
    Thread Scheduling in Multi-core Operating Systems Redha Gouicem To cite this version: Redha Gouicem. Thread Scheduling in Multi-core Operating Systems. Computer Science [cs]. Sor- bonne Université, 2020. English. tel-02977242 HAL Id: tel-02977242 https://hal.archives-ouvertes.fr/tel-02977242 Submitted on 24 Oct 2020 HAL is a multi-disciplinary open access L’archive ouverte pluridisciplinaire HAL, est archive for the deposit and dissemination of sci- destinée au dépôt et à la diffusion de documents entific research documents, whether they are pub- scientifiques de niveau recherche, publiés ou non, lished or not. The documents may come from émanant des établissements d’enseignement et de teaching and research institutions in France or recherche français ou étrangers, des laboratoires abroad, or from public or private research centers. publics ou privés. Ph.D thesis in Computer Science Thread Scheduling in Multi-core Operating Systems How to Understand, Improve and Fix your Scheduler Redha GOUICEM Sorbonne Université Laboratoire d’Informatique de Paris 6 Inria Whisper Team PH.D.DEFENSE: 23 October 2020, Paris, France JURYMEMBERS: Mr. Pascal Felber, Full Professor, Université de Neuchâtel Reviewer Mr. Vivien Quéma, Full Professor, Grenoble INP (ENSIMAG) Reviewer Mr. Rachid Guerraoui, Full Professor, École Polytechnique Fédérale de Lausanne Examiner Ms. Karine Heydemann, Associate Professor, Sorbonne Université Examiner Mr. Etienne Rivière, Full Professor, University of Louvain Examiner Mr. Gilles Muller, Senior Research Scientist, Inria Advisor Mr. Julien Sopena, Associate Professor, Sorbonne Université Advisor ABSTRACT In this thesis, we address the problem of schedulers for multi-core architectures from several perspectives: design (simplicity and correct- ness), performance improvement and the development of application- specific schedulers.
    [Show full text]
  • Design and Analysis of a Scala Benchmark Suite for the Java Virtual Machine
    Design and Analysis of a Scala Benchmark Suite for the Java Virtual Machine Entwurf und Analyse einer Scala Benchmark Suite für die Java Virtual Machine Zur Erlangung des akademischen Grades Doktor-Ingenieur (Dr.-Ing.) genehmigte Dissertation von Diplom-Mathematiker Andreas Sewe aus Twistringen, Deutschland April 2013 — Darmstadt — D 17 Fachbereich Informatik Fachgebiet Softwaretechnik Design and Analysis of a Scala Benchmark Suite for the Java Virtual Machine Entwurf und Analyse einer Scala Benchmark Suite für die Java Virtual Machine Genehmigte Dissertation von Diplom-Mathematiker Andreas Sewe aus Twistrin- gen, Deutschland 1. Gutachten: Prof. Dr.-Ing. Ermira Mezini 2. Gutachten: Prof. Richard E. Jones Tag der Einreichung: 17. August 2012 Tag der Prüfung: 29. Oktober 2012 Darmstadt — D 17 For Bettina Academic Résumé November 2007 – October 2012 Doctoral studies at the chair of Prof. Dr.-Ing. Er- mira Mezini, Fachgebiet Softwaretechnik, Fachbereich Informatik, Techni- sche Universität Darmstadt October 2001 – October 2007 Studies in mathematics with a special focus on com- puter science (Mathematik mit Schwerpunkt Informatik) at Technische Uni- versität Darmstadt, finishing with a degree of Diplom-Mathematiker (Dipl.- Math.) iii Acknowledgements First and foremost, I would like to thank Mira Mezini, my thesis supervisor, for pro- viding me with the opportunity and freedom to pursue my research, as condensed into the thesis you now hold in your hands. Her experience and her insights did much to improve my research as did her invaluable ability to ask the right questions at the right time. I would also like to thank Richard Jones for taking the time to act as secondary reviewer of this thesis.
    [Show full text]
  • Virtual Appliances for Applications High Performance, High Density & Operationally Efficient Java Virtualization
    Virtual Appliances for Applications High Performance, High Density & Operationally Efficient Java Virtualization Axel Grosse Principal Sales Consultant Server Technologies Competence Center – FMW Mitte 1 The following is intended to outline our general product direction. It is intended for information purposes only, and may not be incorporated into any contract. It is not a commitment to deliver any material, code, or functionality, and should not be relied upon in making purchasing decisions. The development, release, and timing of any features or functionality described for Oracle’s products remains at the sole discretion of Oracle. 2 Cloud auf dem Peak der Hype Kurve Source: Gartner "Hype Cycle for Cloud Computing, 2009" Research Note G00168780 3 SaaS, PaaS und IaaS Anwendungen als Service für Software as a Service Endbenutzer im Netzwerk Entwicklungs- und Deployment Platform as a Service Plattformen als Service im Netzwerk Server, Storage und Netzwerk Infrastructure as a Service Hardware samt dazugehöriger Software als Service im Netzwerk 4 Public Clouds und Private Clouds Public Clouds Private Cloud • Externer Anbieter • Eigene IT als Anbieter • Weniger Aufwand I I SaaS SaaS N N • Mehr Aufwand • Weniger Einfluss T T PaaS auf PaaS E R • Mehr Einfluss A R IaaS •Sicherheit N N IaaS •Verfügbarkeit E E T T •... Benutzer 5 Cloud Computing: Oracle’s Perspektive • Basiert auf neuen Ideen und Möglichkeiten, basiert jedoch auf etablierter Technologie • Interessante Vorteile begleitet von ernstzunehmenden Bedenken • Unternehmen werden einen
    [Show full text]
  • Jikes RVM and IBM DK for Java ƒ Understanding System Behavior ƒ Other Issues 4
    Dynamic Compilation and Adaptive Optimization in Virtual Machines Instructor: Michael Hind Material contributed by: Matthew Arnold, Stephen Fink, David Grove, and Michael Hind © 2006 IBM Corporation IBM Research Who am I? Helped build Jikes RVM (1998-2006) – GC Maps, live analysis, dominators, register allocation refactoring – Adaptive optimization system – Management, project promotion, education, etc. Work for IBM, home of 2 other Java VMs – IBM DK for Java, J9 In previous lives, worked on – Automatic parallelization (PTran) – Ada implementation (Phd Thesis) – Interprocedural ptr analysis – Professor for 6 years Excited to share what I know – And learn what I don’t! 2 ACACES’06 | Dynamic Compilation and Adaptive Optimization in Virtual Machines | July 24-28, 2006 © 2006 IBM Corporation IBM Research Course Goals Understand the optimization technology used in production virtual machines Provide historical context of dynamic/adaptive optimization technology Debunk common misconceptions Suggest avenues of future research 3 ACACES’06 | Dynamic Compilation and Adaptive Optimization in Virtual Machines | July 24-28, 2006 © 2006 IBM Corporation IBM Research Course Outline 1. Background 2. Engineering a JIT Compiler 3. Adaptive Optimization 4. Feedback-Directed and Speculative Optimizations 5. Summing Up and Looking Forward 4 ACACES’06 | Dynamic Compilation and Adaptive Optimization in Virtual Machines | July 24-28, 2006 © 2006 IBM Corporation IBM Research Course Outline 1. Background Why software optimization matters Myths, terminology, and historical context How programs are executed 2. Engineering a JIT Compiler What is a JIT compiler? Case studies: Jikes RVM, IBM DK for Java, HotSpot High level language-specific optimizations VM/JIT interactions 3. Adaptive Optimization Selective optimization Design: profiling and recompilation Case studies: Jikes RVM and IBM DK for Java Understanding system behavior Other issues 4.
    [Show full text]
  • Acknowledgments Diagnosing and Tolerating Bugs in Deployed Systems
    The Dissertation Committee for Michael David Bond certifies that this is the approved version of the following dissertation: Acknowledgments Diagnosing and Tolerating Bugs in Deployed Systems I am deeply grateful to Kathryn McKinley for supporting and mentor- by ing me in many ways. Kathryn has provided invaluable technical expertise Diagnosing and Tolerating Bugs in Deployed Systems and feedback even as we worked in new areas. She has been unwaveringly Michael David Bond, B.S., M.C.S. enthusiastic and positive about students and ideas and projects, and she has been a strong advocate, putting my and other students’ interests first. I en- tered graduate school uncertain about getting a Ph.D., and I leave inspired to pursue an academic career because of Kathryn. Committee: Steve Blackburn has been an enthusiastic supporter and has devoted a lot of his time to solving problems and giving advice, both solicited and unso- Copyright DISSERTATION Kathryn S. McKinley, Supervisor licited. Steve, Keshav Pingali, Peter Stone, and Emmett Witchel have given by Presented to the Faculty of the Graduate School of helpful feedback and spent a lot of time reading long documents and attending Michael David Bond Stephen M. Blackburn The University of Texas at Austin long talks. Vitaly Shmatikov and Steve Keckler have provided valuable advice 2008 in Partial Fulfillment and mentoring. David Padua and Craig Zilles were supportive advisors when Keshav Pingali of the Requirements I was a master’s student at the University of Ilinois. for the Degree of The graduate student community at UT has been incredibly support- Peter Stone ive.
    [Show full text]
  • A Comparative Study of Garbage Collection Techniques in Java Virtual Machines
    Sindh Univ. Res. Jour. (Sci. Ser.) Vol.44 (4) 667- 672 (2012) SINDH UNIVERSITY RESEARCH JOURNAL (SCIENCE SERIES) A Comparative Study of Garbage Collection Techniques in Java Virtual Machines S. IQBAL, M.A. KHAN, I.A. MEMON* Department of Computer Science, Bahauddin Zakariya University, Pakistan. Received 07th February 2012 and Revised 12th July 2012 Abstract: Garbage collection mechanisms implemented in Java Virtual Machines (JVMs) are used to reclaim memory allocated to the objects that become inaccessible during program execution. An efficient garbage collection mechanism certainly facilitates in smooth running of an application. This paper performs a comparative analysis of garbage collection techniques implemented in Sun HotSpot, Oracle JRockit and IBM J9 virtual machines. We perform experiments with several benchmarks containing dynamic creation for objects of TreeMap, ArrayList and String array data types. Due to the creation of a large number of big sized memory objects and loss of references, the garbage is created that is collected by the garbage collectors implemented in the virtual machines. The experiments are carried out on different hardware architectures including Intel Core 2 Quad and Intel Xeon based systems. The performance of each garbage collector is computed in terms of the rate of garbage collection that mainly depends upon the strategy implemented for garbage collection. Our results show that the Sun HotSpot garbage collector performs 4.66 times better than the Oracle JRockit and 144.06 times better than the IBM J9 garbage collectors. Keywords: Garbage Collection, Java Virtual Machines, JRockit, IBM J9, Sun HotSpot, Bytecode. This paper contributes towards a performance 1. INTRODUCTION Java is a widely used language that is analysis of the garbage collection approaches implemented on a large variety of platforms ranging implemented in virtual machines.
    [Show full text]
  • Using HPM-Sampling to Drive Dynamic Compilation
    Using HPM-Sampling to Drive Dynamic Compilation Dries Buytaerty Andy Georgesy Michael Hind∗ Matthew Arnold∗ Lieven Eeckhouty Koen De Bosscherey y Department of Electronics and Information Systems, Ghent University, Belgium ∗IBM T.J. Watson Research Center, New York, NY, USA fdbuytaer,[email protected], fhindm,[email protected], fleeckhou, [email protected] Abstract guage require a dynamic execution environment called a All high-performance production JVMs employ an adaptive virtual machine (VM). To achieve high performance, pro- strategy for program execution. Methods are first executed duction Java virtual machines contain at least two modes unoptimized and then an online profiling mechanism is used of execution: 1) unoptimized execution, using interpreta- to find a subset of methods that should be optimized during tion [21, 28, 18] or a simple dynamic compiler [16, 6, 10, 8] the same execution. This paper empirically evaluates the de- that produces code quickly, and 2) optimized execution us- sign space of several profilers for initiating dynamic com- ing an optimizing dynamic compiler. Methods are first ex- pilation and shows that existing online profiling schemes ecuted using the unoptimized execution strategy. An online suffer from several limitations. They provide an insufficient profiling mechanism is used to find a subset of methods to number of samples, are untimely, and have limited accu- optimize during the same execution. Many systems enhance racy at determining the frequently executed methods. We de- this scheme to provide multiple levels of optimized execu- scribe and comprehensively evaluate HPM-sampling, a sim- tion [6, 18, 28], with increasing compilation cost and bene- ple but effective profiling scheme for finding optimization fits at each level.
    [Show full text]
  • Security Policy Level 1
    20.10.20 RSA® BSAFE® Crypto-J JSAFE and JCE Software Module 6.2 and 6.2.1.1 Security Policy Level 1 This document is a non-proprietary security policy for RSA BSAFE Crypto-J JSAFE and JCE Software Module 6.2 and 6.2.1.1 (Crypto-J JSAFE and JCE Software Module) security software. This document may be freely reproduced and distributed whole and intact including the copyright notice. Note: Refer to the Change Summary for the location of the latest change to this document. Contents: Preface .............................................................................................................2 Terminology .............................................................................................2 Document Organization .........................................................................2 1 The Cryptographic Module .........................................................................3 1.1 Introduction ........................................................................................3 1.2 Module Characteristics .....................................................................3 1.3 Module Interfaces ............................................................................10 1.4 Roles and Services ......................................................................... 11 1.5 Cryptographic Key Management ..................................................23 1.6 Cryptographic Algorithms ...............................................................26 1.7 Self-tests ...........................................................................................28
    [Show full text]
  • SIMD Intrinsics on Managed Language Runtimes
    1 56 2 SIMD Intrinsics on Managed Language Runtimes 57 3 58 4 Anonymous Author(s) 59 5 60 6 Abstract speedup, but neither languages like Java, JavaScript, Python, 61 7 Managed language runtimes such as the Java Virtual Ma- or Ruby, nor their managed runtimes, provide direct access 62 8 chine (JVM) provide adequate performance for a wide range to SIMD facilities. This means that SIMD optimizations, if 63 9 of applications, but at the same time, they lack much of the available at all, are left to the virtual machine (VM) and the 64 10 low-level control that performance-minded programmers built-in just-in-time (JIT) compiler to carry out automatically, 65 11 appreciate in languages like C/C++. One important example which often leads to suboptimal code. As a result, developers 66 12 of such a control is the intrinsics interface that expose in- may be pushed to use low-level languages such as C/C++ to 67 13 structions of SIMD (Single Instruction Multiple Data) vector gain access to the intrinsics API. But leaving the high-level 68 14 ISAs (Instruction Set Architectures). In this paper we present ecosystem of Java or other languages also means to abandon 69 15 an automatic approach for including native intrinsics in the many high-level abstractions that are key for the produc- 70 16 runtime of a managed language. Our implementation con- tive and efficient development of large-scale applications, 71 17 sists of two parts. First, for each vector ISA, we automatically including access to a large set of libraries.
    [Show full text]
  • Trace Register Allocation
    Submitted by DI Josef Eisl, BSc. Submitted at Institute for System Software Supervisor and First Examiner o.Univ.-Prof. DI Dr.Dr.h.c. Hanspeter Mössenböck Second Examiner Ao.Univ.-Prof. DI Dr. Andreas Krall October 2018 Trace Register Allocation Doctoral Thesis to obtain the academic degree of Doktor der technischen Wissenschaften in the Doctoral Program Technische Wissenschaften JOHANNES KEPLER UNIVERSITY LINZ Altenbergerstraße 69 4040 Linz, Österreich www.jku.at DVR 0093696 Oracle, Java, HotSpot, and all Java-based trademarks are trademarks or registered trademarks of Oracle in the United States and other countries. All other product names mentioned herein are trademarks or registered trademarks of their respective owners. In memory of my brother. (Wolfgang Eisl, 1973–2016) v Abstract Register allocation, i.e., mapping the variables of a programming language to the physical reg- isters of a processor, is a mandatory task for almost every compiler and consumes a significant portion of the compile time. In a just-in-time compiler, compile time is a particular issue because compilation happens during program execution and contributes to the overall application run time. Compilers often use global register allocation approaches, such as graph coloring or linear scan, which only have limited potential for improving compile time since they process a whole method at once. With growing methods sizes, these approaches are reaching their scalability boundary due to their limited flexibility. We developed a novel trace register allocation framework, which competes with global approaches in both, compile time and code quality. Instead of processing a whole method at once, our al- locator processes linear code segments (traces) independently and is therefore able to (1) select different allocation strategies based on the characteristics of a trace in order to control thetrade- off between compile time and peak performance, and (2) to allocate traces in parallel inorderto reduce compilation latency, i.e., the time until the result of a compilation is available.
    [Show full text]