LLVM-Chimps: Compilation Environment for Fpgas Using LLVM Compiler Infrastructure and Chimps Computational Model

Total Page:16

File Type:pdf, Size:1020Kb

LLVM-Chimps: Compilation Environment for Fpgas Using LLVM Compiler Infrastructure and Chimps Computational Model LLVM-CHiMPS: Compilation Environment for FPGAs Using LLVM Compiler Infrastructure and CHiMPS Computational Model Seung J. Lee1, David K. Raila2, Volodymyr V. Kindratenko1 1) National Center for Supercomputing Applications (NCSA) 2) University of Illinois at Urbana-Champaign (UIUC) [email protected], [email protected], [email protected] Abstract developed by Xilinx, employs a unique approach by imposing a computational model on the FPGA CHiMPS (Compiling High level language to subsystem, treating it as a virtualized hardware Massively Pipelined System) system, developed by architecture that is convenient for high level code Xilinx is gaining popularity due to its convenient compilation. The CHiMPS compiler transforms high computational model and architecture for field level language (HLL) codes, currently written in ANSI programmable gate array computing. The CHiMPS C, to CHiMPS target language (CTL), which is system utilizes CHiMPS target language as an combined with runtime support for the FPGA runtime intermediate representation to bridge between the high environment and assembled using the CHiMPS level language and the data flow architecture assembler and FPGA design tools to a hardware data generated from it. However, currently the CHiMPS flow implementation [2]. The instructions defined in frontend does not provide many commonly used CTL have a close resemblance to a traditional optimizations and has some use restrictions. In this microprocessor instruction set which is convenient for paper we present an alternative compiler environment compiling, optimizing, and should be familiar to based on low level virtual machine compiler programmers [3, 4]. This effectively bridges the gap environment extended to generate CHiMPS target between hardware design and traditional software language code for the CHiMPS architecture. Our development and only imposes minor costs related to implementation provides good support for global runtime overhead support to virtualize the architecture. optimizations and analysis and overcomes many limitations of the original Xilinx CHiMPS compiler. 1.2. LLVM Simulation results from codes based on this approach show to outperform those obtained with the original Low level virtual machine (LLVM) compiler [5], CHiMPS compiler. originally developed at the University of Illinois, is now an open source compiler project that is aimed at 1. Introduction supporting global optimizations. It has many attractive features for program analysis and optimization. LLVM 1.1. CHiMPS has a GCC based C and C++ frontend as well as its own frontend, Clang, that offers modular and reusable Recently, systems have been developed that employ components for building compilers that are target- Field Programmable Gate Arrays (FPGAs) as independent and easy to use. This reduces the time and application accelerators. In order to execute effort to build a particular compiler. Static backends applications on such systems, the application already implemented in LLVM including X86, X86-64, programmer extracts computationally intensive kernel PowerPC 32/64, ARM and others share those from the application, reorganizes data structures to components. A notable feature of LLVM is its low accommodate memory subsystem, and rewrites the level static single assignment (SSA) form virtual kernel in the highly specialized environment for FPGA instruction set which is a low-level intermediate execution. The CHiMPS (Compiling High level representation (IR) that uses RISC-like instructions. language to Massively Pipelined System) system [1], The IR makes LLVM versatile and flexible as well as language-independent and is amenable data flow level source code is translated into a low-level representations. hardware language, such as VHDL or Verilog. The idea of the frontend optimization strategy using 1.3. LLVM in CHiMPS compilation flow LLVM has been proven by the Trident compiler [19]. Trident’s framework is similar to LLVM-CHiMPS The current LCC based frontend of CHiMPS compilation except that Trident compiler is used compiler [1] is known to have some serious drawbacks instead of CHiMPS backend. Also in our LLVM- as shown in Section 2. In the work presented in this CHiMPS compiler a more aggressive set of paper, LLVM is employed to improve these optimization techniques is implemented in the LLVM shortcomings and to investigate the ability to do high frontend whereas in the Trident’s approach many level program transformations that better support the optimizations are pushed down to the Trident’s CHiMPS architecture. Figure 1 illustrates the backend. compilation flows of the original Xilinx CHiMPS. The ROCCC compiler [20] is another example of a LLVM backend part is aimed at substituting for compiler that translates HLL code into highly parallel CHiMPS HLL compiler. In the following sections, the and optimized circuits. Instead of using LLVM implementation details are discussed. frontend and CHiMPS backend, the ROCCC compiler is built on the SUIF2 and Machine-SUIF [29, 30] platforms, with the SUIF2 platform being used to process optimizations while the Machine-SUIF platform is used to further specialize for the hardware and generate VHDL. On the commercial side, several HLL compilers are available including Mitrion-C [26], Impulse-C [25], DIME-C [27], and MAP-C [28]. These compilers, however, are proprietary and support only a few reconfigurable computing systems. In our LLVM-CHiMPS approach, we take advantage of the optimization techniques implemented in the open-source LLVM compiler and of the FPGA hardware knowledge applied to the backend CHiMPS compiler developed by Xilinx. 2. Examples of known limitations in Xilinx CHiMPS compiler Original Xilinx CHiMPS compiler has some limitations that make it inconvenient and difficult for programming. For example, legal high level expressions in HLL code sometimes need to be paraphrased for CHiMPS compilation. Consider the ANSI C code shown in Figure 2. Figure 1. Compilation flow of Xilinx CHiMPS char* foo (int select, char* brTid, char* brFid) { compiler. if (select) return brTid; 1.4. Related work return brFid; } Several research and commercial HLL to FPGA compilers are currently available [19-27]. Although Figure 2. Simple code that fails in Xilinx CHiMPS their implementations differ significantly, the general compiler framework of compilation strategies is the same: high for (i=0; i<=n; i++) { } for (i = 0; i < n; i++) { } for (i=0; i<n; i+=2) { } for (i=1; i<n; i++) { } Figure 3. (a) Supported and (b) unsupported expressions of for loop construct int foo() { Enter foo; define i32 @foo() nounwind { int n; reg k, n entry: int k=0; add 0;k ret i32 45 for (n=0; n<10; n++) reg temp0:1 } k+=n; nfor l0;10;n add k,n;k return k; end l0 } exit foo; k Source code CTL from Xilinx CHiMPS compiler LLVM IR Figure 4. XIlinx CHiMPS compiler does not employ common optimization techniques. This simple code fails because two return own virtual instruction set which is a low-level IR that statements are used in a single function, and although uses RISC-like target independent instructions. Every this is legal in ANSI C, the Xilinx CHiMPS frontend LLVM backend for a specific target is based on this does not support multiple returns. LLVM IR. Figure 3a shows the only form of for statement that In the following sections, overall comparison of is fully supported by the Xilinx CHiMPS compiler. All instructions at two different levels is exposed. While other uses of for statement, such as shown in Figure 3b LLVM instructions are low level, CHiMPS target are not guaranteed to produce correct code [16]. language defines low level, as well as some high level A very significant weakness of the Xilinx CHiMPS instructions such as for and nfor, which are similar to compiler frontend is that no optimizations are carried for statement in C language. The difference between out during compilation [17]. Figure 4 depicts an these two levels is critical in this implementation of example for a simple summation from 0 to 9 that yields LLVM-CHiMPS, and so it needs to be discussed in 45, and is not optimized at all by the CHiMPS frontend. more detail. Readers are directed to refer to the LLVM The CTL code emitted is almost directly related to the and CHiMPS documentation [2, 3, 5] for more detailed source code, although this kind of a simple high level syntax, semantics and examples of the instructions for code is commonly optimized at the compilation time, CHiMPS and LLVM. as shown in the LLVM IR, the right column in Figure 4. These serious drawbacks in CHiMPS can be 3.1. Implementation of low level improved using LLVM as a frontend which offers more representations in CHiMPS stable compilation with numerous global optimizations and the potential for significant program LLVM is intended to support global optimization transformations. which is one of the motivations for using it in this work. Additionally, LLVM’s simple RISC-like 3. From LLVM to CHiMPS instructions are quite similar to CHiMPS instructions which are also similar to traditional microprocessor As shown in Figure 1, our approach is to use LLVM instructions, so there is a good connection between as the frontend to generate CTL code instead of the these models and a good starting point for LLVM- Xilinx CHiMPS frontend. This section describes how CHiMPS. LLVM is modified to meet this goal. LLVM has its Integer arithmetic Binary operations - add, sub, multiply, divide, cmp - add, sub, mul, udiv, sdiv, fdiv, urem, srem, frem Floating-point arithmetic Bitwise binary operations - i2f, f2i, fadd, fsub, fmultiply, fdivide, fcmp - shl, lshr, ashr, and, or, xor Logical operations Other operations - and, or, xor, not, shl, shr - icmp, fcmp, … Conversion operations - sitofp, fptosi, … Figure 5. Arithmetic and logical instructions in (a) CHiMPS and (b) LLVM Pseudo-instructions - reg, enter, exit, assign, foreign, call Figure 6. CHiMPS pseudo-instructions Standard memory access instructions Memory access and addressing operations - memread, memwrite - load, store, … Figure 7. Memory access instructions in (a) CHiMPS and (b) LLVM CHiMPS also has common instructions for call are easily associated with call instruction in LLVM arithmetic and logical operations.
Recommended publications
  • Eee4120f Hpes
    Lecture 22 HDL Imitation Method, Benchmarking and Amdahl's for FPGAs Lecturer: HDL HDL Imitation Simon Winberg Amdahl’s for FPGA Attribution-ShareAlike 4.0 International (CC BY-SA 4.0) HDL Imitation Method Using Standard Benchmarks for FPGAs Amdahl’s Law and FPGA or ‘C-before-HDL approach to starting HDL designs. An approach to ‘golden measures’ & quicker development void mymod module mymod (char* out, char* in) (output out, input in) { { out[0] = in[0]^1; out = in^1; } } The same method can work with Python, but C is better suited due to its typical use of pointer. This method can be useful in designing both golden measures and HDL modules in (almost) one go … It is mainly a means to validate that you algorithm is working properly, and to help get into a ‘thinking space’ suited for HDL. This method is loosely based on approaches for C→HDL automatic conversion (discussed later in the course) HDL Imitation approach using C C program: functions; variables; based on sequence (start to end) and the use of memory/registers operations VHDL / Verilog HDL: Implements an entity/module for the procedure C code converted to VHDL Standard C characteristics Memory-based Variables (registers) used in performing computation Normal C and C programs are sequential Specialized C flavours for parallel description & FPGA programming: Mitrion-C , SystemC , pC (IBM Parallel C) System Crafter, Impulse C , OpenCL FpgaC Open-source (http://fpgac.sourceforge.net/) – does generate VHDL/Verilog but directly to bit file Best to simplify this approach, where possible, to just one module at a time When you’re confident the HDL works, you could just leave the C version behind Getting a whole complex design together as both a C-imitating-HDL program and a true HDL implementation is likely not viable (as it may be too much overhead to maintain) Example Task: Implement an countup module that counts up on target value, increasing its a counter value on each positive clock edge.
    [Show full text]
  • Review of FPD's Languages, Compilers, Interpreters and Tools
    ISSN 2394-7314 International Journal of Novel Research in Computer Science and Software Engineering Vol. 3, Issue 1, pp: (140-158), Month: January-April 2016, Available at: www.noveltyjournals.com Review of FPD'S Languages, Compilers, Interpreters and Tools 1Amr Rashed, 2Bedir Yousif, 3Ahmed Shaban Samra 1Higher studies Deanship, Taif university, Taif, Saudi Arabia 2Communication and Electronics Department, Faculty of engineering, Kafrelsheikh University, Egypt 3Communication and Electronics Department, Faculty of engineering, Mansoura University, Egypt Abstract: FPGAs have achieved quick acceptance, spread and growth over the past years because they can be applied to a variety of applications. Some of these applications includes: random logic, bioinformatics, video and image processing, device controllers, communication encoding, modulation, and filtering, limited size systems with RAM blocks, and many more. For example, for video and image processing application it is very difficult and time consuming to use traditional HDL languages, so it’s obligatory to search for other efficient, synthesis tools to implement your design. The question is what is the best comparable language or tool to implement desired application. Also this research is very helpful for language developers to know strength points, weakness points, ease of use and efficiency of each tool or language. This research faced many challenges one of them is that there is no complete reference of all FPGA languages and tools, also available references and guides are few and almost not good. Searching for a simple example to learn some of these tools or languages would be a time consuming. This paper represents a review study or guide of almost all PLD's languages, interpreters and tools that can be used for programming, simulating and synthesizing PLD's for analog, digital & mixed signals and systems supported with simple examples.
    [Show full text]
  • A Mobile Programmable Radio Substrate for Any‐Layer Measurement and Experimentation
    A Mobile Programmable Radio Substrate for Any‐layer Measurement and Experimentation Kuang-Ching Wang Department of Electrical and Computer Engineering Clemson University, Clemson, SC 29634 [email protected] Version 1.0 September 27 2010 This work is sponsored by NSF grant CNS-0940805. 1 A Mobile Programmable Radio Substrate for Any‐layer Measurement and Experimentation A Whitepaper Developed for GENI Subcontract #1740 Executive Summary Wireless networks have become increasingly pervasive in all aspects of the modern life. Higher capacity, ubiquitous coverage, and robust adaptive operation have been the key objectives for wireless communications and networking research. Latest researches have proposed a myriad of solutions for different layers of the protocol stack to enhance wireless network performance; these innovations, however, are hard to be validated together to study their combined benefits and implications due to the lack of a platform that can easily incorporate new protocols from any protocol layers to operate with the rest of the protocol stack. GENI’s mission is to create a highly programmable testbed for studying the future Internet. This whitepaper analyzes GENI’s potentials and requirements to support any-layer programmable wireless network experiments and measurements. To date, there is a clear divide between the research methodology for the lower layers, i.e., the physical (PHY) and link layers, and that for the higher layers, i.e., from link layer above. For lower layer research, software defined radio (SDR) based on PCs and/or FPGAs has been the technology of choice for programmable testbeds. For higher layer research, PCs equipped with a range of commercial-off-the-shelf (COTS) network interfaces and standard operating systems generally suffice.
    [Show full text]
  • Total Ionizing Dose Effects on Xilinx Field-Programmable Gate Arrays
    University of Alberta Total Ionizing Dose Effects on Xilinx Field-Programmable Gate Arrays Daniel Montgomery MacQueen O A thesis submitted to the Faculty of Graduate Studies and Research in partial fulfillment of the requirements for the degree of Master of Science Department of Physics Edmonton, Alberta FaIl 2000 National Library Bibliothèque nationale I*I of Canada du Canada Acquisitions and Acquisitions et Bibliographie Services services bibliographiques 395 Wellington Street 345. nie Wellington Ottawa ON K1A ON4 Ottawa ON KIA ON4 Canada Canada The author has granted a non- L'auteur a accordé une licence non exclusive licence allowing the exclusive permettant a la National Library of Canada to BLbiiothèque nationale du Canada de reproduce, han, distniute or seil reproduire, prêter, distribuer ou copies of this thesis in microform, vendre des copies de cette these sous paper or electronic formats. la forme de microfiche/nlm, de reproduction sur papier ou sur format PIectronique . The author retains ownership of the L'auteur conserve la propriété du copyxïght in this thesis. Neither the droit d'auteur qui protège cette thèse. thesis nor substantial extracts fi-om it Ni la thèse ni des extraits substantiels may be printed or otherwise de celle-ci ne doivent être imprimés reproduced without the author's ou autrement reproduit. sans son permission. autorisation. Abstract This thesis presents the results of radiation tests of Xilinx XC4036X se- ries Field Programmable Gate Arrays (FPGAs) . These radiation tests investigated the suitability of the XC403GX FPGAs as controllers for the ATLAS liquid argon calorimeter front-end boards. The FPGAs were irradiated with gamma rays from a cobalt-60 source at a average dose rate of 0.13 rad(Si)/s.
    [Show full text]
  • Altera Powerpoint Guidelines
    Ecole Numérique 2016 IN2P3, Aussois, 21 Juin 2016 OpenCL On FPGA Marc Gaucheron INTEL Programmable Solution Group Agenda FPGA architecture overview Conventional way of developing with FPGA OpenCL: abstracting FPGA away ALTERA BSP: abstracting FPGA development Live Demo Developing a Custom OpenCL BSP 2 FPGA architecture overview FPGA Architecture: Fine-grained Massively Parallel Millions of reconfigurable logic elements I/O Thousands of 20Kb memory blocks Let’s zoom in Thousands of Variable Precision DSP blocks Dozens of High-speed transceivers Multiple High Speed configurable Memory I/O I/O Controllers Multiple ARM© Cores I/O 4 FPGA Architecture: Basic Elements Basic Element 1-bit configurable 1-bit register operation (store result) Configured to perform any 1-bit operation: AND, OR, NOT, ADD, SUB 5 FPGA Architecture: Flexible Interconnect … Basic Elements are surrounded with a flexible interconnect 6 FPGA Architecture: Flexible Interconnect … … Wider custom operations are implemented by configuring and interconnecting Basic Elements 7 FPGA Architecture: Custom Operations Using Basic Elements 16-bit add 32-bit sqrt … … … Your custom 64-bit bit-shuffle and encode Wider custom operations are implemented by configuring and interconnecting Basic Elements 8 FPGA Architecture: Memory Blocks addr Memory Block data_out data_in 20 Kb Can be configured and grouped using the interconnect to create various cache architectures 9 FPGA Architecture: Memory Blocks addr Memory Block data_out data_in 20 Kb Few larger Can be configured and grouped using
    [Show full text]
  • System-Level Methods for Power and Energy Efficiency of Fpga-Based Embedded Systems
    ATTENTION: The Singapore Copyright Act applies to the use of this document. Nanyang Technological University Library SYSTEM-LEVEL METHODS FOR POWER AND ENERGY EFFICIENCY OF FPGA-BASED EMBEDDED SYSTEMS PAWEŁ PIOTR CZAPSKI School of Computer Engineering A thesis submitted to the Nanyang Technological University in fulfillment of the requirement for the degree of Doctor of Philosophy 2010 i ATTENTION: The Singapore Copyright Act applies to the use of this document. Nanyang Technological University Library To My Parents ii ATTENTION: The Singapore Copyright Act applies to the use of this document. Nanyang Technological University Library Acknowledgements ACKNOWLEDGEMENTS I would like to express my sincere appreciation to my supervisor Associate Professor Andrzej Śluzek for his continuous interest, infinite patience, guidance, and constant encouragement that was motivating me during this research work. His vision and broad knowledge play an important role in the realization of the whole work. I acknowledge gratefully possibility to conduct this research at the Intelligent Systems Centre, the place with excellent working environment. I would also like to acknowledge the financial support that I received from the Nanyang Technological University and the Intelligent Systems Centre during my studies in Singapore. Finally, I would like to acknowledge my parents and my best friend Maciej for a constant help in these though moments. iii ATTENTION: The Singapore Copyright Act applies to the use of this document. Nanyang Technological University Library Table of Contents TABLE OF CONTENTS Title Page i Acknowledgements iii Table of Contents iv List of Symbols vii List of Abbreviations x List of Figures xiii List of Tables xv Abstract xvii Chapter I Introduction 1 1.1.
    [Show full text]
  • SDCC Compiler User Guide
    SDCC Compiler User Guide SDCC 3.0.4 $Date:: 2011-08-17 #$ $Revision: 6749 $ Contents 1 Introduction 6 1.1 About SDCC.............................................6 1.2 Open Source..............................................7 1.3 Typographic conventions.......................................7 1.4 Compatibility with previous versions.................................7 1.5 System Requirements.........................................8 1.6 Other Resources............................................9 1.7 Wishes for the future.........................................9 2 Installing SDCC 10 2.1 Configure Options........................................... 10 2.2 Install paths.............................................. 12 2.3 Search Paths.............................................. 13 2.4 Building SDCC............................................ 14 2.4.1 Building SDCC on Linux.................................. 14 2.4.2 Building SDCC on Mac OS X................................ 15 2.4.3 Cross compiling SDCC on Linux for Windows....................... 15 2.4.4 Building SDCC using Cygwin and Mingw32........................ 15 2.4.5 Building SDCC Using Microsoft Visual C++ 6.0/NET (MSVC).............. 16 2.4.6 Windows Install Using a ZIP Package............................ 17 2.4.7 Windows Install Using the Setup Program.......................... 17 2.4.8 VPATH feature........................................ 17 2.5 Building the Documentation..................................... 18 2.6 Reading the Documentation....................................
    [Show full text]
  • A TMD-MPI/MPE Based Heterogeneous Video System
    i A TMD-MPI/MPE Based Heterogeneous Video System by Tony Ming Zhou Supervisor: Professor Paul Chow April 2010 ii Abstract: The FPGA technology advancements have enabled reconfigurable large-scale hardware system designs. In recent years, heterogeneous systems comprised of embedded processors, memory units, and a wide-variety of IP blocks have become an increasingly popular solution to building future computing systems. The TMD- MPI project has evolved the software standard message passing interface, MPI, to the scope of FPGA hardware design. It provides a new programming model that enabled transparent communication and synchronization between tasks running on heterogeneous processing devices in the system. In this thesis project, we present the design and characterization of a TMD-MPI based heterogeneous video processing system. The system is comprised of hardware peripheral cores and software video codec. By hiding low-level architectural details from the designer, TMD-MPI can improve development productivity and reduce the level of difficulty. In particular, with the type of abstraction TMD-MPI provides, the software video codec approach is an easy-to-entry point for hardware design. The primary focus is the functionalities and different configurations of the TMD-MPI based heterogeneous design. iii Acknowledgements I would like to thank supervisor Professor Paul Chow for his patience and guidance, and the University of Toronto and the department of Engineering Science for the wonderful five year journey that led me to this point. Special thanks go to the TMD-MPI research group, in particular to Sami Sadaka, Kevin Lam, Kam Pui Tang, and Manuel Saldaña. Last but not the least, I would like to thank my family and my friends for always being there for me.
    [Show full text]
  • SDCC Compiler User Guide
    SDCC Compiler User Guide SDCC 3.0.1 $Date:: 2011-03-02 #$ $Revision: 6253 $ Contents 1 Introduction 6 1.1 About SDCC.............................................6 1.2 Open Source..............................................7 1.3 Typographic conventions.......................................7 1.4 Compatibility with previous versions.................................7 1.5 System Requirements.........................................8 1.6 Other Resources............................................9 1.7 Wishes for the future.........................................9 2 Installing SDCC 10 2.1 Configure Options........................................... 10 2.2 Install paths.............................................. 12 2.3 Search Paths.............................................. 13 2.4 Building SDCC............................................ 14 2.4.1 Building SDCC on Linux.................................. 14 2.4.2 Building SDCC on Mac OS X................................ 15 2.4.3 Cross compiling SDCC on Linux for Windows....................... 15 2.4.4 Building SDCC using Cygwin and Mingw32........................ 15 2.4.5 Building SDCC Using Microsoft Visual C++ 6.0/NET (MSVC).............. 16 2.4.6 Windows Install Using a ZIP Package............................ 17 2.4.7 Windows Install Using the Setup Program.......................... 17 2.4.8 VPATH feature........................................ 17 2.5 Building the Documentation..................................... 18 2.6 Reading the Documentation....................................
    [Show full text]
  • Altera Powerpoint Guidelines
    Ecole Informatique 2016 IN2P3, Ecole Polytechnique, 27 Mai 2016 OpenCL On FPGA Marc Gaucheron INTEL Programmable Solution Group Agenda FPGA architecture overview Conventional way of developing with FPGA OpenCL: abstracting FPGA away ALTERA BSP: abstracting FPGA development Examples 2 A bit of Marketing ? 3 FPGA architecture overview FPGA Architecture: Fine-grained Massively Parallel Millions of reconfigurableLet’s zoom logic in elements I/O Thousands of 20Kb memory blocks Thousands of Variable Precision DSP blocks Dozens of High-speed transceivers Multiple High Speed configurable Memory Controllers I/O I/O Multiple ARM© Cores I/O 5 FPGA Architecture: Basic Elements Basic Element 1-bit configurable 1-bit register operation (store result) Configured to perform any 1-bit operation: AND, OR, NOT, ADD, SUB 6 FPGA Architecture: Flexible Interconnect … Basic Elements are surrounded with a flexible interconnect 7 FPGA Architecture: Flexible Interconnect … … Wider custom operations are implemented by configuring and interconnecting Basic Elements 8 FPGA Architecture: Custom Operations Using Basic Elements 16-bit add 32-bit sqrt … … … Your custom 64-bit bit-shuffle and encode Wider custom operations are implemented by configuring and interconnecting Basic Elements 9 FPGA Architecture: Memory Blocks addr Memory Block data_out data_in 20 Kb Can be configured and grouped using the interconnect to create various cache architectures 10 FPGA Architecture: Memory Blocks addr Memory Block data_out data_in 20 Kb Few larger Can be configured and grouped
    [Show full text]
  • Direct Hardware Generation from High-Level Programming Language
    Budapest University of Technology and Economics Department of Control Engineering and Information Technology DIRECT HARDWARE GENERATION FROM HIGH-LEVEL PROGRAMMING LANGUAGE Ph.D. thesis Csák, Bence Consultant: Prof. Dr. Arató, Péter Full member of Hungarian Academy of Sciences Budapest 2009 Tezisfuzet_EN.doc 1/12 1. INTRODUCTION The ever increasing speed and complexity demand of control applications is driving developers to work out newer and newer solutions for ages. In the course of this arrived mankind to the point where after purely mechanical components, electromechanical, electrical and later electronic components are used in the more and more complex control and computing machines. For the sake of flexibility of control systems even in the times of purely mechanical systems, program control has been created, which then spread over to electromechanical and later electronic systems. Development of programming has accelerated truly in the times of electronic computers. Today programs are written at an abstraction level from which no single step is enough to get to the level, where the first programs were written. This high level ensures that extremely big software systems can operate even safety critical systems without malfunction. The unquenchable desire for even higher computation power - despite of the development of software engineering - forces that certain tasks are solved by purpose built hardware. In case of a given task the desired performance is assured on a cost-effective way by composing the system of application specific hardware and CPU based software. Optimal partition of software and hardware parts is aided by the methodology of hardware-software co-design. Here, system specification is given on a standard symbolic language, from which a given partitioning procedure derives hardware and software components.
    [Show full text]
  • Xilinx/Synopsys Interface Guide— ISE 4 Printed in U.S.A
    Xilinx/ Synopsys Interface Guide Xilinx/Synopsys Interface Guide— ISE 4 Printed in U.S.A. Xilinx/Synopsys Interface Guide R The Xilinx logo shown above is a registered trademark of Xilinx, Inc. CoolRunner, RocketChips, RocketIP, Spartan, StateBENCH, StateCAD, Virtex, XACT, XILINX, XC2064, XC3090, XC4005, and XC5210 are registered trademarks of Xilinx, Inc. The shadow X shown above is a trademark of Xilinx, Inc. ACE Controller, ACE Flash, A.K.A. Speed, Alliance Series, AllianceCORE, Bencher, ChipScope, Configurable Logic Cell, CORE Generator, CoreLINX, Dual Block, EZTag, Fast CLK, Fast CONNECT, Fast FLASH, FastMap, Fast Zero Power, Foundation, Gigabit Speeds...and Beyond!, HardWire, HDL Bencher, IRL, J Drive, JBits, LCA, LogiBLOX, Logic Cell, LogiCORE, LogicProfessor, MicroBlaze, MicroVia, MultiLINX, NanoBlaze, PicoBlaze, PLUSASM, PowerGuide, PowerMaze, QPro, Real-PCI, Rocket I/O, Select I/O, SelectRAM, SelectRAM+, Silicon Xpresso, Smartguide, Smart-IP, SmartSearch, SMARTswitch, System ACE, Testbench In A Minute, TrueMap, UIM, VectorMaze, VersaBlock, VersaRing, Wave Table, WebFITTER, WebPACK, WebPOWERED, XABEL, XACTstep Advanced, XACTstep Foundry, XACT-Floorplanner, XACT-Performance, XAM, XAPP, X-BLOX +, XChecker, XDM, XEPLD, Xilinx Foundation Series, Xilinx XDTV, Xinfo, XSI, XtremeDSP, all XC designated products, and ZERO+ are trademarks of Xilinx, Inc. The Programmable Logic Company is a service mark of Xilinx, Inc. All other trademarks are the property of their respective owners. Xilinx, Inc. does not assume any liability arising out of the application or use of any product described or shown herein; nor does it convey any license under its patents, copyrights, or maskwork rights or any rights of others. Xilinx, Inc. reserves the right to make changes, at any time, in order to improve reliability, function or design and to supply the best product possible.
    [Show full text]