International Journal of Soft Computing and Engineering

Total Page:16

File Type:pdf, Size:1020Kb

International Journal of Soft Computing and Engineering International Journal of Innovative Technology and Exploring Engineering (IJITEE) ISSN: 2278-3075, Volume-8 Issue-8S2, June 2019 Analysis of Various Computer Architectures Kulvinder Singh, Pushpendra Kumar Pateriya, Devinder Kaur, Mritunjay Rai Abstract: With the tremendous acceptance of computers globally in the different fields, there comes the challenge of managing II. ISA instructions, processing, and improving the efficiency of computer systems. Ever since the inception of computers from Instruction Set Architecture (ISA) is a mechanism which earliest ENIAC (developed at the University of Pennsylvania) to governs the interaction of hardware and software. ISA modern supercomputer Sunway TaihuLight (developed at comprises of hardware implemented instructions, defined National Supercomputing Center, China) a broad research along addressing modes, types, and sizes of different operands used with great ideas in the field of computer architecture has laid its as well as registers and input-output control [10]. Broadly way. It has lead to the advances which has made the computers capable to perform computations at lightning speeds. For an speaking, an ISA is a set of operations which defines example, the computing speed of TaihuLight exceeds 93 everything a machine language programmer must be familiar petaflops. It uses 40,960 individual RISC processors to reach such with to be able to program a computer [2]. Every ISA has its high speeds. In this, the architecture plays an indispensable role own set of supported operations and valid instructions format right from accessing memory, allocating instruction to processors which are decoded by the control unit of the processor while for execution, and exploiting the architecture to increase the execution is performed by ALU. The first ISA appears on the efficiency [12]. Moreover, reducing energy consumption and cost IBM System 360 in 1964. computer architecture needs to be highly specific and fast in making decisions. With the passage of time, various ideas have come up to enhance this performance index. In this paper, we III. CLASSIFICATION OF ISA describe most frequently used computer architectures with a Amongst the different possible characteristics, the type of general review and working process which can help the readers to instruction set architecture supported can be the most understand the underlying principles of various computer significant feature. It fundamentally defines how the architectures. computer will be executing the instructions. Although, there Index Terms: Instruction Set Architecture, Load/store exist different approaches for comparing instruction set Architecture, Stack Architecture, Pipelining, NOP architectures, a popular approach is based on the architectural complexity. The restriction of fixed instruction length adds I. INTRODUCTION simplicity, also the number of operations supported by that Before the criticism of David Patterson and John L. particular instruction set get reduced. The architecture Hennessy, computer architecture was entirely referred to the designs which follow similar approach are called Reduced instruction set architecture. While, other features of the Instruction Set Computers (RISC)[4]. architecture were called as the implementation. But, the While, an opposite approach also exists alternative to the criticism shifted the view of the community to accept RISC which makes use of complex instructions. In this, size implementation aspect also in the formal definition of of instructions may vary adding capability of specifying computer architecture. Hence, most general approach defines sophisticated and multi step operations. These architecture computer architecture as the architecture which comprises of designs are called Complex Instruction Set Computers the organization of different components, physical design of (CISC) and were widely used before advent of the RISC in the individual hardware elements, and instruction set architecture 1970s. On the other hand, Very Long Instruction Word [1]. Although, hardware and organization paradigms have got (VLIW) architecture relies on the compiler for rearranging advanced, but still general principles have not changed much the program in advance for exploiting instruction-level in their case. On the other hand, ISA (Instruction Set parallelism. It is contrary to other types discussed above Architecture) has introduced new ways of implementing ideas which uses additional hardware to maximize parallel related to the interface between the hardware and software. execution of instructions [5]. Based on complexity, Through past years various attempts have been made in this instruction sets like Minimal Instruction Set Computer direction. In this paper, we have discussed various standard (MISC) have also been studied which have lesser complexity architectures, ideas, their working, historical perspective, than RISC. challenges, and their significance in present scenario as well Some others are One instruction Set Computer (OISC) as the future scope. and Zero Instruction Set Computer (ZISC) which along with many others are theoretically important for research perspective. But, they have insignificant performance and hence not been commercialized [10]. Several other architectures also exist like Transport Triggered Architecture Revised Manuscript Received on May 22, 2019. (TTA) which deals with data transfer buses to enhance the Kulvinder Singh, Lovely Professional University, Punjab, India. Email: performance with the help of direct buses. [email protected] Pushpendra Kumar Pateriya, Lovely Professional University, Punjab, India. E-mail: [email protected] Devinder Kaur, Lovely Professional University, Punjab, India. E-mail: [email protected] Mritunjay Rai, Lovely Professional University, Punjab, India. E-mail: [email protected] Published By: Blue Eyes Intelligence Engineering Retrieval Number: H11890688S219/19©BEIESP 1113 & Sciences Publication Analysis of Various Computer Architectures A. RISC execution of instruction whereas reducing the efforts in A primary set of operations, a compatible, and scalable programming the problem [11]. design can be used to build a computer which can give The CISC architecture consists of a large number of bits in instruction formats, which specifies the type of operation, the effectively faster results with lower latency. The decoding of number, type and position of the operands etc. Unlike RISC, it the instruction is faster due to the fixed set of addressing doesn’t follow load/store architecture and fetches instruction from modes, while the simplicity of instruction set enhances the memory every time it needs to perform some operation with it, refer execution of most instructions in a single machine cycle. In to Table 1 [6]. Code density, addition, such a machine should have a much shorter design i.e. the memory program takes in memory, is more than in CISC than time [6]. As the name suggests it’s architecture enables usage in RISC [12]. of lesser instruction as compared to CISC, the instructions are TABLE I: RISC VS. CISC [13] precise, simple, and general. Table no. I [7] can be referred to compare these two. RISC strictly follows Load/store Sr no. CISC RISC architecture which means that only load and store instructions have the privilege to access memory. No other instruction can 1. Complex instructions Simple instructions with individually access memory by itself [6]. Since RISC design with variable length fixed length is based on load/store design, it is important to provide a compatible and fast memory methodology to work in sync 2. Dont follow Load/Store Strongly follow design Load/Store design with processor [5]. In a RISC system, the memory methodology comprises of effective use management of 3. Pipelining not efficient Highly efficient with registers and caches to increase performance, and write pipelining buffers. Usually, there is also an on-chip memory management unit [8]. 4. Complex addressing Simple and limited ad- In a RISC design, the capability to predict the approximate modes dressing modes time for processing of the instructions enable pipelines to 5. Focuses on embedding Focuses on execution of work efficiently and faster [9]. The term RISC first appeared large number of one instruction per cycle in a paper by David Patterson of the University of operation on a single California, Berkeley, though similar concepts had appeared processor before [1]. A Similar paradigm of designing microprocessor 6. high number of opcode fixed number of opcodes architecture was discussed by John L. Hennessy of Stanford with variable size with fixed sizes University in 1981, with the name of MIPS project where the emphasis was on a simplified instruction set to get high 7. fetches data or every makes use of principle of performance, which became functioning in 1983, and was time locality for lesser interaction with able to run simple programs by 1984 [10]. Henessey an instruction needs it memory explained that, The RISC architecture is simple both in the 8. limited number of large number of registers instruction set and the hardware needed to implement that registers instruction set. Although the MIPS instruction set has a simple hardware implementation (i.e. it requires a minimal 9. Minimum program Increased program length amount of hardware control), the user level instruction set is length and minimum work on instruction and maximum work on not as straightforward, and the simplicity of the user level instruction instruction
Recommended publications
  • Defense Industry Restructuring in Russia
    S t a n f o r d U n i v e r s i t y C I S A C Center for International Security and Arms Control The Center for International Security and Arms Control, part of Stanford University’s Institute for International Studies, is a multidisciplinary community dedicated to research and train- ing in the field of international security. The Center brings together scholars, policymakers, scientists, area specialists, members of the business community, and other experts to examine a wide range of international security issues. CISAC publishes its own series of working papers and reports on its work and also sponsors a series, Studies in International Se- curity and Arms Control, through Stanford University Press. Center for International Security and Arms Control Stanford University 320 Galvez Street Stanford, California 94305-6165 (415) 723-9625 http://www-leland.stanford.edu/group/CISAC/ Contents Acknowledgments iv Executive Summary v I Introduction 1 Section One: Case Studies II The Central Aerohydrodynamic Research Institute (TsAGI) 9 III ELVIS+ and The Moscow Center for SPARC Technology (MCST) 28 IV Impuls 45 V The Mashinostroenie Enterprise 59 VI The Saratov Aviation Plant 79 Section Two: Analysis VII Privatization at Four Enterprises 111 VIII Organizational Restructuring 137 IX Principal Differences in Accounting Systems in Russia 163 and the United States X Reallocation of the Social Services 183 XI Conclusion 207 Glossary 216 1 Acknowledgments Many people have contributed to this report, and still more have contributed to the research leading up to it. In writing this report, we have not attempted to reach consensus among the authors on the interpretations to be drawn from the data.
    [Show full text]
  • CS 2210: Theory of Computation
    CS 2210: Theory of Computation Spring, 2019 Administrative Information • Background survey • Textbook: E. Rich, Automata, Computability, and Complexity: Theory and Applications, Prentice-Hall, 2008. • Book website: http://www.cs.utexas.edu/~ear/cs341/automatabook/ • My Office Hours: – Monday, 6:00-8:00pm, Searles 224 – Tuesday, 1:00-2:30pm, Searles 222 • TAs – Anjulee Bhalla: Hours TBA, Searles 224 – Ryan St. Pierre, Hours TBA, Searles 224 What you can expect from the course • How to do proofs • Models of computation • What’s the difference between computability and complexity? • What’s the Halting Problem? • What are P and NP? • Why do we care whether P = NP? • What are NP-complete problems? • Where does this make a difference outside of this class? • How to work the answers to these questions into the conversation at a cocktail party… What I will expect from you • Problem Sets (25%): – Goal: Problems given on Mondays and Wednesdays – Due the next Monday – Graded by following Monday – A learning tool, not a testing tool – Collaboration encouraged; more on this in next slide • Quizzes (15%) • Exams (2 non-cumulative, 30% each): – Closed book, closed notes, but… – Can bring in 8.5 x 11 page with notes on both sides • Class participation: Tiebreaker Other Important Things • Go to the TA hours • Study and work on problem sets in groups • Collaboration Issues: – Level 0 (In-Class Problems) • No restrictions – Level 1 (Homework Problems) • Verbal collaboration • But, individual write-ups – Level 2 (not used in this course) • Discussion with TAs only – Level 3 (Exams) • Professor clarifications only Right now… • What does it mean to study the “theory” of something? • Experience with theory in other disciplines? • Relationship to practice? – “In theory, theory and practice are the same.
    [Show full text]
  • The Problem with Threads
    The Problem with Threads Edward A. Lee Electrical Engineering and Computer Sciences University of California at Berkeley Technical Report No. UCB/EECS-2006-1 http://www.eecs.berkeley.edu/Pubs/TechRpts/2006/EECS-2006-1.html January 10, 2006 Copyright © 2006, by the author(s). All rights reserved. Permission to make digital or hard copies of all or part of this work for personal or classroom use is granted without fee provided that copies are not made or distributed for profit or commercial advantage and that copies bear this notice and the full citation on the first page. To copy otherwise, to republish, to post on servers or to redistribute to lists, requires prior specific permission. Acknowledgement This work was supported in part by the Center for Hybrid and Embedded Software Systems (CHESS) at UC Berkeley, which receives support from the National Science Foundation (NSF award No. CCR-0225610), the State of California Micro Program, and the following companies: Agilent, DGIST, General Motors, Hewlett Packard, Infineon, Microsoft, and Toyota. The Problem with Threads Edward A. Lee Professor, Chair of EE, Associate Chair of EECS EECS Department University of California at Berkeley Berkeley, CA 94720, U.S.A. [email protected] January 10, 2006 Abstract Threads are a seemingly straightforward adaptation of the dominant sequential model of computation to concurrent systems. Languages require little or no syntactic changes to sup- port threads, and operating systems and architectures have evolved to efficiently support them. Many technologists are pushing for increased use of multithreading in software in order to take advantage of the predicted increases in parallelism in computer architectures.
    [Show full text]
  • Computer Architecture: Dataflow (Part I)
    Computer Architecture: Dataflow (Part I) Prof. Onur Mutlu Carnegie Mellon University A Note on This Lecture n These slides are from 18-742 Fall 2012, Parallel Computer Architecture, Lecture 22: Dataflow I n Video: n http://www.youtube.com/watch? v=D2uue7izU2c&list=PL5PHm2jkkXmh4cDkC3s1VBB7- njlgiG5d&index=19 2 Some Required Dataflow Readings n Dataflow at the ISA level q Dennis and Misunas, “A Preliminary Architecture for a Basic Data Flow Processor,” ISCA 1974. q Arvind and Nikhil, “Executing a Program on the MIT Tagged- Token Dataflow Architecture,” IEEE TC 1990. n Restricted Dataflow q Patt et al., “HPS, a new microarchitecture: rationale and introduction,” MICRO 1985. q Patt et al., “Critical issues regarding HPS, a high performance microarchitecture,” MICRO 1985. 3 Other Related Recommended Readings n Dataflow n Gurd et al., “The Manchester prototype dataflow computer,” CACM 1985. n Lee and Hurson, “Dataflow Architectures and Multithreading,” IEEE Computer 1994. n Restricted Dataflow q Sankaralingam et al., “Exploiting ILP, TLP and DLP with the Polymorphous TRIPS Architecture,” ISCA 2003. q Burger et al., “Scaling to the End of Silicon with EDGE Architectures,” IEEE Computer 2004. 4 Today n Start Dataflow 5 Data Flow Readings: Data Flow (I) n Dennis and Misunas, “A Preliminary Architecture for a Basic Data Flow Processor,” ISCA 1974. n Treleaven et al., “Data-Driven and Demand-Driven Computer Architecture,” ACM Computing Surveys 1982. n Veen, “Dataflow Machine Architecture,” ACM Computing Surveys 1986. n Gurd et al., “The Manchester prototype dataflow computer,” CACM 1985. n Arvind and Nikhil, “Executing a Program on the MIT Tagged-Token Dataflow Architecture,” IEEE TC 1990.
    [Show full text]
  • Ece585 Lec2.Pdf
    ECE 485/585 Microprocessor System Design Lecture 2: Memory Addressing 8086 Basics and Bus Timing Asynchronous I/O Signaling Zeshan Chishti Electrical and Computer Engineering Dept Maseeh College of Engineering and Computer Science Source: Lecture based on materials provided by Mark F. Basic I/O – Part I ECE 485/585 Outline for next few lectures Simple model of computation Memory Addressing (Alignment, Byte Order) 8088/8086 Bus Asynchronous I/O Signaling Review of Basic I/O How is I/O performed Dedicated/Isolated /Direct I/O Ports Memory Mapped I/O How do we tell when I/O device is ready or command complete? Polling Interrupts How do we transfer data? Programmed I/O DMA ECE 485/585 Simplified Model of a Computer Control Control Data, Address, Memory Data Path Microprocessor Keyboard Mouse [Fetch] Video display [Decode] Printer [Execute] I/O Device Hard disk drive Audio card Ethernet WiFi CD R/W DVD ECE 485/585 Memory Addressing Size of operands Bytes, words, long/double words, quadwords 16-bit half word (Intel: word) 32-bit word (Intel: doubleword, dword) 0x107 64-bit double word (Intel: quadword, qword) 0x106 Note: names are non-standard 0x105 SUN Sparc word is 32-bits, double is 64-bits 0x104 0x103 Alignment 0x102 Can multi-byte operands begin at any byte address? 0x101 Yes: non-aligned 0x100 No: aligned. Low order address bit(s) will be zero ECE 485/585 Memory Operand Alignment …Intel IA speak (i.e. word = 16-bits = 2 bytes) 0x107 0x106 0x105 0x104 0x103 0x102 0x101 0x100 Aligned Unaligned Aligned Unaligned Aligned Unaligned word word Double Double Quad Quad address address word word word word -----0 address address address address -----00 ----000 ECE 485/585 Memory Operand Alignment Why do we care? Unaligned memory references Can cause multiple memory bus cycles for a single operand May also span cache lines Requiring multiple evictions, multiple cache line fills Complicates memory system and cache controller design Some architectures restrict addresses to be aligned Even in architectures without alignment restrictions (e.g.
    [Show full text]
  • 10. Assembly Language, Models of Computation
    10. Assembly Language, Models of Computation 6.004x Computation Structures Part 2 – Computer Architecture Copyright © 2015 MIT EECS 6.004 Computation Structures L10: Assembly Language, Models of Computation, Slide #1 Beta ISA Summary • Storage: – Processor: 32 registers (r31 hardwired to 0) and PC – Main memory: Up to 4 GB, 32-bit words, 32-bit byte addresses, 4-byte-aligned accesses OPCODE rc ra rb unused • Instruction formats: OPCODE rc ra 16-bit signed constant 32 bits • Instruction classes: – ALU: Two input registers, or register and constant – Loads and stores: access memory – Branches, Jumps: change program counter 6.004 Computation Structures L10: Assembly Language, Models of Computation, Slide #2 Programming Languages 32-bit (4-byte) ADD instruction: 1 0 0 0 0 0 0 0 1 0 0 0 0 0 1 0 0 0 0 1 1 0 0 0 0 0 0 0 0 0 0 0 opcode rc ra rb (unused) Means, to the BETA, Reg[4] ß Reg[2] + Reg[3] We’d rather write in assembly language: Today ADD(R2, R3, R4) or better yet a high-level language: Coming up a = b + c; 6.004 Computation Structures L10: Assembly Language, Models of Computation, Slide #3 Assembly Language Symbolic 01101101 11000110 Array of bytes representation Assembler 00101111 to be loaded of stream of bytes 10110001 into memory ..... Source Binary text file machine language • Abstracts bit-level representation of instructions and addresses • We’ll learn UASM (“microassembler”), built into BSim • Main elements: – Values – Symbols – Labels (symbols for addresses) – Macros 6.004 Computation Structures L10: Assembly Language, Models
    [Show full text]
  • Winter 2017/18
    Newsletter - Calibrated for Creative Communications Vol. 15, no. 4, Winter 2017/18 Schneeberg seen from the Wienerwald, Nov. 2017 Vantage Point IT STAR representatives nd-of-year provides an excellent vantage point to assess Eachievements and shortcomings during the exiting year and Austria/OCG-R. Bieber, Bulgaria/BAS- I. Dimov, to look forward into the future. Croatia/CITA-M. Frkovic, Cyprus/CCS-P. Masouras, Czech Rep./CSKI-J. Stuller, Greece/GCS-S. Katsikas, Based on feedback from our readers, we can tick the 2017 per- Hungary/NJSZT-B. Domolki, Italy/AICA-G. Occhini, formance of the Newsletter as successful. Our Honorary Advi- Lithuania/LIKS-E. Telešius, Macedonia/MASIT- sory Board is growing and we are happy to have a stronger fe- P. Indovski, Poland/PIPS-M. Holynski, Romania/ male representation of the informatics profession. ATIC-V. Baltac, Serbia/JISA-D. Dukic, Slovakia/SSCS- I. Privara, Slovenia/SSI-N. Schlamberger In this last 2017 Issue we are pleased to introduce our new mem- bers of the NL Advisory Board, to brief you on the Russian Vir- Contents tual Computer Museum and to provide an update on develop- New members of the Advisory Board ........................... 2 ments related to Big Data and Competences. IT STAR WS and BM in Sofia ...................................... 3 Big Data: The Data Scientist ........................................ 6 Regarding IT STAR, 2017 was rather challenging and efforts were made to build upon successful activities so as to recalibrate Russian Virtual Computer Museum ............................. 9 the Association to the benefit of its members and the larger ICT ITU Information Society Report 2017 ......................... 13 community. MS News ...................................................................
    [Show full text]
  • Actor Model of Computation
    Published in ArXiv http://arxiv.org/abs/1008.1459 Actor Model of Computation Carl Hewitt http://carlhewitt.info This paper is dedicated to Alonzo Church and Dana Scott. The Actor model is a mathematical theory that treats “Actors” as the universal primitives of concurrent digital computation. The model has been used both as a framework for a theoretical understanding of concurrency, and as the theoretical basis for several practical implementations of concurrent systems. Unlike previous models of computation, the Actor model was inspired by physical laws. It was also influenced by the programming languages Lisp, Simula 67 and Smalltalk-72, as well as ideas for Petri Nets, capability-based systems and packet switching. The advent of massive concurrency through client- cloud computing and many-core computer architectures has galvanized interest in the Actor model. An Actor is a computational entity that, in response to a message it receives, can concurrently: send a finite number of messages to other Actors; create a finite number of new Actors; designate the behavior to be used for the next message it receives. There is no assumed order to the above actions and they could be carried out concurrently. In addition two messages sent concurrently can arrive in either order. Decoupling the sender from communications sent was a fundamental advance of the Actor model enabling asynchronous communication and control structures as patterns of passing messages. November 7, 2010 Page 1 of 25 Contents Introduction ............................................................ 3 Fundamental concepts ............................................ 3 Illustrations ............................................................ 3 Modularity thru Direct communication and asynchrony ............................................................. 3 Indeterminacy and Quasi-commutativity ............... 4 Locality and Security ............................................
    [Show full text]
  • Computer Architectures an Overview
    Computer Architectures An Overview PDF generated using the open source mwlib toolkit. See http://code.pediapress.com/ for more information. PDF generated at: Sat, 25 Feb 2012 22:35:32 UTC Contents Articles Microarchitecture 1 x86 7 PowerPC 23 IBM POWER 33 MIPS architecture 39 SPARC 57 ARM architecture 65 DEC Alpha 80 AlphaStation 92 AlphaServer 95 Very long instruction word 103 Instruction-level parallelism 107 Explicitly parallel instruction computing 108 References Article Sources and Contributors 111 Image Sources, Licenses and Contributors 113 Article Licenses License 114 Microarchitecture 1 Microarchitecture In computer engineering, microarchitecture (sometimes abbreviated to µarch or uarch), also called computer organization, is the way a given instruction set architecture (ISA) is implemented on a processor. A given ISA may be implemented with different microarchitectures.[1] Implementations might vary due to different goals of a given design or due to shifts in technology.[2] Computer architecture is the combination of microarchitecture and instruction set design. Relation to instruction set architecture The ISA is roughly the same as the programming model of a processor as seen by an assembly language programmer or compiler writer. The ISA includes the execution model, processor registers, address and data formats among other things. The Intel Core microarchitecture microarchitecture includes the constituent parts of the processor and how these interconnect and interoperate to implement the ISA. The microarchitecture of a machine is usually represented as (more or less detailed) diagrams that describe the interconnections of the various microarchitectural elements of the machine, which may be everything from single gates and registers, to complete arithmetic logic units (ALU)s and even larger elements.
    [Show full text]
  • Computer Architecture Lecture 1: Introduction and Basics Where We Are
    Computer Architecture Lecture 1: Introduction and Basics Where we are “C” as a model of computation Programming Programmer’s view of a computer system works How does an assembly program end up executing as Architect/microarchitect’s view: digital logic? How to design a computer that meets system design goals. What happens in-between? Choices critically affect both How is a computer designed the SW programmer and using logic gates and wires the HW designer to satisfy specific goals? HW designer’s view of a computer system works CO Digital logic as a model of computation 2 Levels of Transformation “The purpose of computing is insight” (Richard Hamming) We gain and generate insight by solving problems How do we ensure problems are solved by electrons? Problem Algorithm Program/Language Runtime System (VM, OS, MM) ISA (Architecture) Microarchitecture Logic Circuits Electrons 3 The Power of Abstraction Levels of transformation create abstractions Abstraction: A higher level only needs to know about the interface to the lower level, not how the lower level is implemented E.g., high-level language programmer does not really need to know what the ISA is and how a computer executes instructions Abstraction improves productivity No need to worry about decisions made in underlying levels E.g., programming in Java vs. C vs. assembly vs. binary vs. by specifying control signals of each transistor every cycle Then, why would you want to know what goes on underneath or above? 4 Crossing the Abstraction Layers As long as everything goes well, not knowing what happens in the underlying level (or above) is not a problem.
    [Show full text]
  • Open Access Proceedings Journal of Physics: Conference Series
    IOP Conference Series: Earth and Environmental Science PAPER • OPEN ACCESS Issues of compatibility of processor command architectures To cite this article: T R Zmyzgova et al 2020 IOP Conf. Ser.: Earth Environ. Sci. 421 042006 View the article online for updates and enhancements. This content was downloaded from IP address 85.143.35.46 on 13/02/2020 at 06:53 AGRITECH-II-2019 IOP Publishing IOP Conf. Series: Earth and Environmental Science 421 (2020) 042006 doi:10.1088/1755-1315/421/4/042006 Issues of compatibility of processor command architectures T R Zmyzgova, A V Solovyev, A G Rabushko, A A Medvedev and Yu V Adamenko Kurgan State University, 62 Proletarskaya street, Kurgan, 640002, Russia E-mail: [email protected] Abstract. Modern computers and computing devices are based on the principle of open architecture, according to which the computer consists of several sufficiently independent devices that perform a certain function. These devices must meet certain standards of interaction with each other. Existing standards relate to both the technical characteristics of the devices and the content of the signals exchanged between them. The article considers the issue of creating a universal architecture of processor commands. The brief analysis of the components of devices and interrelations between them (different hardware features of processors, architecture of memory models and registers of peripheral devices, mechanisms of operand processing, the number of registers and processed data types, interruptions, exceptions, etc.) is carried out. The problem of architecture standardization, which generalizes the capabilities of the most common architectures and is suitable for high-performance emulation on most computer architectures, is put.
    [Show full text]
  • Background Optimization in Full System Binary Translation
    Background Optimization in Full System Binary Translation Roman A. Sokolov Alexander V. Ermolovich MCST CJSC Intel CJSC Moscow, Russia Moscow, Russia Email: [email protected] Email: [email protected] Abstract—Binary translation and dynamic optimization are architecture codes fully utilizing all architectural features widely used to provide compatibility between legacy and promis- introduced to support binary translation. Besides, dynamic ing upcoming architectures on the level of executable binary optimization can benefit from utilization of actual information codes. Dynamic optimization is one of the key contributors to dynamic binary translation system performance. At the same about executables behavior which static compilers usually time it can be a major source of overhead, both in terms of don’t possess. CPU cycles and whole system latency, as long as optimization At the same time dynamic optimization can imply sig- time is included in the execution time of the application under nificant overhead as long as optimization time is included translation. One of the solutions that allow to eliminate dynamic in the execution time of application under translation. Total optimization overhead is to perform optimization simultaneously with the execution, in a separate thread. In the paper we present optimization time can be significant but will not necessarily implementation of this technique in full system dynamic binary be compensated by the translated codes speed-up if application translator. For this purpose, an infrastructure for multithreaded run time is too short. execution was implemented in binary translation system. This Also, the operation of optimizing translator can worsen the allowed running dynamic optimization in a separate thread latency (i.e., increase pause time) of interactive application or independently of and concurrently with the main thread of execution of binary codes under translation.
    [Show full text]