The Virtualization Spectrum from Hyperthreads to Grids†

Total Page:16

File Type:pdf, Size:1020Kb

The Virtualization Spectrum from Hyperthreads to Grids† THE VIRTUALIZATION SPECTRUM FROM HYPERTHREADS TO GRIDS† Neil J. Gunther Performance Dynamics Company, Castro Valley, California, USA www.perfdynamics.com Modern virtual machines (VMs) can be opaque to conventional performance management tools because VM technology has surpassed standard measurement paradigms. We attempt to ameliorate that problem by (1) observing that disparate types of VMs lie on a discrete spectrum bounded by hyperthreading at one extreme and GRID-like services at the other, and (2) recognizing that poll-based scheduling is the common architectural element in many VM implementations. The associ- ated polling frequency (from GHz to µHz) positions each VM in its respective region of the VM-spectrum. Several case studies are analyzed to illustrate how this insight can make VMs more visible to performance management techniques. 1 INTRODUCTION The inclusion of GRIDs and P2P under the umbrella of virtualization is an unusual step, but our intent is to use the VM-spectrum both as a classification scheme to Virtualization remains a hot topic because of the opacity organize the discussion of an otherwise bewildering array of virtual systems from the standpoint of conventional of VM architectures, and as a quantitative framework for performance management. These difficulties are further explaining previously reported performance anomalies in exacerbated by the appearance of so many disparate im- a variety of controlled measurements on VM systems. plementations employing the adjective virtual [See e.g., Sin04]. Moreover, modern computer architectures that Most previous discussions have tended to organize VM create virtual resources and services out of physical re- capacity planning issues according to whether they are sources are quite distinct from the more familiar virtual- implemented in hardware or software [See e.g., Joh03, ization paradigms e.g., virtual memory or virtual storage. DBK03, Fri03, Bra05, Fer05]. Macro-VMs have not fea- tured in such discussions, as far as I am aware. Con- This paper proposes a more unified picture of mod- sequently, the inability to achieve full virtual capacity ern virtualization by recognizing that many of these ap- in hyperthreaded hardware and anomalous performance parently disparate forms of virtualized resources or vir- outcomes in software hypervisors, have been presented tual machines (VMs) can be considered to lie on a dis- as distinct performance effects. crete spectrum—the virtual machine spectrum or VM- spectrum—comprised of three principal regions: The VM-spectrum paradigm, however, views these ef- 1. Micro-VMs: represented by the hyperthreaded mul- fects as arising from an architectural feature that is com- ticore processors discussed in Section 3. mon to both hardware and software implementations; a form of scheduling which we define in Section 2.2 as 2. Meso-VMs: represented by virtual machine moni- proportional polling. In particular, it provides a simple tors and hypervisors discussed in Section 4. explanation for the observed Missing MIPS problem in 3. Macro-VMs: represented by the GRID services and Sect. 3.2.2. The VM-spectrum also leads to the no- peer-to-peer (P2P) architectures in Section 5. tion that performance management of modern VMs is a † Copyright c 2006 Performance Dynamics Company and function of the time and distance scales on which their CMG, Inc. Accepted for publication in the Proceedings of the respective polling mechanisms operate. Computer Measurement Group (CMG) 32nd International Confer- ence, Reno NV, December 3–8, 2006. Version of September 11, The paper is organized as follows. Section 2 defines the 2006 The Virtualization Spectrum rationale for the VM-spectrum and its three principal re- gions. Each of these VM-regions is examined in detail, Micro P2P starting in Section 3 with the micro-VM scale. The best waves known VM implementations in this region are virtual pro- 1. Micro-VM: represented by cessors or hyperthreaded CPUs. Virtual processors also GRIDs Macro Scale: constitute the first example of a polling-based scheduler Long wavelengths, operating in the GHz to kHz frequency range. The as- hyperthreaded multicore sociated performance case studies and recommendations IR PVM Low frequencies are presented in Section 3.2. Section 4 moves up the VM-spectrum (Fig. 1) to slower polling meso-VMs repre- sented by virtual machine monitors or hypervisors. These processors VMs constitute the second example of a polling-based Sysplex scheduler operating in the kHz (kilohertz) to mHz (mil- lihertz) frequency range. The associated performance analysis case studies and recommendations are presented VMWare in Section 4.2. Section 5 discusses the slowest polling macro-VM services represented by GRIDs and P2P net- works. At this scale, different types of state information Xen are collected via polling mechanisms which can have op- 2. Meso-VM: represented by erating periods in the range of days! Obviously, these Meso Scale: frequency ranges can have a significant impact on the Visible FSS Medium wavelengths, performance and scalability of virtual services. The as- virtual machine monitors and sociated performance analysis case studies are presented Visible frequencies in Section 5.2. Section 6 presents conclusions and rec- TSS ommendations. operating systems LPAR 2 VIRTUALIZATION SPECTRUM Virtualization is about creating illusions. In particular, z/VM modern computer systems are now sufficiently powerful to present users with the illusion that one physical ma- UV Mach 3. Macro-VM: represented by chine is really multiple virtual machines, each one run- Micro Scale: ning a separate instances of a different operating system Short wavelengths, (OS). This is one reason for the resurgence of interest High frequencies GRID services and P2P in virtualization technologies. In the interests of space, Micro we forego any review of the history or relative merits of code various VM projects and products, which are adequately discussed elsewhere [See e.g., Sin04, Fer05, and refer- architectures ences therein]. X-rays HTT As many authors have already noted, the idea of creat- ing virtual resources e.g., software emulators and virtual memory, is not new. For the purposes of this paper, how- Figure 1: The continuous EM-spectrum compared with 1 ever, we draw a distinction between modern VMs and the discrete VM-spectrum. In both kinds of spectra, older concepts of virtualized resources. Virtual mem- the relative location of each region is determined by ory, for example, is fundamentally different from VMs their respective frequency scales (Table 1). For the VM- in that it is intrinsically unstable and subject to thrash- spectrum, the frequency is set by the VM polling rate. ing [Gun95]. The isolation of modern VMs often helps Like the invisible regions of the EM-spectrum, the mi- to inhibit this kind of instability. (See Sect. 4) cro and macro regions of the VM-spectrum are also less visible to standard performance management tools. The distinctive architectural feature which modern VMs have in common is some form of polling mechanism by the underlying scheduling subsystem. To make this Table 1: VM-spectrum Scales statement more quantitative, we refer to a case where Spectral Distance Polling the polling periods are well documented [Gun99]: meso- Region Scale (m) Period Frequency VM scheduling (see Sect. 4.1). The polling period Tp, for Macro 102 to 106 min to day mHz to µHz the scheduler to associate physical resource consumption Meso 100 to 102 ms to min kHz to mHz with a each active OS instance (software VM), is once −6 −3 Micro 10 to 10 ns to µs GHz to MHz every 4000 ms or Tp = 4 s. The frequency is therefore f = 1/Tp = 0.25 cycles per second or 250 mHz. to accomplish resource sharing. Polling algorithms are We assume (because it is not documented) that the intrinsically stable and fair; even round-robin polling micro-VM polling period lies in the range of ns (the exhibits intrinsic fairness. The potential performance processor GHz clock frequency) to µs (MHz frequency). penalty arises from the sequential nature of polling be- Macro-VMs can take minutes or days to detect active ing worse than asynchronous communications. Based peer horizons. These frequencies are key VM perfor- on this distinction, we can classify the variety of VM mance determinants. For those who prefer to think in manifestations on a discrete spectrum, analogous to the terms of size, a distance scale d (in meters) can be loosely continuous electromagnetic or EM-spectrum (Fig. 1). related to the period Tp by d = vTp where v is the phase velocity of the communication signal. Typically, v = c; the speed of light. All these scales are summarized in 2.1 Principal VM Regions Table 1. Just as the EM-spectrum can be grouped into the ultra- violet (UV), visible and infra-red (IR) regions, the VM- 3 MICRO-SCALE: HYPERTHREADS spectrum can be similarly grouped into the micro-VM, meso-VM, and macro-VM spectral regions. The relative We begin a detailed analysis of VMs starting with the position of each VM is defined by their respective polling highest frequency (smallest size) scale on the bottom of rate or frequency scale in Table 1. the VM-spectrum in Fig. 1; virtualization of physical pro- cessing resources. Intel, for example, refers to this form The so-called visible region on the EM-spectrum is an of processor virtualization as hyper-threading technology anthropocentric term. Certain snakes can see in the IR (HTT)) or multithreading (MT) on its XeonTM and Pen- (detect heat) and bees can see in UV light. Similarly, tium 4TM product lines. The rationale is to maximize only meso-VMs are “visible” to us via conventional per- throughput performance by utilizing idle cycles. Part of formance management tools, in the sense of providing an the current confusion over hyperthreading performance immediate view of performance and thereby some level stems from two possible views of what HTT offers: of potential control.
Recommended publications
  • Cluster, Grid and Cloud Computing: a Detailed Comparison
    The 6th International Conference on Computer Science & Education (ICCSE 2011) August 3-5, 2011. SuperStar Virgo, Singapore ThC 3.33 Cluster, Grid and Cloud Computing: A Detailed Comparison Naidila Sadashiv S. M Dilip Kumar Dept. of Computer Science and Engineering Dept. of Computer Science and Engineering Acharya Institute of Technology University Visvesvaraya College of Engineering (UVCE) Bangalore, India Bangalore, India [email protected] [email protected] Abstract—Cloud computing is rapidly growing as an alterna- with out any prior reservation and hence eliminates over- tive to conventional computing. However, it is based on models provisioning and improves resource utilization. like cluster computing, distributed computing, utility computing To the best of our knowledge, in the literature, only a few and grid computing in general. This paper presents an end-to- end comparison between Cluster Computing, Grid Computing comparisons have been appeared in the field of computing. and Cloud Computing, along with the challenges they face. This In this paper we bring out a complete comparison of the could help in better understanding these models and to know three computing models. Rest of the paper is organized as how they differ from its related concepts, all in one go. It also follows. The cluster computing, grid computing and cloud discusses the ongoing projects and different applications that use computing models are briefly explained in Section II. Issues these computing models as a platform for execution. An insight into some of the tools which can be used in the three computing and challenges related to these computing models are listed models to design and develop applications is given.
    [Show full text]
  • Grid Computing: What Is It, and Why Do I Care?*
    Grid Computing: What Is It, and Why Do I Care?* Ken MacInnis <[email protected]> * Or, “Mi caja es su caja!” (c) Ken MacInnis 2004 1 Outline Introduction and Motivation Examples Architecture, Components, Tools Lessons Learned and The Future Questions? (c) Ken MacInnis 2004 2 What is “grid computing”? Many different definitions: Utility computing Cycles for sale Distributed computing distributed.net RC5, SETI@Home High-performance resource sharing Clusters, storage, visualization, networking “We will probably see the spread of ‘computer utilities’, which, like present electric and telephone utilities, will service individual homes and offices across the country.” Len Kleinrock (1969) The word “grid” doesn’t equal Grid Computing: Sun Grid Engine is a mere scheduler! (c) Ken MacInnis 2004 3 Better definitions: Common protocols allowing large problems to be solved in a distributed multi-resource multi-user environment. “A computational grid is a hardware and software infrastructure that provides dependable, consistent, pervasive, and inexpensive access to high-end computational capabilities.” Kesselman & Foster (1998) “…coordinated resource sharing and problem solving in dynamic, multi- institutional virtual organizations.” Kesselman, Foster, Tuecke (2000) (c) Ken MacInnis 2004 4 New Challenges for Computing Grid computing evolved out of a need to share resources Flexible, ever-changing “virtual organizations” High-energy physics, astronomy, more Differing site policies with common needs Disparate computing needs
    [Show full text]
  • Cloud Computing Over Cluster, Grid Computing: a Comparative Analysis
    Journal of Grid and Distributed Computing Volume 1, Issue 1, 2011, pp-01-04 Available online at: http://www.bioinfo.in/contents.php?id=92 Cloud Computing Over Cluster, Grid Computing: a Comparative Analysis 1Indu Gandotra, 2Pawanesh Abrol, 3 Pooja Gupta, 3Rohit Uppal and 3Sandeep Singh 1Department of MCA, MIET, Jammu 2Department of Computer Science & IT, Jammu Univ, Jammu 3Department of MCA, MIET, Jammu e-mail: [email protected], [email protected], [email protected], [email protected], [email protected] Abstract—There are dozens of definitions for cloud Virtualization is a technology that enables sharing of computing and through each definition we can get the cloud resources. Cloud computing platform can become different idea about what a cloud computing exacting is? more flexible, extensible and reusable by adopting the Cloud computing is not a very new concept because it is concept of service oriented architecture [5].We will not connected to grid computing paradigm whose concept came need to unwrap the shrink wrapped software and install. into existence thirteen years ago. Cloud computing is not only related to Grid Computing but also to Utility computing The cloud is really very easier, just to install single as well as Cluster computing. Cloud computing is a software in the centralized facility and cover all the computing platform for sharing resources that include requirements of the company’s users [1]. software’s, business process, infrastructures and applications. Cloud computing also relies on the technology II. CLUSTER COMPUTING of virtualization. In this paper, we will discuss about Grid computing, Cluster computing and Cloud computing i.e.
    [Show full text]
  • “Grid Computing”
    VISHVESHWARAIAH TECHNOLOGICAL UNIVERSITY S.D.M COLLEGE OF ENGINEERING AND TECHNOLOGY A seminar report on “Grid Computing” Submitted by Nagaraj Baddi (2SD07CS402) 8th semester DEPARTMENT OF COMPUTER SCIENCE ENGINEERING 2009-10 1 VISHVESHWARAIAH TECHNOLOGICAL UNIVERSITY S.D.M COLLEGE OF ENGINEERING AND TECHNOLOGY DEPARTMENT OF COMPUTER SCIENCE ENGINEERING CERTIFICATE Certified that the seminar work entitled “Grid Computing” is a bonafide work presented by Mr. Nagaraj.M.Baddi, bearing USN 2SD07CS402 in a partial fulfillment for the award of degree of Bachelor of Engineering in Computer Science Engineering of the Vishveshwaraiah Technological University Belgaum, during the year 2009-10. The seminar report has been approved as it satisfies the academic requirements with respect to seminar work presented for the Bachelor of Engineering Degree. Staff in charge H.O.D CSE (S. L. DESHPANDE) (S. M. JOSHI) Name: Nagaraj M. Baddi USN: 2SD07CS402 2 INDEX 1. Introduction 4 2. History 5 3. How Grid Computing Works 6 4. Related technologies 8 4.1 Cluster computing 8 4.2 Peer-to-peer computing 9 4.3 Internet computing 9 5. Grid Computing Logical Levels 10 5.1 Cluster Grid 10 5.2 Campus Grid 10 5.3 Global Grid 10 6. Grid Architecture 11 6.1 Grid fabric 11 6.2 Core Grid middleware 12 6.3 User-level Grid middleware 12 6.4 Grid applications and portals. 13 7. Grid Applications 13 7.1 Distributed supercomputing 13 7.2 High-throughput computing 14 7.3 On-demand computing 14 7.4 Data-intensive computing 14 7.5 Collaborative computing 15 8. Difference: Grid Computing vs Cloud Computing 15 9.
    [Show full text]
  • Computer Systems Architecture
    CS 352H: Computer Systems Architecture Topic 14: Multicores, Multiprocessors, and Clusters University of Texas at Austin CS352H - Computer Systems Architecture Fall 2009 Don Fussell Introduction Goal: connecting multiple computers to get higher performance Multiprocessors Scalability, availability, power efficiency Job-level (process-level) parallelism High throughput for independent jobs Parallel processing program Single program run on multiple processors Multicore microprocessors Chips with multiple processors (cores) University of Texas at Austin CS352H - Computer Systems Architecture Fall 2009 Don Fussell 2 Hardware and Software Hardware Serial: e.g., Pentium 4 Parallel: e.g., quad-core Xeon e5345 Software Sequential: e.g., matrix multiplication Concurrent: e.g., operating system Sequential/concurrent software can run on serial/parallel hardware Challenge: making effective use of parallel hardware University of Texas at Austin CS352H - Computer Systems Architecture Fall 2009 Don Fussell 3 What We’ve Already Covered §2.11: Parallelism and Instructions Synchronization §3.6: Parallelism and Computer Arithmetic Associativity §4.10: Parallelism and Advanced Instruction-Level Parallelism §5.8: Parallelism and Memory Hierarchies Cache Coherence §6.9: Parallelism and I/O: Redundant Arrays of Inexpensive Disks University of Texas at Austin CS352H - Computer Systems Architecture Fall 2009 Don Fussell 4 Parallel Programming Parallel software is the problem Need to get significant performance improvement Otherwise, just use a faster uniprocessor,
    [Show full text]
  • Strategies for Managing Business Disruption Due to Grid Computing
    Strategies for managing business disruption due to Grid Computing by Vidyadhar Phalke Ph.D. Computer Science, Rutgers University, New Jersey, 1995 M.S. Computer Science, Rutgers University, New Jersey, 1992 B.Tech. Computer Science, Indian Institute of Technology, Delhi, 1989 Submitted to the MIT Sloan School of Management in Partial Fulfillment of the Requirements for the Degree of Master of Science in the Management of Technology at the Massachusetts Institute of Technology June 2003 © 2003 Vidyadhar Phalke All Rights Reserved The author hereby grants to MIT permission to reproduce and to distribute publicly paper and electronic copies of this thesis document in whole or in part Signature of Author: MIT Sloan School of Management 9 May 2003 Certified By: Starling D. Hunter III Theodore T. Miller Career Development Assistant Professor Thesis Supervisor Accepted By: David A. Weber Director, Management of Technology Program 2 Strategies for managing business disruption due to Grid Computing by Vidyadhar Phalke Submitted to the MIT Sloan School of Management on May 9 2003 in partial fulfillment of the requirements for the degree of Master of Science in the Management of Technology ABSTRACT In the technology centric businesses disruptive technologies displace incumbents time and again, sometimes to the extent that incumbents go bankrupt. In this thesis we would address the issue of what strategies are essential to prepare for and to manage disruptions for the affected businesses and industries. Specifically we will look at grid computing that is poised to disrupt (1) certain Enterprise IT departments, and (2) the software industry in the high-performance and web services space.
    [Show full text]
  • Taxonomy of Flynn (1966)
    Taxonomy of Flynn (1966). To describe these non-von Neumann or parallel architectures, a generally accepted taxonomy is that of Flynn (1966). The classification is based on the notion of two streams of information flow to a processor: instructions and data. These two streams can be either single or multiple, giving four classes of machines: 1. Single instruction single data (SISD) 2. Single instruction multiple data (SIMD) 3. Multiple instruction single data (MISD) 4. Multiple instruction multiple data (MIMD) Table 5.4 shows the four primary classes and some of the architectures that fit in those classes. Most of these architectures will be briefly discussed. 5.4.2.1 Single Instruction Single Data The SISD architectures encompass standard serial von Neumann architecture computers. In a sense, the SISD category is the base metric for Flynn’s taxonomy. 5.4.2.2 Single Instruction Multiple Data The SIMD computers are essentially array processors. This type of parallel computer architecture has n-processors, each executing the same instruction, but on different data streams. Often each element in the array can only communicate with its nearest neighbour. Computer architectures that are usually classified as SIMD are the systolic and wave-front array computers. In both types of processor, each processing element executes the same (and only) instruction, but on different data. Hence these architectures are SIMD. SIMD machines are widely used for such imaging computation as matrix arithmetic and convolution . 5.4.2.3 Multiple Instruction Single Data The MISD computer architecture lends itself naturally to those computations requiring an input to be subjected to several operations, each receiving the input in its original form.
    [Show full text]
  • Computer Architectures
    Parallel (High-Performance) Computer Architectures Tarek El-Ghazawi Department of Electrical and Computer Engineering The George Washington University Tarek El-Ghazawi, Introduction to High-Performance Computing slide 1 Introduction to Parallel Computing Systems Outline Definitions and Conceptual Classifications » Parallel Processing, MPP’s, and Related Terms » Flynn’s Classification of Computer Architectures Operational Models for Parallel Computers Interconnection Networks MPP’s Performance Tarek El-Ghazawi, Introduction to High-Performance Computing slide 2 Definitions and Conceptual Classification What is Parallel Processing? - A form of data processing which emphasizes the exploration and exploitation of inherent parallelism in the underlying problem. Other related terms » Massively Parallel Processors » Heterogeneous Processing – In the1990s, heterogeneous workstations from different processor vendors – Now, accelerators such as GPUs, FPGAs, Intel’s Xeon Phi, … » Grid computing » Cloud Computing Tarek El-Ghazawi, Introduction to High-Performance Computing slide 3 Definitions and Conceptual Classification Why Massively Parallel Processors (MPPs)? » Increase processing speed and memory allowing studies of problems with higher resolutions or bigger sizes » Provide a low cost alternative to using expensive processor and memory technologies (as in traditional vector machines) Tarek El-Ghazawi, Introduction to High-Performance Computing slide 4 Stored Program Computer The IAS machine was the first electronic computer developed, under
    [Show full text]
  • Difference Between Grid Computing Vs. Distributed Computing
    Difference between Grid Computing Vs. Distributed Computing Definition of Distributed Computing Distributed Computing is an environment in which a group of independent and geographically dispersed computer systems take part to solve a complex problem, each by solving a part of solution and then combining the result from all computers. These systems are loosely coupled systems coordinately working for a common goal. It can be defined as 1. A computing system in which services are provided by a pool of computers collaborating over a network. 2. A computing environment that may involve computers of differing architectures and data representation formats that share data and system resources. Distributed Computing, or the use of a computational cluster, is defined as the application of resources from multiple computers, networked in a single environment, to a single problem at the same time - usually to a scientific or technical problem that requires a great number of computer processing cycles or access to large amounts of data. Picture: The concept of distributed computing is simple -- pull together and employ all available resources to speed up computing. The key distinction between distributed computing and grid computing is mainly the way resources are managed. Distributed computing uses a centralized resource manager and all nodes cooperatively work together as a single unified resource or a system. Grid computingutilizes a structure where each node has its own resource manager and the system does not act as a single unit. Definition of Grid Computing The Basic idea between Grid Computing is to utilize the ideal CPU cycles and storage of millions of computer systems across a worldwide network function as a flexible, pervasive, and inexpensive accessible pool that could be harnessed by anyone who needs it, similar to the way power companies and their users share the electrical grid.
    [Show full text]
  • Introduction to GRID Computing and Overview of the European Data Grid
    Introduction to GRID com p uting a nd ov e rv ie w of th e E urop e a n Da ta Grid P roj e ct The European DataGrid Project http: / / w w w . edg . org Overview What is GRID computing ? What is a GRID ? Why GRIDs ? GRID pr oj e cts w or l d w id e T he E ur ope an Data Gr id Overview of EDG goals and organization Overview of th e EDG m iddleware c om p onents Introduction to GRID Computing and the EDG 2 The Grid Vision The GRID: networked data p roc es s i ng c entres and Res earc hers p erf orm thei r ” m i ddl eware” s of tware as the ac ti v i ti es reg ardl es s “ g l u e” of res ou rc es . g eog rap hi c al l oc ati on, i nterac t wi th c ol l eag u es , s hare and ac c es s data c i enti f i c i ns tru m ents and e! p eri m ents p rov i de hu g e am ou nt of data Federico.C a rm in a t i@cern .ch Introduction to GRID Computing and the EDG 3 What is GRID computing : coordinated resource sharing and problem solving in dy namic, multi-institutional virtual organiz ations. [ I.Foster] A V O i s a c ol l ec ti on of sers sh a ri n g si m i l a r n eed s a n d re! i rem en ts i n th ei r a c c ess to p roc essi n g " d a ta a n d d i stri # ted reso rc es a n d p rs i n g si m i l a r g oa l s.
    [Show full text]
  • Multi-Core Cpus, Clusters, and Grid Computing: a Tutorial∗
    Multi-core CPUs, Clusters, and Grid Computing: a Tutorial∗ Michael Creel Department of Economics and Economic History Edifici B, Universitat Autònoma de Barcelona 08193 Bellaterra (Barcelona) Spain [email protected] and William L. Goffe Department of Economics SUNY—Oswego 416 Mahar Hall Oswego, NY 13126 [email protected] June, 2005 Abstract The nature of computing is changing and it poses both challenges and opportunities for economists. Instead of increasing clock speed, future microprocessors will have “multi-cores” with separate execution units. “Threads” or other multi-processing techniques that are rarely used today are required to take full advantage of them. Beyond one machine, it has become easy to harness multiple computers to work in clusters. Besides dedicated clusters, they can be made up of unused lab comput- ers or even your colleagues’ machines. Finally, grids of computers spanning the Internet are now becoming a reality and one is ready for use by economists. ∗We would like to thank, without implicating, Ken Judd, Aaron Reece, and Robert J. Tetlow. 1 Introduction For nearly two decades, economists have used desktop computers with a single pro- cessor for most of their computing needs. Each year their power has increased and economists have been able to tackle more difficult problems. This era is coming to an end. Future improvements in microprocessors will come from adding additional pro- cessors to each chip; the power of a single processor is unlikely to substantially change. This transition will not be transparent to programmers or sophisticated users. Soon economists with a computational bent will be programming “multi-core” microproces- sors with tools they likely have not used before.
    [Show full text]
  • Chapter 1 Introduction
    Chapter 1 Introduction 1.1 Parallel Processing There is a continual demand for greater computational speed from a computer system than is currently possible (i.e. sequential systems). Areas need great computational speed include numerical modeling and simulation of scientific and engineering problems. For example; weather forecasting, predicting the motion of the astronomical bodies in the space, virtual reality, etc. Such problems are known as grand challenge problems. On the other hand, the grand challenge problem is the problem that cannot be solved in a reasonable amount of time [1]. One way of increasing the computational speed is by using multiple processors in single case box or network of computers like cluster operate together on a single problem. Therefore, the overall problem is needed to split into partitions, with each partition is performed by a separate processor in parallel. Writing programs for this form of computation is known as parallel programming [1]. How to execute the programs of applications in very fast way and on a concurrent manner? This is known as parallel processing. In the parallel processing, we must have underline parallel architectures, as well as, parallel programming languages and algorithms. 1.2 Parallel Architectures The main feature of a parallel architecture is that there is more than one processor. These processors may communicate and cooperate with one another to execute the program instructions. There are diverse classifications for the parallel architectures and the most popular one is the Flynn taxonomy (see Figure 1.1) [2]. 1 1.2.1 The Flynn Taxonomy Michael Flynn [2] has introduced taxonomy for various computer architectures based on notions of Instruction Streams (IS) and Data Streams (DS).
    [Show full text]