The Graphics Supercomputer: a New Class of Computer

Total Page:16

File Type:pdf, Size:1020Kb

The Graphics Supercomputer: a New Class of Computer INFORMATION PROCESSING 89, G.X. Ritter (ed.) Elsevier Science Publishers B.V. (North-Holland) O IFIP, 1989 THE GRAPHICS SUPERCOMPUTER: A NEW CLASS OF COMPUTER Gordon BELL and William S. WORLEY Jr. Ardent Computer Company, Sunnyvale, California, U.S.A. Invited Paper In 1988 a new class of computer, the graphics supercom- tion-set, large memory, microprogrammed, cache mem- puter, was introduced by two start-up companies. As a ory, virtual memory operation, multiprocessors for multi- member of both the supercomputer and workstation fami- programmed use, and, finally, vector processing); to the lies, the graphics supercomputer enables both high-per- VAX "minicomputer"; and to the truly interactive "per- formance computation and high-speed, threedimensional, sonal computers" and "workstations." interactive visualization to be performed on a single, inte- grated system. In Ardent's TITAN system, high floating peak computational power as measured in millions or point operation rates are provided by combining the fastest billions of floating point operations per second for scien- RISC microprocessors with traditional supercomputer tific use. This line has been characterized by the Seymour components such as multiple, pipelined vector processors Cray designs which employ the fastest clocks, hardwired and interleaved memory systems. This has resulted in much logic, and extensive parallelism through pipelining and more cost-effective computation and a clear diseconomy of multiple functional units. These systems evolved to vec- scale over supercomputers and minisupercomputers. For tor processing, multiple processors for multiprogram- users ofworkstations that lack the benefits of supercomputer med use, and finally to the parallel processing of a single components, the graphics supercomputer can be used sim- computation. Lower cost supercomputer lines based on ply as a workstation which provides fiveto-ten times more extensive parallelism were extended in the early 80's with performance for important classes of applications. the "minisupercomputer" by Alliant and Convex, and with the introduction this year of the "graphicssupercom- Ardent has introduced a next generation, extendible, object- puter" by Ardent and Stellar. oriented dynamic rendering environment, Dort, as a de- facto standard for the graphics supercomputer computing The importance of the UNIX operating system to the evolu- paradigm, Borden [I]. The graphics supercomputer en- tion of computing should be emphasized. It has permitted ables new applications in computational chemistry, compu- users of one class of system to have compatibility with ma- tational fluid dynamics, real time simulation, animation, chines of other classes of systems. We believe that both mechanical design and engineering, interactive visualiza- horizontal (machines within a class) and vertical (machines tion, and image processing as typified in medicine, surveil- of different classes) compatibility are essential for accelerat- lance, and petroleum exploration. ing the growth of computer use. Machine Classes How Is The Graphics Supercomputer Used? Supercomputers, mainframes, minicomputers, super-mini- Unless a new class of computer engenders a new type, style, computers, mini-supercomputers,workstations,and personal logistics, or economics of use which differentiates it from computers have f&med and evolved as distinct computer other classes of comuuters. it is unlikelv to be successful. classes, differentiated primarily by price. Each new class that Although the graphics su&computer 'has been on the formed has been based upon key new technologies; a market market for less than one year, radical new uses are emerging of users with corresponding application and style-ofcomput- in science, engineering: medicine, finance, arts, training ing needs; and start-up companies who built and introduced (real time simulators), and intelligence/surveillance based the new type of computer. While the record shows that on coupled high computation and interactive visualization established companies do not innovate fundamentally new capabilities. products, they frequently are a source of key people and serve as initial market outlets for those companies that do inno- At the same time, many of the uses of a graphics supercom- vate. Ultimately, if the market is a large one, they adopt the puter are simply evolutions of the uses of structures that have new class idea from a start-up. Examples include IBM's been in operation on ordinary workstations. In effect, a adoptions of the minicomputer from DEC and the PC from graphics supercomputer can be viewed as a very high per- Apple, and DEC's recent adoption of engineering worksta- formance workstation-the quantitative differences in its tions from Apollo and Sun. performance capabilities result in qualitative differences in the applications and manner in which it is used. In this Figure 1 shows how various members of established com- regard, compatibility at many levels is essential for training puter classes have occurred in time, following two "main and the preservation of software investments. This includes lines" of development: common hardware busses, adapters, and protocols; Local Area networks; data formats; programming languages and general purpose processing for both commercial and compilers; operating systems, including the system, window- scientific use, as characterized by the thread of develop ing, and user interfaces; graphics languages; and end-user ment from the Univac I "mainframe" in 1950, to the IBM applications. S/360, S/370 ... 3090 evolution (general complex instruc- C.G. Bell and W.S. Worley EVOLUTION OF COMPUTING STYLES Supercomputing for Science and Engineering ....................................................... Ordinary General Purpose Office, Transaction, and Data Processing Computing 1980 Figure 1 In other cases, the graphics supercomputer simply can sub- tion of graphics supercomputers. Very large scale integrated stitute for an ordinary computer of less power, such as a CMOS circuits can be used to build both large and fast super-minicomputer (e.g. VAX 8600), a mini-supercom- memories; fast, simple, pipelined scalar processors (i.e. RISC puter with equivalent power and higher price (e.g. Alliant or microprocessors); pipelined floating-point arithmetic units; Convex), or a mainframe (where the performance is often an and large, complex application-specific ICs. Large gate ar- order of magnitude worse than a supercomputer). In some rays provide the control for the complex vector units and instances, the graphics supercomputer is used together with graphics processors which compose the heart of the graphics a supercomputer, as a workstation for visualization and for supercomputer. The main architecturalingredients of graph- ancillary high speed computations. ics supercomputers are RISC microprocessors, vector proc- essors, and specialized graphics processors. Components The most interesting and perhaps most important aspect of realizing each of these architectural elements became pos- the graphics supercomputer is that it is the avant garde for sible due to large scale integrated circuits. personal supercomputing. We believe that the productivity of scientific researchers can significantly be augmented by Few now debate the pro's and con's of RISC architectures, the availability to every scientist and engineer of his own, the RISC revolution is over. RISC won. Today, RISC micro- personal supercomputer. A graphics supercomputer of the processors set the performance/price standards for worksta- power of the Ardent Titan System, which is usually sufficient tions, technical computation, and even some commercial LOsubstitute for a share of a central supercomputer, at a cost systems. They simply cannot be equaled by CISC microprc- below $50,000, could make this possible. The next genera- cessors, mainframes, or minicomputers. tion graphics supercomputers are expected to reach this performance/price, enabling the transition to full distrib- The Dower of vector Drocessors derives from the vector uted supercomputingwhere every scientist and engineer has paradigm that is very generally applicable and equivalent to one on a desk, instead of timesharing a large, single system. concurrent issue and execution of many scalar instructions. The Ardent Titan vector processor, for example, simultane- New Technologies Have Enabled the Graphics ously executes two loadvector register pipelines (pipes),one Supercomputer Clans to Form store pipe, and a one-tc-three input operation pipe. Each completed pipe operation includes executing its main op- Each of the computer classes cited earlier was formed ini- eration, advancing register and memory addresses, decre- tially, and evolved subsequently, based on new circuit, and menting a vector length, testing for completion, and con- software technologies. The large scale integrated circuit is tinuingoperation until complete. In sequential form, if each the basic hardware technology that has enabled the construc- of these elements were done by a single instruction, and The Graphics Supercomputer assuming self-addressincrementingload/store instructions, Applications Needs are Critical for Market Formation each Titan vector unit cycle would require approximately sixteen scalar instructions. In effect, specializedvector hard- While technology lies at the root of a new computer class, ware permits sustained, pipelined execution of sequences of corresponding applications' needs must exist and be in highly generic operations.
Recommended publications
  • 4: Gigabit Research
    Gigabit Research 4 s was discussed in chapter 3, the limitations of current networks and advances in computer technology led to new ideas for applications and broadband network design. This in turn led to hardware and software developmentA for switches, computers, and other network compo- nents required for advanced networks. This chapter describes some of the research programs that are focusing on the next step-the development of test networks.1 This task presents a difficult challenge, but it is hoped that the test networks will answer important research questions, provide experience with the construction of high-speed networks, and demonstrate their utility. Several “testbeds” are being funded as part of the National Research and Education Network (NREN) initiative by the Advanced Research Projects Agency (ARPA) and the National Science Foundation (NSF). The testbed concept was first proposed to NSF in 1987 by the nonprofit Corporation for National Research Initiatives (CNRI). CNRI was then awarded a planning grant, and solicited proposals or white papers" from The HPCC prospective testbed participants. A subsequent proposal was then reviewed by NSF with a focus on funding levels, research program’s six objectives, and the composition of the testbeds. The project, cofunded by ARPA and NSF under a cooperative agreement with testbeds will CNRI, began in 1990 and originally covered a 3-year research program. The program has now been extended by an additional demonstrate fifteen months, through the end of 1994. CNRI is coordinat- gigabit 1 Corporation for National Research Initiatives, ‘‘A Brief Description ofthe CNRJ Gigabit lkstbed Initiative,” January 1992; Gary StiX, “Gigabit Conn=tiom” Sci@@ net-working.
    [Show full text]
  • NVIDIA Tesla Personal Supercomputer, Please Visit
    NVIDIA TESLA PERSONAL SUPERCOMPUTER TESLA DATASHEET Get your own supercomputer. Experience cluster level computing performance—up to 250 times faster than standard PCs and workstations—right at your desk. The NVIDIA® Tesla™ Personal Supercomputer AccessiBLE to Everyone TESLA C1060 COMPUTING ™ PROCESSORS ARE THE CORE is based on the revolutionary NVIDIA CUDA Available from OEMs and resellers parallel computing architecture and powered OF THE TESLA PERSONAL worldwide, the Tesla Personal Supercomputer SUPERCOMPUTER by up to 960 parallel processing cores. operates quietly and plugs into a standard power strip so you can take advantage YOUR OWN SUPERCOMPUTER of cluster level performance anytime Get nearly 4 teraflops of compute capability you want, right from your desk. and the ability to perform computations 250 times faster than a multi-CPU core PC or workstation. NVIDIA CUDA UnlocKS THE POWER OF GPU parallel COMPUTING The CUDA™ parallel computing architecture enables developers to utilize C programming with NVIDIA GPUs to run the most complex computationally-intensive applications. CUDA is easy to learn and has become widely adopted by thousands of application developers worldwide to accelerate the most performance demanding applications. TESLA PERSONAL SUPERCOMPUTER | DATASHEET | MAR09 | FINAL FEATURES AND BENEFITS Your own Supercomputer Dedicated computing resource for every computational researcher and technical professional. Cluster Performance The performance of a cluster in a desktop system. Four Tesla computing on your DesKtop processors deliver nearly 4 teraflops of performance. DESIGNED for OFFICE USE Plugs into a standard office power socket and quiet enough for use at your desk. Massively Parallel Many Core 240 parallel processor cores per GPU that can execute thousands of GPU Architecture concurrent threads.
    [Show full text]
  • Annual Reports of FCCSET Subcommittee Annual Trip Reports To
    Annual Reports of FCCSET Subcommittee Annual trip reports to supercomputer manufacturers trace the changes in technology and in the industry, 1985-1989. FY 1986 Annual Report of the Federal Coordinating Council on Science, Engineering and Technology (FCCSET). by the FCCSET Ocnmittee. n High Performance Computing Summary During the past year, the Committee met on a regular basis to review government and industry supported programs in research, development, and application of new supercomputer technology. The Committee maintains an overview of commercial developments in the U.S. and abroad. It regularly receives briefings from Government agency sponsored R&D efforts and makes such information available, where feasible, to industry and universities. In addition, the committee coordinates agency supercomputer access programs and promotes cooperation with particular emphasis on aiding the establish- ment of new centers and new communications networks. The Committee made its annual visit to supercomputer manufacturers in August and found that substantial progress had been made by Cray Research and ETA Systems toward developing their next generations of machines. The Cray II and expanded Cray XMP series supercomputers are now being marketed commercially; the Committee was briefed on plans for the next generation of Cray machines. ETA Systems is beyond the prototype stage for the ETA-10 and planning to ship one machine this year. A ^-0 A 1^'Tr 2 The supercomputer vendors continue to have difficulty in obtaining high performance IC's from U.S. chip makers, leaving them dependent on Japanese suppliers. In some cases, the Japanese chip suppliers are the same companies, e.g., Fujitsu, that provide the strongest foreign competition in the supercomputer market.
    [Show full text]
  • Chapter 1: Computer Abstractions and Technology 1.1 – 1.4: Introduction, Great Ideas, Moore’S Law, Abstraction, Computer Components, and Program Execution
    Chapter 1: Computer Abstractions and Technology 1.1 – 1.4: Introduction, great ideas, Moore’s law, abstraction, computer components, and program execution ITSC 3181 Introduction to Computer Architecture https://passlab.github.io/ITSC3181/ Department of Computer Science Yonghong Yan [email protected] https://passlab.github.io/yanyh/ Lectures for Chapter 1 and C Basics Computer Abstractions and Technology ☛• Lecture 01: Chapter 1 – 1.1 – 1.4: Introduction, great ideas, Moore’s law, abstraction, computer components, and program execution • Lecture 02 - 03: C Basics; Compilation, Assembly, Linking and Program Execution • Lecture 03 - 04: Chapter 1 – 1.6 – 1.7: Performance, power and technology trends • Lecture 04 - 05: Memory and Binary Systems • Lecture 05: – 1.8 - 1.9: Multiprocessing and benchmarking 2 § 1.1 Introduction 1.1 The Computer Revolution • Progress in computer technology – Underpinned by Moore’s Law • Every two years, circuit density ~= increasing frequency ~= performance, double • Makes novel applications feasible – Computers in automobiles – Cell phones – Human genome project – World Wide Web – Search Engines • Computers are pervasive 3 Generation Of Computers https://solarrenovate.com/the-evolution-of-computers/ 4 New School Computer 5 Classes of Computers • Personal computers (PC) --> computers are PCs today – General purpose, variety of software – Subject to cost/performance tradeoff • Server computers – Network based – High capacity, performance, reliability – Range from small servers to building sized 6 Classes of Computers
    [Show full text]
  • Resource Guide: Workstation Productivity
    Resource Guide: Workstation Productivity Contents Beyond the Desktop: Workstation Productivity ……………………………………………………2 Find out how to get the highest productivity possible from high performance applications that require more power than traditional desktop computers Top IT Considerations for Mobile Workstations …………………...............................................4 Discover the top IT considerations for organization that require mobile workstations designed to provide a balance of performance and portability Maximum Management and Control with Virtualized Workstations ....…………………...…….6 Explore the benefits of virtualization to ensure the greatest possible management and control of high performance workstations for the most demanding workloads and applications Sponsored by: Page 1 of 6 Copyright © MMIX, CBS Broadcasting Inc. All Rights Reserved. For more downloads and a free TechRepublic membership, please visit http://techrepublic.com.com/2001-6240-0.html TechRepublic Resource Guide: Workstations Productivity Beyond the Desktop: Workstation Productivity Find out how to get the highest productivity possible from high performance applications that require more power than traditional desktop computers Desktop computers have come a long way in terms of value and performance for the average user but more advanced applications now require even more in terms of processing, graphics, memory and storage. The following is a brief overview of the differences between standard desktops and workstations as well as the implications for overall performance and productivity as compiled from the editorial resources of CNET, TechRepublic and ZDNet.com. Think Inside the Box A lot of people tend to think of desktop computers and computer workstation as one and the same but that isn’t always the case in terms of power and performance. In fact, many of today’s most demanding workloads and applications for industries such as engineering, entertainment, finance, manufacturing, and science actually require much higher functionality than what traditional desktops have to offer.
    [Show full text]
  • 2200, Canmath 201.Qxd
    An Introduction to the Computer Age Computers are changing our world. The small amount of training. Amazingly invention of the internal combustion engine enough, as the size has decreased, the power and the harnessing of electricity have had a has increased and prices have plummeted. profound effect on the way society operates. How does the computer age affect the The widespread use of the computer is Christian? What is the history behind the having a similar and perhaps even greater computer? What are the fundamental parts impact on our society. of a computer and how do they work In only decades, computers have shrunk together? What are the uses, advantages, from mammoth, room-filling machines that and limitations of computers? Do I need a only the highly educated could operate to computer? This LightUnit and those follow - tiny devices held in the palm of the hand ing it will begin to answer some of these that the average person can use after a questions for you. Section 1 Computer Background Any study must be based on definitions very specialized field, new words have come about the subject. If definitions are not into being, and many common words have understood, there is little hope that much acquired new definitions. Some of the words can be learned about the subject. The goal of may already be familiar to you, but in the the first section of this LightUnit is to pro - context of computers, they may take on a vide a basis for the rest of the course by different meaning. Therefore, do not assume defining computer and many terms associ - you know the definition even if the word is ated with computers.
    [Show full text]
  • Cpu Performance
    LECTURE 1 Introduction CLASSES OF COMPUTERS • When we think of a “computer”, most of us might first think of our laptop or maybe one of the desktop machines frequently used in the Majors’ Lab. Computers, however, are used for a wide variety of applications, each of which has a unique set of design considerations. Although computers in general share a core set of technologies, the implementation and use of these technologies varies with the chosen application. In general, there are three classes of applications to consider: desktop computers, servers, and embedded computers. CLASSES OF COMPUTERS • Desktop Computers (or Personal Computers) • Emphasize good performance for a single user at relatively low cost. • Mostly execute third-party software. • Servers • Emphasize great performance for a few complex applications. • Or emphasize reliable performance for many users at once. • Greater computing, storage, or network capacity than personal computers. • Embedded Computers • Largest class and most diverse. • Usually specifically manufactured to run a single application reliably. • Stringent limitations on cost and power. PERSONAL MOBILE DEVICES • A newer class of computers, Personal Mobile Devices (PMDs), has quickly become a more numerous alternative to PCs. PMDs, including small general-purpose devices such as tablets and smartphones, generally have the same design requirements as PCs with much more stringent efficiency requirements (to preserve battery life and reduce heat emission). Despite the various ways in which computational technology can be applied, the core concepts of the architecture of a computer are the same. Throughout the semester, try to test yourself by imagining how these core concepts might be tailored to meet the needs of a particular domain of computing.
    [Show full text]
  • Copyrighted Material
    1 The Function of Computation As in music theory, we cannot discuss the microprocessor without positioning it in the context of the history of the computer, since this component is the integrated version of the central unit. Its internal mechanisms are the same as those of supercomputers, mainframe computers and minicomputers. Thanks to advances in microelectronics, additional functionality has been integrated with each generation in order to speed up internal operations. A computer1 is a hardware and software system responsible for the automatic processing of information, managed by a stored program. To accomplish this task, the computer’s essential function is the transformation of data using computation, but two other functions are also essential. Namely, these are storing and transferring information (i.e. communication). In some industrial fields, control is a fourth function. This chapter focuses on the requirements that led to the invention of tools and calculating machines to arrive at the modern version of the computer that we know today. The technological aspect is then addressed. Some chronological references are given. Then several classification criteria are proposed. The analog computer, which is then described, was an alternative to the digital version. Finally, the relationship between hardware and software and the evolution of integration and its limits are addressed. NOTE.– This chapter does not attempt to replace a historical study. It gives only a few key dates and technical benchmarks to understand the technological evolution of the field. COPYRIGHTED MATERIAL 1 The French word ordinateur (computer) was suggested by Jacques Perret, professor at the Faculté des Lettres de Paris, in his letter dated April 16, 1955, in response to a question from IBM to name these machines; the English name was the Electronic Data Processing Machine.
    [Show full text]
  • The Paramountcy of Reconfigurable Computing
    Energy Efficient Distributed Computing Systems, Edited by Albert Y. Zomaya, Young Choon Lee. ISBN 978-0-471--90875-4 Copyright © 2012 Wiley, Inc. Chapter 18 The Paramountcy of Reconfigurable Computing Reiner Hartenstein Abstract. Computers are very important for all of us. But brute force disruptive architectural develop- ments in industry and threatening unaffordable operation cost by excessive power consumption are a mas- sive future survival problem for our existing cyber infrastructures, which we must not surrender. The pro- gress of performance in high performance computing (HPC) has stalled because of the „programming wall“ caused by lacking scalability of parallelism. This chapter shows that Reconfigurable Computing is the sil- ver bullet to obtain massively better energy efficiency as well as much better performance, also by the up- coming methodology of HPRC (high performance reconfigurable computing). We need a massive cam- paign for migration of software over to configware. Also because of the multicore parallelism dilemma, we anyway need to redefine programmer education. The impact is a fascinating challenge to reach new hori- zons of research in computer science. We need a new generation of talented innovative scientists and engi- neers to start the beginning second history of computing. This paper introduces a new world model. 18.1 Introduction In Reconfigurable Computing, e. g. by FPGA (Table 15), practically everything can be implemented which is running on traditional computing platforms. For instance, recently the historical Cray 1 supercomputer has been reproduced cycle-accurate binary-compatible using a single Xilinx Spartan-3E 1600 development board running at 33 MHz (the original Cray ran at 80 MHz) 0.
    [Show full text]
  • Lenovo Thinkpad P15 Gen 1
    ThinkPad P15 Gen 1 Business professionals can now get extreme power in a mobility-friendly package. This mobile workstation supports advanced workloads with an Intel Xeon or 10th Gen Core processors and powerful NVIDIA Quadro RTX5000 graphics. Choose the 4K OLED or LCD display with Dolby Vision HDR for unmatched clarity and comfort. Users can connect up to three additional displays, and transfer large files at unprecedented speeds with twin ThunderBolt 3 connectors. PROFESSIONAL MACHINES FOR REASONS TO BUY ADVANCED USERS Built to the highest thermal engineering standards, it stays cool and quiet while maintaining the power users need for advanced workloads. Support for WiFi 6 ensures internet connectivity is faster and more secure while users can stay connected on the move with optional 4G WWAN. Certified or recommended for use with: Adobe Premiere Pro, Adobe Photoshop, Adobe After Effects Configure with an 8-core Intel Xeon or Core i9 processor, fast M.2 SSD storage and up to 128GB of DDR4 memory; enabling multitasking of large spreadsheets, databases or complex statistical software. This 15.6" display ThinkPad P Series mobile workstations offer a choice of professional graphics and processor options. most.lenovo.com ThinkPad P15 Gen 1 Recommended for this KEY SPECIFICATIONS CONNECTIVITY device Processor up to Intel Xeon processor W-10885M or 10th Gen Intel Core i9 I/O Ports 1x USB 3.2 Type-C (power delivery and DisplayPort), 2x USB 3.2 processor Type-C Gen 2 / Thunderbolt 3 (power delivery and DisplayPort), 2x USB-A 3.2 Gen 1, HDMI 2.0,
    [Show full text]
  • High Performance Computing for Science
    Index Algorithms: 8, 15, 21, 24 Design, high-performance computers: 21 Applications programs: 8 Digital libraries and archives: 8 ARPANET: 5, 10 Digital libraries, Global Digital Library: 8 Artificial intelligence: 8, 26 Education: 8, 10, 13 Bardon/Curtis Report: 9 EDUCOM, Networking and Telecommunications Bibliographic services: 10 Task Force: 10 Black hole: 5, 6 Electronic Bulletin boards: 4, 6, 10, 11, 12 information technologies: 4, 5, 6 Journals: 6 California Institute of Technology: 29 mail: 6, 10, 12 Central processing units (CPU): 22,24, 30 Executive Office of the President: 11 Committee on Science, Engineering, and Public Policy (COSEPUP): 9 Federal Coordinating Council for Science, Engineering, Computational science: 21, 22 and Technology (FCCSET): 1, 3 Computer Federal Government Alliant: 6 budget and funding: 12, 17, 18, 19, 20, 22 Apple Macintosh: 31 funding, individual research: 12, 19 connection machines: 31 policy: 12, 16, 20, 23, 26 Control Data ETA: 22 procurement regulations: 16 Control Data 6600: 26 research and development (R&D): 3, 4, 5, 8, 11, 15, 16, Cray 1: 31 17, 18, 24,25, 26, 32 Cray X-MP computer: 7, 19 responsibilities, 11, 13, 15 data flow processors: 31 FederaI High Performance Computing and Communications design: 22, 28, 29, 30, 32 Program (HPCC): 2, 18 fuzzy logic: 31 Fifth Generation Project: 2 hypercube: 29, 31 Floating point operations (FLOPS): 31 IBM Stretch: 27 Florida State University (FSU): 7, 18 IBM 3090: 30 Fluid flow: 6 IBM 3090 computer: 6 IBM 360: 27 Gallium Arsenide: 28-29 manufacture: 22
    [Show full text]
  • 2 CLASSIFICATION of COMPUTERS.Pdf
    CLASSIFICATION OF COMPUTERS Computers can be classified in the following methods: I . Computational Method I. Size and Capability I. Classification based on Computational method: Based on the way a system performs the computations, a computer can be classified as follows: • Digital • Analog • Hybrid Digital computer: A digital computer can count and accept numbers and letters through various input devices. The input devices convert the data into electronic pulses, and perform arithmetical operations on numbers in discrete form. In addition to performing arithmetical operations, they are also capable of:- 1. Storing data for processing 2. Performing logical operations 3. Editing or deleting the input data. One of the main advantages in the use of digital computers is that any desired level of accuracy can be achieved by considering as many places of decimal as are necessary and hence are most suitable for business application. The main disadvantage is their high cost, even after regular reductions in price and the complexity in programming. Example: To calculate the distance travelled by a car in a particular time interval, you might take the diameter of the tyre to calculate the periphery, take into consideration the number of revolutions of the wheel per minute, take the time in minutes and multiply them all to get the distance moved. This is called digital calculation. A computer using the principle of digital calculations can be called a digital computer. Analog Computer: Analog computers process data input in a continuous form. Data such as voltage, resistance or temperature are represented in the computer as a continuous, unbroken flow of information, as in engineering and scientific applications, where quantities to be processed exists as waveforms or continually rising and falling voltages, pressure and so on.
    [Show full text]