Massively Parallel Artificial Intelligence
Total Page:16
File Type:pdf, Size:1020Kb
Massively Parallel Artificial Intelligence Hiroaki Kitano (Chairperson) Carnegie Mellon University NEC Corporation Pittsburgh, PA, 15213 Tokyo 108, Japan [email protected] James Hendler Tetsuya Higuchi University of Maryland, USA Electrotechnical Laboratory, Japan [email protected] [email protected] Dan Moldovan David Waltz University of Southern California, USA Thinking Machines Corporation, USA and [email protected] Brandeis University, USA [email protected] Abstract exploring the full potential of the massive parallelism offered on currently available machines. One of the causes of this Massively Parallel Artificial Intelligence is a new is that little communication has occurred between hardware and growing area of AI research, enabled by the architects and AI researchers. Hardware architects design emergence of massively parallel machines. It is a without actually recognizing the processing, memory, and new paradigm in AI research. A high degree of par performance requirements of AI algorithms. AI researchers allelism not only affects computing performance, have developed their theories and models assuming idealiza but also triggers drastic change in the approach to tions of massive parallelism. Further, with few exceptions, ward building intelligent systems; memory-based the AI community has often taken parallelism as a mere "im reasoning and parallel marker-passing are examples plementation detail" and has not yet come up with algorithms of new and redefined approaches. These new ap and applications which take full advantage of the massively proaches, fostered by massively parallel machines, parallelism available. offer a golden opportunity for AI in challenging the The panel intends to rectify this situation by inviting pan vastness and irregularities of real - world data that are elists knowledgeable and experienced in both hardware and encountered when a system accesses and processes application aspects of massively parallel computing in artifi- Very Large Data Bases and Knowledge Bases. This cial intelligence. There are two interrelated issues which will article describes the current status of massively par be addressed by the panel: (1) the design of massively parallel allel artificial intelligence research and positions of hardware for artificial intelligence, and (2) the potential ap each panelist. plications, algorithms and paradigms aimed at fully exploring the power of massively parallel computers for symbolic AL 1 Introduction 2 Current Research in Massively Parallel AI The goal of the panel is to highlight current accomplishments and future issues in the use of massively parallel machines for 2.1 Massively Parallel Machines artificial intelligence research, a field generally called mas• Currently, there are a few research projects involving the de sively parallel artificial intelligence. The importance of mas velopment of the massively parallel machines and a few com sively parallel artificial intelligence has been recognized in mercially available machines being used for symbolic AI. recent years due to three major reasons: Three projects of particular importance are: 1. increasing availability of massively parallel machines, ♦ The CM-2 Connection Machine (Thinking Machines 2. increasing interest in memory-based reasoning and other Corporation), highly-parallel AI approaches, ♦ The Semantic Network Array Processor (University of 3. development efforts on Very Large Knowledge Bases Southern California) (VLKB). ♦ The IXM2 Associative Memory Processor (Electrotech nical Laboratory, Japan). Despite wide recognition of massively parallel computing as an important aspect of high performance computing and These machines provide an extremely high-level of par general interest in the AI community on highly parallel pro allelism (8K - 256K) and promise even more in the future. cessing, only a small amount of attention has been paid to Table 1 shows the specification of these machines. Kitano, et al. 557 There are two major approaches to designing a massively attractive approach to AI on massively parallel machines due parallel machine: the Array Processor and the Associative to the memory-intensive and data-parallel nature of its opera Processor. CM-2 and SNAP arc examples of the array proces tion. Traditional AI work has been largely constrained by the sor architecture, and IXM2 is an example of the associative performance characteristics of serial machines. Thus for ex processor architecture. While the array processor architec ample, the memory efficiency and optimization of serial rule ture attains parallelism by the number of physical processors application has been regarded as a central issue in expert sys available, the associative processor attains parallelism by the tems design. However, massively parallel machines may take associative memory assigned to each processor. Thus, the away such constraints by the use of highly parallel operations parallelism attained by the associative processor architecture based on the idea of data-parallelism. The memory-based is beyond the number of processors in the machine, whereas reasoning fits perfectly with this idea. the array processor attains parallelism equal to the number of Another approach is marker-passing. In the Marker- actual processors. This is why the IXM2 attains 256K paral Passing approach, knowledge is stored in semantic net lelism with 64 processors. However, operations carried out works, and objects called markers are propagated, in par by associative memories are limited to bit-marker passing and allel, to perform the inference. Marker-passing is a powerful relatively simple arithmetic operations. When more complex method of performing inferencing on large semantic network operations are necessary, the parallelism will be equal to the knowledge-bases on massively parallel machines, due to the number of processors. high degree of parallelism that can be attained. One obvious Regarding the parallelism, the next version of the IXM2 application of this approach is the processing of Very Large (may be called IXM3) will aim at over one million paral Knowledge Bases (VLKB) such as MCC's CYC [Lcnat and lelism using up-to-data processors and high density associa Guha, 1989], EDR's electric dictionaries [EDR, 1988] (both tive memory chips. The SNAP project is planning to develop a of which are expected to require millions of network links), custom VLSI to attain a one million processor-scale machine. and ATR's dialogue database tEhara et. al., 1990]. It is clear DARPA (Defense Advanced Research Projects Agency) is that as KBs grow substantially large (over a million concepts) fundings project to attain TeraOps by 1995 [Waltz, 1990]. the complex (and often complete) searches used in many tra ditional inferencing systems will have to give way to heuristic 2.2 Massively Parallel Al Paradigm solutions unless a high degree of parallelism can be exploited. Thus, the use of massively parallel computers for VLKB pro In addition to designing new hardware architectures, the cessing is clearly warranted. strategies and perhaps even the paradigms for designing and Table 2 shows massively parallel Al systems developed so building AI systems may need to be changed in order to take far. The list is by no means exhaustive, only lists the major advantage of the full potential of massively parallel machines. systems. Also, there are many other models which match Emergence of massively parallel machines offers a new op well with massively parallel machines. But, we only list the portunity for AI in that large-scale DB/KB processing can be systems that are actually implemented on massively parallel made possible in real-time. Walu's talk at AAAI-90 [Waltz, machines. 1990] envisioned the challenge of massively parallel Al. Two of the major ideas that play central roles in massively parallel AI are; memory-based reasoning and marker-passing. Memory-based reasoning and case-based reasoning assume that memory is a foundation of intelligence. Systems based on this approach store a large number of memory instances of past cases, and modify them to provide solutions to new prob lems. Computationally, the memory-based reasoning is an 558 Panels 3 Associative Memory Architecture of two orders of magnitude in execu- tion time be tween Cray-XMP and CM-2 for 64K data, and a Tetsuya Higuchi difference of three orders between Cray-XMP and Electrotechnical Laboratory, Japan IXM2. In this talk, we consider the architectural requirements for Example 2. Marker Propagation: Marker propaga massively parallel AI applications, based on our experiences tion is intensively used in processing is-a hierarchy of developing a parallel associative processor IXM2 and ap know- ledge base. It actually traverses links of the plications for IXM2, In addition, we introduce the current network structured data. A marker propagation pro- status of the Electric Dictionary Research project in Japan gram written in C, which uses recursive procedure which is a real example of a very large knowledge base con call for traversing links, was run both on SUN-4 and taining 400,000 concepts. Cray-XMP. In spite of the exactly same program, We have developed a parallel associative processor IXM2 Cray was slightly slower than SUN-4. The main which enables 256K parallel operations using a large associa reasons for this are that the overhead of recursive tive memory. IXM2 consists of 64 associative processors with procedure calls is heavy, and that network struc 256K word large associative