Metacomputing Across Intercontinental Networks S.M

Metacomputing Across Intercontinental Networks S.M

Future Generation Computer Systems 17 (2001) 911–918 Metacomputing across intercontinental networks S.M. Pickles a,∗, J.M. Brooke a, F.C. Costen a, Edgar Gabriel b, Matthias Müller b, Michael Resch b, S.M. Ord c a Manchester Computing, The University of Manchester, Oxford Road, Manchester M13 9PL, UK b High Performance Computing Center, Stuttgart, Allmandring 30, D-70550 Stuttgart, Germany c NRAL, Jodrell Bank, Macclesfield, Cheshire SK11 9DL, UK Abstract An intercontinental network of supercomputers spanning more than 10 000 miles and running challenging scientific appli- cations was realized at the Supercomputing ’99 (SC99) conference in Portland, OR using PACX-MPI and ATM PVCs. In this paper, we describe how we constructed the heterogeneous cluster of supercomputers, the problems we confronted in terms of multi-architecture and the way several applications handled the specific requirements of a metacomputer. © 2001 Elsevier Science B.V. All rights reserved. Keywords: Metacomputing; High-performance networks; PACX-MPI 1. Overview of the SC99 global network • national high speed networks in each partner coun- try, During SC99 a network connection was set up that • the European high speed research network connected systems in Europe, the US and Japan. For TEN-155, the experiments described in this paper, four super- • a transatlantic ATM connection between the Ger- computers were linked together. man research network (DFN) and the US research network Abilene, • A Hitachi SR8000 at Tsukuba/Japan with 512 pro- • a transpacific ATM research network (TransPAC) cessors (64 nodes). that was connecting Japanese research networks to • A Cray T3E-900/512 at Pittsburgh/USA with 512 STAR-TAP at Chicago. processors. • A Cray T3E-1200/576 at Manchester/UK with 576 Details of the network configuration can be seen in processors. Fig. 1. Using ATM, PVCs were set up with a band- • A Cray T3E-900/512 at Stuttgart/Germany with 512 width of 10 MBit/s in the European networks. On all processors. other connections bandwidth was shared with other users. The sustained bandwidths and the latencies be- Together this virtual supercomputer had a peak per- tween sites were as described in Table 1. formance of roughly 2.1 Teraflops. The network was based on: 2. Problems of multi-architecture Heterogeneous metacomputing introduces some ∗ Corresponding author. problems that are similar to those well known from 0167-739X/01/$ – see front matter © 2001 Elsevier Science B.V. All rights reserved. PII: S0167-739X(01)00032-2 912 S.M. Pickles et al. / Future Generation Computer Systems 17 (2001) 911–918 Fig. 1. Network configuration as set up during SC99. cluster computing and some that are very specific [2]. then collect information about the amount of load The most important ones are imbalance. Based on these data and on the charac- teristics of the network connection, the application • different data representations on each system, would then decide how much effort should be spent • different processor speeds on each system, to redistribute the load. Such an approach, however, • different communication speeds for internal mes- is based on the assumption that the basic parame- sages on each system, ters (processor load, network speed, application load) • different communication speeds of messages inter- remain constant. Allowing these parameters to vary nal to a system and messages between systems, — as is necessary if working in a non-dedicated en- • lack of a common file system, vironment or using adaptive methods — will further • lack of resource management. complicate the problem. The problem of load imbalance due to different The problem of scheduling resources is a conti- processor speed has to be tackled on the application nuing research topic. For metacomputing it is level. So far there is no standard way for an applica- necessary to make sure that resources are available tion to determine the speed of a processor. Moreover, concurrently. That means that the scheduling of sev- it is difficult for a given application to get information eral systems and of networks must be coordinated. No about its sustained performance. An approach that is automatic mechanism was available to us at SC99. currently being investigated is to start the application The problem of distributed file systems was not tack- with a standard distribution. The application would led in our approach. We assumed that data would be in place whenever necessary. Table 1 All other problems relate mainly to communication Network bandwidths and latencies between sites during SC99 and have to be tackled by the message passing library that is used. To specially focus on the problems of Latency (ms) Bandwidth (MBit/s) metacomputing HLRS has developed an MPI library HLRS-CSAR 20 6.4 called PACX-MPI [1,3] that couples big systems into a HLRS-PSC 78 1.6 single resource from the communication point of view. CSAR-PSC 58 4.0 Based on the experience of several other projects the S.M. Pickles et al. / Future Generation Computer Systems 17 (2001) 911–918 913 radio telescope). This presents particular challenges to our metacomputer since the distribution of the data makes severe demands on the intercontinental band- width. This experiment involved coupling of the three T3E machines only. 3.1. The Jodrell bank de-dispersion pulsar search code [6] The application searches for pulsars in the radio sig- nal from a globular cluster. The observed radio signal Fig. 2. Basic concept of PACX-MPI. Two systems integrated into from a pulsar manifests at the telescope as a periodic one single MPI COMM WORLD. increase in broadband radio noise. In order to observe the pulse with a high signal to noise ratio, we need to observe across a wide band of radio frequencies. library relies on three main concepts: As the space between the earth and the pulsar (the in- • Single MPI COMM WORLD for an application. terstellar medium) is slightly charged it is dispersive When starting an application with PACX-MPI and therefore different frequencies propagate at dif- a local MPI COMM WORLD is created on ferent velocities. To reconstruct the pulsar signal it is each system in the metacomputer. These are necessary to first determine the relationship between combined by PACX MPI to form a single a given radio frequency and the time lag in the arrival MPI COMM WORLD. of the pulse (called the dispersion measure, DM) and • Usage of communication daemons. On each system then use this to reconstruct the signal. Since a glob- in the metacomputer two daemons take care of com- ular cluster is a compact object, the DM will be ap- munication between systems. This allows bundling proximately constant for all pulsars in the cluster and of communication and to avoid having thousands of reconstructing at several trial DMs in a narrow range open connections between processes. In addition it yields an effective method of searching. allows handling of security issues centrally. • Splitting of internal and external communication. 3.2. Implementation on the metacomputer For all internal communication PACX-MPI makes use of the native library of the system. It only imple- The total data from a continuous sequence of mea- ments communication for external messages, that surements from the telescope is first split into lengths is messages that have to be sent between systems. at least twice as long as the dispersive delay time This has the advantage that optimized communica- across the observing bandwidth. Each length is then tion inside a system can be combined with portable subjected to a Fourier transform and multiplied by the communication protocols between systems. Fourier transform of the inverse filter. The resultant data is then transformed back into the time domain In Fig. 2 global numbering denotes the global and processed to produce a de-dispersed time series. MPI COMM WORLD, while local numbering corre- The next length is then processed, and the resultant sponds to the local communicators. Processes 0 and series are then concatenated together. In practice, the 1 on each system are acting as daemons for external input data has to be forward transformed only once, communication. but transformed back into the time domain as many times as we have trial DMs. 3. Use of experimental data This process defines the challenge for the metacom- puter. If the data is read onto one T3E (in this case at In this section, we describe a novel use of meta- Manchester) then subsets of the data have to be sent to computing, namely the processing of data from an the remote T3Es via a connection, that is two orders of experimental facility (in this case the Jodrell Bank magnitude slower than the inter-processor connection 914 S.M. Pickles et al. / Future Generation Computer Systems 17 (2001) 911–918 within a T3E. This might seem to invalidate the whole 3.4. Static load-balancing algorithm metacomputing concept, but the requirement that many trial DMs be applied to each length means that Let N be the number of hosts, and let ni be we can perform sufficient processing on the remote the number of client processors to use on host i, machines to cover the cost of transmission. More- discounting the extra two processors required by over, since our ultimate goal is real-time processing PACX-MPI and those processors on host 1 which of experimental data (i.e. processing directly after the have been assigned to server duties. Denote the band- observations via network links to a large computing width from host 1 (where the input data resides) to facility), we would find that the remote host has far host i by βi. The rate at which data can be read more processing power than the machines onto which from disk on host 1 is denoted by r; it is assumed the data is initially read.

View Full Text

Details

  • File Type
    pdf
  • Upload Time
    -
  • Content Languages
    English
  • Upload User
    Anonymous/Not logged-in
  • File Pages
    8 Page
  • File Size
    -

Download

Channel Download Status
Express Download Enable

Copyright

We respect the copyrights and intellectual property rights of all users. All uploaded documents are either original works of the uploader or authorized works of the rightful owners.

  • Not to be reproduced or distributed without explicit permission.
  • Not used for commercial purposes outside of approved use cases.
  • Not used to infringe on the rights of the original creators.
  • If you believe any content infringes your copyright, please contact us immediately.

Support

For help with questions, suggestions, or problems, please contact us