IBM Systems Group
Total Page:16
File Type:pdf, Size:1020Kb
IBM STG Deep Computing IBM Systems Group Deep Computing with IBM Systems Barry Bolding, Ph.D. IBM Deep Computing SciComp 2005 Confidential | Systems Group 2004 © 2004 IBM Corporation IBM Systems and Technology Group Deep Computing Components High Performance Computing Leadership Research and Innovation Systems Expertise – pSeries – xSeries – Storage – Networking Innovative Systems © 2004 IBM Corporation IBM Systems and Technology Group Deep Computing Focus Government Research Labs – Energy and Defense Weather/Environmental – Weather Forecasting Centers – Climate Modeling Higher Education/Research Universities Life Sciences – Pharma, BioTech, Chemical Aero/Auto Petroleum Business Intelligence, Digital Media, Financial Services, On Demand HPC © 2004 IBM Corporation IBM STG Deep Computing Deep Computing Teams and Organization Confidential | Systems Group 2004 © 2004 IBM Corporation IBM Systems and Technology Group Deep Computing Technical Team Kent Winchell Technical Team Deep Computing Barry Bolding Jeff Zais Technical Manager Technical Manager Public Sector Industrial Sector Farid Parpia John Bauer Martin Feyereisen Doug Petesch HPC Applications HPC Storage Auto/Aero Auto/Aero Life Sciences Government, HPC Business Intelligence “Suga” Sugavanam Wei Chen Charles Grassl Guangye Li EDA Government HPC Applications Auto/Aero Asia Pacific HPC HIgher Ed. BlueGene/L Stephen Behling Ray Paden Si MacAlester Harry Young Higher Ed GPFS CFD HPC Storage Digitial Media Digital Media Joseph Skovira James Abeles Scott Denham Janet Shiu Schedulers Weather/Environment CSM Petroleum Petroleum, Visualization Marcus Wagner Len Johnson Government Life Sciences Digital Media/Storage © 2004 IBM Corporation IBM STG Deep Computing IBM Deep Computing Summary of Technology Directions Confidential | Systems Group 2004 © 2004 IBM Corporation IBM Systems and Technology Group HPC Cluster System Direction Segmentation Based on Implementation Off Roadmap Segment High End High Value Segment Midrange Systems High Volume 'Good Enough' Segment Blades Blades - Density - Segment 2004 2005 2006 2007 © 2004 IBM Corporation IBM Systems and Technology Group HPC Cluster Directions Limited Configurability (Memory Size, Bisection) Power (w/Accelerators?) P Linux, HCAs E PF BlueGene Extended Configurability R C Power, Linux, HCAs S 100TF Power, AIX, Federation Machines Capability Performance Linux Clusters Less Demanding Communication Power, Intel, BG, Cell Nodes usters Capacity Cl Blades 2004 2005 2006 2007 2008 2009 2010 © 2004 IBM Corporation IBM Systems and Technology Group Deep Computing Architecture User Community Gateways, Webservers, Firewalls, On-Demand Access SAN switch Large-Memory High Density Emerging BW driven Computing Technologies HPC Network Backbone Network Storage Network Shared Storage © 2004 IBM Corporation IBM Systems and Technology Group Deep Computing Architecture (Multicluster GPFS) User Community Gateways, Webservers, Firewalls, On-Demand Access HPC Network Backbone Network Storage Network Large-Memory High Density Emerging BW driven Computing Technologies Shared Storage Shared Storage Shared Storage © 2004 IBM Corporation IBM Offerings are Deep and Wide IBM Systems and Technology Group pSeries, eServer1600 IBM Power4 and Power5 chip AIX/Linux HPC Clusters, Workstations Software, expertise Storage, Networking, and Busines System Grids, "to tie it all together for your HPC solution" Blades Tools Management, s Partners xSeries/eServer Intel Xeon AMD Opteron Blad 1350 Linux, Server2003 eCent © 2004 IBM Corporation er IBM Systems and Technology Group Processor Directions Power Architectures – Power4 Æ Power5 ÆPower6 Æ – PPC970 Æ Power6 technology – BlueGene/L Æ BlueGene/P – Cell Architectures (Sony, Toshiba, IBM) Intel – IA32 Æ EM64T (NOCONA) Æ AMD Opteron – Single-core Æ dual-core © 2004 IBM Corporation IBM Systems and Technology Group System Design Power Consumption (not heat dissipation) Chips might only be 10-20% of the power on a system/node New metrics – Power/ft^2 – Performance/ft^2 – Total cost of ownership (including power/cooling) Power5 clusters (p575) = 96 cpu/rack 1U rack optimized clusters = 128 cpu/rack Bladecenter(PPC/Intel/AMD) = 168 cpu/rack (dual core will increase this) BlueGene = 2048 cpu/rack © 2004 IBM Corporation IBM Systems and Technology Group Systems Directions Optimizing Nodes – 2,4,8,16 CPU nodes – Large SMPs – Rack Optimized Servers and BladeCenter Optimizing Interconnects – Higher Performance Networks – HPS, Myrinets, Infiniband, Quadrics, 10GigE – Utility Networks – Ethernet, Gigabit, 10GigE Optimizing Storage – Global Filesystems (MultiCluster GPFS) – Avoiding Bottlenecks (NFS, Spindle counts, FC adapters and switches) Optimizing Grid Infrastructure © 2004 IBM Corporation IBM Systems and Technology Group Systems Directions pSeries – Power4 Systems (p-6xx) – 2,4,8,16 way Power5 clusters (p-5xx, OpenPower-7xx) – 32,64 way Power5 SMPs (p-595) – BladeCenter cluster (JS20) xSeries – Intel EM64T, Rack Optimized and BladeCenter – x335,x336,HS20,HS40 – AMD Opteron Rack Optimized – x325,x326,LS20 BlueGene/L Interconnects – HPS, Myrinet, IB, GIGE, 10GIGE © 2004 IBM Corporation IBM Systems and Technology Group Software Directions System Software – Unix (AIX, Solaris) – Linux – Linux on POWER – Linux on Intel and Opteron – Windows HPC Software – Same Software on AIX and Linux on POWER – Compilers, Libraries, Tools – Same HPC Infrastructure on Linux/Intel/Opteron and POWER – GPFS, Loadleveler, CSM – MultiCluster GPFS – Grid Software – Backup and Storage Management © 2004 IBM Corporation IBM Systems and Technology Group Linux Software Matrix Kernels (not even considering distros) – 2.4, 2.6 Interconnects – IB (3 different vendors), Myrinet, Quadrics, GigE (mpich and lam) 32 and 64-bit binaries and libraries Compiler options (Intel, Pathscale, PGI, gcc) Geometric increase in number of binaries and sets of libraries that any code developer might need to support. © 2004 IBM Corporation IBM Systems and Technology Group There are passengers and there are drivers! IBM is a Driver – POWER (www.power.org) – Linux on Power and Intel/Opteron, LTC – BlueGene/L – STI Cell Architectures – Open Platform Support HP, SGI, SUN, Cray are passengers – Rely primarily on external innovations (HP, SGI, SUN, Cray). © 2004 IBM Corporation IBM Systems and Technology Group Introducing IBM’s Deep Computing Organization Government Weather Forecasting Crash Analysis • Clear #1 position in High Performance Computing (Top500, Gartner, IDC, …) Financial Services • “Our goal is to solve consistently larger and more complex problems more quickly and at lower cost.” Petroleum Exploration Drug Chip Design Discovery Digital Media © 2004 IBM Corporation IBM Systems and Technology Group The CAE World is in Flux ƒHardware vendors ƒSoftware vendors ƒOperating systems ƒCluster computing ƒMicroprocessors Most users are seeing dramatic changes in their CAE environment © 2004 IBM Corporation IBM Systems and Technology Group Evolution of Hardware: drive towards commonality Clusters (~2002) RISC SMPs (~1994) Cluster architecture (Unix & Linux) now dominates crash and Vectors CFD environments (~1983) SMP architecture was often first introduced in the CFD MainFrames department and helped (~1979) push parallel computing. Beginning in 1986 crash simulation drove CAE compute requirements Mostly MSC.Nastran © 2004 IBM Corporation IBM Systems and Technology Group Transition of the CAE environment Structural Analysis 100 90 80 70 Serial 60 SMP 50 40 4-30 CPUs 30 >30 CPUs 20 Percent Workload 10 0 1998 2000 2002 2004 Crash Simulation 70 60 50 Serial rkload o 40 SMP W 30 4-30 CPUs 20 >30 CPUs Percent 10 0 1998 2000 2002 2004 CFD Simulation 80 70 60 Serial rkload 50 o SMP W 40 4-30 CPUs 30 >30 CPUs Percent 20 10 0 1998 2000 2002 2004 © 2004 IBM Corporation IBM Systems and Technology Group Recent Trends – Top 20 Automotive Sites 100% S 90% P 80% MIPS FLO ga 70% i d G 60% e l POWER al 50% t Alpha s n 40% I PA-RISC SPARC of t 30% n e c r 20% e P IA-32 10% Vector Other IA-64 0% 1997 1998 1999 2000 2001 2002 2003 Source: TOP500 website http://www.top500.org/lists/2003/11/ © 2004 IBM Corporation IBM STG Deep Computing IBM Power Technology and Products Confidential | Systems Group 2004 © 2004 IBM Corporation IBM Systems and Technology Group POWER : The Most Scaleable Architecture POWER5 POWER4+ POWER4 s POWER3 ervers POWER2 S Binary Compatibility Binary PPC PPC 970FX PPC op 750GX Deskt PPC PPC 750FX ames PPC 750 750CXe G 603e PPC PPC 440GX 440GP ded PPC Embedd 405GP PPC 401 © 2004 IBM Corporation IBM Systems and Technology Group IBM powers Mars exploration IBM returns to MARS PowerPC is at the heart of the BAE Systems RAD6000 Single Board Computer, a specialized system enabling the Mars Rovers — Spirit and Opportunity — to explore, examine and even photograph the surface of Mars. In fact, a new generation of PowerPC based space computers is ready for the next trip to another planet. The RAD750, also built by BAE Systems, is powered by a licensed radiation-hardened PowerPC 750 microprocessor that will power space exploration and Department of Defense applications in the years in the come. © 2004 IBM Corporation IBM Systems and Technology Group IBM OpenPower / eServerTM p5 Server Product Line No Compromises...... High-end Linux POWER5 Midrange Systems p5-595 Std & Turbo POWER4+ PPC970+ p5-590 p5-570 p5-575 Systems Systems Express, Std & Turbo Entry p5-550 P590 Towers Express & Std p550 p520 p570 p575 p595 p575 p5-520 p590 Express & Std p595 p50 p5-510 Entry