Intel Cluster Ready Appro Xtreme-X Computers with Mellanox QDR Infiniband AA PP PP RR OO II NN TT EE RR NN AA TT II OO NN AA LL II NN CC Steve Lyness Vice President, HPC Solutions Engineering [email protected] Company Overview :: Corporate Snapshot • Computer Systems manufacturer founded in 1991 • From OEM to branded products and solutions in 2001 – Headquarters in Milpitas, CA – Regional Office in Houston, TX – Global Presence via International Resellers and Partners – Manufacturing and Hardware R&D in Asia – 100+ employees worldwide • Leading developer of high performance servers, clusters and supercomputers. – Balanced Architecture – Delivering a Competitive Edge – Shipped over thousands of clusters and servers since 2001 – Solid Growth from 69.3M in 2006 to 86.5M in 2007 • Target Markets: - Electronic Design Automation - Financial Services - Government / Defense - Manufacturing - Oil & Gas *Source: IDC WW Quarterly PC Tracker SLIDE | 2 SLIDE A P P R O H P C P R E S E N T A T I O N Company Overview :: Why Appro High Performance Computing Expertise – Designing Best in Class Supercomputers Leadership in Price/Performance Energy Efficient Solution Best System Scalability and Manageability High Availability Features SLIDE | 3 SLIDE A P P R O H P C P R E S E N T A T I O N Company Overview :: Our Customers SLIDE | 4 SLIDE A P P R O H P C P R E S E N T A T I O N Company Overview :: Solid Growth +61% Compound Annual Revenue Growth $86.5M $69.3M $33.2M 2005 2006 2007 SLIDE | 5 SLIDE A P P R O H P C P R E S E N T A T I O N HPC Experience :: Recent Major Installations September 2006 June 2007 NOAA Cluster LLNL Minos Cluster • 1,424 processor cores • 6,912 processor cores • 2.8TB System Memory • 13.8TB System Memory • 15 TFlops • 33 TFlops RecentRecent InstallationsInstallations November 2006 February 2008 LLNL Atlas Cluster DE Shaw Research Cluster • 9,216 processor cores • 4,608 processor cores • 18.4TB System Memory • 9.2TB System Memory • 44 TFlops • 49 TFlops SLIDE | 6 SLIDE A P P R O H P C P R E S E N T A T I O N HPC Experience :: Current Major Projects April - June 2008 July 2008 TLCC Clusters Renault F1 CFD Cluster • Total 426TFlops project • Total 38TFlops • LLNL, LANL, SNL • Dual-rail IB • 8 Clusters CurrentCurrent ProjectsProjects June 2008 U of Tsukuba Cluster • Total 95TFlops • Quad-rail IB SLIDE | 7 SLIDE A P P R O H P C P R E S E N T A T I O N HPC/Supercomputing Trends :: What are the pain points? • Technology is driven by customer’s needs • Needs are driven by desire to reduce pain points – Clusters are still hard to use – as clusters get bigger, the problem grows exponentially – Power, cooling and floor space are major issues – Weak interconnect performance – RAS is a growing issue/problem SLIDE | 8 SLIDE A P P R O H P C P R E S E N T A T I O N Intel Cluster Ready and Intel Multi-Core Processors AA PP PP RR OO II NN TT EE RR NN AA TT II OO NN AA LL II NN CC Intel® Cluster Ready Intel® Cluster Ready Platform Solutions Provider Solution Deployment End-Users Deliver SKU that is a Certified Solution Intel® Cluster Ready Platform Certification Appro Xtreme-X1 ISV-Applications Supercomputer Intel® Cluster Ready Application Registration Complete Intel® Cluster Ready Solution Intel® Cluster Ready Hardware Certification Component Certification Mellanox Appro HW and SW ConnectX Certified Solution from a specific Platform Integrator SLIDE | 10 SLIDE A P P R O H P C P R E S E N T A T I O N Developer Considerations Intel® Software Tools for Building HPC Applications Multi-Threading Æ Å Message Passing Performance | Compatibility | Support | Productivity | Cross-Platform http://www.intel.com/software | 11 SLIDE A P P R O H P C P R E S E N T A T I O N Benefits of Choosing an Intel® Cluster Ready Certified System •In production faster • Simplified selection of compatible systems and applications • Faster and higher confidence initial deployment •Less downtime • Ability to detect and correct problems soon after they occur • Better support from ISVs and your system vendor – Lower initial and ongoing costs – Higher ongoing productivity Demand an Intel® Cluster Ready certified system for your next cluster SLIDE | 12 SLIDE A P P R O H P C P R E S E N T A T I O N HPC/Supercomputing Trends New 45nm Processors Quad-Core Intel® Xeon® processor 5400 series Technology Rich Compelling IT Benefits Enhanced Greater Intel Core Performance at Microarchitecture Given Frequency and Higher Larger Caches Frequencies New SSE4 Instructions Greater Energy Efficiency 45 nm High-k Process Technology 2nd Generation Intel Quad-Core also available Dual-Core Intel® Xeon® processor 5200 series | 13 SLIDE A P P R O H P C P R E S E N T A T I O N Nehalem Based System Architecture Up to 25.6 Gb/sec bandwidth per link Benefits Nehalem Nehalem • More application performance • Improved energy efficiency QPI I/O • End to end HW assist Hub – Improved Virtualization Technology PCI DMI Functional •Stable IT image Express* system – Software compatible Gen 1, 2 ICH demonstrated Sept 2007 IDF – Live migration compatible with today’s dual and quad-core Intel® Core™ Key Technologies Extending Today’s Leadership New 45nm Intel® Microarchitecture New Intel® QuickPath interconnect Production Q4’08 Integrated Memory Controller Volume Ramp 1H’09 Next Generation Memory (DDR3) PCI Express Gen 2 SLIDE | 14 SLIDE All future products, dates, and figures are preliminary and are subject to change without any notice. A P P R O H P C P R E S E N T A T I O N Appro Xtreme-X Supercomputers AA PP PP RR OO II NN TT EE RR NN AA TT II OO NN AA LL II NN CC Xtreme-X Supercomputer :: Reliability by Design • Redundancy where Required – Disk Drives • Diskless architecture – Fan Trays • N+1 Redundancy at the blade – Power Supplies • Hot Swappable N+1 Power supplies – Network components • Redundant InfiniBand and Ethernet networks – Servers • Management Nodes RAID6 Disk storage arrays • Management nodes - redundant pair (active/standby) • Sub-Management Nodes - redundant pair (active/active) • Compute nodes are hot swappable in sub-racks and are easily removed without removing any cabling from other nodes SLIDE | 16 SLIDE A P P R O H P C P R E S E N T A T I O N Xtreme-X Supercomputer :: The Thermal Problem • A 100C increase in ambient temperature produces a 50% reduction in the long-term reliability of electronic equipment • A 50C reduction in temperature can triple the life expectancy of electronic equipment. • MTBF and Availability depend on cooling. Source: BCC, Inc. Report CB-185R SLIDE | 17 SLIDE A P P R O H P C P R E S E N T A T I O N Xtreme-X Supercomputer :: Current Datacenter Cooling Methodology Top View Hot Aisle / Cold Aisle Hot Isle Cold Isle • As Much as 40% or the cold air flow by passes the equipment • Since the air is not directly pushed into the equipment, the cold air from the chillers are mixed with the hot air reducing the efficiency to cool the datacenter equipment. SLIDE | 18 SLIDE A P P R O H P C P R E S E N T A T I O N Xtreme-X Supercomputer :: Directed Airflow Cooling Methodology Top View • Up to 30% Improvement in Density with Lower Power per Equipment Rack • Delivers Cold Air directly to the equipment for optimum rack cooling efficiency. • Delivers comfortable temperature to the room for return to Chillers • Back-to-Back Rack configuration saves floor space in the datacenter and eliminates hot isles • FRU and maintenance is done in the front side of the blade rack cabinet | 19 SLIDE A P P R O H P C P R E S E N T A T I O N Xtreme-X Supercomputer :: Power/Cooling Efficiency & Density Xtreme-X Supercomputer Back-To-Back Configuration Rack 1 Rack 2 Sub-RackRack Sub-Rack Sub-Rack Sub-Rack Sub-Rack Sub-Rack Sub-Rack Sub-Rack SIDE VIEW SIDE VIEW FLOOR UNDERCOLD FLOOR AIR COLDFROM AIR FROM CHILLERS 30% IMPROVEMENT IN SPACE BY ELIMINATING HOT ISLES AND HAVING BACK-TO-BACK RACK CONFIGURATION SLIDE | 20 SLIDE A P P R O H P C P R E S E N T A T I O N Xtreme-X Supercomputer :: Based on Scalable Units • A new way to approach building Supercomputing Linux clusters • Multiple clusters of various sizes can be built and deployed into production very rapidly. • The SU concept is very flexible by accommodating clusters from very small node counts to thousands of nodes. • This new approach yields many benefits to dramatically lower total cost of ownership. Awarded Supercomputing Online 2007 Product of the Year SLIDE | 21 SLIDE A P P R O H P C P R E S E N T A T I O N Appro Cluster Engine AA PP PP RR OO II NN TT EE RR NN AA TT II OO NN AA LL II NN CC Appro Cluster Engine :: ACE: Scalable SW Architecture • Issues with today’s cluster environments – Change Operating systems EASILY and QUICKLY – Run jobs on the cluster efficiently – Monitor nodes for health – Simple cluster management (Ex. IP Address allocation) Cluster Management Software turns a Collection of Servers into a, Functional, Usable, Reliable, Available, computing system SLIDE | 23 SLIDE A P P R O H P C P R E S E N T A T I O N Appro Cluster Engine • Total Management Package 9 Network 9 Servers 9 Host 9 Scheduling 9 RAS • Two Tier Management Architecture 9 Scalable to large numbers of Servers 9 Minimal Overhead 9 Complete Remote Lights-Out Control • Redundancy provides both High Availability and High Performance 9 Redundant Management Servers 9 Redundant Networks 9 Active-Active Operation Supported SLIDE | 24 SLIDE A P P R O H P C P R E S E N T A T I O N Appro Cluster Engine :: ACE: Scalable SW Architecture Global File System Storage Server Mgmt : InfiniBand Node for Computing : 10GbE Storage Operation Server N GbE 2x GbE : GbE Operation
Details
-
File Typepdf
-
Upload Time-
-
Content LanguagesEnglish
-
Upload UserAnonymous/Not logged-in
-
File Pages35 Page
-
File Size-