Boosting Shared-Memory Supercomputing
Total Page:16
File Type:pdf, Size:1020Kb
casE STUDY Intel® Xeon® Processor 7500 Family Enterprise Server High-Performance Computing Performance for Data-Intensive Computing Boosting Shared-Memory Supercomputing NCSA doubles performance and nearly triples memory capacity with SGI Altix* UV system based on the Intel® Xeon® processor 7500 family As one of the premier high-performance computing (HPC) research institutions in the U.S., the National Center for Supercomputing Applications (NCSA) provides HPC resources for a wide range of scientific applications, including many whose performance needs require highly parallel shared-memory architectures. NCSA recently replaced its previous shared-memory system, an SGI Altix* supercomputer based on the Intel® Itanium® 2 processor, with an SGI Altix UV supercomputer powered by the Intel® Xeon® processor 7500 family. NCSA says the new system, which it calls Ember, consumes half the power while delivering double the performance and nearly triple the memory capacity. CHALLEngES • Rising demand, aging system. At nearly six years old, NCSA’s Intel Itanium 2 processor- based shared-memory supercomputer couldn’t keep up with rising demand. SOLUTIOns • Technologies to maximize data-intensive performance. NCSA chose an SGI Altix UV system with 256 six-core Intel Xeon processors 7500 family. The system has 8 TB of RAM and runs Red Hat Enterprise Linux*. IMPacT • More speed and capacity. The new supercomputer provides 16 teraflops (TF) of peak performance and 13.5 GB per second of I/O bandwidth, helping scientists accelerate progress on a variety of complex and important problems. • Cost and energy savings. The system consumes half as much power as the previous platform while delivering double the performance—the equivalent of four times more performance per watt. • Increased density. The system has just four racks of compute nodes, making it four “Our expectation was that we’d times denser than its predecessor and freeing valuable space in the data center. see a baseline of maybe a 20 to 50 or 70 percent increase in Resources for Memory-Intensive Six years ago, NCSA replaced an SGI shared-memory system that used propri- application performance over Applications Parallel shared-memory supercomputers etary processors, with an SGI Altix system the previous system. What we are crucial resources for HPC applications based on the Intel Itanium processor. NCSA later upgraded the system to the found is that many applications in areas such as computational chemistry, fluid dynamics, and solid mechanics— Intel Itanium 2 processor, but by 2010, are seeing a factor of two and the demand for those resources is that system was failing to keep pace with soaring demand. “Our allocation requests increase in performance, and growing. Few organizations understand that demand better than NCSA, a for shared-memory systems consistently sometimes much more.” national leader in providing advanced exceeded the available resources—by as infrastructure for open scientific much as seven times, in one review of requests,” says John Towns, director of – John Towns. discovery through the TeraGrid* program persistent infrastructure at NCSA and Director of Persistent Infrastructure and its successor, the eXtreme Digital* NCSA (XD*) program. chair of the TeraGrid Forum. The Intel® Xeon® processor’s energy SPOTLigHT ON NCSA The National Center for efficiency helps give NCSA four Supercomputing Applications (NCSA) times more performance per watt was established in 1986 as one of the original sites of the National Science Foundation’s Supercomputer Centers Program. Located at the University of Illinois at Urbana-Champaign, Towns says Ember is an essential tool for Individual applications run on up to 384 NCSA is supported by the state of shared-memory applications, particularly processor cores on the Ember system, Illinois, the University of Illinois, the for those where memory bandwidth can giving researchers access to a 2 TB pool National Science Foundation, and be a major constraint on performance. of shared memory. To handle the massive grants from other federal agencies. His team chose the Intel Xeon processor storage requirements, the system NCSA is a leader in deploying robust X7542 at 2.66 GHz and configured the connects to two SGI InfiniteStorage* HPC resources and in working with system to provide maximum memory 15000 controllers with 700 TB of capacity. research communities to develop new bandwidth per core for parallel shared- computing and software technologies. memory computing with large datasets. Advancing Research and Industry “We made several choices that were In production since November 2010, important to support parallel shared Ember provides additional resources to researchers across the nation’s TeraGrid applications in a shared memory As for NCSA’s future: “The new E7 environment more effectively,” Towns and XD programs. In addition, NCSA works with a number of private companies platform will definitely be on our radar explains. “The specific processor, the as we look to deploy and maintain a configuration, the DIMM size, the core to accelerate their use of innovative architectures, and Ember makes it easier next-generation evergreen system,” count, and other factors are all aimed says Towns. at maximizing memory bandwidth. for those companies to adapt shared- Ember also provides researchers with memory architectures with higher core Find a solution that is right for your approximately five and one-third counts and use their performance to drive organization. Contact your Intel gigabytes per core, which is more economic growth. representative or visit the Reference memory per core than any other system “The industry is typically several years Room at www.intel.com/references. that’s available across TeraGrid.” behind where we are, but we work with Towns says the Intel Xeon processor several technology-savvy companies X7542 at 2.66 GHz lets researchers whose goal is to be just six months take full advantage of Intel® QuickPath behind us,” Towns says. “Bringing in Interconnect (Intel® QPI). Intel QPI platforms like Ember is of great interest optimizes shared-memory performance to them because they know this is in their with an integrated memory controller and future. Working with us to prototype and high-speed, point-to-point connections test their next-generation applications from one processor to another and and often to run production versions of between processors and memory. “With their applications is very beneficial in this processor, we’re at the highest QPI helping them take advantage of the clocking and QPI bandwidth,” he states. latest advances.” “That improves communication between the processors and the node boards within the shared memory environment and supports our parallel applications more effectively.” Performance tests and ratings are measured using specific computer systems and/or components and reflect the approximate performance of Intel® products as measured by those tests. Any difference in system hardware or software design or configuration may affect actual performance. Intel may make changes to specifications, product descriptions, and plans at any time, without notice. This document and the information given are for the convenience of Intel’s customer base and are provided “AS IS” WITH NO WARRANTIES WHATSOEVER, EXPRESS OR IMPLIED, INCLUDING ANY IMPLIED WARRANTY OF MERCHANTABILITY, FITNESS FOR A PARTICULAR PURPOSE, AND NONINFRINGEMENT OF INTELLECTUAL PROPERTY RIGHTS. Receipt or possession of this document does not grant any license to any of the intellectual property described, displayed, or contained herein. Intel® products are not intended for use in medical, lifesaving, life-sustaining, critical control, or safety systems, or in nuclear facility applications. Copyright © 2011, Intel Corporation. All rights reserved. Intel, the Intel logo, Intel Core, Intel Itanium, Intel vPro, Intel Xeon, and Xeon Inside are trademarks of Intel Corporation in the U.S. and other countries. *Other names and brands may be claimed as the property of others. Printed in USA 0611/YMB/CMD/PDF Please Recycle 325613-001US.