Servers & Systems: The Right Compute
1752340 Members
5963 Online
108787 Solutions
New Article ๎ฅ‚
BillMannel

NASA achieves optimal energy efficiency with its first modular supercomputer

NASA has adopted a novel approach to cooling that not only enhances data center performance, but also conserves electricity and water. NASAโ€™s first modular supercomputer, Electra, is changing the game for researchers.

shutterstock_image (15).jpgAs organizations tackle increasingly complex and data-heavy challenges, high performance computing (HPC) systems are working overtime to quickly execute workloads and streamline data center operations. As a result, developers are striving to deploy a new breed of HPC system that combines extreme speed and density with superior energy efficiency. Organizations with large-scale data centers are beginning to adopt a new approach to energy usage, leveraging powerful and eco-friendly solutions to turbocharge operational performance.

As the needs and densities of servers and data centers increase, organizations are turning to advanced methods for cooling to bolster their HPC environments. Today, NASA is achieving new levels of water and power efficiency and performance of the Modular Supercomputing Facility (MSF) at NASAโ€™s Ames Research Center in Silicon Valley. Using technologies from Hewlett Packard Enterprise (HPE) and other partners, NASAโ€™s first module-based supercomputer, Electra, ranked 33rd on the November 2017 TOP500 list of the worldโ€™s most powerful supercomputers. The MSF uses a combination of natural resources and adiabatic technology, consuming less than 10% of the energy used in traditional supercomputing facilities.

Optimizing energy efficiency for HPC

NASA has adopted a novel approach to cooling that not only enhances data center performance, but also conserves electricity and water. Electraโ€™s new module employs a combination of outdoor air and adiabatic coolers on the roof to rapidly cool the system. More specifically, warm air is drawn through water-moistened pads, and as the water evaporates, the air is chilled and pushed out. Additionally, the new module adds four HPE E-cells to deliver even greater efficiency. An E-cell is a sealed unit which uses a closed-loop cooling technology to release heated air outside the data center in order to ensure 100% heat removal. By transporting facility-supplied water into the system, E-cells help to rapidly and continuously cool the system, without compromising on performance or cost.

In its first year, Electra used 95% less water than a traditional data center environment, and it is expected to save 1,000,000 kilowatts per hour and 1,300,00 gallons of water each year. This can provide users an additional 280 million hours of compute time each year as well as an additional 685 million hours of compute time to augment its sister system, Pleiades.

According to Bill Thigpen, Chief of the Advanced Computing Branch at Amesโ€™ NASA Advanced Supercomputing (NAS) Facility, โ€œThis is a different way for NASA to do supercomputing in a cost-effective manner. It makes it possible for us to be flexible and add computing resources as needed, and we can save about $35 million dollarsโ€”about half the cost of building another big facility.โ€ Thigpen is also Deputy Project Manager for NASA's High-End Computing Capability Project.

โ€œThis is a different way for NASA to do supercomputing in a cost-effective manner. It makes it possible for us to be flexible and add computing resources as needed, and we can save about $35 million dollarsโ€”about half the cost of building another big facility.โ€ - Bill Thigpen, Chief of the Advanced Computing Branch at Amesโ€™ NASA Advanced Supercomputing (NAS) Facility,

Deploying energy-optimized solutions

Electraโ€™s new module is based on the HPE SGI 8600, a scalable, high-density clustered supercomputer that utilizes liquid cooling to achieve maximum efficiency and substantial savings in energy usage. This leading-edge system is based on E-cells, each containing two 42U-high E-racks separated by a cooling rack. The E-cells are comprised of 1,152 nodes with dual 20-core Intelยฎ Xeonยฎ Gold 6148 processors, increasing Electraโ€™s theoretical peak performance from 1.23 petaflops to 4.79 petaflops. And with 24 racks (or 2,304 nodes), 78,336 total cores, and 368 terabytes of memory, Electra is engineered with the speed and robust compute capabilities to handle NASAโ€™s most challenging workloads.

Based on the success of the new module, NASA is considering an expansion of up to 16x the current capabilities of the modular environment. This effort would enable scientists and engineers nationwide to harness Electra for their research supporting NASA missions.

To learn more about the benefits of liquid cooling versus. air cooling, I invite you to visit me on Twitter at @Bill_Mannel. And check out @HPE_HPC for the latest news and updates in HPC innovation.

If you want to learn more about HPC and advanced technologies at work today:

 

Featured articles:

 

About the Author

BillMannel

As VP & GM for HPC, I lead worldwide business execution and commercial HPC focus for one of the fastest growing market segments in Hewlett Packard Enterpriseโ€™s Hybrid IT (HIT) group that includes the recent Cray acquisition and the HPE Apollo portfolio.