NERSC's next supercomputer will be a Cray pre-exascale system named “Perlmutter” in honor of Saul Perlmutter, an astrophysicist at Berkeley Lab who shared the 2011 Nobel Prize in Physics for his contributions to research showing that the expansion of the universe is accelerating. Dr. Perlmutter has been a NERSC user for many years, and part of his Nobel Prize-winning work was carried out on NERSC machines and the system name reflects and highlights NERSC's commitment to advancing scientific research.
Perlmutter, a Cray system based on the “Shasta” platform, will be a heterogeneous system comprising both CPU-only and GPU-accelerated nodes, with a performance of 3-4 times Cori, NERSC’s current platform. It will include a number of innovations designed to meet the diverse computational and data analysis needs of NERSC’s user base and speed their scientific productivity. The new system derives performance from advances in hardware and software, including a new Cray system interconnect, code-named Slingshot, which is designed for data-centric computing. Slingshot’s Ethernet compatibility, advanced adaptive routing, first-of-a-kind congestion control, and sophisticated quality of service capabilities improve system utilization and performance and scalability of supercomputing and AI applications and workflows. The system will also feature NVIDIA GPUs with new Tensor Core technology, direct liquid cooling and will be NERSC’s first supercomputer with an all-flash scratch filesystem. Developed by Cray to accelerate I/O, the 35-petabyte Lustre filesystem will move data at a rate of more than 5 terabytes/sec.
The system is scheduled to be delivered in two phases: Phase 1, with 12 GPU-accelerated cabinets and 35 PB of All-Flash storage, will be delivered in late 2020, and Phase 2 with 12 CPU cabinets will be delivered in mid 2021.
Each of Phase 1's GPU-accelerated nodes will have 4 of NVIDIA A100 Tensor Core GPUs based on the NVIDIA Ampere GPU architecture, along with 256GB of memory for a total of over 6000 GPUs. In addition, the Phase 1 nodes will each have a single AMD Milan CPU.
Each of Phase 2's CPU nodes will have 2 AMD Milan CPUs with 512 GB of memory per node. The system will contain over 3000 CPU-only nodes.
More detail on the CPUs and GPUs will be available in the coming weeks.
To ensure that our users are able to utilize the new technology in Perlmutter, NERSC has implemented a robust application readiness plan for simulation, data and learning applications through the NERSC Exascale Science Applications Program (NESAP). Support for complex workflows through new scheduling techniques and support for Exascale Computing Project (ECP) software is also planned on the new system.
The new system will be located in Wang Hall at Berkeley Lab, the site of NERSC's current supercomputers. NERSC is the DOE Office of Science’s (SC’s) mission high performance computing facility, supporting more than 7,000 scientists and 700 projects annually. The Perlmutter system represents SC’s ongoing commitment to extreme-scale science, developing new energy sources, improving energy efficiency, discovering new materials and analyzing massive data sets from scientific experimental facilities.
- NERSC Played Key Role in Nobel Laureate's Discovery
- NERSC Nobel Lecture Series: Saul Perlmutter Lecture, “Data, Computation, and the Fate of the Universe,” June 11, 2014
- Discovery of Dark Energy Ushered in a New Era in Computational Cosmology
- NERSC and the Fate of the Universe
- Computational Cosmology Comes of Age