NERSCPowering Scientific Discovery Since 1974

CS Chang

FES Requirements Worksheet

1.1. Project Information - Center for Plasma Edge Simulation

Document Prepared By

CS Chang

Project Title

Center for Plasma Edge Simulation

Principal Investigator

CS Chang

Participating Organizations

New York University, ORNL, PPPL, LBNL, MIT, Columbia U., Rutgers U. Lehigh U., Georgia Tech, Auburn U., U. Colorado, U. California at Irvine, Caltech, Hinton Associates

Funding Agencies

 DOE SC  DOE NSA  NSF  NOAA  NIH  Other:

2. Project Summary & Scientific Objectives for the Next 5 Years

Please give a brief description of your project - highlighting its computational aspect - and outline its scientific objectives for the next 3-5 years. Please list one or two specific goals you hope to reach in 5 years.

Develop the XGC large scale edge kinetic codes further for higher fidelity simulation of the electromagnetic multiscale edge physics in ITER. Using the kinetic codes, perform integrated simulation among kinetic, MHD, neutral particles, and atomic physics for higher fidelity understanding of the multiscale edge physics and the wall heat-load.

3. Current HPC Usage and Methods

3a. Please list your current primary codes and their main mathematical methods and/or algorithms. Include quantities that characterize the size or scale of your simulations or numerical experiments; e.g., size of grid, number of particles, basis sets, etc. Also indicate how parallelism is expressed (e.g., MPI, OpenMP, MPI/OpenMP hybrid)

Our current primary codes are XGC0 and XGC1 particle-in-cell codes. Particle motions are described by Largrangian equation of motion in 3D cylindrical coordinate system on realistic toroidal geometry with magnetic separatrix. Either Runge-Kutta or predictor-corrector methods are used. The field quantities are evaluated on grid mesh using linear multigrid PETSc solvers. The scale of the simulation is characterized by grid size, which then determines the number of particles. The parallelism is expressed by MPI/OpenMP hybrid. 

3b. Please list known limitations, obstacles, and/or bottlenecks that currently limit your ability to perform simulations you would like to run. Is there anything specific to NERSC?

Current limitation we are trying to overcome is in the algorithm which enables the fully electromagnetic turbulence simulation within the 5D gyrokinetic formalism. This is related to the time-step resolution of the fast electron motions, hence to the computing power. Presently, 5D ion full-f simulation of the DIII-D edge plasma requires 20 hours of simulation on 100,000 processor cores. Other machines can require more computing power. Fluid-kinetic or split-weight electron simplification technique can significantly reduce the computing-power requirement, by demanding only 2 times more number of processor cores instead of factor of 60 (full electron kinetics in deuteron plasmas). The electromagnetic simulation of DIII-D edge plasma thus requires 200K processor cores on Franklin for one-day completion. With higher computing power, the electrons can be simulated in full-velocity function, instead of the simplified velocity space function. ITER simulation will require about 1 million processor cores for a 20 hour run, assuming that the current linear scalability holds. 

3c. Please fill out the following table to the best of your ability. This table provides baseline data to help extrapolate to requirements for future years. If you are uncertain about any item, please use your best estimate to use as a starting point for discussions.

Facilities Used or Using

 NERSC  OLCF  ACLF  NSF Centers  Other:  

Architectures Used

 Cray XT  IBM Power  BlueGene  Linux Cluster  Other:  

Total Computational Hours Used per Year

 65,000,000 Core-Hours

NERSC Hours Used in 2009

 8,000,000 Core-Hours

Number of Cores Used in Typical Production Run

 15,000 - 170,000

Wallclock Hours of Single Typical Production Run

 20-100

Total Memory Used per Run

 40 GB

Minimum Memory Required per Core

0.3 GB

Total Data Read & Written per Run

 5,000 GB

Size of Checkpoint File(s)

1,000 GB

Amount of Data Moved In/Out of NERSC

 10 GB per  day

On-Line File Storage Required (For I/O from a Running Job)

 4 TB and 3,000 Files

Off-Line Archival Storage Required

 1 TB and 30 Files

Please list any required or important software, services, or infrastructure (beyond supercomputing and standard storage infrastructure) provided by HPC centers or system vendors.

As the system becomes larger, early detection for the processor failure is an important issue for an HPC code such as XGC1. A detection software, which can be compiled and run together with an HPC code can be quite helpful. If the software can replace the faulty node with another one, it will be even better. Otherwise, the code run can be stopped for a restart. 

4. HPC Requirements in 5 Years

4a. We are formulating the requirements for NERSC that will enable you to meet the goals you outlined in Section 2 above. Please fill out the following table to the best of your ability. If you are uncertain about any item, please use your best estimate to use as a starting point for discussions at the workshop.

Computational Hours Required per Year

 500,000,000

Anticipated Number of Cores to be Used in a Typical Production Run

 1,000,000

Anticipated Wallclock to be Used in a Typical Production Run Using the Number of Cores Given Above

 20-100

Anticipated Total Memory Used per Run

 100,000 GB

Anticipated Minimum Memory Required per Core

 0.1 GB

Anticipated total data read & written per run

 25,000 GB

Anticipated size of checkpoint file(s)

 5,000 GB

Anticipated Amount of Data Moved In/Out of NERSC

50 GB per  day

Anticipated On-Line File Storage Required (For I/O from a Running Job)

5 TB and 3,000 Files

Anticipated Off-Line Archival Storage Required

10 TB and 100 Files

4b. What changes to codes, mathematical methods and/or algorithms do you anticipate will be needed to achieve this project's scientific objectives over the next 5 years.

A new physics algorithm is needed, and currently under development, which utilizes the full-f ions and the delta-f electrons. Fully parallelized particle and grid data are needed for enhanced data locality and reduced memory requirement.

4c. Please list any known or anticipated architectural requirements (e.g., 2 GB memory/core, interconnect latency < 1 μs).

0.1 GB memory/core

4d. Please list any new software, services, or infrastructure support you will need over the next 5 years.

Since XGC1 is anticipated to be running at near the maximal capacity of the new machines over the next 5 years, efficient fault tolerance services will be needed. 

4e. It is believed that the dominant HPC architecture in the next 3-5 years will incorporate processing elements composed of 10s-1,000s of individual cores, perhaps GPUs or other accelerators. It is unlikely that a programming model based solely on MPI will be effective, or even supported, on these machines. Do you have a strategy for computing in such an environment? If so, please briefly describe it.

We have been quite successful in adapting to the current level of multi-core architecture by using the MPI/OpenMP hybrid mode. We find that the all-OpenMP operation per node is not the optimal solution on 12 core XT5. Instead, the solution was two Open-MP processes per node. As the core numbers increase per processing element, our low level strategy is to find the highest-performance mixture between OpenMP and MPI per node. At a higher level, our 3-5 years strategy is to develop asynchronous algorithms for effective utilization of many heterogeneous cores. A run-time scheduler such as StarPU will be used to coordinate and map threads to computational resources. Another approach is the incorporation of partitioned global address space (PGAS) languages to offer means for expressing locality of data. 
 
We are also looking into the GPGPUs. Sparse matrix-vector multiply has been demonstrated on GPGPU and is supported by optimized library from Nvidia. Similarly, multi-grid has been demonstrated to have efficient implementation on GPGPU. These are expected to be useful for the XGC paricle-in-cell code. 

New Science With New Resources

To help us get a better understanding of the quantitative requirements we've asked for above, please tell us: What significant scientific progress could you achieve over the next 5 years with access to 50X the HPC resources you currently have access to at NERSC? What would be the benefits to your research field if you were given access to these kinds of resources?

Please explain what aspects of "expanded HPC resources" are important for your project (e.g., more CPU hours, more memory, more storage, more throughput for small jobs, ability to handle very large jobs).

We will have higher fidelity 5D gyrokinetic simulation of the ITER plasma including kinetic electrons and electromagnetic turbulence in realistic diverted magnetic field geometry. This will allow us to understand many unknown, but necessary, physical phenomena for successful ITER research, which have not been possible using the present reduced models. 
 
Ability to handle very large jobs and more CPU hours are important for our project.