All times are Mountain Time (MST)
Monday, Nov. 13
8:00 p.m
Todd Gamblin, Lawrence Livermore National Laboratory
“Introducing the High Performance Software Foundation (HPSF)”
Abstract
The Linux Foundation is announcing the intent to form the High Performance Software Foundation (HPSF), which aims to build, promote, and advance a portable core software stack for high performance computing (HPC) by increasing adoption, lowering barriers to contribution, and supporting development efforts. Join us for remarks from the founding members and projects of HPSF!
8:15 p.m.
Rick Stevens, Argonne National Laboratory
“Introducing the Trillion Parameter Consortium”
Abstract
A global consortium of scientists from federal laboratories, research institutes, academia, and industry has formed to address the challenges of building large-scale artificial intelligence (AI) systems and advancing trustworthy and reliable AI for scientific discovery. The Trillion Parameter Consortium (TPC) brings together teams of researchers engaged in creating large-scale generative AI models to address key challenges in advancing AI for science.
Tuesday, Nov. 14
10:45 a.m.
Stephen Lin, Sandia National Laboratories
“Performant simulations of laser-metal manufacturing processes using adaptive, interface-conformal meshing techniques”
Abstract
Parts manufactured using laser-metal manufacturing techniques, such as laser welding and laser powder bed fusion additive manufacturing, are highly sensitive to the process details, which govern the formation of defects and distortions. Prediction of process outcomes requires models capable of representing both the complex dynamics present at the liquid metal-gas interface at the mesoscale and the overall macroscopic thermal response of the part. This work describes a model implemented in the Sierra mechanics code suite which uses mesh adaptivity and dynamic, interface conformal elements to capture both the high-fidelity physics at the mesoscale and the macroscopic thermal response of the part-scale. For the high-fidelity mesoscale model, the explicit representation of the interface in the dynamic mesh allows for simple and stable implementation of laser energy deposition, thermo-capillary convection, and vapor recoil pressure effects. For the part-scale macroscopic thermal response model, the build process for the entire geometry is resolved and restrictions on element sizes due to both the process layer height and laser spot size cause nearly intractable element counts without adaptivity. Improved adaptive meshing and load balancing strategies allow for scalable performance across thousands of cores and the efficient bridging of scales between laser and workpiece. Comparisons between our high-fidelity mesoscale model results and experimental data are presented as well as demonstrations of the model’s ability to predict pore formation, a key process-induced defect. Samples of the observed speed up and resource reduction for our part-scale models are also presented.
11:30 a.m.
Moderator: Meifeng Lin, Brookhaven National Laboratory — Panelists: Kerstin Kleese van Dam (Brookhaven National Laboratory), Eric Stahlberg (Frederick National Laboratory for Cancer Research) and Peter Coveney (University College London)
“Digital Twins and the Rise of the Virtual Human”
Abstract
Biomedical Digital Twins, an innovative fusion of medical science and information technologies, are poised to transform medicine and improve patient care. These new approaches to medicine are just beginning to impact the fundamental understanding of biology and influence technologies available for preventing, detecting, diagnosing, and treating disease. Notably, the availability of Biomedical Digital Twins is opening exciting new avenues for exploring more personalized approaches to address specific health, wellness, and well-being needs of individual patients. This panel presentation will report on the outcomes of the recent Virtual Human Global Summit 2023 (https://www.bnl.gov/virtual-human-global-summit/), presenting an overview of state of the art in this domain, from academia and national labs to industry, and outlining future research directions.
1:00 p.m.
Dominic Manno, Los Alamos National Laboratory
“Leveraging Computational Storage for Simulation Science Storage System Design”
Abstract
LANL has been engaged in how to exploit computation near data storage devices for simulation analysis. The activities in this area will be outlined with some early results and directions for the future will be covered.
1:45 p.m.
Lori Diachin, LLNL, ECP Director; Andrew Siegel, ANL, Director of ECP Application Development; Michael Heroux, SNL, Director of ECP Software Technology; Richard Gerber, NERSC, Director of ECP Hardware and Integration
“Delivering a Capable Exascale Computing Ecosystem – for Exascale and Beyond”
Abstract
A close look at the massive co-design, collaborative effort to build the world’s first capable computing ecosystem for exascale and beyond. With the delivery of the U.S. Department of Energy’s (DOE’s) first exascale system, Frontier, in 2022, and the upcoming deployment of Aurora and El Capitan systems by next year, researchers will have the most sophisticated computational tools at their disposal to conduct groundbreaking research. Exascale machines, which can perform more than a quintillion operations per second, are 1,000 times faster and more powerful than their petascale predecessors, enabling simulations of complex physical phenomena in unprecedented detail to push the boundaries of scientific understanding well beyond its current limits. But the benefits of the exascale computing era will be far-reaching, impacting numerous aspects of HPC. This incredible feat of research, development, and deployment has been made possible through a national effort sponsored by two Department of Energy organizations, the Office of Science and the National Nuclear Security Administration, to maximize the benefits of high-performance computing (HPC) for strengthening U.S. economic competitiveness and national security.
2:30 p.m.
Sudip Dosanjh, Lawrence Berkely National Laboratory
“The Next 50 Years: How NERSC is Evolving to Support the Changing Mission Space and Technology Landscape”
Abstract
As NERSC heads into 2024 and its 50th anniversary, the center already has its sights set on the next 50 years. This is second nature for NERSC, which is always looking ahead to what is coming next in science, research, and technology as well as planning, installing, and refining its systems and services in response – and anticipation. Now, as we contemplate the future, NERSC is already evolving and expanding, with an eye toward both near- and long-term goals. These include: NERSC-10: Work on NERSC’s next supercomputer, NERSC-10, is underway, scheduled for arrival in 2026. The NERSC-10 system will accelerate end-to-end DOE Office of Science workflows and enable new modes of scientific discovery through the integration of experiment, data analysis, and simulation. AI and ML: HPC centers are preparing for a shift toward new AI-enhanced workflows, recognizing the need for future systems to leverage new technologies and support emerging needs in AI and experimental/observational science in order to accelerate workflows and enable novel scientific discovery. Preparation for NERSC-10 and future system planning are aligned with these imperatives. Superfacility: Since 2019, NERSC has led the Berkeley Lab Superfacility Project, formalizing the work of supporting this model and enabling better communication across teams working on related topics. In 2022, NERSC’s superfacility team participated in the DOE ASCR program’s Integrated Research Infrastructure Architecture Blueprint Activity, a project designed to lay the groundwork for a coordinated, integrative research ecosystem going forward. Quantum: Quantum information science and quantum computing hold promise for tackling complex computational problems. With the NERSC user base in mind, NERSC has been developing staff expertise and engaging scientists, QIS researchers, and quantum computing companies, initially through the QIS@Perlmutter program.
3:15 p.m.
Mahantesh Halappanavar, Pacific Northwest National Laboratory
“ExaGraph: Graph and combinatorial methods for enabling exascale applications”
Abstract
Combinatorial algorithms in general and graph algorithms in particular play a critical enabling role in numerous scientific applications. However, the irregular memory access nature of these algorithms makes them one of the hardest algorithmic kernels to implement on parallel systems. With tens of billions of hardware threads and deep memory hierarchies, the exascale computing systems in particular pose extreme challenges in scaling graph algorithms. The codesign center on combinatorial algorithms, ExaGraph, was established to design and develop methods and techniques for efficient implementation of key combinatorial (graph) algorithms chosen from a diverse set of exascale applications. Algebraic and combinatorial methods have a complementary role in the advancement of computational science and engineering, including playing an enabling role on each other. In this presentation, we survey the algorithmic and software development activities performed under the auspices of ExaGraph and detail experimental results from GPU-accelerated preexascale and exascale systems.
4:00 p.m.
Marc Day, National Renewable Energy Laboratory
“Adaptive computing and multi-fidelity strategies for control, design and scale-up of renewable energy applications”
Abstract
We describe our ongoing research in adaptive computing and multi-fidelity modeling strategies. Our goal is to use a combination of low- and high-fidelity simulation models to enable computationally efficient optimization and uncertainty quantification. We develop optimization formulations that take into account the compute resources currently available, which act as a constraint with regards to the fidelity level simulation we can run while maximizing information gain. These strategies are being implemented into a software framework with a generalized API allowing its application to a broad range of applications, from power grid stability and buildings control to material synthesis and biofuels processing. We will discuss a few examples from these applications that can benefit from this approach, especially when considering challenges arising in scaling up experiments and simulations.