2022 ECP Annual Meeting
Each year, the Exascale Computing Project, or ECP, holds an annual meeting to highlight the many technical accomplishments of its talented research teams and to provide a collaborative working forum. Through a combination of talks, workshops, tutorials, and gatherings for planning and co-design, the meeting fosters project understanding, team building, and continued progress.
Among the attendees are hundreds of researchers, domain scientists, mathematicians, computer and computational scientists, US high-performance computing (HPC) vendor participants, and project management experts who support ECP’s research focus areas of Application Development, Software Technology, and Hardware and Integration. Also present are members of ECP’s external advisory group, the ECP Industry and Agency Council, and the project’s sponsors and stakeholder program managers from the Department of Energy, with representation from the Advanced Scientific Computing Research program of the Office of Science and the Advanced Simulation and Computing program of the National Nuclear Security Administration.
Behind the collaborative spirit of the ECP Annual Meeting is inspiration borne of a vision shared by technology leaders, industry luminaries, computational pioneers, and forward-thinking application, software, and hardware experts to push the boundaries of HPC to shape the nation’s future exascale ecosystem.
The objective of sharing the following content is to provide insights from the ECP 2022 Annual Meeting to the broad HPC community.
Affiliation Abbreviations:
- AL: Ames Laboratory
- AMD: Advanced Micro Devices
- ANL: Argonne National Laboratory
- BNL: Brookhaven National Laboratory
- CAL: University of California, Berkeley
- CBIT: Center for Biomedical Informatics and Information Technology
- CMU: Carnegie Mellon University
- CU Denver: University of Colorado Denver
- CU: University of Colorado Boulder
- DOD: Department of Defense
- GE: General Electric
- GFDL: Geophysical Fluid Dynamics Laboratory
- HPE: Hewlett Packard Enterprise
- ICL: Innovative Computing Laboratory
- ISU: Iowa State University
- LANL: Los Alamos National Laboratory
- LBNL: Lawrence Berkeley National Laboratory
- LLNL: Lawrence Livermore National Laboratory
- MIT CCSE: MIT Center for Computational Science and Engineering
- NASA: National Aeronautics and Space Administration
- NERSC: National Energy Research Scientific Computing Center
- NETL: National Energy Technology Laboratory
- NOAA: National Oceanic and Atmospheric Administration
- NREL: National Renewable Energy Laboratory
- NU: Northwestern University
- ORNL: Oak Ridge National Laboratory
- PPPL: Princeton Plasma Physics Laboratory
- PU: Purdue University
- RU: Rice University
- SBU: Stony Brook University
- SFSU: San Francisco State University
- SHI: Sustainable Horizons Institute
- SLAC: SLAC National Accelerator Laboratory
- SMU: Southern Methodist University
- SNL: Sandia National Laboratory
- TUD: Technical University – Dresden, Germany
- UA: University of Alabama
- UB: University of Buffalo
- UC: University of Chicago
- UDEM: Montreal university
- UIUC: University of Illinois
- UO: University of Oregon
- UofU: University of Utah
- UT: University of Tennessee
- UW: University of Wisconsin
- VT: Virginia Polytechnic Institute and State University
2022 Annual Meeting Tutorials
Title | Date | Materials (Slides and/or Recordings) | Session Duration (Hours) | Organizers/Speakers |
---|---|---|---|---|
Performance Tuning with the Roofline Model on GPUs and CPUsThe HPC community is on a never ending quest for better performance and scalability. Performance models and tools are an integral component in the optimization process as they quantify performance relative to machine capabilities, track progress towards optimality, and identify performance bottlenecks. The Roofline performance model offers an insightful and intuitive method for extracting the key computational characteristics of HPC applications and comparing them against the performance bounds of CPUs and GPUs. Its capability to abstract the complexity of memory hierarchies and identify the most profitable optimization techniques has made Roofline-based analysis increasingly popular in the HPC community. This 180-minute tutorial is centered around four components. First, we will introduce the Roofline model and discuss how changes in data locality and arithmetic intensity visually manifest in the context of the Roofline model. Next, we will introduce and demonstrate the use of Roofline analysis in NVIDIA Nsight Compute and discuss several real-world use cases of Roofline in the NERSC NESAP effort. After a short break, we will return and introduce and demonstrate the use of AMD’s new Roofline tool for analysis of AMD GPU-accelerated applications running at OLCF. Finally, we will conclude by introducing and demonstrating Intel Advisor’s capability for Roofline analysis of applications running on Intel GPUs and CPUs as well as Roofline use cases at ALCF. | Mon, May 2 | 3.5 | Samuel Williams (LBNL), Neil Mehta (LBNL), Noah Wolfe (AMD), Xiaomin Lu (AMD), JaeHyuk Kwack (ANL) | |
Using HDF5 Efficiently on HPC SystemsHDF5 is a data model, file format, and I/O library that has become a de facto standard for HPC applications to achieve scalable I/O and for storing and managing big data from computer modeling, large physics experiments, and observations. Several ECP applications are currently using or planning to use HDF5 for I/O. Many new features, such as caching and prefetching, asynchronous I/O, log structured I/O, etc., have been developed in ECP for HDF5 applications’ taking advantage of exascale storage subsystems. | Mon, May 2 | Recording | 3.5 | Suren Byna (LBNL), Scot Breitenfeld (The HDF Group), Houjun Tang (LBNL), Huihuo Zheng (ANL), Jordan Henderson (The HDF Group), Qiao Kang, Neil Fortner (The HDF Group), Wei-keng Liao (NU), Michael Brim (ORNL), Kaiyuan Hou (NU), Dana Robinson (The HDF Group) |
GPU Capable Sparse Direct SolversIn this tutorial we illustrate the use of the sparse direct solvers and factorization based preconditioners SuperLU and STRUMPACK on modern HPC systems with GPU accelerators. Sparse direct solvers rely on LU factorization and the accompanying triangular solution. They are indispensable tools for building various algebraic equation solvers. They can be used as direct solvers, as coarse-grid solvers in multigrid, or as preconditioners for iterative solvers. Particular benefits of direct solvers are their robustness and their efficiency when solving systems with multiple right-hand sides. | Mon, May 2 | Recording | 3.5 | Sherry Li (LBNL), Yang Liu (LBNL), Pieter Ghysels (LBNL) |
CMake TutorialAn interactive tutorial on how to write effective build systems using CMake. | Mon, May 2 | 3.5 | Zack Galbreath (Kitware, Inc.), Betsy McPhail (Kitware, Inc.), Julien Jomier (Kitware, Inc.) | |
E4S for ECP ST and AD teamsWith the increasing complexity and diversity of the software stack and system architecture of high performance computing (HPC) systems, the traditional HPC community is facing a huge productivity challenge in software building, integration and deployment. Recently, this challenge has been addressed by software build management tools such as Spack that enable seamless software building and integration. Container based solutions provide a versatile way to package software and are increasingly being deployed on HPC systems. The DOE Exascale Computing Project (ECP) Software Technology focus area is developing an HPC software ecosystem that will enable the efficient and performant execution of exascale applications. Through the Extreme-scale Scientific Software Stack (E4S) [https://e4s.io], it is developing a curated, Spack-based, comprehensive and coherent software stack that will enable application developers to productively write highly parallel applications that can portably target diverse exascale architectures. E4S provides both source builds through the Spack platform and a set of containers (base, full featured for three GPU architectures) that feature a broad collection of HPC software packages. E4S exists to accelerate the development, deployment, and use of HPC software, lowering the barriers for HPC and AI/ML users. It provides container images, build manifests, and turn-key, from-source builds of popular HPC software packages developed as Software Development Kits (SDKs). This effort includes a broad range of areas including programming models and runtimes (MPICH, Kokkos, RAJA, OpenMPI), development tools (TAU, PAPI), math libraries (PETSc, Trilinos), data and visualization tools (Adios, HDF5, Paraview), and compilers (LLVM), all available through the Spack package manager. The tutorial will describe the community engagements and interactions that led to the many artifacts produced by E4S, and will introduce the E4S containers that are being deployed at the HPC systems at DOE national laboratories. Specifically, the participants will learn: | Mon, May 2 | Recording | 1.5 | Sameer Shende (UO) |
Using Spack to Accelerate Developer WorkflowsThe modern scientific software stack includes thousands of packages, from C, C++, and Fortran libraries, to packages written in interpreted languages like Python and R. HPC applications may depend on hundreds of packages spanning all of these ecosystems. To achieve high performance, they must also leverage low-level and difficult-to-build libraries such as MPI, BLAS, and LAPACK. Integrating this stack is extremely challenging. The complexity can be an obstacle to deployment at HPC sites and deters developers from building on each other’s work. | Mon, May 2 | Recording | 3.5 | Todd Gamblin (LLNL), Greg Becker (LLNL), Richarda Butler (LLNL), Tamara Dahlgren (LLNL), Adam Stewart (UIUC) |
TAU Performance SystemThe TAU Performance System is a powerful and highly versatile profiling and tracing tool ecosystem for performance analysis of parallel programs at all scales. Developed over the last two decades, TAU has evolved with each new generation of HPC systems and supports performance evaluation of applications targeting GPU platforms from NVIDIA, Intel, and AMD. The tutorial will demonstrate how to generate profiles and traces with TAU using profiling interfaces including Kokkos, OpenACC, OpenMP Tools Interface, CUPTI (CUDA), OpenCL, Intel Level Zero, AMD Rocprofiler, and Roctracer. TAU's 3D profile browser, ParaProf and PerfExplorer tools will be used in the hands-on to analyze the profile data. The tutorial will also show integration with LLVM compiler toolchain, including the TAU plugin for selective instrumentation for compiler-based instrumentation at the routine and file level. It will also demonstrate TAU's use in container based environments using the Extreme-scale Scientific Software Stack (E4S) containers. | Mon, May 2 | 1.5 | Sameer Shende (UO), Kevin Huck (UO) | |
AMD Software Tools for Exascale ComputingThis tutorial will be a set of live sessions and tutorials demonstrating the capabilities of various software tools that support AMD CPU and GPU hardware. These tools will be available on OLCF's upcoming system, Frontier. The following Radeon Open Compute (ROCm) integrated tools will be presented in AMD HPC Tools session: | Mon, May 2 | Recording | 3.5 | Bill Williams (TUD), Bert Wesarg, Giuseppe Congiu (ICL), Michel Dagenais (UDEM), Arnaud Fiorini (UDEM), Yoann Heitz (UDEM), John Mellor-Crummey (RU), Barton Miller (UW), Sameer Shende (UO), John Del-Signore (Perforce Software, Inc.), Beau Paisley, Louise Spellacy (Arm Ltd), Bill Burns (Perforce), Karlo Kraljic (HPE) |
Performance Autotuning of ECP Applications with Gaussian Process-Based and Cloud Database-Enhanced GPTune PackageAgenda for the day: https://bit.ly/37MYJtf | Mon, May 2 | Recording | 3.5 | Yang Liu (LBNL), Younghyun Cho (CAL), James Demmel (CAL), Sherry Li (LBNL) |
Dense Linear Algebra and FFT Libraries: SLATE, MAGMA, heFFTeThis tutorial focuses on the SLATE and MAGMA linear algebra libraries and the heFFTe multidimensional FFT library. These libraries are designed to harness today's supercomputers with multicore CPUs and multiple GPU accelerators. The tutorial covers practical aspects in setting up matrices and calling the libraries in your application. No prior knowledge is required. | Mon, May 2 | Recording | 1.5 | Mark Gates (UT), Stan Tomov (UT) |
Developing a Testing and Continuous Integration Strategy for your TeamA thorough and robust testing regime is central to the productive development, evolution, and maintenance of quality, trustworthy scientific software. Continuous integration, though much discussed, is just one element of such a testing regime. Most project teams feel that they could (and should) do a “better job” of testing. In many cases, designing and implementing a strong testing strategy can seem so daunting that it is hard to know where to start. | Mon, May 2 | Recording | 1.5 | Greg Watson (ORNL), David Rogers (ORNL) |
Overview and Use of New Features in the SUNDIALS Suite of Nonlinear and Differential/Algebraic Equation SolversSUNDIALS is a SUite of Nonlinear and DIfferential/ALgebraic equation Solvers consisting of six packages which provide robust time integrators and nonlinear solvers. The suite is designed to be easily incorporated into existing simulation codes and to require minimal information from the user. The modular implementation allows users to easily supply their own data structures underneath SUNDIALS packages and to easily incorporate outside solver packages or user-supplied linear solvers and preconditioners. SUNDIALS consists of the following six packages: CVODE, solves initial value problems for ordinary differential equation (ODE) systems with linear multi-step methods; CVODES, solves ODE systems and includes sensitivity analysis capabilities (forward and adjoint); ARKODE, solves initial value ODE problems with explicit, implicit, or IMEX additive Runge-Kutta methods; IDA, solves initial value problems for differential-algebraic equation (DAE) systems with BDF methods; IDAS, solves DAE systems and includes sensitivity analysis capabilities (forward and adjoint); KINSOL, solves nonlinear algebraic systems. | Mon, May 2 | Recording | 1.5 | Carol S. Woodward (LLNL), Cody J. Balos (LLNL), David J. Gardner (LLNL), Daniel Reynolds (SMU) |
OpenMP Tutorial for ECPWhile most HPC developers have OpenMP experience, many are not familiar with the latest OpenMP features and how to use them in their codes. Modern OpenMP can be used in creative ways to map parallelism to current and emerging parallel architectures. Yet it is much broader and more complex than OpenMP was even just three years ago, and a significant amount of training is needed if programmers are to be able to exploit it fully. | Mon, May 2 | Recording | 3.5 | Vivek Kale (BNL), Dossay Oryspayev (BNL), Johannes Doerfert (ANL), Thomas Scogland (LLNL), Colleen Bertoni (ANL) |
Transfer learning and online tuning with PROTEAS-TUNE/ytoptytopt is a machine-learning-based search software package developed within the ECP PROTEAS-TUNE project. It consists of sampling a small number of input parameter configurations of a given kernel/library/application, evaluating them, and progressively fitting a surrogate model over the input-output space until exhausting the user-defined time or the maximum number of evaluations. In this tutorial, we will present ytopt’s new transfer learning and online tuning capabilities. The transfer learning in autotuning is useful to reduce the tuning time when dealing with related autotuning tasks. This is achieved by training a model to learn the promising regions of the search space in one autotuning setting and using the learnt model to bias the search for the related autotuning settings. The online tuning is useful when high-performing parameter configurations need to be generated with severe tuning budget constraints. To that end, in the offline mode, ytopt trains a model with the high-performing configurations on various representative inputs; in the online mode, given a new unseen input, the trained model is used to generate high-performing configurations directly without any search. | Mon, May 2 | Recording | 1.5 | Prasanna Balaprakash (ANL), Xingfu Wu (ANL), Michael Kruse, Brice Videau (ANL), Paul Hovland (ANL), Mary Hall (UofU) |
ECP Software for In Situ Visualization and AnalysisThis 90-minute tutorial will cover ECP software for in situ visualization and analysis, which is an important capability for addressing slow I/O on modern supercomputers. The majority of the time will be spent on two products: Ascent and Cinema. For Ascent, an in situ library supported by ECP ALPINE, topics covered will include how to integrate into a simulation code, data model, build and linking issues, and an overview of capabilities. For Cinema, topics covered will include both examples of use cases and information on how to get started. Finally, we will give short presentations on other ECP software projects for in situ processing. One short presentation will be on Catalyst, which is the in situ form of the popular ParaView tool and which is also supported by ECP ALPINE. The other will be on VTK-m, which is a many-core visualization library used by Ascent and Catalyst, and can also be directly incorporated into simulation codes as well. | Mon, May 2 | Recording | 1.5 | Hank Childs (UO), Cyrus Harrison (LLNL), Terry Turton (LANL), David Rogers (ORNL), Kenneth Moreland, Berk Geveci (Kitware Inc.) |
Automated Application Performance Analysis with Caliper, SPOT, and HatchetAt LLNL, we have developed a workflow enabling users to automate application performance analysis and pinpoint bottlenecks in their codes. Our workflow leverages three open-source tools – Caliper, SPOT, and Hatchet – to provide a wholistic suite for integrated performance data. In this tutorial, we will provide an overview of each of the tools, and demonstrate how to profile your applications with Caliper, how to visualize your performance data in SPOT, and how to programmatically analyze your data with Hatchet. | Mon, May 2 | Recording | 1.5 | David Boehme (LLNL), Stephanie Brink (LLNL), Olga Pearce (LLNL) |
The SENSEI Generic In Situ Interface – A TutorialSENSEI, which is part of the DOE Exascale Computing Project (ECP) Data-Vis SDK, is an open-source, generic in situ interface that allows parallel simulations or other data producers to code-couple to parallel third-party endpoints. These endpoints may be applications or tools/methods, including user-written code in C++ or Python and a growing class of data-intensive capabilities accessible through Python, capabilities such as AI/ML. Once a data producer is instrumented with the SENSEI interface, changing to a different endpoint is as simple as modifying an XML configuration file with a text editor. SENSEI fully supports configurations where the simulation and endpoint are run at differing levels of concurrency. SENSEI will manage the potentially tricky process of partitioning and moving data from the M producer ranks to the N consumer ranks. The movement of data may be bidirectional, which means that a simulation sends data to an endpoint and has access to results computed from the endpoint. While initially designed and implemented for analysis, visualization, and other data-intensive in situ tasks, SENSEI's design and implementation support the coupling of arbitrary code. This tutorial will present an overview of SENSEI, along with several hands-on/follow-along coding examples to introduce you to key capabilities. | Tue, May 3 | Recording | 1.5 | E. Wes Bethel (SFSU), Silvio Rizzi (ANL), Burlen Loring (LBNL) |
Accelerating your application I/O with UnifyFSUnifyFS is a user-level file system that highly-specialized for fast shared file access on HPC systems with distributed burst buffers. In this tutorial, we will present users with an introductory overview of the lightweight UnifyFS file system that can be used to improve the I/O performance of HPC applications. We will begin at a high level describing how UnifyFS works with burst buffers and how users can incorporate it into their jobs. Following this, we will dive into more detail on what kinds of I/O UnifyFS currently supports and what we expect it to support in the future. We will also discuss the interoperation of UnifyFS with HDF5 and MPI-IO. | Wed, May 4 | Recording | 1 | Michael Brim (ORNL), Cameron Stanavige (LLNL) |
Tools for Floating-Point Debugging and Analysis: FPChecker and FLiTWith the proliferation of heterogeneous architectures, multiple compilers and programing models, numerical inconsistencies and errors are not uncommon. This tutorial will show how to debug numerical issues using two tools, FPChecker and FLiT. FPChecker performs detection of floating-point exceptions and other anomalies, reporting the location of such errors (file and line number) in a comprehensive report. It can check for errors in CPUs (via the clang/LLVM compiler) and in NVIDIA GPUs. FLiT identifies the location of code that is affected by compiler optimizations and can produce inconsistent numerical results with different compilation flags. The tutorial will provide hands-on exercises to demonstrate the tools. | Wed, May 4 | Recording | .75 | Ignacio Laguna (LLNL), John Jacobson (UofU), Cayden Lund (UofU), Ganesh Gopalakrishnan (UofU) |
FFTX: Next-Generation Open-Source Software for Fast Fourier TransformsThe FFTX project has two goals: | Thu, May 5 | Recording | 1.5 | Phillip Colella (LBNL), Franz Franchetti (CMU), Patrick Broderick, Peter McCorquodale (LBNL) |
Kokkos EcoSystem New Features TutorialThis Kokkos Tutorial will teach attendees to leverage important new capabilities in the KokkosSystem. Attendees are assumed to have a basic understanding of the Kokkos EcoSystem, specifically this tutorial will not be an introduction to capabilities covered in the Kokkos lectures (https://kokkos.link/the-lectures). | Thu, May 5 | 1.5 | Christian Trott (SNL), Damien Lebrun-Grandie (ORNL), Siva Rajamanickam (SNL), Luc Berger-Vergiat (SNL) | |
ExaGraph: Combinatorial Methods for Enabling Exascale ApplicationsWe are working on several applications in the domain of combinatorial scientific computing, spanning infection modeling, numerical analysis, and multi-omics analytics. Most of these applications are challenging to efficiently scale (both strong and weak cases) due to the irregular nature of computation required to process sparse inputs in distributed-memory. Therefore, it is vital to identify the trade-offs that can be exposed such that a user can select an option that can yield the desired performance at the expense of quality/accuracy. Also, a number of combinatorial applications are memory access intensive, and scaling up to larger inputs may require developing new algorithms or heuristics. | Fri, May 6 | 1.5 | Sayan Ghosh (PNNL), Marco Minutoli (PNNL), Mahantesh M. Halappanavar (PNNL), Aydin Buluc (LBNL), Alex Pothen (PU), Erik G. Boman (SNL) | |
Updates to ADIOS 2.8: Storage and in situ I/OAs concurrency and complexity continue to increase on high-end machines, I/O performance is rapidly becoming a fundamental challenge to achieving exascale computing. To address this challenge, wider adoption of higher-level I/O abstractions will be critically important. The ADIOS I/O framework provides data models, portable APIs, storage abstractions, in situ infrastructure, and self-describing data containers. These capabilities enable reproducible science, allow for more effective data management throughout the data life cycle, and facilitate parallel I/O with high performance and scalability. In this tutorial, participants will learn about the ADIOS concept and APIs and we will show a pipeline of a simulation, analysis and visualization, using both storage I/O and in situ data staging. Participants will also learn how to use state-of-the art compression techniques with ADIOS. We will introduce the new BP5 file format and show new functionalities around handling many steps in a single dataset efficiently. | Fri, May 6 | Recording | 3.5 | Scott Klasky (ORNL), Norbert Podhorszki (ORNL) |
Debugging and Performance Profiling for FrontierThis half-day tutorial will walk through practical examples of debugging and performance profiling MPI programs on Crusher, an OLCF computer available now with nodes identical to Frontier. The tutorial will cover the range of tools provided by both the HPE Cray Programming Environment and the AMD ROCm Platform. In particular, topics will include the following. | Fri, May 6 | Recording | 3.5 | Trey White (HPE), Stephen Abbott (HPE), Constantinos Makrides (HPE) |
Porting PETSc-based application codes to GPU-accelerated architecturesThe Portable, Extensible Toolkit for Scientific Computation (PETSc) provides scalable solvers for nonlinear time-dependent differential and algebraic equations and for numerical optimization. In this tutorial, we will demonstrate, through case studies, how to port PETSc application codes to perform the computations on GPUs and to understand the expected performance gains. We will present examples illustrating how users can employ their preferred GPU programming model for application “physics” code, independent of the GPU programming model used for PETSc solvers, and we will examine an end-to-end case study that combines finite element assembly (via the libCEED library) with PETSc algebraic solvers, all running on GPUs. | Fri, May 6 | Recording | 1.5 | Richard Mills (ANL), Mark F. Adams, Matthew Knepley (UB), Todd Munson, Barry Smith (Flatiron Institute), Junchao Zhang (ANL), Jed Brown (CU) |
How To Leverage New MPI Features for Exascale ApplicationsTo prepare for the upcoming Exascale supercomputers, an number of new features have been designed and implemented in MPI to better support Exascale. In this tutorial we will focus on partitioned communication, MPI Sessions, and Open MPI’s support for user-level threading. Partitioned communication in a new interface designed for efficient and performant data transfer when using multithreaded and accelerator-based communication. In this tutorial we will explain the new partitioned communication interface as it exists in MPI 4.0, present examples with the interface in use, present proposed and in development changes to these designs, and share insights based on experience converting codes to use partitioned communication. Next, we will present MPI sessions, a new initialization scheme that allows for multiple instances of MPI to be used concurrently in a single process. This feature enables scientific libraries to easily maintain data separation by initializing an independent ‘session’ of MPI. Additionally, we will present the recently added support of user-level threading libraries (ULT) significantly improves the interaction between Open MPI and threading runtimes that rely on cooperative multitasking. In this session we also showcase a hybrid application that relies on this functionality for correctness and performance and give insight into future development in this area. | Fri, May 6 | Recording | 1.5 | Matthew Dosanjh (SNL), Howard Prichard, W. Pepper Marts, Jan Ciesko (SNL) |
Using Containers to Accelerate HPCWithin just the past few years, the use of containers has revolutionized the way in which industries and enterprises have developed and deployed computational software and distributed systems. The containerization model has gained traction within the HPC community as well with the promise of improved reliability, reproducibility, portability, and levels of customization that were previously not possible on supercomputers. This adoption has been enabled by a number of HPC Container runtimes that have emerged including Singularity, Shifter, Enroot, Charliecloud and others. | Fri, May 6 | Recording | 1.5 | Andrew Younge (SNL), Sameer Shende (UO), Shane Canon |
2022 Annual Meeting Breakouts, BoFs, and Panels
Title | Date | Session Type | Materials (Slides and/or Recordings) | Session Duration (Hours) | Organizers/Speakers |
---|---|---|---|---|---|
Advances in Science and Engineering Enabled by ECP Applications IThree ECP Application Development projects (GAMESS, Combustion-PELE, EXAALT) temporarily turn their focus away from code development and optimization to highlight recent scientific and engineering accomplishments in their respective domains. The motivating science drivers and the anticipated scientific impact of exascale computing on their application areas will be discussed. | Tue, May 3 | Breakout | Recording | 2 | Mark Gordon, Jackie Chen (SNL), Danny Perez (LANL) |
CEED: High-Order Methods, Applications and Performance for ExascaleThis session will present a progress update on the research and development activities in the CEED co-design center, targeting state-of-the-art high-order finite-element algorithms for high-order applications on GPU-accelerated platforms. We will discuss the GPU developments in several components of the CEED software stack, including the MFEM, Nek, libCEED and libParanumal projects, and report performance and capability improvements in several CEED-enabled miniapps and applications on both NVIDIA and AMD GPU systems. We are interested in getting feedback from the whole community, so please join us if you'd like to learn more about GPU-performant high-order methods, or have questions for the CEED team. | Tue, May 3 | Breakout | Recording | 2 | Tzanio Kolev (LLNL), Paul Fischer (UIUC), Tim Warburton (VT), Misun Min (ANL), Noel Chalmers (AMD), Damon McDougall (AMD), Malachi Phillips (SNL), Jed Brown (CU) |
ECP Broader Engagement InitiativeThis session will begin with remarks by Dr. Julie A. Carruthers, Senior Science and Technology Advisor, Office of the Principal Deputy Director and Acting Director for the Office of Scientific Workforce Diversity, Equity, and Inclusion, DOE, who will speak about advancing policies and practices for promoting diversity, equity, and inclusion at DOE. We will then introduce the Broader Engagement Initiative, which has the mission of establishing a sustainable plan to recruit and retain a diverse workforce in the DOE HPC community by fostering a supportive and inclusive culture within the computing sciences at DOE national laboratories. We will describe key activities within three complementary thrusts: establishing a Workforce Development and Retention Action Group, creating accessible ‘Intro to HPC’ training materials, and establishing the Sustainable Research Pathways for High-Performance Computing (SRP-HPC) internship/mentoring program. We are leveraging ECP’s unique multi-lab partnership to work toward sustainable collaboration across the DOE community, with the long-term goal of changing the culture and demographic profile of DOE computing sciences. The session will include interactive (and fun!) activities so that attendees can learn more about the initiative’s goals and how each person can contribute to the success of the program. | Tue, May 3 | Breakout | 2 | Julie Carruthers (DOE), Ashley Barker (ORNL), Lois Curfman McInnes (ANL), Suzanne Parete-Koon (ORNL), Sreeranjani "Jini" Ramprakash (ANL), Mary Ann Leung (SHI), Daniel Martin (LBNL) | |
Facility Deployment of E4S at ALCF, OLCF, and NERSCThe Extreme Scale Scientific Software Stack (E4S) is a collection of open source software packages for development and running scientific software on HPC systems. E4S software collection is a subset of spack packages that are built and tested regularly. E4S software stack is available as a spack configuration (spack.yaml), container images in docker and singularity image format, a spack buildcache, and AWS AMI image. | Tue, May 3 | Breakout | 1 | Shahzeb Siddiqui (LBNL), Jamie Finney (ORNL), Matt Belhorn (ORNL), Frank Willmore (ANL) | |
Checkpointing with VELOC: Challenges, Use Cases, IntegrationEfficient checkpoint-restart is a capability that many ECP applications need. However, they rely on custom solutions that are burdened by many issues: (1) performance and scalability of the underlying parallel file system where the checkpoints need to be saved in a reliable fashion; (2) resource-efficiency concerns (e.g., extra memory overhead) needed to achieve high performance and scalability; (3) complexity of interacting with a heterogeneous storage hierarchy on future Exascale machines (e.g., many vendor-specific external storage options like burst buffers with custom APIs) that leads to loss of productivity. The ECP VeloC project aims to solve these issues through a specialized middleware that aims to deliver high performance and scalability with minimal overhead on resource utilization, while exposing a simple API that hides the complexity of interacting with a heterogeneous storage layer. Several ECP applications supported by VeloC and various other groups have been experimenting with VeloC in various checkpoint-restart scenarios tailored to their specific needs. This breakout session aims highlight the latest features available in VeloC and to share the experience VeloC users with the ECP community, in the hope of raising awareness of what benefits can be expected from VeloC and what is the best way to integrate with it. | Tue, May 3 | Breakout | Recording | 1 | Bogdan Nicolae (ANL), Salman Habib (ANL), Murat Keceli (ANL), Keita Terianishi (SNL) |
The Programming Environments Landscape at the ASCR FacilitiesThe three ASCR computing facilities (ALCF, NERSC, and OLCF) are in the process of fielding their next generation of systems (Aurora, Perlmutter, and Frontier). All three will be based on GPU accelerators from different vendors, and those vendors are offering different “preferred” programming environments. However, there are a number of efforts underway to broaden the availability of those environments and provide an increased degree of commonality across the facilities to facilitate application portability. | Tue, May 3 | Breakout | Recording | 1 | David E. Bernholdt (ORNL), Jack Deslippe (LBNL), Scott Parker (ANL) |
Everything Proxy Apps: Platform/Application Readiness and The Future | Tue, May 3 | Breakout | Recording | 2 | Jeanine Cook (SNL), Vinay Ramakrishnaiah (LANL), Ramesh Balakrishnan, Saumil Patel (ANL), Jamaludin Mohd Yusof (LANL), Alvaro Mayagoitia (ANL) |
Compression for Scientific Data for ECP applications (part 1)Abstract: Large-scale numerical simulations, observations, and experiments are generating very large datasets that are difficult to analyze, store and transfer. Data compression is an attractive and efficient technique to significantly reduce the size of scientific datasets. The overall goal of the breakout session will be to present the technologies developed in ECP for compression of scientific data, to present success stories of lossy compression for ECP applications and discuss open questions and future directions. Specifically, this breakout session will review the state of the art in lossy compression of scientific datasets and discusses in detail the compression technologies developed in ECP: 1) the SZ, ZFP, MGARD compressors and 2) the Libpressio unified compression API, and the Z-checker and Foresight error analysis tools. The Breakout session will also present ECP application use cases in different science domains (e.g., Cosmology, Crystallography, Quantum Chemistry, Molecular Dynamics, and more). A large part of the presentations and discussion will be devoted to lossy compression error quantification, analysis, and understanding. The breakout session will examine how lossy compressors affect scientific data and posthoc analysis. The presentation will be given by compression technology and application experts. The break-out session will encourage discussions and interactions between the speakers and the audience. | Wed, May 4 | Breakout | Recording | 2 | Franck Cappello (ANL), Sheng Di (ANL), Robert Underwood (ANL), Peter Lindstrom (LLNL), Pascal Grosset (LANL), Ben Whitney, Chun Hong Yoon (SLAC), Danny Perez (LANL), Houjun Tang (LBNL), Qian Gong (ORNL), Katrin Heitmann (ANL) |
Advances in Science and Engineering Enabled by ECP Applications IIThree ECP Application Development projects (WDMApp, ExaWind, ExaBiome) temporarily turn their focus away from code development and optimization to highlight recent scientific and engineering accomplishments in their respective domains. The motivating science drivers and the anticipated scientific impact of exascale computing on their application areas will be discussed. | Wed, May 4 | Breakout | 2 | Amitava Bhattacharjee (PNNL), Michael Sprague (NREL), Kathy Yelick (LBNL) | |
Preparing AMReX and Applications for Frontier and Aurora (part 1)The goal of this breakout session is for the community of AMReX users and developers within ECP to discuss new features, plans for future development, and readiness for Frontier and Aurora. The session will start with a brief overview of the AMReX framework with an emphasis on new developments and integrations, followed by updates from several AMReX-supported application development projects within the ECP community. Status on the early access systems for Frontier and Aurora will be emphasized. The session will end with an informal discussion among participants, with the goal of identifying areas for future improvement in the AMReX framework, as well as areas for future collaboration and integration. | Wed, May 4 | Breakout | Recording | 2 | Andrew Myers (LBNL), Erik Palmer (MIT CCSE), Michael Zingale (SBU), Jean Sexton (LBNL), Axel Huebl (LBNL), Roberto Porcu (NETL), Akash Druv, Marc Day (LBNL), Jon Rood (NREL) |
Compression for Scientific Data for ECP applications (part 2)Abstract: Large-scale numerical simulations, observations, and experiments are generating very large datasets that are difficult to analyze, store and transfer. Data compression is an attractive and efficient technique to significantly reduce the size of scientific datasets. The overall goal of the breakout session will be to present the technologies developed in ECP for compression of scientific data, to present success stories of lossy compression for ECP applications and discuss open questions and future directions. Specifically, this breakout session will review the state of the art in lossy compression of scientific datasets and discusses in detail the compression technologies developed in ECP: 1) the SZ, ZFP, MGARD compressors and 2) the Libpressio unified compression API, and the Z-checker and Foresight error analysis tools. The Breakout session will also present ECP application use cases in different science domains (e.g., Cosmology, Crystallography, Quantum Chemistry, Molecular Dynamics, and more). A large part of the presentations and discussion will be devoted to lossy compression error quantification, analysis, and understanding. The breakout session will examine how lossy compressors affect scientific data and posthoc analysis. The presentation will be given by compression technology and application experts. The break-out session will encourage discussions and interactions between the speakers and the audience. | Wed, May 4 | Breakout | Recording | 2 | Franck Cappello (ANL), Sheng Di (ANL), Robert Underwood (ANL), Peter Lindstrom (LLNL), Pascal Grosset (LANL), Ben Whitney, Jullie Bessac (ANL), Katrin Heitmann (ANL), Chun Hong Yoon (SLAC), Danny Perez (LANL), Houjun Tang (LBNL), Qian Gong (ORNL) |
Performance Tools for Emerging Exascale Platforms: Status and ChallengesThe DOE’s emerging GPU-accelerated exascale platforms will have the potential to compute at enormous rates. To realize that potential, application developers will need to identify and ameliorate scaling bottlenecks and inefficiencies in their codes. Development of three performance tools has been funded as part of the Exascale Computing Project: Exa-PAPI, HPCToolkit, and TAU. In this breakout, the project teams will provide an update about the current status of these tools on ECP testbeds and the challenges ahead for these tools to meet the needs of application developers trying to harness the power of exascale platforms. | Thu, May 5 | Breakout | Recording | 2 | John Mellor-Crummey (RU), Xiaozhu Meng (RU), Heike Jagode (UT), Anthony Danalis (UT), Sameer Shende (UO) |
NESAP Success Stories with ECP AppsThe NERSC Exascale Science Applications Program (NESAP) has partnered with 10 ECP applications teams as well additional ECP Software Technology projects to help prepare those teams for deployment on Frontier and Aurora utilizing the NERSC Perlmutter system as a waypoint. In this breakout section, we will hear from the leads of multiple NESAP-ECP collaborations on the outcomes of the readiness activities, science and performance results being achieved on Perlmutter as well as activities the collaboration is pursuing that target Frontier and Aurora directly. | Thu, May 5 | Breakout | Recording | 1 | Ronnie Chatterjee, Muaaz Awan (NERSC), Brandon Cook (LBNL), Rahul Gayatri, Neil Mehta (LBNL), Jack Deslippe (LBNL) |
Preparing AMReX and Applications for Frontier and Aurora (part 2)The goal of this breakout session is for the community of AMReX users and developers within ECP to discuss new features, plans for future development, and readiness for Frontier and Aurora. The session will start with a brief overview of the AMReX framework with an emphasis on new developments and integrations, followed by updates from several AMReX-supported application development projects within the ECP community. Status on the early access systems for Frontier and Aurora will be emphasized. The session will end with an informal discussion among participants, with the goal of identifying areas for future improvement in the AMReX framework, as well as areas for future collaboration and integration. | Wed, May 4 | Breakout | Recording | 2 | Andrew Myers (LBNL), Erik Palmer (MIT CCSE), Michael Zingale (SBU), Jean Sexton (LBNL), Axel Huebl (LBNL), Roberto Porcu (NETL), Akash Druv, Marc Day (LBNL), Jon Rood (NREL) |
Performance of Particle Applications on Early Exascale HardwareThe ECP Co-design Center for Particle Applications (CoPA) addresses the challenges for particle-based applications to run on upcoming exascale computing architectures. Several applications from within CoPA will give updates on their progress and challenges preparing for Frontier, with emphasis on performance results from the OLCF Crusher testbed. | Wed, May 4 | Breakout | Recording | 2 | Stan Moore, Adrian Pope (ANL), Peter McCorquodale (LBNL), Michael Wall (LANL), Samuel Reeve (RONL), Aaron Scheinberg (Jubilee Development) |
On-Ramp to ECP Software and ApplicationsThis session will provide potential users of ECP software and applications an overview of obtaining, building and getting help with ECP packages. ST and AD leadership will describe their philosophy and approach to enabling wide use of ECP software and applications. After this, PIs from ST and AD projects will provide concrete examples of how their projects make it easy for those outside ECP to use their code. The session will end with a Q&A time for the audience to get more details and provide feedback. | Wed, May 4 | Breakout | Recording | 2 | Mike Heroux (SNL), Andrew Siegel (ANL), Pete Bradley (Raytheon Technologies) |
The OpenMP Ecosystem in LLVM - Features, Functionality, and FlagsThe LLVM/OpenMP environment has developed beyond a functional implementation of the OpenMP specification. Among the features introduced as part of ECP we find zero-overhead debugging capabilities (e.g., assertions), interoperability with CUDA code (device + host), link-time-optimizations (LTO), just-in-time (JIT) compilation, offloading to virtual and remote GPUs, and much more. | Wed, May 4 | Breakout | Recording | 2 | Joseph Huber (ORNL), Johannes Doerfert (ANL), Shilei Tian (SBU), Michael Kruse, Giorgis Georgakoudis (LLNL) |
Best Practices #somycodewillseethefuture with BSSw FellowsBetter Scientific Software does not happen without learning and implementing BETTER – Planning, Development, Performance, Reliability, Collaboration, and Skills. But how do we learn to be BETTER? The Better Scientific Software (BSSw) Fellowship program fosters and promotes practices, processes and tools to improve developer productivity and software sustainability of scientific codes. BSSw Fellows are chosen annually to develop a funded activity that promotes better scientific software, such as organizing a workshop, preparing a tutorial, or creating content to engage the scientific software community. Their expertise is diverse and they are often leaders in their own community. Learn from BSSw Fellows and discover how their work will help your scientific software project be BETTER. | Wed, May 4 | Breakout | Recording | 1 | Jeff Carver (UA), Amy Roberts (CU Denver) |
SUNDIALS User ExperiencesThe goal of this breakout session is to help connect SUNDIALS software developers with stakeholders from ECP applications and software technology projects. The session will start with an overview of the SUNDIALS suite emphasizing new features. The session will continue with brief presentations from SUNDIALS application users, including Don Willcox of the AMReX team, Steve DeWitt on the ExaAM team, and Lucas Esclapez on the Pele team. The goals of this session will be: | Wed, May 4 | Breakout | Recording | 2 | Carol S. Woodward (LLNL), David J. Gardner (LLNL), Stephen DeWitt (ORNL), Lucas Esclapez, Don Willcox |
Applications IntegrationIn this breakout, members of the ECP application integration teams will provide an overview and update on applications activities at the facilities along with short talks on engagements with select applications development projects outlining some key technical achievements and lessons learned. This will be followed by a short discussion of applications integration focused topics, in which the participants can share their experiences on topics related to preparing applications for the pre-exascale and exa-scale platforms. | Thu, May 5 | Breakout | Recording | 1 | Scott Parker (ANL), Balint Joo, Deborah Bard (NERSC), Stephen Nichols (ORNL), Christopher J. Knight (ANL), Venkat Vishwanath, Murali Emani, John Gounley (ORNL), Yasaman Ghadar (ANL), Rahul Gayatri, Neil Mehta (LBNL), Abhishek Bagusetty (ANL), Hannah Ross, Matt Norman (ORNL) |
US Federal Agency's Exascale Applications and Software NeedsThis session will feature leaders from US Federal Agencies detailing their exascale needs and discussing ideas about working with ECP to move forward. This will be an opportunity for ECP projects to connect with potential users and understand how ECP software and applications might be used more widely. | Wed, May 4 | Breakout | Recording | 2 | Mike Heroux (SNL), Andrew Siegel (ANL), David Martin (ANL), Suzy Tichenor (ORNL), Jeff Durachta (NOAA/GFDL), Piyush Mehrotra (NASA), Frances Hill (DOD), Emily Greenspan (CBIT) |
Crusher Early User ExperiencesCrusher the Test and Development System has been made available to ECP Teams at the start of 2022 prior to the first Crusher Training Workshop on January 13. Since then we have had 2 Crusher Hackathons with more being planned for the future. This BoF session will provide a way to share some of the earlier Application and ST Project experiences with the wider ECP Community, by showcasing a mixture of ECP Teams chosen to represent a variety of different programming models and languages. We envisage 4 talks, of 10-12 minutes each followed by brief Q&A in the first hour and a brief panel discussion for 30 minutes at the end of the session with the panel comprised of the 4 presenters augmented with facilities and potentially Vendor Center of Excellence Staff. | Wed, May 4 | BoF | Recording | 2 | Balint Joo |
Updates and Roadmap for the SYCL CommunityWith an increased reliance on accelerators such as graphics processing units (GPUs), achieving performance portability across a diverse set of computing platforms is non-trivial. SYCL is an open, single-source, multi-vendor, multi-architecture high-level programming model builds on the underlying abstractions of OpenCL and embraces modern features of C++-17 standards. This BoF provides updates on the salient features of SYCL 2020 specifications, cross platform abstractions using SYCL for GPUs from Nvidia, AMD and Intel. GPU backend specification covering details about mapping of the platform, execution, memory models and interoperability will be discussed. Some potential proposals that can steer adaptation of the upcoming SYCL Next specifications will also be discussed. | Wed, May 4 | BoF | 1 | Thomas Applencourt (ANL), Nevin Liber (ANL), Kevin Harms (ANL), Brandon Cook (LBNL), David E. Bernholdt (ORNL), Abhishek Bagusetty (ANL) | |
Early Experiences Preparing for AuroraAs Argonne National Laboratory prepares for the arrival of Aurora, application and software developers are busy readying their codes to enable new scientific breakthroughs via advances in simulation, data, and learning capabilities. Throughout the development cycle, ALCF and Intel staff have been engaged with developers assisting with a myriad of tasks ranging from porting code to one of several GPU programming models, debugging issues in software and hardware, and helping to understand observed performance on the available Intel GPU testbeds. After a high-level overview of the Aurora system, presenters will discuss a range of topics focused on their experiences, lessons learned, and best-practices for preparing codes for Aurora and plans moving forward. | Wed, May 4 | BoF | Recording | 2 | Christopher J. Knight (ANL), Yasaman Ghadar (ANL), Abhishek Bagusetty (ANL), Ye Luo (ANL), Timothy J. Williams |
Early Experience of Application Developers With OpenMP OffloadingThe next generation of supercomputers will consist of heterogeneous CPU+GPU systems from three different hardware vendors. To target these systems, OpenMP is one possible programming model that can be used to take advantage of the massive parallelism available on GPUs. OpenMP is a portable programming model which will be supported at all three DOE computing facilities, with the ability to run on GPUs from Intel, NVIDIA, and AMD. In this session, developers who are using OpenMP to target GPUs currently in their applications (spanning a variety of domains and including Fortran and C/C++) will give short talks about how they are using OpenMP, any lessons learned or best practices discovered, and any feedback they have for any of the current OpenMP implementations. From these case studies, attendees can learn which features of the OpenMP specification current developers are using, any specific strategies they employed, and best practices developed. We are also looking for feedback on important OpenMP 5.x features applications developers plan to use or are using on current implementations. | Wed, May 4 | BoF | 2 | JaeHyuk Kwack (ANL), Rahulkumar Gayatri (NERSC), Mauro Del Ben (LBNL), Austin Harris (ORNL), Reuben Budiardja (ORNL), Xingqiu Yuan (ANL), Jean-Luc Fattebert (ORNL), Ye Luo (ANL), John R. Tramm (ANL), Vivek Kale (BNL), Yun He (LBNL), Christopher Daley (LBNL), Yasaman Ghadar (ANL), Stephen Nichols (ORNL), Melisa Alkan (ISU), Buu Pham (AL), Dossay Oryspayev (BNL), Swaroop S. Pophale (ORNL), Tosaporn Sattasathuchana (ISU), Peng Xu (AL/ISU), Veronica Melesse Vergara (ORNL), Swen Boehm (ORNL), Nick Hagerty, Philip Thomas, Colleen Bertoni (ANL), Brandon Cook (LBNL), Meifeng Lin (BNL), Timothy J. Williams (ANL) | |
BSSw Fellowship BoFThis Birds-of-a-Feather session provides an opportunity to learn more about how BSSw Fellowship recipients and Honorable Mentions from 2018 to present are impacting better scientific software and engaging with the ECP community. | Wed, May 4 | BoF | Recording | 1 | Elsa Gonsiorowski (LLNL), Hai Ah Nam (LBNL) |
Stochastic Applications in HPC: Challenges and SolutionsStochastic HPC applications face unique algorithmic and performance challenges. Parallel random number generation, frequent branching, non unit stride memory access patterns, load imbalance between threads, and reproducibility are just a few of the issues that stochastic applications can face when running on modern HPC architectures like GPUs. In this cross-cutting session, GPU application teams from a variety of scientific fields will give short presentations detailing the challenges they face stemming from stochastic methods and what strategies they are using to overcome them. While a particular emphasis is placed on GPU performance in many of the talks, some discussions will also cover more general issues faced by stochastic methods in HPC such as stochastic multiphysics convergence as well as pseudorandom number generation algorithms. Applications discussed will include: OpenMC, QMCPACK, HACC, Mercury, Imp, and Shift. | Thu, May 5 | BoF | Recording | 2 | John R. Tramm (ANL), Ye Luo (ANL), Salman Habib (ANL), Michael McKinley (LLNL), Steven Hamilton (ORNL), Nick Frontiere (ANL) |
NNSA/ATDM Application BoF | Thu, May 5 | BoF | 1.5 | Marianne Francois (LANL), Robert Rieben (LLNL), Chris Malone (LANL), Curtis Ober (SNL) | |
Panel on Sustainability of ECP Software and ApplicationsDOE’s expectation is that ECP software and applications will be used long past the end of the ECP project. To ensure this, ECP leadership is putting mechanisms in place to continue support, maintenance and updates for the long term. In this session, ECP leadership will share their philosophy toward long-term sustainability and discuss concrete mechanisms to ensure longevity. In addition, members of the NITRD High End Computing Interagency Working Group will share the group's analysis of sustainability work across a number of federal agencies. The session will be arranged as a panel with short talks by each of the speakers, followed by a moderated discussion with the speakers and audience. | Wed, May 4 | Panel | Recording | 2 | Andrew Siegel (ANL), Mike Heroux (SNL), Hal Finkel (DOE), David Kepczynski (GE) |
Application Experiences with Online Data Analysis and ReductionThis panel will present and discuss practical experiences with online data analysis and reduction methods from the perspectives of a range of ECP applications. Participants will hear about varied application motivations for online data analysis and reduction, the technologies that are being used to couple application components and to perform data reduction, and results achieved in different settings. | Wed, May 4 | Panel | Recording | 1 | CS Chang (PPPL), Ian Foster (UC), Scott Klasky (ORNL), Axel Huebl (LBNL), Robert Jacob (ANL), Arvind Ramanathan (ANL) |
Panel on Revisiting Predictions from the IESP (and other Exascale) WorkshopsDOE is on the verge of deploying its Exascale systems, so what better time then now to revisit our Exascale predictions from a decade ago. In this panel, prominent members from the early Exascale reports will discuss their prediction hits, misses, and omissions. Our panelists will also venture their predictions for 2032. Audience questions are encouraged. | Tue, May 3 | Panel | 1.25 | Jeffrey Vetter (ORNL), Jack Dongarra (UT), Pete Beckman (ANL), Kathy Yelick (LBNL), Bob Lucas (Livermore Software Technology) | |
What Can Be Learned from Applying Team of Teams Principles to the ECP projects PETSc, Trillinos, xSDK, and E4S?The ECP core mission is to develop a capable exascale computing ecosystem that accelerates scientific discovery and supports addressing critical challenges in energy, earth systems, materials, data, and national security. The very nature of this mission has drawn a wide range of talented and successful scientists with diverse backgrounds to work together in new ways toward this goal. In this breakout/panel session, we build on lessons learned after the “Team of Teams” and “Fostering a Culture of Passion and Productivity in ECP Teams” break-out sessions presented at past ECP Annual Meetings as well as the Collegeville 21 whitepapers, “The Community is the Infrastructure” and “Challenges of and Opportunities for a Large Diverse Software Team” to discuss the experiences of the PETSc, Trilinos, xSDK, and E4S communities as framed by the construct, “Team of Teams.” We consider how, why, and when each of these teams may or may not function as Teams of Teams and where the Team of Teams principles might provide a benefit to the projects. We present strategies centered around developing engaged and productive virtual software teams and offer a deeper dive into these communities. We explore how developing a capable exascale ecosystem depends on meeting technical, social, and cultural challenges. | Thu, May 5 | Panel | Recording | 2 | Reed Milewicz (SNL), David Moulton (LANL), Miranda Mundt (SNL), Todd Munson, Elaine Raybourn (SNL), Benjamin Sims (LANL), Jim Willenbring (SNL), Greg Watson (ORNL), Ulrike Yang (LLNL) |