Loading…
May 5-8, 2025
Chicago, IL
View More Details & Registration

The Sched app allows you to build your schedule but is not a substitute for your event registration. You must be registered for the event to participate in the sessions. If you have not registered but would like to join us, please go to the event registration page to find out more information.

This schedule is automatically displayed in Central Time (UTC/GMT -6 hours). To see the schedule in your preferred timezone, please select from the drop-down menu to the right, above "Filter by Date."

IMPORTANT NOTE: Timing of sessions and room locations are subject to change.

Monday, May 5
 

8:00am CDT

Registration & Badge Pick-Up
Monday May 5, 2025 8:00am - 6:00pm CDT
Monday May 5, 2025 8:00am - 6:00pm CDT
Ballroom Meeting Foyer

9:00am CDT

Welcome Remarks - Christian Trott, Sandia National Laboratories
Monday May 5, 2025 9:00am - 9:10am CDT
Welcome to the inaugural HPSF Conference! These remarks will provide an overview of the conference and what to expect. Join us for the opening of this new chapter of HPSF.
Speakers
CT

Christian Trott

Distinguished Member of Technical Staff, Sandia National Laboratories
Christian Trott is a High Performance Computing expert at Sandia National Laboratories, where he co-leads the Kokkos core team, developing performance portability solutions for engineering and science applications. He heads Sandia's delegation to the ISO C++ committee and is a principal... Read More →
Monday May 5, 2025 9:00am - 9:10am CDT
Chicago River Ballroom

9:10am CDT

HPSF State of the Project Overview - Todd Gamblin, Lawrence Livermore National Laboratory
Monday May 5, 2025 9:10am - 9:20am CDT
Join us for an engaging session that highlights the remarkable progress of the High Performance Software Framework (HPSF) in its inaugural year. We will explore the expansion of its vibrant community, the increasing number of members, and the diverse projects that have emerged. Attendees will gain insights into HPSF's strategic goals and future initiatives aimed at advancing the landscape of high-performance software. Discover how HPSF is poised to shape the future of software development in high-performance computing and learn how you can contribute to this exciting journey.
Speakers
TG

Todd Gamblin

Distinguished Member of Technical Staff, Lawrence Livermore National Laboratory
Monday May 5, 2025 9:10am - 9:20am CDT
Chicago River Ballroom

9:20am CDT

Governing Board Introduction - Heidi Poxon, AWS
Monday May 5, 2025 9:20am - 9:35am CDT
Learn about the work of the HPSF governing board (GB). The GB is responsible for the overall governance of HPSF including member admission, budget decisions and general policies. This session will provide attendees with an overview of the GBs work, how it makes decisions and what its priorities are for steering the foundation. 
Speakers
HP

Heidi Poxon

Principal Member of Technical Staff, AWS
Monday May 5, 2025 9:20am - 9:35am CDT
Chicago River Ballroom

9:35am CDT

TAC Introduction - Bill Hoffman, Kitware
Monday May 5, 2025 9:35am - 9:50am CDT
Join us for an informative session on the role of the Technical Advisory Committee (TAC) in shaping our project's landscape. The TAC oversees project admissions, develops a comprehensive project life-cycle approach, and facilitates coordination among member projects. This presentation will provide an overview of the TAC's responsibilities and detail the onboarding process for new projects, offering valuable insights into how we ensure successful integration and collaboration within our community. Don't miss this opportunity to understand the vital functions of the TAC and how they contribute to our collective success.
Speakers
avatar for Bill Hoffman

Bill Hoffman

CTO, Kitware
Mr. Hoffman is a founder of Kitware and currently serves as Chairman of the Board, Vice President, and Chief Technical Officer (CTO). He is the original author and lead architect of CMake, an open source, cross-platform build and configuration tool that is used by hundreds of projects... Read More →
Monday May 5, 2025 9:35am - 9:50am CDT
Chicago River Ballroom

9:50am CDT

Working Groups Introduction - Damien Lebrun-Grandie, Oak Ridge National Laboratory
Monday May 5, 2025 9:50am - 10:00am CDT
HPSF conducts project cross-cutting activities in working groups. This presentation will give a short overview of the existing working groups. Attendees will get the information to decide which breakout session to attend on day two of the HPSF conference, to get involved in the important work of the foundation.
Speakers
DL

Damien Lebrun-Grandie

Senior Computational Scientist, Oak Ridge National Laboratory
Monday May 5, 2025 9:50am - 10:00am CDT
Chicago River Ballroom

10:00am CDT

Coffee Break
Monday May 5, 2025 10:00am - 10:30am CDT
Monday May 5, 2025 10:00am - 10:30am CDT

10:30am CDT

Project Updates
Monday May 5, 2025 10:30am - 12:30pm CDT
This session offers a fast-paced overview of HPSF's cutting-edge software projects, showcasing the innovative capabilities and the vibrant community driving these initiatives. Attendees will gain insights into the diverse range of projects and their potential applications. Join us to explore new avenues for collaboration and discover how you can leverage HPSF software projects to enhance your own work and contribute to the community's growth.

Project Updates: Systemtools
10:25 - 10:40 AM: HPCToolkit - Jonathon Anderson
10:40 - 10:55 AM: Apptainer - Dave Dykstra, Fermi National Accelerator Laboratory
10:55 - 11:10 AM: Charliecloud - Reid Priedhorsky, Los Alamos National Laboratory
11:10 - 11:25 AM: Spack - Greg Becker, Lawrence Livermore National Laboratory
11:25 - 11:40 AM: E4S - Sameer Shende, Paratools

Project Updates: Applications
11:40 - 11:55 AM: WarpX - Edoardo Zoni, Lawrence Berkeley National Laboratory

Project Updates: Scientific Libraries  Viskores (12:00 - 12:15 PM) Trilinos (12:15 - 12:30 PM)
Speakers
GB

Greg Becker

Software Developer, Lawrence Livermore National Laboratory
DD

Dave Dykstra

Fermi National Accelerator Laboratory
avatar for Ken Moreland

Ken Moreland

Oak Ridge National Laboratory
avatar for Curtis Ober

Curtis Ober

Distinguished Member of Technical Staff, Sandia National Laboratories
Curt has been with Sandia for 30 years and is currently a Distinguished Member of the Technical Staff.  His career has spanned many projects and missions at the laboratory, including computational fluid dynamics (CFD) for re-entry vehicles, shock hydrodynamics, time integration (Trilinos... Read More →
RP

Reid Priedhorsky

Scientist, Los Alamos National Laboratory
I am a staff scientist at Los Alamos National Laboratory. Prior to Los Alamos, I was a research staff member at IBM Research. I hold a Ph.D. in computer science from the University of Minnesota and a B.A., also in computer science, from Macalester College.My work focuses on large-scale... Read More →
avatar for Sameer Shende

Sameer Shende

Research Professor and Director, Performance Research Lab, U. Oregon, University of Oregon
Sameer Shende serves as a Research Professor and the Director of the Performance Research Lab at the University of Oregon and the President and Director of ParaTools, Inc. (USA) and ParaTools, SAS (France). He serves as the Technical Lead of the Extreme-scale Scientific Software Stack... Read More →
EZ

Edoardo Zoni

Research Software Engineer, Lawrence Berkeley National Laboratory
Monday May 5, 2025 10:30am - 12:30pm CDT
Chicago River Ballroom

12:30pm CDT

Lunch (Provided for Attendees)
Monday May 5, 2025 12:30pm - 2:00pm CDT
Monday May 5, 2025 12:30pm - 2:00pm CDT
Atrium

2:00pm CDT

Project Updates
Monday May 5, 2025 2:00pm - 3:00pm CDT
This session offers a fast-paced overview of HPSF's cutting-edge software projects, showcasing the innovative capabilities and the vibrant community driving these initiatives. Attendees will gain insights into the diverse range of projects and their potential applications. Join us to explore new avenues for collaboration and discover how you can leverage HPSF software projects to enhance your own work and contribute to the community's growth.

Project Updates: Libraries / Programming Systems
1:55 - 2:10 PM: AMReX - Andrew Myers, Lawrence Berkeley National Laboratory
2:10 - 2:25 PM: Kokkos - Christian Trott, Sandia National Laboratories


Project Updates: New and Prospective Projects
2:25 - 2:35 PM: Chapel - Brad Chamberlain, HPE
2:35 - 2:45 PM: HPX - Hartmut Kaiser, LSU
2:45 - 2:55 PM: OpenHPC - Adrian Reber, Red Hat
Speakers
HK

Hartmut Kaiser

STE||AR Group, LSU
Hartmut is a member of the faculty at the CS department at Louisiana State University (LSU) and a senior research scientist at LSU's Center for Computation and Technology (CCT). He received his doctorate from the Technical University of Chemnitz (Germany) in 1988. He is probably best... Read More →
avatar for Adrian Reber

Adrian Reber

Senior Principal Software Engineer, Red Hat
Adrian is a Senior Principal Software Engineer at Red Hat and is migrating processes at least since 2010. He started to migrate processes in a high performance computing environment and at some point he migrated so many processes that he got a PhD for that. Most of the time he is... Read More →
avatar for Brad Chamberlain

Brad Chamberlain

Distinguished Technologist, HPE
Brad Chamberlain is a Distinguished Technologist at Hewlett Packard Enterprise (formerly Cray Inc.) who has spent his career focused on user productivity for high-performance computing (HPC) systems, particularly through the design and development of the Chapel parallel programming... Read More →
AM

Andrew Myers

Computer Systems Engineer, Lawrence Berkeley National Laboratory
CT

Christian Trott

Distinguished Member of Technical Staff, Sandia National Laboratories
Christian Trott is a High Performance Computing expert at Sandia National Laboratories, where he co-leads the Kokkos core team, developing performance portability solutions for engineering and science applications. He heads Sandia's delegation to the ISO C++ committee and is a principal... Read More →
Monday May 5, 2025 2:00pm - 3:00pm CDT
Chicago River Ballroom

3:00pm CDT

Coffee Break
Monday May 5, 2025 3:00pm - 3:30pm CDT
Monday May 5, 2025 3:00pm - 3:30pm CDT

3:30pm CDT

Panel Discussion: Status and Trends in the HPC Landscape - Moderated by Todd Gamblin, Lawrence Livermore National Laboratory
Monday May 5, 2025 3:30pm - 5:00pm CDT
High-performance computing (HPC) systems have undergone significant evolution over the years. Today, HPC professionals utilize not only traditional on-premise large clusters but also cloud-based solutions and appliance-like platforms that bridge the gap between workstations and larger clusters. Join this session to hear from experts who are at the forefront of deploying the largest exascale systems, along with representatives from leading cloud vendors. Gain insights into the current landscape of HPC platforms and learn about the systems you should prepare for in the near future to stay ahead in this rapidly changing field.
Moderators
TG

Todd Gamblin

Distinguished Member of Technical Staff, Lawrence Livermore National Laboratory
Speakers
avatar for Dan Stanzione

Dan Stanzione

Director, Texas Advanced Computing Center
Dr. Dan Stanzione, Associate Vice President for Research at The University of Texas at Austin since 2018 and Executive Director of the Texas Advanced Computing Center (TACC) since 2014, is a nationally recognized leader in high performance computing. He serves on the National Artificial... Read More →
HP

Heidi Poxon

Principal Member of Technical Staff, AWS
SC

Sara Campbell

Program Manager, DOE/NNSA
avatar for Jayesh Badwaik

Jayesh Badwaik

HPC Software Engineer, Jülich Supercomputing Centre
NF

Nur Fadel

Head of Scientific Computing Unit, CSCS
AJ

Andrew Jones

Engineering Leader, Future AI & HPC Capabilities, Microsoft
DJ

Doug Jacobsen

HPC Software Engineer, Google Cloud
Monday May 5, 2025 3:30pm - 5:00pm CDT
Chicago River Ballroom

5:00pm CDT

Attendee Reception
Monday May 5, 2025 5:00pm - 7:00pm CDT
Monday May 5, 2025 5:00pm - 7:00pm CDT
Ballroom Meeting Foyer

5:00pm CDT

Poster Sessions
Monday May 5, 2025 5:00pm - 7:00pm CDT
Featured Posters:

  • Porting Legacy Codes to Kokkos - Trévis Morvany, CEA
  • Designing a Usable Architecture for Geometric Particle-In-Cell Methods with AMReX - Emil Poulsen & Nils Schild, Max Planck Institute for Plasma Physics
  • Kubernetes Turbo Charging Productivity in Era of Accelerated Computing - Tobi Knaup, Nutanix
  • The Evolution of Virtualization: Adding some Xen to HPC - Cody Zuschlag, Vates
  • High-performance Phase-field Solver Based on AMReX Software Framework - Akash Shinde & Nasir Attar, Centre for Development of Advanced Computing 
  • HPSF Project Resource: Center for Open-Source Research Software Advancement (CORSA) - Daniel S. Katz, University of Illinois Urbana-Champaign & Elaine M. Raybourn, Sandia National Laboratories
  • PESO: Partnering for Scientific Software Ecosystem Stewardship Opportunities - James Willenbring, Sandia National Laboratories
  • Introduction to Charliecloud and its Weirdness - Reid Priedhorsky, LANL / Charliecloud, Angela Loshak, Los Alamos National Laboratory & Megan Phinney, Los Alamos National Laboratory
  • Boosting ANUGA Performance by GPU Porting - Samir Shaikh & Harsha Ugave, Centre for Development of Advanced Computing (C-DAC), India
  • AI Governance in High-Performance Computing: Ensuring Compliance, Efficiency, and Security - Rohith Vangalla, Optum Services
  • DaggerMPI:Seamless MPI operations and Scheduling - Felipe De Alcantara Tome, MIT
  • Compiler Dependencies in Spack v1.0 - Gregory Becker, LLNL 
  • The Kokkos Performance Portability EcoSystem - Christian Trott, Sandia National Laboratories; Damien Lebrun-Grandie, Lawrence Livermore National Laboratory; Luc Berger-Vergiat & Siva Rajamanickam, Sandia National Laboratories


Speakers
avatar for Daniel S. Katz

Daniel S. Katz

Chief Scientist, NCSA, University of Illinois Urbana-Champaign
Dan's interest is in the development and use of advanced cyberinfrastructure to solve challenging problems at multiple scales. This includes applications, algorithms, fault tolerance, and programming in parallel and distributed computing, including HPC, Grid, Cloud, etc., as well... Read More →
RP

Reid Priedhorsky

Scientist, Los Alamos National Laboratory
I am a staff scientist at Los Alamos National Laboratory. Prior to Los Alamos, I was a research staff member at IBM Research. I hold a Ph.D. in computer science from the University of Minnesota and a B.A., also in computer science, from Macalester College.My work focuses on large-scale... Read More →
avatar for Elaine M. Raybourn

Elaine M. Raybourn

Principal Member of the Technical Staff, Sandia National Laboratories
Elaine M. Raybourn is a social scientist at Sandia National Laboratories. She has worked in the UK (British Telecom), Germany (Fraunhofer FIT), and France (INRIA) as a Fellow of the European Research Consortium in Informatics and Mathematics (ERCIM). She supports the DOE Office of... Read More →
avatar for Cody Zuschlag

Cody Zuschlag

Developer Relations Evangelist, Vates
With a clear focus on open-source solutions, Cody is deeply committed to shaping technology for the greater good. Cody has championed the benefits of full-stack and decentralized applications, underscoring the significance of open-source technologies. Through his work, presentations... Read More →
avatar for Megan Phinney

Megan Phinney

Scientist, Los Alamos National Laboratory
AL

Angelica Loshak

Student, Los Alamos National Laboratory
avatar for Samir Shaikh

Samir Shaikh

Scientist, Centre for Developement of Advanced Computing (C-DAC)
Samir Shaikh is an HPC specialist at C-DAC, Pune, optimizing large-scale workloads, parallel computing, and system architecture. As a Scientist C, he enhances HPC performance for AI/ML, scientific computing, and NSM supercomputers. An IIT Guwahati M.Tech graduate, he has contributed... Read More →
avatar for Akash Shinde

Akash Shinde

Project Engineer, Center for Development of Advanced Computing
Akash Shinde, Project Engineer, C-DAC Basically I working at C-DAC as Scientific Software Devloper.
avatar for Felipe De Alcantara Tome

Felipe De Alcantara Tome

Research Software Engineer, MIT
Felipe Tomé is a Brazilian Research Software Engineer passionate about high-performance computing (HPC), parallel computing, and scalable algorithms. At MIT, he contributed to Dagger.jl and DLA.jl, optimizing MPI, GPU acceleration, and numerical linear algebra. His research, published... Read More →
avatar for Harsha Ugave

Harsha Ugave

HPC Engineer, Centre for Development of Advanced Computing (C-DAC), India
Harsha Ugave is an HPC Engineer at C-DAC Pune, specializing in performance portability, parallel computing, and system optimization. She plays a key role in deploying and tuning HPC applications under the National Supercomputing Mission (NSM). Her work ensures efficient execution... Read More →
avatar for James Willenbring

James Willenbring

Senior Member of Technical Staff, Sandia National Laboratories
James M. Willenbring is a senior member of R&D Technical Staff in the Software Engineering and Research department at Sandia National Laboratories. His research interests include software sustainability and the application of software engineering methodologies for high-performance... Read More →
avatar for Nasir Attar

Nasir Attar

Project Engineer, Centre for Development of Advanced Computing
I’m Nasir Attar, and for the past three years, I’ve been with the Centre for Development of Advanced Computing (C-DAC), Pune, India. I lead the development of a high-performance phase-field solver using the AMReX software framework as part of a collaboration with leading Indian... Read More →
avatar for Nils Schild

Nils Schild

PhD student, Max Planck Institute for Plasma Physics
After studying physics and working on solvers for sparse eigenvalue problems in quantum mechanics at the University of Bayreuth, he moved to the Max Planck Institute for Plasma Physics in Garching (Germany). During his Ph.D., he started implementing the software BSL6D, a solver for... Read More →
avatar for Rohith Vangalla

Rohith Vangalla

Lead Software Engineer, Optum Technologies (UnitedHealth Group), Optum services
I am Dr. Rohith Vangalla, a Lead Software Engineer at Optum Technologies, a subsidiary of UnitedHealth Group. I specialize in high-performance computing (HPC) and AI-driven healthcare solutions, focusing on scalable, cloud-native architectures and regulatory compliance. With a PhD... Read More →
avatar for Tobi Knaup

Tobi Knaup

VP & GM Cloud Native, Nutanix
As GM Cloud Native at Nutanix, Tobi leads development of the company’s Kubernetes and Cloud Native portfolio, used by the likes of AWS and Microsoft Azure. A cloud-native pioneer, Tobi co-founded D2iQ – the leading independent Kubernetes company relied on by 30% of the Fortune... Read More →
GB

Greg Becker

Software Developer, Lawrence Livermore National Laboratory
DL

Damien Lebrun-Grandie

Senior Computational Scientist, Oak Ridge National Laboratory
CT

Christian Trott

Distinguished Member of Technical Staff, Sandia National Laboratories
Christian Trott is a High Performance Computing expert at Sandia National Laboratories, where he co-leads the Kokkos core team, developing performance portability solutions for engineering and science applications. He heads Sandia's delegation to the ISO C++ committee and is a principal... Read More →
avatar for Trévis Morvany

Trévis Morvany

Research engineer, CEA
Holder of a Master’s Degree in High Performance Computing and Simulation from Paris-Saclay University, Trévis Morvany joined the CExA team as a developer in January 2025.
avatar for Emil Poulsen

Emil Poulsen

Post Doc, Max Planck Institute for Plasma Physics
Dr. Poulsen has more than six years of experience with scientific high performance computing in topics as diverse as quantum many-body physics, micromagnetism and plasma physics using mainly C++ and Fortran in combination with CUDA, MPI and OpenMP.
LB

Luc Berger-Vergiat

Sandia National Laboratories
SR

Siva Rajamanickam

Sandia National Laboratories
Monday May 5, 2025 5:00pm - 7:00pm CDT
Chicago River Ballroom
 
Tuesday, May 6
 

8:30am CDT

Registration & Badge Pick-Up
Tuesday May 6, 2025 8:30am - 5:00pm CDT
Tuesday May 6, 2025 8:30am - 5:00pm CDT
Ballroom Meeting Foyer

9:00am CDT

Welcome & Overview of Day
Tuesday May 6, 2025 9:00am - 9:05am CDT
Tuesday May 6, 2025 9:00am - 9:05am CDT
Chicago River Ballroom

9:05am CDT

Performance, Usability and Issues on Current Systems - Kevin Huck, University of Oregon & Chris Siefert, Sandia National Laboratories
Tuesday May 6, 2025 9:05am - 10:30am CDT
This session offers a comprehensive overview of the current landscape of exascale supercomputing technologies from the user perspective. Attendees will gain valuable insights into the nuances of transitioning between various system architectures and software stacks, focusing on usability, performance, and the common challenges faced by users. By exploring real-world experiences and best practices, participants will be better equipped to navigate the complexities of exascale computing environments, ultimately enhancing their ability to leverage these advanced technologies for their research and applications.
Speakers
CS

Chris Siefert

R&D Staff, Sandia National Laboratories
avatar for Kevin Huck

Kevin Huck

Senior Research Associate, University of Oregon
Kevin Huck is a Senior Research Associate in the Oregon Advanced Computing Institute for Science and Society (OACISS) at the University of Oregon. He is interested in the unique problems of performance analysis of large HPC applications as well as automated methods for diagnosing... Read More →
Tuesday May 6, 2025 9:05am - 10:30am CDT
Chicago River Ballroom

10:30am CDT

Coffee Break
Tuesday May 6, 2025 10:30am - 11:00am CDT
Tuesday May 6, 2025 10:30am - 11:00am CDT

11:00am CDT

Panel Discussion: Processor Trends and What They Mean for Software - Speakers to be Announced
Tuesday May 6, 2025 11:00am - 12:30pm CDT
The rapid acceleration of processor innovation in recent years has introduced both opportunities and challenges for developers of high-performance software. This expert panel will delve into emerging hardware trends that are poised to shape the future of software development. Join us for an insightful discussion as we explore the implications of these advancements, the new challenges that lie ahead, and how our community can collaboratively address them. Gain valuable perspectives on what to expect in the evolving landscape of high-performance computing and how to adapt our strategies for success.
Moderators
CT

Christian Trott

Distinguished Member of Technical Staff, Sandia National Laboratories
Christian Trott is a High Performance Computing expert at Sandia National Laboratories, where he co-leads the Kokkos core team, developing performance portability solutions for engineering and science applications. He heads Sandia's delegation to the ISO C++ committee and is a principal... Read More →
Tuesday May 6, 2025 11:00am - 12:30pm CDT
Chicago River Ballroom

12:30pm CDT

Lunch (Provided for Attendees)
Tuesday May 6, 2025 12:30pm - 2:00pm CDT
Tuesday May 6, 2025 12:30pm - 2:00pm CDT
Atrium

2:00pm CDT

Working Group Breakouts
Tuesday May 6, 2025 2:00pm - 3:00pm CDT
Join our working group breakout sessions for an interactive opportunity to engage in face-to-face discussions with fellow community members. Discover the ongoing activities of each working group and contribute to shaping future initiatives. Topics include continuous integration, benchmarking, community outreach, and efforts to enhance software project interoperability. Brief reports from each breakout will summarize key insights and activities, ensuring that the entire community stays informed and connected. Your participation is vital in driving collaboration and innovation within our community!
Tuesday May 6, 2025 2:00pm - 3:00pm CDT
Chicago River Ballroom

3:00pm CDT

Working Group Breakout Reports
Tuesday May 6, 2025 3:00pm - 3:25pm CDT
Tuesday May 6, 2025 3:00pm - 3:25pm CDT
Chicago River Ballroom

3:25pm CDT

Coffee Break
Tuesday May 6, 2025 3:25pm - 3:50pm CDT
Tuesday May 6, 2025 3:25pm - 3:50pm CDT

3:50pm CDT

HPSF Community BOF
Tuesday May 6, 2025 3:50pm - 5:20pm CDT
Join us for an engaging Birds of a Feather (BoF) session focused on the current state and future direction of the HPSF community. Key community leaders will discuss successes and identify areas for improvement, fostering an open dialogue about our collective vision. This interactive session will provide ample opportunities for audience questions and feedback, allowing you to contribute your insights. Together, let’s shape the HPSF community into a valuable resource that meets the needs of all its members!
Tuesday May 6, 2025 3:50pm - 5:20pm CDT
Chicago River Ballroom

5:20pm CDT

Closing Remarks
Tuesday May 6, 2025 5:20pm - 5:35pm CDT
Tuesday May 6, 2025 5:20pm - 5:35pm CDT
Chicago River Ballroom
 
Wednesday, May 7
 

8:30am CDT

Registration & Badge Pick-Up
Wednesday May 7, 2025 8:30am - 4:30pm CDT
Wednesday May 7, 2025 8:30am - 4:30pm CDT
Ballroom Meeting Foyer

9:00am CDT

Welcome and Overview - Todd Gamblin, Lawrence Livermore National Laboratory
Wednesday May 7, 2025 9:00am - 9:10am CDT
Speakers
TG

Todd Gamblin

Distinguished Member of Technical Staff, Lawrence Livermore National Laboratory
Wednesday May 7, 2025 9:00am - 9:10am CDT
Salon E-G

9:00am CDT

Numerical Relativity with AMReX - Miren Radia, University of Cambridge, Lawrence Berkeley National Laboratory
Wednesday May 7, 2025 9:00am - 9:20am CDT
Einstein’s theory of General Relativity revolutionised Physics over a century ago. Despite this, the number of known analytical solutions to the equations, particularly in the dynamical strong-field case is very small. Numerical relativity is often the only tool that can be used to investigate this regime.
Speakers
MR

Miren Radia

Research Software Engineer, University of Cambridge
Wednesday May 7, 2025 9:00am - 9:20am CDT
Rock River 1 & 2

9:00am CDT

The History of Apptainer - One Dave's Perspective, Dave Godlove, CIQ
Wednesday May 7, 2025 9:00am - 9:20am CDT
Speakers
Wednesday May 7, 2025 9:00am - 9:20am CDT
Mississippi River 1 & 2

9:00am CDT

Introduction to Containers and Charliecloud - Reid Priedhorsky, Los Alamos National Laboratory
Wednesday May 7, 2025 9:00am - 9:40am CDT
Charliecloud is a lightweight, fully unprivileged, open source container implementation for HPC applications. It can handle the entire container workflow, including building images, pushing/pulling to registries, and dealing with accelerators like GPUs. We take a unique approach to containerization while remaining compatible with the broader container ecosystem where it matters.

Notable distinctions include: fully unprivileged end-to-end workflow, small code base with minimal dependencies, delegation of all security boundaries, layer-free build cache, zero-consistency root emulation based on seccomp, and a spirit of openness and public service.

This talk will motivate containers for HPC applications, then explain Charliecloud’s design philosophy and how we address these needs. We’ll discuss Charliecloud’s “weirdness” when compared to other implementations (i.e., its design philosophy and key distinctions), provide a brief overview of the code’s structure, and summarize recent and upcoming changes of interest.

LA-UR-25-22140
Speakers
RP

Reid Priedhorsky

Scientist, Los Alamos National Laboratory
I am a staff scientist at Los Alamos National Laboratory. Prior to Los Alamos, I was a research staff member at IBM Research. I hold a Ph.D. in computer science from the University of Minnesota and a B.A., also in computer science, from Macalester College.My work focuses on large-scale... Read More →
Wednesday May 7, 2025 9:00am - 9:40am CDT
Illinois River

9:00am CDT

Updates from the Kokkos Team
Wednesday May 7, 2025 9:00am - 10:20am CDT
1. Welcome and overview, Damien Lebrun-Grandie, Oak Ridge National Laboratory (10 minutes)

2. Update on the Ecosystem and Community, Christian Trott, Sandia National Laboratories (20 minutes)

3. Kokkos Core update, Damien Lebrun-Grandie, Oak Ridge National Laboratory (30 minutes)

4. Kokkos-Kernels update, Luc Berger-Vergiat, Sandia National Laboratories (20 minutes)
Speakers
LB

Luc Berger-Vergiat

Sandia National Laboratories
DL

Damien Lebrun-Grandie

Senior Computational Scientist, Oak Ridge National Laboratory
CT

Christian Trott

Distinguished Member of Technical Staff, Sandia National Laboratories
Christian Trott is a High Performance Computing expert at Sandia National Laboratories, where he co-leads the Kokkos core team, developing performance portability solutions for engineering and science applications. He heads Sandia's delegation to the ISO C++ committee and is a principal... Read More →
Wednesday May 7, 2025 9:00am - 10:20am CDT
Salon A-C

9:10am CDT

State of the Spack Community, Todd Gamblin, Lawrence Livermore National Laboratory
Wednesday May 7, 2025 9:10am - 9:45am CDT
Speakers
TG

Todd Gamblin

Distinguished Member of Technical Staff, Lawrence Livermore National Laboratory
Wednesday May 7, 2025 9:10am - 9:45am CDT
Salon E-G

9:20am CDT

Cosmological Discoveries Enabled by the Nyx Code Zarija Lukic - Lawrence Berkeley National Laboratory
Wednesday May 7, 2025 9:20am - 9:40am CDT
Speakers
ZL

Zarija Lukic

Staff Scientist and Group Lead, Lawrence Berkeley National Laboratory
Wednesday May 7, 2025 9:20am - 9:40am CDT
Rock River 1 & 2

9:20am CDT

Apptainer in Lima - Anders Björklund, Cloud Native Computing Foundation
Wednesday May 7, 2025 9:20am - 9:40am CDT
Speakers
AB

Anders Björklund

Cloud Native Computing Foundation
Wednesday May 7, 2025 9:20am - 9:40am CDT
Mississippi River 1 & 2

9:40am CDT

Quokka: A New Multiphysics Code for Star Formation and Astrophysics - Ben Wibking, Michigan State University
Wednesday May 7, 2025 9:40am - 10:00am CDT
I will describe Quokka, a new AMR code for astrophysics, aimed at problems in star formation and galaxy formation. Quokka is a Newtonian radiation hydrodynamics code with support for particles, chemistry, self-gravity, and (soon) magnetic fields based on a method-of-lines formulation. We use AMReX's unique capabilities to enable many of these physics features, from AMReX's OpenBC Poisson solver for self-gravity, to AMReX-Astro Microphysics for the cell-by-cell chemistry network integration with symbolic code generation. I will talk about our current code development focused on particles and a constrained-transport implementation of magnetohydrodynamics, and briefly mention applications in simulating galactic winds and the formation of the first stars.
Speakers
BW

Ben Wibking

Research Associate, Michigan State University
Wednesday May 7, 2025 9:40am - 10:00am CDT
Rock River 1 & 2

9:40am CDT

Using Apptainer in a Pilot-based Distributed Workload - Marco Mambelli, Fermi National Accelerator Laboratory
Wednesday May 7, 2025 9:40am - 10:00am CDT
GlideinWMS is a pilot and pressure-based workload manager for distributed scientific computing. Many experiments like CMS and Fermilab’s Neutrino experiments use it to provision elastic clusters for their analysis and simulations, split in close to a million concurrent jobs. The majority of the user jobs require containers and the pilots use Apptainer to setup the desired platform. For the pilots, that run as regular batch jobs, Apptainer is safer, lighter and easier to use than other containerization solutions. Many images used by the pilots are expanded SIF images distributed via the CernVMFS: this combination is very efficient.
Speakers
MM

Marco Mambelli

Fermi National Accelerator Laboratory
Wednesday May 7, 2025 9:40am - 10:00am CDT
Mississippi River 1 & 2

9:40am CDT

Charliecloud Workshop: Key Workflow Operation - Pull - Megan Phinney, Los Alamos National Laboratory
Wednesday May 7, 2025 9:40am - 10:00am CDT
This workshop will provide participants with background and hands-on experience to use basic Charliecloud containers for HPC applications. Participants will build toy containers and a real HPC application, and then run them in parallel on a supercomputer. This will be a highly interactive workshop with lots of Q&A.

This section will walk through the key workflow operation pull.

LA-UR-25-22140
Speakers
avatar for Megan Phinney

Megan Phinney

Scientist, Los Alamos National Laboratory
Wednesday May 7, 2025 9:40am - 10:00am CDT
Illinois River

9:45am CDT

Spack v1.0 - Greg Becker, Lawrence Livermore National Laboratory
Wednesday May 7, 2025 9:45am - 10:20am CDT
Speakers
GB

Greg Becker

Software Developer, Lawrence Livermore National Laboratory
Wednesday May 7, 2025 9:45am - 10:20am CDT
Salon E-G

10:00am CDT

Leveraging SUNDIALS to Accelerate AMReX Simulations - Andy Nonaka, Lawrence Berkeley National Laboratory
Wednesday May 7, 2025 10:00am - 10:20am CDT
In this talk I discuss the role of the SUNDIALS (SUite of Nonlinear and DIfferential/ALgebraic equation Solvers) package in three different AMReX-based applications. In this context, SUNDIALS provides support for the overall temporal integration scheme of the model equations. We are particularly interested in the multirate infinitesimal (MRI) schemes, where different physical processes can be advanced using different time steps. Our initial application is in the context of combustion, where we show that we are able to achieve increased accuracy over our traditional spectral deferred corrections approach with greater computational efficiency. Our second application is in micromagnetic memory and storage devices where we demonstrate efficiency by leveraging the non-stiff nature of the computationally-expensive demagnetization processes. Our final application is low-power ferroelectric transistors where we also effectively leverage the non-stiff nature of the computationally-expensive Poisson process. In each case we show significant computational savings over our baseline codes.
Speakers
AN

Andy Nonaka

Staff Scientist and CCSE Group Lead, Lawrence Berkeley National Laboratory
Wednesday May 7, 2025 10:00am - 10:20am CDT
Rock River 1 & 2

10:00am CDT

Creating Apptainer Workflows with Docker-Compose-like Utilities - Brandon Biggs, Idaho National Laboratory
Wednesday May 7, 2025 10:00am - 10:20am CDT
Speakers
BB

Brandon Biggs

Idaho National Laboratory
Wednesday May 7, 2025 10:00am - 10:20am CDT
Mississippi River 1 & 2

10:00am CDT

Charliecloud Workshop: Containers Are Not Special Alpine via Tarball - Megan Phinney, Los Alamos National Laboratory
Wednesday May 7, 2025 10:00am - 10:20am CDT
This workshop will provide participants with background and hands-on experience to use basic Charliecloud containers for HPC applications. Participants will build toy containers and a real HPC application, and then run them in parallel on a supercomputer. This will be a highly interactive workshop with lots of Q&A.

This section will explain that containers are not special by running an Alpine container via tarball.

LA-UR-25-22140
Speakers
avatar for Megan Phinney

Megan Phinney

Scientist, Los Alamos National Laboratory
Wednesday May 7, 2025 10:00am - 10:20am CDT
Illinois River

10:20am CDT

Coffee Break
Wednesday May 7, 2025 10:20am - 10:45am CDT
Wednesday May 7, 2025 10:20am - 10:45am CDT

10:45am CDT

ExaEpi: Agent-Based Modeling for Epidemiology using AMReX - Andrew Myers, Lawrence Berkeley National Laboratory
Wednesday May 7, 2025 10:45am - 11:05am CDT
The adaptive mesh and particle capabilities of AMReX have been used to implement a wide range of numerical methods, targeting combustion, plasma physics, earth systems modeling, cosmology, and more. This talk will show how they can also be used to implement a quite different type of algorithm: an agent-based model (ABM) for the spread of respiratory diseases called ExaEpi. ABMs are valuable because they provide a fundamental and natural description of the system and are able to capture emergent phenomena. However, their use in forecasting and control is limited by the difficulty in calibrating and quantifying the uncertainty associated with a large number of parameters. By leveraging AMReX, ExaEpi can help address these limitations by enabling many large ensembles to run quickly on exascale compute facilities. 
Speakers
AM

Andrew Myers

Computer Systems Engineer, Lawrence Berkeley National Laboratory
Wednesday May 7, 2025 10:45am - 11:05am CDT
Rock River 1 & 2

10:45am CDT

Breaking Barriers in HPC: How Apptainer is Changing Software Deployment in NSM - Parikshit Ardhapurkar & Samir Shaikh, Centre for Development of Advanced Computing
Wednesday May 7, 2025 10:45am - 11:05am CDT
Managing software on supercomputers is challenging due to diverse architectures, dependency complexities, and security constraints. Traditional tools like Spack help with software compilation, but they lack portability and require system-specific configurations. Apptainer (formerly Singularity) solves these issues by enabling secure, portable, and reproducible containerized environments tailored for HPC. Unlike Docker, Apptainer runs without root access, ensuring security and compliance in multi-user clusters. It supports MPI, GPUs, and Slurm, allowing high-performance workloads to run seamlessly across systems. By integrating Apptainer into NSM supercomputers, researchers benefit from faster deployments, reproducible results, and reduced administrative overhead. This session explores Apptainer’s role in transforming HPC workflows, its advantages over traditional software management, and best practices for adopting containers in large-scale computing environments.
Speakers
avatar for Parikshit Ardhapurkar

Parikshit Ardhapurkar

HPC Engineer, Centre for Development of Advanced Computing, Pune
Parikshit Ardhapurkar is an HPC Engineer at C-DAC, India, specializing in parallel computing, software optimization, and performance tuning. He plays a key role in deploying Apptainer for secure and portable HPC environments under the National Supercomputing Mission (NSM). With expertise... Read More →
avatar for Samir Shaikh

Samir Shaikh

Scientist, Centre for Developement of Advanced Computing (C-DAC)
Samir Shaikh is an HPC specialist at C-DAC, Pune, optimizing large-scale workloads, parallel computing, and system architecture. As a Scientist C, he enhances HPC performance for AI/ML, scientific computing, and NSM supercomputers. An IIT Guwahati M.Tech graduate, he has contributed... Read More →
Wednesday May 7, 2025 10:45am - 11:05am CDT
Mississippi River 1 & 2

10:45am CDT

Charliecloud Workshop: Key Workflow Operation - Build from Dockerfile - Megan Phinney, Los Alamos National Laboratory
Wednesday May 7, 2025 10:45am - 11:05am CDT
This workshop will provide participants with background and hands-on experience to use basic Charliecloud containers for HPC applications. Participants will build toy containers and a real HPC application, and then run them in parallel on a supercomputer. This will be a highly interactive workshop with lots of Q&A.

This section will walk through the key workflow operation, building from a Dockerfile.

LA-UR-25-22140
Speakers
avatar for Megan Phinney

Megan Phinney

Scientist, Los Alamos National Laboratory
Wednesday May 7, 2025 10:45am - 11:05am CDT
Illinois River

10:45am CDT

Optimizing Spack: Multi-Package Parallel Builds for Faster Installation - Kathleen Shea, Lawrence Livermore National Laboratory
Wednesday May 7, 2025 10:45am - 11:05am CDT
Spack 1.0 will feature faster builds through multi-package parallelism. In this talk, I’ll describe how I accelerated Spack’s package installation process by parallelizing its main installer loop. By enabling Spack to spawn multiple package builds concurrently, this feature increases available parallelism and significantly reduces overall build times for multi-package installations. I'll talk about the design decisions, tradeoffs, and performance gains achieved, providing valuable lessons for optimizing package builds on platforms from laptops to large-scale HPC environments. I’ll also talk about how, even as a a relatively new Spack user, I was able to come up to speed quickly and contribute meaningful improvements to the project.

Speakers
avatar for Kathleen Shea

Kathleen Shea

Software Developer, Lawrence Livermore National Laboratory
Kathleen Shea graduated from Colorado College with a Bachelor of Arts in Computer Science in 2024. She then started her career at Lawrence Livermore National Lab contributing to both Center for Applied Scientific Computing and Livermore Computing. She specializes in core feature development... Read More →
Wednesday May 7, 2025 10:45am - 11:05am CDT
Salon E-G

10:45am CDT

Kokkos in Applications
Wednesday May 7, 2025 10:45am - 12:05pm CDT
1. FleCSI Applications, Ben Bergen & Hyun Lim, Los Alamos National Laboratory (10 minutes)
The Flexible Computational Science Infrastructure (FleCSI) programming system provides a clutter-free environment that allows developers to focus on the arithmetic operations of their methods without the distraction of computer science details that are often visible in legacy simulation codes. To this end, FleSCI provides light-weight wrappers over the raw Kokkos interface that resemble native C++ keywords, e.g., forall. Using this design philosophy, we have been able to evolve our support to cover various Kokkos policies and execution spaces. HARD is a FleCSI-based application for radiation hydrodynamics that is performance portable across a variety of systems, e.g., El Capitan, Venado, and Crossroads, and inherits FleCSI’s support for multiple distributed-memory and tasking backends, e.g., Legion, HPX, and MPI. In this talk, we will demonstrate the basic data-parallel interface with implementation and usage examples. We will also present results for several test problems in inertial confinement fusion with comparisons between different backends and performance assessments in different heterogeneous computing environments.

2. DDC: A Performance Portable Library Abstracting Computation on Discrete Domains, Thomas Padioleau, CEA Paris-Saclay (10 minutes)
The Discrete Domain Computation (DDC) library is a modern C++ library that aims to offer to the C++ world an equivalent to the xarray.DataArray Python environment. The Xarray library introduces labeled multidimensional arrays, enabling more intuitive data manipulation by associating dimensions with user-provided names rather than relying on positional indexing. This approach simplifies indexing, slicing, and broadcasting while reducing common indexing errors. Inspired by these ideas, DDC extends the Kokkos library providing zero-overhead dimension labeling for multidimensional arrays along with performance-portable multidimensional algorithms. This labeling mechanism enables compile-time detection of indexing and slicing errors, ensuring safer and more expressive array operations in C++. In this presentation, we will introduce the core concepts of DDC and demonstrate its usage through a simple example that highlights its key features.

3. TChem-atm - A Performance Portable Chemistry Solver for Atmospheric Chemistry, Oscar Diaz-Ibarra, Sandia National Laboratories (20 minutes)
TChem-atm (https://github.com/PCLAeroParams/TChem-atm) is a performance-portable software library designed to support atmospheric chemistry applications, specifically computing source term Jacobian matrices. The software utilizes Kokkos as its portability layer, preparing it for next-generation computing architectures. The software interface employs a hierarchical parallelism design to leverage the massive parallelism available on modern computing platforms, including model parallelism, batch parallelism, and nested parallelism for each problem instance. Additionally, TChem-atm is designed to be coupled with third-party libraries that may be used to advance the state of gas and particle species over time, notably interfacing with the Tines, Kokkos-kernels, and Sundials libraries. We have tested TChem-atm in two scenarios: using a typical reaction mechanism in atmospheric science and an example involving multiple aerosol particles. This testing framework allows us to evaluate our code by varying the number of evaluations and the size of the source term (right-hand side). Finally, we report performance measurements using the CUDA, HIP, and OpenMP back ends.

4. GPU Porting of the TRUST CFD Platform with Kokkos, Rémi Bourgeois, French Atomic Energy Commission (CEA) (20 minutes)
TRUST is a High Performance Computing thermohydraulic platform for Computational Fluid Dynamics developed at the French Atomic Energy Commission (CEA). This software is designed for massively parallel (MPI) simulations of conduction, incompressible single-phase, and Low Mach Number (LMN) flows with a Weakly-Compressible multi-species solver and compressible multi-phase flows. It is used as the basis for many specialised applications in the nuclear and new energy fields across CEA. The code is being progressively ported to support GPU acceleration (Nvidia/AMD/Intel) thanks to the Kokkos library, as it is one of the demonstrators of the CExA project. In this talk we will go over our experience using Kokkos to progressively port our large code base. We will cover our enabled GPU features and performances. We will mention some of the difficulties we encountered as well as the strategies we had to adopt that sometimes differ from standard good practices due to the specificity of our application.

5. Omega: Towards a Performance-portable Ocean Model using Kokkos, Maciej Waruszewski, Sandia National Laboratories (20 minutes)
High-resolution simulations of the Earth system require resources available only on the world's largest supercomputers, which are increasingly based on GPUs. However, CPU-based systems are still frequently used to conduct simulations at coarse resolutions. To be able to take advantage of all compute platforms, we are developing Omega: the Ocean Model for E3SM Global Applications, a new ocean model written in C++ using Kokkos for performance portability. Omega will replace MPAS-Ocean to become the new ocean component of the DOE’s Energy Exascale Earth System Model (E3SM). Omega is an unstructured mesh ocean model based on the same finite-volume scheme as the current ocean component. Work on Omega began in 2023. Currently, Omega is a layered shallow water model with passive tracers. While still simple, this initial version can run on realistic size meshes and contains computational kernels representative of the full model horizontal numerics. After briefly describing Omega, this talk will go into our experiences with Kokkos and present initial performance results from a variety of compute platforms.)
Speakers
avatar for Ben Bergen

Ben Bergen

Scientist, Los Alamos National Laboratory
Ben Bergen is a computational scientist working on runtime systems, data structures, and applications development.
avatar for Hyun Lim

Hyun Lim

Scientist, Los Alamos National Laboratory
Hyun Lim is a staff scientist in CCS-7. Hyun has a background in theoretical and computational astrophysics, gravitational physics, and numerical methods.
avatar for Maciej Waruszewski

Maciej Waruszewski

R&D Computer Science, Sandia National Laboratories
Maciej is a computer scientist at Sandia National Laboratories. He is one of the developers of the DOE’s Energy Exascale Earth System Model (E3SM). He holds a PhD in atmospheric physics from the University of Warsaw.
avatar for Oscar Diaz-Ibarra

Oscar Diaz-Ibarra

Senior member of the technical staff, Sandia National Laboratories
Oscar is a senior member of the technical staff at Sandia National Laboratories, specializing in high-performance applications for atmospheric chemistry using Kokkos and modern C++. He holds a Ph.D. in chemical engineering from the University of Utah and has over 7 years of experience... Read More →
avatar for Rémi Bourgeois

Rémi Bourgeois

Researcher / Engineer, French Atomic Energy Commission (CEA)
Rémi Bourgeois is a French researcher/engineer at CEA Saclay, specializing in HPC and numerical analysis for the TRUST platform, a massively parallel thermo-hydraulic simulation tool. He earned his PhD at CEA, focusing on MHD convection, developing finite-volume methods and GPU-based... Read More →
avatar for Thomas Padioleau

Thomas Padioleau

Engineer-Researcher, CEA
Dr. Thomas Padioleau is a CEA Engineer-Researcher at Maison de la Simulation. He leads the DDC project and also works on Voice++.
Wednesday May 7, 2025 10:45am - 12:05pm CDT
Salon A-C

11:05am CDT

Updates on the WarpX Project: Experiences with AMReX, Lessons Learned, and Future Challenges - Edoardo Zoni, Lawrence Berkeley National Laboratory
Wednesday May 7, 2025 11:05am - 11:25am CDT
The WarpX project is advancing the modeling and simulation of a wide range of physics applications, including particle accelerators and nuclear fusion, through high-performance scientific computing. This talk will provide a comprehensive update on WarpX, focusing on our experiences with the AMReX software framework. We will highlight key achievements, share valuable lessons learned, and discuss the future challenges we anticipate. The presentation will cover the critical role of AMReX in enabling high performance and portability, showcase significant milestones, and provide insights into the challenges faced and solutions developed.
Speakers
EZ

Edoardo Zoni

Research Software Engineer, Lawrence Berkeley National Laboratory
Wednesday May 7, 2025 11:05am - 11:25am CDT
Rock River 1 & 2

11:05am CDT

Leveraging Apptainer for Scalable and Secure Distributed Computing - Om Jadhav & Vamshi Krishna, C-DAC India
Wednesday May 7, 2025 11:05am - 11:25am CDT
This session delves into the advanced utilization of Apptainer (formerly Singularity) for distributed computing in high-performance computing (HPC) environments. It will cover containerized execution methodologies, security enhancements, and scalability optimizations critical for scientific computing. The discussion will emphasize MPI integration, multi-node orchestration, and performance considerations for HPC workloads within Apptainer containers. Researchers and scientists will gain a comprehensive understanding of deploying, managing, and optimizing containerized scientific applications across distributed HPC infrastructures.
Speakers
avatar for Om Jadhav

Om Jadhav

Scientist, C-DAC, India
Mr. Om Jadhav is positioned as Scientist-D, C-DAC, Ministry of Electronics and Information Technology, Government of India. He is associated with the HPC-Technologies team, CDAC Pune. His areas of expertise include HPC application optimization and management on HPC clusters. He has... Read More →
avatar for Vamshi Krishna

Vamshi Krishna

HPC Application Expert, C-DAC India
Dr. Vamshi Krishna, an HPC Application Expert & Scientist, has 20+ years of experience in HPC, Embedded Systems, Robotics, IoT, and AI. He plays a key role in deploying supercomputers across India under the NSM initiative, with 15 systems deployed and 9 more planned using indigenous... Read More →
Wednesday May 7, 2025 11:05am - 11:25am CDT
Mississippi River 1 & 2

11:05am CDT

Charliecloud Workshop: Key Workflow Operation - Push - Megan Phinney, Los Alamos National Laboratory
Wednesday May 7, 2025 11:05am - 11:25am CDT
This workshop will provide participants with background and hands-on experience to use basic Charliecloud containers for HPC applications. Participants will build toy containers and a real HPC application, and then run them in parallel on a supercomputer. This will be a highly interactive workshop with lots of Q&A.

This section will walk through the key workflow operation push.

LA-UR-25-22140
Speakers
avatar for Megan Phinney

Megan Phinney

Scientist, Los Alamos National Laboratory
Wednesday May 7, 2025 11:05am - 11:25am CDT
Illinois River

11:05am CDT

Fast Binary Installation with Spack Splicing - John Gouwar, Northeastern University
Wednesday May 7, 2025 11:05am - 11:25am CDT
Binary package managers allow for the fast installation of binary artifacts, but limit configurability to ensure compatibility between binaries due to rigid ABI requirements. Source package managers allow for more flexibility in building software, since binaries are compiled on demand, but compilation can take a considerable amount of time. Spack has existing a mechanism for mixing source and precompiled packages; however, because Spack does not model ABI compatibility between packages, all transitive dependencies of a binary package must have been built at the same time as that package in order to maintain ABI compatiblity. We present an extension to Spack, which we call splicing, that models ABI compatibility in the package ecosystem and allows seamless mixing of source and binary distribution of packages. This extension augments both the packaging language and dependency resolution engine of Spack in order to maximize reused binaries while maintaining the flexibility of source based management. Through empirical evaluation, we show that our extension incurs minimal performance overhead to dependency resolution while greatly extending the modeling capability of Spack.
Speakers
avatar for John Gouwar

John Gouwar

Doctoral Student, Northeastern University
John Gouwar is a doctoral student at the Khoury College of Computer Sciences at Northeastern University, advised by Arjun Guha. His doctoral research, which he began in 2021 and expects to complete in 2026, focuses on programming languages and package management. Gouwar is broadly... Read More →
Wednesday May 7, 2025 11:05am - 11:25am CDT
Salon E-G

11:25am CDT

Using AMReX’s Embedded Boundaries to Support MFIX-Exa’s Geometry Capabilities - Deepak Rangarajan, Battelle Memorial Institute
Wednesday May 7, 2025 11:25am - 11:45am CDT
Speakers
DR

Deepak Rangarajan

Data Scientist, Battelle Memorial Institute
Wednesday May 7, 2025 11:25am - 11:45am CDT
Rock River 1 & 2

11:25am CDT

Recent Apptainer Developments - Dave Dykstra, Fermi National Accelerator Laboratory
Wednesday May 7, 2025 11:25am - 11:45am CDT
The lead developer of Apptainer will discuss recent changes to Apptainer.
Speakers
DD

Dave Dykstra

Fermi National Accelerator Laboratory
Wednesday May 7, 2025 11:25am - 11:45am CDT
Mississippi River 1 & 2

11:25am CDT

Charliecloud Workshop: MPI Hello World - Megan Phinney, Los Alamos National Laboratory
Wednesday May 7, 2025 11:25am - 11:45am CDT
This workshop will provide participants with background and hands-on experience to use basic Charliecloud containers for HPC applications. Participants will build toy containers and a real HPC application, and then run them in parallel on a supercomputer. This will be a highly interactive workshop with lots of Q&A.

This section will include a multi-node MPI hello world example.

LA-UR-25-22140
Speakers
avatar for Megan Phinney

Megan Phinney

Scientist, Los Alamos National Laboratory
Wednesday May 7, 2025 11:25am - 11:45am CDT
Illinois River

11:25am CDT

Spack on Windows - John Parent, Kitware, Inc.
Wednesday May 7, 2025 11:25am - 11:45am CDT
While primarily run, tested, and developed on Unix-like operating systems, in particular HPC systems, scientific software is often deployed on Windows. Spack's recent expansion to Windows marks a significant milestone, enabling its powerful package management capabilities on a new platform.This talk will provide technical insight into the process of adapting Spack to support a platform orthogonal in design to anything Spack has supported in the past. We cover the path to initial support, the current state of Spack on Windows, and look at the roadmap for future Windows development. We’ll explore unique challenges supporting Windows and their solutions. This talk will address design, new features introduced to support Windows development, and how Spack brings needed robust package management to the Windows ecosystem. We’ll cover best practices for porting new or existing packages to Windows, deploying and managing Spack environments in real-world scenarios, and standardizing your Windows development workflows with a focus on common pitfalls for Windows developers. This talk will provide a pathway for attendees interested in supporting Windows to go forth and Spack!

Speakers
avatar for John Parent

John Parent

Senior R&D Engineer, Kitware, Inc.
John Parent is a senior research and development engineer on the Software Solutions Team at Kitware, Inc., where he is the primary developer of the Spack package manager’s Windows support. His other work covers contributions to CMake, establishing complex CI systems and C++ /Python... Read More →
Wednesday May 7, 2025 11:25am - 11:45am CDT
Salon E-G

11:45am CDT

Solid Mechanics in Alamo/AMReX: From Optimized Trusses to Burning Propellants - Brandon Runnels, Iowa State University
Wednesday May 7, 2025 11:45am - 12:05pm CDT
The phase field (PF) method is used to simulate a wide range of problems in mechanics, ranging from crack propagation to topology optimization. As a diffuse interface method, PF requires AMR to be computationally feasible, but few PF methods leverage block-structured AMR. Alamo is an AMReX-based code designed to solve phase field equations with implicit elastic solves. It features a variety of phase field methods, material models, and a strong-form nonlinear elastic solver based on AMReX's MLMG. In this talk, we give a high-level overview of some of the applications of Alamo, including deflagration of solid rocket propellant, topology optimization of mechanical structures, phase field fracture, microstructure evolution, and solid-fluid interaction.
Speakers
BR

Brandon Runnels

Associate Professor of Aerospace Engineering, Iowa State University
Wednesday May 7, 2025 11:45am - 12:05pm CDT
Rock River 1 & 2

11:45am CDT

Session to be Announced
Wednesday May 7, 2025 11:45am - 12:05pm CDT
Wednesday May 7, 2025 11:45am - 12:05pm CDT
Mississippi River 1 & 2

11:45am CDT

Charliecloud Workshop: Wrap-up - Megan Phinney, Los Alamos National Laboratory
Wednesday May 7, 2025 11:45am - 12:05pm CDT
This workshop will provide participants with background and hands-on experience to use basic Charliecloud containers for HPC applications. Participants will build toy containers and a real HPC application, and then run them in parallel on a supercomputer. This will be a highly interactive workshop with lots of Q&A.

This section will wrap-up the workshop with final Q&A.

LA-UR-25-22140
Speakers
avatar for Megan Phinney

Megan Phinney

Scientist, Los Alamos National Laboratory
Wednesday May 7, 2025 11:45am - 12:05pm CDT
Illinois River

11:45am CDT

Spack CI: Past, Present, and Future - Ryan Krattiger, Kitware, Inc.
Wednesday May 7, 2025 11:45am - 12:05pm CDT
Spack's Continuous Integration (CI) system is essential for building and distributing reliable HPC software packages. It has dramatically scaled from building hundreds to hundreds of thousands of packages weekly. Leveraging GitLab, Spack CI has evolved from a Linux-only build system to orchestrating runners across diverse providers, architectures, and platforms, supporting multiple domain stacks like E4S, HEP, and Radiuss. Through enhanced monitoring and data-driven methodologies, including machine learning, Spack CI has gained insights to optimizing resource allocation and analyzing failure modes. Improvements to binary caching reduce storage and prevent race conditions. The core goal is to maintain a resilient and efficient CI ecosystem, ensuring the reliability of Spack's HPC software for its expanding community.
Speakers
avatar for Ryan Krattiger

Ryan Krattiger

Senior Research Engineer, Kitware, Inc.
I am a CFD researcher turned software solutions engineer. Starting at a private CFD company building out HPC frameworks for handling multi-phyiscs and FSI simulations to run on heterogeneous systems. Built CI and testing workflows out of necessity and became and HPC build systems... Read More →
Wednesday May 7, 2025 11:45am - 12:05pm CDT
Salon E-G

12:05pm CDT

Lunch (Provided for Attendees)
Wednesday May 7, 2025 12:05pm - 1:35pm CDT
Wednesday May 7, 2025 12:05pm - 1:35pm CDT
Atrium

1:35pm CDT

In Situ Integration of AMReX with Python Ecosystem via pyAMReX - Bhargav Siddani, Lawrence Berkeley National Laboratory
Wednesday May 7, 2025 1:35pm - 1:55pm CDT
The past decade has seen a rapid increase in the development and use of Python-based open-source libraries for data-driven methods and machine learning. The widespread adoption of these libraries across scientific and engineering disciplines highlights their growing prominence in accelerating simulations. This talk presents an in-situ (in-memory/online) workflow, which requires minimal modifications, designed for mature AMReX codes to leverage the rich Python-ecosystem. The workflow enables language-interoperable data transfer through runtime coupling of AMReX and pyAMReX via Multiple Program Multiple Data (MPMD) mode of Message Passing Interface (MPI). This capability is demonstrated through multiscale modeling of granular flows, which involves coupling low-fidelity continuum and high-fidelity particle-based methods. The computational intractability of straightforward coupling between low- and high-fidelity methods is addressed using adaptively (on-the-fly) evolving neural network ensembles, implemented in PyTorch Distributed Data Parallel, as a surrogate for the expensive high-fidelity solver. The scalability of the current approach across multiple GPUs will also be discussed.
Speakers
BS

Bhargav Siddani

Postdoc, Lawrence Berkeley National Laboratory
Wednesday May 7, 2025 1:35pm - 1:55pm CDT
Rock River 1 & 2

1:35pm CDT

Key Charliecloud Innovation - Build Cache - Reid Priedhorsky, Los Alamos National Laboratory
Wednesday May 7, 2025 1:35pm - 1:55pm CDT
Container images are built by interpreting instructions in a machine-readable recipe, typically a Dockerfile, which is often faster with a build cache that stores instruction results for re-use. The standard approach (used e.g. by Docker and Podman) is a many-layered union filesystem, encoding differences between layers as tar archives.

This talk describes a new approach, as implemented in Charliecloud: store changing images in a Git repository. Our experiments show this performs similarly to layered caches on both build time and disk usage, with a considerable advantage for many-instruction recipes. Our approach also has structural advantages: better diff format, lower cache overhead, and better file de-duplication. These results show that a Git-based cache for layer-free container implementations is not only possible but may outperform the layered approach on important dimensions.

LA-UR-25-22140
Speakers
RP

Reid Priedhorsky

Scientist, Los Alamos National Laboratory
I am a staff scientist at Los Alamos National Laboratory. Prior to Los Alamos, I was a research staff member at IBM Research. I hold a Ph.D. in computer science from the University of Minnesota and a B.A., also in computer science, from Macalester College.My work focuses on large-scale... Read More →
Wednesday May 7, 2025 1:35pm - 1:55pm CDT
Illinois River

1:35pm CDT

E4S and Spack - Sameer Shende, University of Oregon
Wednesday May 7, 2025 1:35pm - 1:55pm CDT
This talk will describe how the Extreme-Scale Scientific Software Stack is developed and maintained. E4S is an Ecosystem for Science and is layered upon the Spack package manager. This talk will focus on the maintenance of E4S and the continuous integration (CI) on the Frank system at the University of Oregon.
Speakers
avatar for Sameer Shende

Sameer Shende

Research Professor and Director, Performance Research Lab, U. Oregon, University of Oregon
Sameer Shende serves as a Research Professor and the Director of the Performance Research Lab at the University of Oregon and the President and Director of ParaTools, Inc. (USA) and ParaTools, SAS (France). He serves as the Technical Lead of the Extreme-scale Scientific Software Stack... Read More →
Wednesday May 7, 2025 1:35pm - 1:55pm CDT
Salon E-G

1:35pm CDT

Trilinos Introduction and Overview - Curtis Ober, Sandia National Laboratories
Wednesday May 7, 2025 1:35pm - 1:55pm CDT
In this presentation, we will offer a concise introduction and overview of the Trilinos project. Established in 2001, Trilinos is a collaborative initiative focused on developing algorithms and technologies to address complex multi-physics engineering and scientific challenges on high-performance computing (HPC) architectures. This includes a range of capabilities such as linear and nonlinear solvers, optimization techniques, and sensitivity analysis. A key mission of Trilinos is to cultivate a vibrant community of developers and users who contribute to a modular software framework comprised of diverse computational packages. This collaborative environment facilitates resource sharing among library developers, enhancing software compatibility and overall efficiency amortization of costs.
Speakers
avatar for Curtis Ober

Curtis Ober

Distinguished Member of Technical Staff, Sandia National Laboratories
Curt has been with Sandia for 30 years and is currently a Distinguished Member of the Technical Staff.  His career has spanned many projects and missions at the laboratory, including computational fluid dynamics (CFD) for re-entry vehicles, shock hydrodynamics, time integration (Trilinos... Read More →
Wednesday May 7, 2025 1:35pm - 1:55pm CDT
Mississippi River 1 & 2

1:35pm CDT

Adopting Kokkos
Wednesday May 7, 2025 1:35pm - 3:15pm CDT
1. A Brief Overview of LANL's use of Kokkos - Daniel Holladay, Los Alamos National Laboratory (20 minutes)
Since the commissioning of the first petascale machine, Roadrunner, in 2009 at the Los Alamos
National Laboratory (LANL), the ability for physics codes at LANL to take advantage of accelerators
has provided utility and productivity improvements for code users. The ability to take advantage of an
accelerator, and more specifically general purpose graphics processing units (GPGPUs), will quickly
move from a productivity enhancement to absolutely necessary as more than 90% of the compute
capability of the El Capitan supercomputer at Lawrence Livermore National Laboratory (LLNL) will
only be accessed through effective use of its GPGPUs, a task which has traditionally been
accomplished with vendor specific software extensions such as CUDA or HIP. Many projects with
code bases ranging from large and established FORTRAN codes to new c++ based projects have
made the decision to use Kokkos as the tool that will enable effective use of LLNL's El Capitan
compute resources as well as future machines which could likely benefit from Kokkos's capabilities.
In this talk I will give an overview of several physics code projects at LANL and their usage of Kokkos.

2. Enhancing Fortran Code for Operational Weather Forecasting with Kokkos: Results and Lessons Learned - Timothy Sliwinski, Cooperative Institute for Research in the Atmosphere (CIRA) (20 minutes)
At NOAA, much of the code for numerical weather prediction (NWP) and operational weather forecasting is built upon Fortran, into which decades of scientific research knowledge and expertise has been invested. Therefore, moving away from Fortran and potentially breaking what has been a highly reliable system for many years is a significant challenge.
To demonstrate new methods to modernize NOAA’s NWP models, Kokkos was selected due to its ability to work across multiple GPUs and CPUs with a single source code and the presence of the Fortran Language Compatibility Layer (FLCL), easing development of the interface between Fortran and C++ Kokkos kernels. As a first step, the YSU Planetary Boundary Layer (PBL) scheme was chosen as the target and a prototype with Kokkos was developed, tested, and performance benchmarked. In this presentation, we report the performance of this new Kokkos-enhanced Fortran code on CPU and an Nvidia GPU, the challenges of the C/Fortran interface, potential future prospects for the use of Kokkos at NOAA, and overall lessons learned from this project for anyone else interested in using Kokkos with existing Fortran source codes.

3. Using Umpire's Memory Management Capabilities with Kokkos - Kristi Belcher, LLNL (20 minutes)
Umpire is an open-source data and memory management library created at Lawrence Livermore National Laboratory (LLNL). Although Umpire is part of the RAJA Portability Suite, it was made to be modular and can therefore be used with Kokkos and other performance portability abstractions. Umpire provides memory pools which avoid expensive calls to the underlying device-specific API making allocations, large or small, performant in HPC environments. Umpire provides numerous types of memory resources and allocators (i.e. Device, Host, Unified Memory, IPC Shared Memory, etc.). In this talk, I will discuss key Umpire features and capabilities and showcase a Kokkos example with Umpire.

4. Early Experiences Using Kokkos for Multi-Resolution Analysis - Joseph Schuchart, Stony Brook University (20 minutes)
MADNESS is a framework for multi-resolution analysis with application in quantum chemistry. In this talk, we will present some early experiences in using Kokkos in a port of MADNESS to the TTG data-flow programming model, which includes both a restructuring of the existing program flow and a port to accelerators.
Speakers
avatar for Daniel Holladay

Daniel Holladay

Computational Physicist, Los Alamos National Laboratory
Daniel Holladay is the deputy project leader for computer science for the project that maintains the FLAG Lagrangian multi-physics code at the Los Alamos National Laboratory (LANL). He received a Ph.D. in Nuclear Engineering from Texas A&M University in 2018 while working as a LANL... Read More →
avatar for Joseph Schuchart

Joseph Schuchart

Senior Research Scientist, Stony Brook University
Joseph Schuchart is a Senior Research Scientist at the Institute for Advanced Computational Science at Stony Brook University. He has been working on distributed data flow programming models and communication models, currently working at the intersection with computational chemistry... Read More →
avatar for Kristi Belcher

Kristi Belcher

Software Developer, LLNL
Kristi is a Software Developer at Lawrence Livermore National Laboratory working primarily on Umpire, an open source library that supports parallel data and memory management on HPC platforms, and MARBL, a large multi-physics simulation code. Kristi also works on the RADIUSS project... Read More →
avatar for Timothy Sliwinski

Timothy Sliwinski

HPC Software Developer, Cooperative Institute for Research in the Atmosphere (CIRA)
Dr. Timothy Sliwinski is an atmospheric scientist with the Cooperative Institute for Research in the Atmosphere at Colorado State University. Working directly with NOAA Global System Laboratory federal scientists in the Scientific Computing Branch, Dr. Sliwinski has worked on multiple... Read More →
Wednesday May 7, 2025 1:35pm - 3:15pm CDT
Chicago River Ballroom

1:55pm CDT

A High-Performance, GPU-Enabled, Data-Driven Micromagnetics Solver for Spintronic Systems - Yingheng Tang, Lawrence Berkeley National Laboratory
Wednesday May 7, 2025 1:55pm - 2:15pm CDT
To comprehensively investigate multiphysics coupling in spintronic devices, GPU acceleration is essential to address the spatial and temporal disparities inherent in micromagnetic simulations. Beyond traditional numerical methods, machine learning (ML) offers a powerful approach to replace and accelerate computationally expensive routines, particularly in evaluating demagnetization fields. Leveraging AMReX and python-based ML workflows, we developed an open-source micromagnetics tool that integrates ML-driven surrogate models to enhance computational efficiency. By replacing costly demagnetization field calculations with neural network-based approximations, the framework significantly accelerates simulations while maintaining accuracy. In addition to supporting key magnetic interactions—including Zeeman, demagnetization, anisotropy, exchange, and Dzyaloshinskii-Moriya interactions—it is validated on µMAG standard problems, widely accepted DMI benchmarks, and Skyrmion-based applications. This ML-accelerated approach improves computational performance and enables large-scale, data-driven micromagnetics simulations, advancing the study of spintronic and electronic systems.
Speakers
YT

Yingheng Tang

Postdoc, Lawrence Berkeley National Laboratory
Wednesday May 7, 2025 1:55pm - 2:15pm CDT
Rock River 1 & 2

1:55pm CDT

Key Charliecloud Innovation - SIF - Krishna Chilleri, Los Alamos National Laboratory
Wednesday May 7, 2025 1:55pm - 2:15pm CDT
The Singularity Image Format (SIF) is a compressed read-only SquashFS filesystem that includes everything needed to run a containerized application. Leveraging its SquashFS-based architecture, Charliecloud has integrated support for the reading and execution of this image format. This talk will provide details about how the ch-run functionality identifies SIF files, analyzes the file header structure, and retrieves the SquashFS partition offset needed for container image execution.

Additionally, the talk is designed for anyone looking to gain a clearer understanding of SIF, especially if you find the upstream documentation challenging to navigate.

LA-UR-25-22166
Speakers
KC

Krishna Chilleri

Student, Los Alamos National Laboratory
Wednesday May 7, 2025 1:55pm - 2:15pm CDT
Illinois River

1:55pm CDT

Spack-Stack: An Interagency Collaboration and Spack Extension - Dom Heinzeller, NRL / UCAR
Wednesday May 7, 2025 1:55pm - 2:15pm CDT
Numerical weather prediction (NWP) systems are complex, coupled models with several distinct components. They are often developed by different organizations: national agencies, research institutes, universities, and the community. Naturally, they require extensive third-party libraries and run on a wide range of computational platforms, from the largest high-performance computing systems (HPCs) to low-spec GitHub action runners.

To address this challenge of ever-increasing complexity, in 2021 the Joint Center for Satellite Data Assimilation (JCSDA) and NOAA’s Environmental Modeling Center (EMC) created spack-stack, a novel, collaborative effort geared towards universal and portable NWP software environments based on the spack package manager for supercomputers developed at LLNL. The spack-stack collaboration has grown since its inception, and as of 2025 includes the United States Naval Research Lab (NRL) and NOAA’s Earth System Prediction Center (EPIC). The spack-stack software environments are also increasingly used at NASA’s Global Modeling and Assimilation Office (GMAO), making it a core component of some of the Nation’s flagship NWP systems.
Speakers
avatar for Dom Heinzeller

Dom Heinzeller

Computational Scientist, NRL / UCAR
Dom Heinzeller graduated with a PhD in Astronomy from the University of Heidelberg, Germany. Following a postdoctoral position on the evolution of protoplanetary disks at Kyoto University, Japan, he moved from Astrophysics to Numerical Weather Prediction. In his NWP career, he worked... Read More →
Wednesday May 7, 2025 1:55pm - 2:15pm CDT
Salon E-G

1:55pm CDT

An Overview of the Trilinos Core Products - Roger Pawlowski, Sandia National Laboratories
Wednesday May 7, 2025 1:55pm - 2:15pm CDT
The Trilinos core area includes packages that form the basic building blocks for Trilinos capabilities. This talk
will briefly introduce each package and demonstrate how these can be pulled together into an application. We will
cover the following packages. The Kokkos performance portability library provides data structures and parallel
algorithms that are performant on both CPU and GPU architectures. The Kokkos-Kernels library provides
performance portable BLAS and LAPACK routines. The Tpetra library provides MPI-based parallel distributed linear
algebra data structures built on Kokkos. The Teuchos library provides utilities for a common look and feel
across Trilinos packages, including a reference counted smart pointer, print utilities, timers and a parameter list
for user input. The Zoltan library provides parallel distributed load balancing tools. The PyTrilinos package
provides python wrappers for Trilinos capabilities. The Thyra package provides abstraction layers for writing
linear algebra, linear solvers and nonlinear solvers. The talk will conclude with examples of Trilinos capabilities
integrated into the Empire electromagnetic particle-in-cell code.

Speakers
RP

Roger Pawlowski

Computational Scientist R&D, Sandia National Laboratories
Wednesday May 7, 2025 1:55pm - 2:15pm CDT
Mississippi River 1 & 2

2:15pm CDT

Exascale Modeling with ML-Based Surrogates for Magnon-Photon Dynamics in Hybrid Quantum System - Zhi Jackie Yao, Lawrence Berkeley National Laboratory
Wednesday May 7, 2025 2:15pm - 2:35pm CDT
We introduce a hybrid HPC–ML framework for efficient modeling of magnon–photon interactions. The HPC component uses an explicit FDTD leap-frog Maxwell–LLG solver (second-order accurate), solving Maxwell’s equations in nonmagnetic regions and adding the LLG equation where ferromagnets are present. Parallelization leverages AMReX’s MPI+X model for multicore CPUs or GPUs, partitioning the domain among MPI ranks. Data collected from nine points in the ferromagnet feed a Long Expressive Memory (LEM) encoder–decoder, trained with a composite loss function (reconstruction, prediction, and physics) and guided by Curriculum Learning. During training, we begin with shorter sequences, no physics enforcement, and a higher learning rate, then move to longer sequences, physics constraints, and a lower rate. Using just 1 ns of high-fidelity simulation data, the ML surrogate accurately predicts the magnetic-field evolution and matches frequency responses (13–18 GHz) under various DC biases. With physics constraints included, errors remain low even for longer sequences. The model reproduces transmission spectra and captures both primary and dual resonances (1800–2200 Oe) with high precision, achieving errors below 2.5% and demonstrating robust spatial and spectral generalization.
Speakers
ZJ

Zhi Jackie Yao

Research Scientist, Lawrence Berkeley National Laboratory
Wednesday May 7, 2025 2:15pm - 2:35pm CDT
Rock River 1 & 2

2:15pm CDT

Key Charliecloud Innovation - squashfs - Megan Phinney, Los Alamos National Laboratory
Wednesday May 7, 2025 2:15pm - 2:35pm CDT
The typical filesystem image formats for Charliecloud are SquashFS and tar archives. SquashFS is a compressed, read-only filesystem that unprivileged users can mount in user space with SquashFUSE; it is the preferred image format due to its various efficiencies. The previous SquashFS workflow was non-ideal due to user complexity and difficulties with HPC job schedulers.

We have designed a workflow that requires us to link SquashFUSE to Charliecloud to enable the new mount/unmount procedure of only needing a single user command. Also, an additional persistent process is needed to service the FUSE requests called the FUSE loop: once the containerized application process finishes, it unmounts the SquashFS and ends the FUSE loop. Our SquashFS workflow is currently in production with it being more user friendly, cleans up after itself and is more compatible with HPC job schedulers. We were able to reduce user commands from 3 to 1, increase reliability and decrease mount/unmount time by more than 50%.

LA-UR-25-22140
Speakers
avatar for Megan Phinney

Megan Phinney

Scientist, Los Alamos National Laboratory
Wednesday May 7, 2025 2:15pm - 2:35pm CDT
Illinois River

2:15pm CDT

Packaging for HPC and HTC in High Energy and Nuclear Physics: Comparing Spack to Other Solutions - Wouter Deconinck, University of Manitoba
Wednesday May 7, 2025 2:15pm - 2:35pm CDT
High Energy Physics (HEP) and Nuclear Physics (NP) experiments at accelerators facilities use modular software stacks which evolve over the decades-long lifetimes of these efforts. These software stacks have complex dependency trees, often reaching depths of O(50) levels and containing O(1000) nodes, with detailed versioning constraints. Reproducibility requirements demand that previous versions and their dependencies can be recalled on newer hardware architectures. The software stacks are deployed on computing centers around the world, as part of computing grids for high throughput computing and (increasingly) high performance computing. Current tools for managing these stacks have been around for years, but do not always have support for newer computing practices (containerization, integration in development workflows, heterogeneous architectures, deployment to shared file systems such as CernVM-FS). In this session, I will give an overview of the packaging solutions used in HEP and NP, and compare them with Spack in terms of functionality, adaptability, and usability.
Speakers
avatar for Wouter Deconinck

Wouter Deconinck

Associate Professor, University of Manitoba
Wouter Deconinck is an Associate Professor of Physics at the University of Manitoba. His research activities focus on experimental nuclear physics, in particular precision measurements of quantities that test our current best theory of fundamental particles and their interactions... Read More →
Wednesday May 7, 2025 2:15pm - 2:35pm CDT
Salon E-G

2:15pm CDT

Why Your Science Application Should Be Using Trilinos Linear Solvers - Jonathan Hu, Sandia National Laboratories
Wednesday May 7, 2025 2:15pm - 2:35pm CDT
Trilinos provides a variety of high-performance sparse iterative and direct linear solvers that are portable across CPU and GPU architectures. In this talk, I will provide motivation for why scientists and engineers developing a high-performance application should consider using Trilinos solvers. I'll give an overview of current solver capabilities across a spectrum of science applications, and discuss ongoing research as well as future directions.
Speakers
JH

Jonathan Hu

Computational Scientist R&D, Sandia National Laboratories
Wednesday May 7, 2025 2:15pm - 2:35pm CDT
Mississippi River 1 & 2

2:35pm CDT

AMReX for Renewable Energy Applications at NREL - Marc Day, National Renewable Energy Laboratory
Wednesday May 7, 2025 2:35pm - 2:55pm CDT
Speakers
MD

Marc Day

Group Manager, National Renewable Energy Laboratory
Wednesday May 7, 2025 2:35pm - 2:55pm CDT
Rock River 1 & 2

2:35pm CDT

Key Charliecloud Innovation - seccomp - Reid Priedhorsky, Los Alamos National Laboratory
Wednesday May 7, 2025 2:35pm - 2:55pm CDT
Do Linux distribution package managers need the privileged operations they request to actually happen? Apparently not, at least when building container images for HPC applications. Charliecloud uses this observation to implement a root emulation mode using a Linux seccomp filter that intercepts some privileged system calls, does nothing, and returns success to the calling program. This approach provides no consistency whatsoever but is sufficient to build a wide selection of Dockerfiles, including some that Docker itself cannot build, simplifying fully-unprivileged workflows needed for HPC application containers. This talk will detail the approach along its advantages and disadvantages. 

LA-UR-25-22140
Speakers
RP

Reid Priedhorsky

Scientist, Los Alamos National Laboratory
I am a staff scientist at Los Alamos National Laboratory. Prior to Los Alamos, I was a research staff member at IBM Research. I hold a Ph.D. in computer science from the University of Minnesota and a B.A., also in computer science, from Macalester College.My work focuses on large-scale... Read More →
Wednesday May 7, 2025 2:35pm - 2:55pm CDT
Illinois River

2:35pm CDT

Developing and Distributing HEP Software Stacks with Spack - Kyle Knoepfel & Marc Paterno, Fermi National Accelerator Laboratory
Wednesday May 7, 2025 2:35pm - 2:55pm CDT
The Computational Science and AI Directorate at Fermilab is using Spack to support the development efforts of a large number of scientific programmers, in many independent projects and experiments. While independent, these projects share many dependencies. They are typically under continuous and fairly rapid development. They have to support deployment on diverse hardware. This is a different context than is typical for the management of HPC software, where Spack was born. To support our community, we have created a model that enables users to develop code with greater efficiency than is possible with Spack’s current development facilities.

In this talk we will present:
- a brief introduction to the science we support (particle physics)
how the code we work with is naturally organized into several layers of packages
- how we are using Spack to manage those layers
- how we leverage the layering to provide efficient support for developers, using our Spack extension “MPD”.
- some suggestions for changes or additions to Spack to make such work easier.
Speakers
avatar for Kyle Knoepfel

Kyle Knoepfel

Senior software developer, Fermi National Accelerator Laboratory
I am a senior software developer at Fermilab, responsible for developing and leading computing framework efforts to meet the data-processing needs of many of Fermilab's experiments.
avatar for Marc Paterno

Marc Paterno

Computer Science Researcher, Fermi National Accelerator Laboratory
Marc Paterno is a Computer Science Researcher at the Fermi National Accelerator Laboratory. His research interests include the design and development of frameworks and libraries for large-scale collection, simulation, and analysis of data in High Energy Physics. He has a Ph.D. in... Read More →
Wednesday May 7, 2025 2:35pm - 2:55pm CDT
Salon E-G

2:35pm CDT

Introduction to Trilinos Discretization and Analysis Capabilities - Mauro Perego, Sandia National Laboratories
Wednesday May 7, 2025 2:35pm - 2:55pm CDT
In this presentation, we introduce the discretization and analysis tools available in the Trilinos software suite. Using real-world scientific applications as exemplars, we briefly explain how to efficiently implement portable high-order finite element discretizations on unstructured grids for a wide variety of problems, including leveraging time integration schemes for transient problems. Additionally, we describe how to utilize the automatic differentiation and adjoint capabilities for solving nonlinear problems and computing sensitivities. Finally, we present the simulation-constrained optimization capabilities available through the ROL package, highlighting how different Trilinos tools work together to efficiently solve large-scale inference problems.
Speakers
MP

Mauro Perego

Computational Scientist R&D, Sandia National Laboratories
Wednesday May 7, 2025 2:35pm - 2:55pm CDT
Mississippi River 1 & 2

2:55pm CDT

Advancing High-speed Compressible Flow Simulations: Moving Bodies and Multiphase Flows with AMReX - Mahesh Natarajan, Lawrence Berkeley National Laboratory
Wednesday May 7, 2025 2:55pm - 3:15pm CDT
High-fidelity simulations of high-speed, compressible flows require accurately capturing complex features for precise results. AMReX - an exascale, block-structured adaptive mesh refinement (AMR) framework, enables high-resolution simulations for a range of applications. This talk explores two key challenges in high-speed flows: moving bodies and liquid-gas flows. For moving bodies, a ghost-cell method is developed within the Compressible Navier-Stokes (CNS) framework of AMReX to compute fluxes on moving embedded boundary (EB) faces. A third-order least-squares formulation improves wall velocity gradient accuracy, enhancing skin friction coefficient estimation. The method is validated using inviscid and viscous test cases. For liquid-gas flows, an all-Mach multiphase algorithm solves the compressible flow equations using an unsplit volume-of-fluid (VOF) method with piecewise linear interface calculation (PLIC) for liquid-gas interface reconstruction. Simulations include a liquid jet in supersonic crossflow and spray atomization with acoustic excitation.
Speakers
MN

Mahesh Natarajan

Computer Systems Engineer III, Lawrence Berkeley National Laboratory
Wednesday May 7, 2025 2:55pm - 3:15pm CDT
Rock River 1 & 2

2:55pm CDT

Key Charliecloud Innovation - CDI - Reid Priedhorsky, Los Alamos National Laboratory
Wednesday May 7, 2025 2:55pm - 3:15pm CDT
An ongoing challenge for HPC containers is how to make host resources such as GPU devices and proprietary interconnects performantly available inside a container. In Charliecloud, the key requirement is placing shared libraries (.so files) into the container and then running ldconfig(8) to update its cache. (Other implementations also have to deal with device files and their associated permissions and symlinks, but because Charliecloud bind-mounts host /dev into the container, this is not needed).

Charliecloud has done this for some time using ch-fromhost(1), which is a large, reverse-engineered shell script that copies the needed files into a writeable image. It is difficult to maintain, does not support SquashFS or other write-only images, and adds a workflow step.

Other implementations typically use “OCI hooks”, which are arbitrary vendor-provided or custom programs run at various phases during container setup. These also present maintainability / bit-rot problems, can be opaque, and because their interface is solely “do it for me”, any invalid assumptions that hooks make can be difficult or impossible to work around.

A different approach is the emerging Container Device Interface (CDI) standard, with contributors from nVidia, Intel, Los Alamos, and others. This gives prescriptive JSON/YAML descriptions of what is needed. Charliecloud has implemented CDI in its runtime ch-run(1), bind-mounting requested files and using an unprivileged tmpfs overlay (available since Linux 5.11 in February 2021) to avoid modifying the image. This is a considerably simpler and more maintainable way to make host resources available inside a container.

This talk will provide an overview of CDI, adaptations of the standard to our fully unprivileged workflow, and our C implementation. We will also demonstrate the functionality for nVidia GPUs and HPC/Cray Slingshot interconnect.

LA-UR-25-22140
Speakers
RP

Reid Priedhorsky

Scientist, Los Alamos National Laboratory
I am a staff scientist at Los Alamos National Laboratory. Prior to Los Alamos, I was a research staff member at IBM Research. I hold a Ph.D. in computer science from the University of Minnesota and a B.A., also in computer science, from Macalester College.My work focuses on large-scale... Read More →
Wednesday May 7, 2025 2:55pm - 3:15pm CDT
Illinois River

2:55pm CDT

GAIA: A Software Deployment Strategy, Ordeals, Success, and General Applicability - Etienne Malaboeuf, CEA
Wednesday May 7, 2025 2:55pm - 3:15pm CDT
As High-Performance Computing (HPC) transitions from petascale to exascale, we have observed significant changes and new challenges in HPC software management. This presentation details the choices made for building and deploying HPC applications and tooling on the HPE Cray EX Adastra machine at the French National Computer Science Center for Higher Education (CINES). We present the GAIA project, a home-grown solution on top of Spack, allowing us to deploy software stacks for each of our partitions. We expose our requirements, the trade-offs we faced, the choices we made, and the strategy employed to address them. We explain how our users can leverage the Spack configuration provided by CINES to build their own software and the limitations of the approach. We finally take a look at the general applicability and open-sourcing to our site's user base.
Speakers
avatar for Etienne Malaboeuf

Etienne Malaboeuf

HPC Engineer, CINES/CEA
I focus on improving the performance of projects related to real-time and high-performance computing, while providing various forms of support to researchers using French supercomputers. I have worked on numerical simulation software in an HPC context, on supercomputers and on game... Read More →
Wednesday May 7, 2025 2:55pm - 3:15pm CDT
Salon E-G

2:55pm CDT

Trilinos CI Testing/Contribution Overview - Samuel E. Browne, Sandia National Laboratories
Wednesday May 7, 2025 2:55pm - 3:15pm CDT
The Trilinos project is one of many in the scientific software community that employs the concept of Continuous Integration and thusly has a testing process that ensures code functionality and quality as part of its contribution process. This talk is aimed at software developers and researchers interested in enhancing their understanding of Trilinos’ CI practices. We will examine the current processes that contributors must navigate, alongside emerging paradigms shaping the evolution of these practices. Key topics will include the implementation of containerized software testing environments, requirements for running tests on Trilinos compute resources, and objective standards that contributions must meet to guarantee software quality. Attendees will also gain insights into the challenges of providing a highly configurable set of software packages while maintaining user-friendly build processes. By the end of the session, participants will have a clearer understanding of how to effectively contribute to Trilinos, and the direction that DevSecOps efforts in Trilinos are heading in the future.
Speakers
SE

Samuel E. Browne

Principal R&D Computer Scientist, Sandia National Laboratories
Wednesday May 7, 2025 2:55pm - 3:15pm CDT
Mississippi River 1 & 2

3:15pm CDT

Coffee Break
Wednesday May 7, 2025 3:15pm - 3:40pm CDT
Wednesday May 7, 2025 3:15pm - 3:40pm CDT

3:40pm CDT

Closing Gaps in Spack for Software Application DevOps Infrastructure - Phil Sakievich, Sandia National Laboratories
Wednesday May 7, 2025 3:40pm - 3:50pm CDT
As the demand for efficient DevOps infrastructure for software applications continues to grow, key components such as build processes, build time tests, regression testing, and deployment mechanisms have become critical to successful project delivery. Spack is increasingly being adopted as an orchestration architecture for DevOps, offering a framework that can adapt to various project needs. However, while the overarching patterns among projects may be similar, the specific implementation details can differ significantly, leading to challenges in achieving seamless integration. In this talk, we will present the gaps identified by researchers at Sandia National Laboratories in Spack's current services and our ongoing efforts to address these challenges.Key topics will include advancements in binary caches, binary provenance, reporting test results, enhancing build performance, and improving the overall developer user experience. Attendees will gain valuable insights into successful initiatives that have effectively closed certain gaps, as well as ongoing issues that remain open for the community to tackle.
Speakers
avatar for Phil Sakievich

Phil Sakievich

Senior Computer Scientist R&D, Sandia National Laboratories
Phil comes from a high-performance computing and fluid mechanics background. He became involved with Spack during the ExaScale computing project and author of the Spack-Manager project. Phil is an active member of the Spack technical steering committee and currently leads several... Read More →
Wednesday May 7, 2025 3:40pm - 3:50pm CDT
Salon E-G

3:40pm CDT

A Mixed Formulation Vorticity-velocity Solver using AMReX Framework for Wind Farm Modeling - Balaji Muralidharan, Continuum Dynamics
Wednesday May 7, 2025 3:40pm - 4:00pm CDT
Wind energy plays a crucial role in meeting the electricity demands of the U.S.; however, high maintenance costs highlight the need for accurate predictions of unsteady loading caused by turbine layout and off-design wind conditions. Existing design tools often neglect fluid-structure interactions that drive costly fatigue loads, prompting the research community to leverage high-performance computing (HPC) solvers to study these effects. Unfortunately, such tools remain too complex and costly for industrial applications, particularly due to challenges in grid generation and setup. To address this, CDI is developing a Cartesian-based hybrid solver that integrates an incompressible vorticity-based far-field formulation with a compressible primitive variable solver in the near field. The framework is built on the AMReX library, enabling block-structured mesh refinement for efficient computation. This talk will explore both the computational and mathematical aspects of coupling these two solvers, highlighting advancements in predictive modeling for wind turbine aerodynamics.
Speakers
BM

Balaji Muralidharan

Continuum Dynamics
Wednesday May 7, 2025 3:40pm - 4:00pm CDT
Rock River 1 & 2

3:40pm CDT

The Role of Trilinos in 4C: Advancing Coupled Multiphysics Simulations - Matthias Mayr, University of the Bundeswehr Munich
Wednesday May 7, 2025 3:40pm - 4:00pm CDT
The 4C (Comprehensive Computational Community Code) multiphysics simulation framework has been developed to address complex physical phenomena across various scientific and engineering domains. From its inception, 4C has relied on the Trilinos project, an open-source software library for scalable numerical computations, as its backbone for sparse linear algebra and MPI-parallel computing. This integration enhances 4C's computational capabilities and, more importantly, allows the 4C developers to focus on their core research interest: the numerical modeling of multiphysics systems. The synergy between 4C's physics models and Trilinos' numerical solvers facilitates the simulation of coupled multiphysics systems with improved accuracy and performance. Over the years, this synergy has — in parts — evolved to a co-development of both software frameworks. This presentation will delve into the methodologies employed to incorporate Trilinos into the 4C framework, discuss software and development challenges, and showcase application case studies that demonstrate the practical benefits of this integration in simulating complex multiphysics systems such as fluid-solid interactions, contact mechanics or beam-solid interaction.
Speakers
avatar for Matthias Mayr

Matthias Mayr

Head of Data Science & Computing Lab, University of the Bundeswehr Munich
Wednesday May 7, 2025 3:40pm - 4:00pm CDT
Mississippi River 1 & 2

3:40pm CDT

Charliecloud Office Hours - Reid Priedhorsky, Los Alamos National Laboratory
Wednesday May 7, 2025 3:40pm - 5:00pm CDT
Members of the Charliecloud team will be available for office hours to listen to feedback/suggestions, answer questions, and/or help debug issues.

LA-UR-25-22140
Speakers
RP

Reid Priedhorsky

Scientist, Los Alamos National Laboratory
I am a staff scientist at Los Alamos National Laboratory. Prior to Los Alamos, I was a research staff member at IBM Research. I hold a Ph.D. in computer science from the University of Minnesota and a B.A., also in computer science, from Macalester College.My work focuses on large-scale... Read More →
Wednesday May 7, 2025 3:40pm - 5:00pm CDT
Illinois River

3:40pm CDT

Lightning Talks
Wednesday May 7, 2025 3:40pm - 5:00pm CDT
1. Experience Porting a Scientific Code from YAKL to Kokkos - James Foucar, Sandia National Labs (10 minutes)
The DoE climate code E3SM recently ported a medium sized scientific code, RRTMGP (computes radiative fluxes in planetary atmospheres), from a kernel launcher called YAKL to Kokkos. We'd like to share tips and pain points from this effort, particularly the struggle to get to performance parity with YAKL. We found that a 1:1 port (YAKL API is very similar to Kokkos) was not nearly sufficient to achieve good performance. The main issues were how to allocate temporary views and dealing with MDRangePolicy.

2. Benchmarking Lattice QCD Staggered Fermion Kernel Written in Kokkos - Simon Schlepphorst, Forschungszentrum Juelich GmbH (10 minutes)
Lattice quantum chromodynamics (QCD) is a numerical approach to studying the interactions of quarks and gluons, where the fundamental eqautions governing their interactions are discretized to a four dimension spacetime lattice. One of the most costly computations is the inversion of the lattice Dirac operator, a large sparse matrix. Calculating this inversion with iterative solvers leads to many applications of that operator. This study builds on previous work where we implemented the staggered fermion Dirac operator as a benchmark in Kokkos. We investigate the effects of the tiling size in combination with the use of a 4D MDRangePolicy and 7D Views.

3. Leveraging Liaisons in Your Network for Software Sustainability - Elaine M. Raybourn, Sandia National Laboratories (10 minutes)
Open source software project sustainability is a sociotechnical endeavor that often extends beyond the efforts of individual projects. HPSF and the Linux Foundation offer rich resources of expertise across communities in industry, academia, and agencies. Leveraging this collective knowledge and experience is vital to enhance project practices, especially in early identification of challenges and potential issues. This lightning talk explores the value of leveraging liaisons — key individuals who are actively participating in cross-team networks, to accelerate project sustainability. Liaisons can bridge gaps, share tacit knowledge and incentivize collaborative efforts across communities, go assist in breaking down silos. The value of leveraging liaisons was identified during the DOE Exascale Computing Project to foster strategic project alignment and outreach. Whether a small team, or a larger network of teams of teams, identifying liaisons early on can foster trust and transparency both within and across teams.

4. Vertex-CFD: A Multi-Physics Solver for Fusion Applications - Marc Olivier Delchini & Daniel Arndt, Oak Ridge National Laboratory (10 minutes)
In this talk we will introduce Vertex-CFD, a multiphysics solver that is being developed in response to needs by Oak Ridge National Laboratory (ORNL) to have accurate simulation software for use in modeling of a fusion blanket problem. Vertex-CFD is built upon Trilinos and Kokkos libraries for compatibility with CPU and GPU platforms. It is designed to generate high-fidelity solutions of multiphysics problems in complex geometries by leveraging state-of-the art computing methods and technologies. We will describe how we leverage Kokkos and Trilinos to solve the governing equations by employing a finite element method and high-order implicit temporal integrators.

5. Toucan: Revolutionizing Microstructure Prediction - Benjamin Stump, ORNL (10 minutes)
Going to describe my code, what it is doing (physically), what I need it to do computationally, how I achieved it using Kokkos and optimized it algorithmically.

6. Performance-Portable Spectral Ewald Summation with PyKokkos - Gabriel K Kosmacher, Oden Institute, The University of Texas at Austin (10 minutes)
We present a performance portable implementation of the Spectral Ewald method, employing shared memory and streaming parallelism to rapidly evaluate periodic two-body potentials in Stokes flow. The method splits dense particle evaluation into near-field and far-field components, where the near-field is singular and the far-field decays rapidly in Fourier space. Far-field interactions resemble a Nonuniform Fast Fourier Transform: source potentials are interpolated onto a uniform grid (p2g), an ndFFT is applied, Fourier potentials are scaled, an ndIFFT is applied, and the potentials are interpolated back (g2p). The p2g, g2p, and near-field (p2p) interactions use Kokkos hierarchical parallelism with scratch-pad memory and thread-vector range reductions.

7. Empowering NSM Supercomputers with Kokkos for Scalable HPC - Harsha Ugave & Samir Shaikh, Centre for Developement of Advanced Computing (C-DAC) (10 minutes)
Kokkos is transforming how high-performance applications run on National Supercomputing Mission (NSM) systems. With NSM deploying a mix of CPUs, GPUs, and other accelerators, ensuring software runs efficiently across all these platforms can be challenging. Kokkos simplifies this by providing a single, flexible programming model that adapts to different hardware without requiring major code changes. It supports multiple backends like CUDA, HIP, SYCL, and OpenMP, making it easier for developers to write performance-portable applications. For NSM’s large-scale supercomputers, Kokkos ensures better performance and scalability, allowing applications to make full use of processors, GPUs, and memory hierarchies. It also optimizes energy efficiency by improving memory access and reducing unnecessary data movement, helping to make supercomputing more sustainable. Since Kokkos is open-source and backed by an active community, it keeps up with emerging technologies, ensuring seamless adoption of next-generation NSM systems and preparing them for the future of exascale computing.

8. Real-Time Performance Characterization of the ADIOS2 Library When Kokkos Is Enabled - Ana Gainaru, Oak Ridge National Laboratory (10 minutes)
Modern performance analysis tools are increasingly capable of capturing a high volume of metrics at ever-finer granularity. This abundance of information presents an opportunity to move beyond post-mortem analysis and leverage data streaming for real-time performance monitoring and decision-making. By streaming performance data, applications can provide immediate feedback, enabling dynamic adjustments and optimizations during execution. Furthermore, this streamed data can be directed to individual scientist workstations, facilitating on-the-fly health checks and user-driven interventions to steer the application's behavior. We will demonstrate the practical application of these concepts within the ADIOS2 library, showcasing how data streaming enables detailed monitoring and analysis of an HPC application during large-scale runs.

9. Cabana: Particles, Structured Grids, and Extensions to Unstructured with Kokkos - Sam Reeve, ORNL (10 minutes)
We discuss updates to Cabana, a Kokkos+MPI library for building particle applications. Cabana was created through the U.S. Department of Energy Exascale Computing Project to enable particle simulation across methods on current and future exascale supercomputers. Cabana includes particle and structured grid parallelism, data structures, algorithms, communication, and interfaces to additional libraries, all extending and working alongside Kokkos. We focus in particular on recent efforts to integrate Cabana particles within Trilinos unstructured grids for broader support of scientific applications. We will highlight further recent Cabana development, performance and portability, and application-level demonstrations.
Speakers
avatar for Elaine M. Raybourn

Elaine M. Raybourn

Principal Member of the Technical Staff, Sandia National Laboratories
Elaine M. Raybourn is a social scientist at Sandia National Laboratories. She has worked in the UK (British Telecom), Germany (Fraunhofer FIT), and France (INRIA) as a Fellow of the European Research Consortium in Informatics and Mathematics (ERCIM). She supports the DOE Office of... Read More →
avatar for Daniel Arndt

Daniel Arndt

Large Scale Computational Scientist, Oak Ridge National Laboratory
Daniel Arndt is a computational scientist at Oak Ridge National Laboratory. He is also a mathematician by training specializing on finite element simulations. His research focuses on supporting new backends in Kokkos.
avatar for Samir Shaikh

Samir Shaikh

Scientist, Centre for Developement of Advanced Computing (C-DAC)
Samir Shaikh is an HPC specialist at C-DAC, Pune, optimizing large-scale workloads, parallel computing, and system architecture. As a Scientist C, he enhances HPC performance for AI/ML, scientific computing, and NSM supercomputers. An IIT Guwahati M.Tech graduate, he has contributed... Read More →
avatar for Ana Gainaru

Ana Gainaru

Computer Scientist, Oak Ridge National Laboratory
Ana Gainaru is a computer scientist in the CSM division at Oak Ridge National Laboratory, working on performance optimization for large scale scientific applications and on profiling, managing, and analyzing large-scale data. She received her PhD from the University of Illinois at... Read More →
avatar for Benjamin Stump

Benjamin Stump

Technical Staff, ORNL
Benjamin Stump works at the Oak Ridge National Laboratory's Manufacturing Demonstration Facility on Additive Manufacturing problems.
avatar for Gabriel K Kosmacher

Gabriel K Kosmacher

Graduate Student, Oden Institute, The University of Texas at Austin
Gabriel is a PhD student at the Oden Institute for Computational Engineering & Sciences, where he is advised by George Biros. His research interests lie at the intersection of numerical analysis and scientific computing and is particularly interested in fast numerical methods for... Read More →
avatar for Harsha Ugave

Harsha Ugave

HPC Engineer, Centre for Development of Advanced Computing (C-DAC), India
Harsha Ugave is an HPC Engineer at C-DAC Pune, specializing in performance portability, parallel computing, and system optimization. She plays a key role in deploying and tuning HPC applications under the National Supercomputing Mission (NSM). Her work ensures efficient execution... Read More →
avatar for James Foucar

James Foucar

Software Engineer, Sandia National Labs
I've been a software developer for Sandia for nearly 20 years. For the last 10 yeas, I've been doing software-focussed tasks for E3SM (DoE climate model).
avatar for Marc Olivier Delchini

Marc Olivier Delchini

CFD developer and analyst, Oak Ridge National Laboratory
CFD analyst and developer at Oak Ridge National Laboratory for 10 years. Obtained his PhD in nuclear engineering from Texas A&M University.
avatar for Sam Reeve

Sam Reeve

Staff Scientist, ORNL
Sam Reeve is a staff scientist at ORNL, working at the intersection of materials and computational science. Current focuses include performance portability and software development for physics applications and simulation of mesoscale material phenomena. He leads the development of... Read More →
avatar for Simon Schlepphorst

Simon Schlepphorst

Research Software Engineer, Forschungszentrum Juelich GmbH
After graduating with a Master's degree in physics from the University of Bonn, Simon became a Research Software Engineer at the Juelich Supercomputing Centre developing Lattice QCD codes for current and upcoming accelerators.
Wednesday May 7, 2025 3:40pm - 5:00pm CDT
Chicago River Ballroom

3:50pm CDT

Development of Complex Software Stacks with Spack - Cedric Chevalier, CEA
Wednesday May 7, 2025 3:50pm - 4:00pm CDT
In this presentation, we will describe how we manage to develop multi-physics applications with a software stack deployed with Spack.

We will describe how we have designed a workflow, first using Spack features for development, such as "setup," "dev-build," and "develop." And how we ended up creating a Spack plugin to generate custom CMake presets.

CMake presets are a portable way to set up CMake configurations. They can describe several configurations. And they can be exploited by different tools, from the command line to IDE.
Generating these presets from Spack concretizations allows users to exploit their classical environment, benefiting both from a correct installation of their dependencies as well as advanced features of the IDE that do not need to integrate explicitly with Spack.

We will present our journey between the different spack solutions, the development of a CMake "preload cache"-based answer, and we will illustrate use cases of why we ultimately switched to CMake presets.
Speakers
avatar for Cedric Chevalier

Cedric Chevalier

Research Scientist, CEA
Cédric Chevalier is a research scientist at CEA in France. He is interested in developing libraries for HPC simulation codes, particularly in Linear Algebra and Mesh/Graph partitioning. His work at CEA is led by providing practical ways to exploit newer hardware, use new programming... Read More →
Wednesday May 7, 2025 3:50pm - 4:00pm CDT
Salon E-G

4:00pm CDT

Feedback on Using Spack to Deploy a Development Environment for the Gyselalibxx Library - Thomas Padioleau, CEA
Wednesday May 7, 2025 4:00pm - 4:10pm CDT
In this presentation I will present the feedback of packaging and deploying the dependencies of the open source library `gyselalibxx` using Spack on a local cluster and on the national supercomputer Adastra. I will explain the challenges faced to deploy the library and why we chose Spack.
Speakers
avatar for Thomas Padioleau

Thomas Padioleau

Engineer-Researcher, CEA
Dr. Thomas Padioleau is a CEA Engineer-Researcher at Maison de la Simulation. He leads the DDC project and also works on Voice++.
Wednesday May 7, 2025 4:00pm - 4:10pm CDT
Salon E-G

4:00pm CDT

ExaWind: Leveraging AMReX for High Fidelity Wind Farm Simulations - Marc Henry de Frahan, National Renewable Energy Laboratory
Wednesday May 7, 2025 4:00pm - 4:20pm CDT
AMR-Wind is a high-fidelity computational-fluid-dynamics solver for simulating wind farm flow physics. The solver enables predictive simulations of the atmospheric boundary layer and wind plants by leveraging a block-structured, adaptive-mesh, incompressible-flow solver. AMR-Wind is designed for scalability on high performance computing systems, with an emphasis on performance portability for graphical processing units (GPUs). These flow solver capabilities and performance characteristics were enabled using the AMReX library. In this talk, we present AMR-Wind, its capabilities and performance characteristics. We detail the numerical implementation, verification and validations efforts, as well as demonstrate AMR-Wind for large eddy simulations of wind farm physics. A demonstration simulation is presented for a 12-turbine wind farm operating in a turbulent atmospheric boundary layer with realistic wake interactions. We also discuss AMR-Wind in the wider context of the ExaWind suite of codes, including (1) its integration as a background solver to Nalu-Wind with overset methods to perform geometry-resolved simulations of wind turbines and (2) its use coupled to a mesoscale weather simulation code, ERF.
Speakers
MH

Marc Henry de Frahan

Computational Scientist, National Renewable Energy Laboratory
Wednesday May 7, 2025 4:00pm - 4:20pm CDT
Rock River 1 & 2

4:00pm CDT

Automated Preconditioner Design in the Trilinos/Teko Package - Malachi Phillips, Sandia National Laboratories
Wednesday May 7, 2025 4:00pm - 4:20pm CDT
The solution of multiphysics problems depends on the construction of robust preconditioners which significantly influence the convergence rate and computational performance of linear solvers. We employ physics-based block preconditioning strategies from the Trilinos/Teko package developed by Cyr, Shadid, and Tuminaro [SIAM Journal on Scientific Computing, 38(5):S307–S331, 2016]. Teko preconditioner parameter selection, however, remains a difficult task for users. This talk presents an automated approach for the selection of multiphysics preconditioner parameters in the Trilinos/Teko package. We discuss design principles for preconditioners, emphasizing the importance of physics blocking, subblock ordering, and subblock solver choices based on equation types and coupling strengths. Finally, we present an extension to the Trilinos/Teko package in the form of a user-friendly software tool for automatic multiphysics preconditioner design with minimal user input.
Speakers
MP

Malachi Phillips

Postdoctoral Appointee, Sandia National Laboratories
Wednesday May 7, 2025 4:00pm - 4:20pm CDT
Mississippi River 1 & 2

4:10pm CDT

Spack at the Linac Coherent Light Source: Progress, Success, and Challenges - Valerio Mariani, SLAC National Accelerator Laboratory
Wednesday May 7, 2025 4:10pm - 4:20pm CDT
The Linac Coherent Light Source (LCLS) at the SLAC National Accelerator Laboratory takes X-ray snapshots of atoms and molecules in action, revealing fundamental processes in materials, technology and living things. Advanced data and computing systems are playing an increasingly vital role in LCLS’s operation: data acquisition software developed at LCLS allows the collection of very high throughput data, and makes sure that all the collected data is stored to disk. Additionally, LCLS provides scientists with sophisticated data analysis frameworks and helps them run their own bleeding-edge scientific analysis tools. With new detectors and analysis techniques being introduced all the time, new software is constantly being developed and maintaining compatibility with old hardware while supporting new frameworks and programs is an everyday challenge. Spack seems to offer a solid foundation for the software development process and is starting to be used at LCLS. This talk will discuss how Spack is making the development process easier and what challenges the LCLS Data Analysis group is still facing in making Spack a useful tool for everyday work.
Speakers
avatar for Valerio Mariani

Valerio Mariani

LCLS Data Analysis Department Head, SLAC National Accelerator Laboratory
Dr. Valerio Mariani received his PhD in Biophysics at the University of Basel in Switzerland, and is currently Data Analysis Department Head at Linac Coherent Light Source (LCLS), part of the SLAC National Accelerator Laboratory. He has in the past collaborated with world-scale user... Read More →
Wednesday May 7, 2025 4:10pm - 4:20pm CDT
Salon E-G

4:20pm CDT

Towards a Zero-Install Programming Environment - Mike Kiernan & Victor Gamayunov, Microsoft
Wednesday May 7, 2025 4:20pm - 4:30pm CDT
So, we got bored with installing stuff. Our project aims to accelerate time to results, improve reproducibility, and reduce reliance on proprietary programming environments and manual installs. Built on Spack, our tooling enables the rapid deployment of versioned programming environments to globally distributed HPC clusters, ensuring consistency across clusters and regions. This talk will present our solution and the problems it solves for us, discuss its benefits for HPC productivity, and invite community feedback on its broader applicability.
Speakers
avatar for Mike Kiernan

Mike Kiernan

Principle Technical Program Manager, Microsoft
Mike Kiernan leads the Public Sector HPC and AI Customer Solutions and Innovation Team at Microsoft, and is based in Cambridge, UK. Joining Mike are Victor Gamayunov and Trevor Cooper-Chadwick, both Technical Program Managers in Mike's team, also based in the UK.
avatar for Victor Gamayunov

Victor Gamayunov

Senior Technical Program Manager, Microsoft
Victor is a Senior TPM for Azure HPC and AI Customer Solutions and Innovation team at Microsoft. Prior to that he spent 2 decades at Intel in HPC and application engineering.
Wednesday May 7, 2025 4:20pm - 4:30pm CDT
Salon E-G

4:20pm CDT

Load Balancing in AMReX: Advancements and Discussing Future Goals - Kevin Gott, Lawrence Berkeley National Laboratory - NERSC
Wednesday May 7, 2025 4:20pm - 4:40pm CDT
Load balancing is an evergreen topic in HPC mesh-and-particle codes like AMReX. Appropriate, fast and effective load balancing strategies are critical to ensure AMReX codes are capable of making the best use of available computational resources and maximizing the size and type of simulations that can be performed. I have begun investigating potential short and long term advancements in load balancing that may be of use to the AMReX community. In this talk, I will give an overview of the current load balancing options in AMReX plus an algorithmic investigation performed last summer by NERSC summer interns which found potential improvements in our current Knapsack and SFC algorithms. Then, I will present an overview of possible next steps, long-term advancements, and where I believe these advances would be most helpful. Hopefully, this will generate a discussion around which investigations would be most helpful to the AMReX community to determine the best targets for this summer’s interns and beyond.
Speakers
KG

Kevin Gott

NERSC Computational Engineer, Lawrence Berkeley National Laboratory - NERSC
Wednesday May 7, 2025 4:20pm - 4:40pm CDT
Rock River 1 & 2

4:20pm CDT

Keeping Trilinos Running Performantly Everywhere Every Night - Chris Siefert, Sandia National Laboratories
Wednesday May 7, 2025 4:20pm - 4:40pm CDT
The Trilinos scientific software library is a key enabling technology for application codes in a variety of physics and engineering areas across the US Department of Energy and beyond. This presentation will "pull back the curtain" and describe the process by which the Trilinos Tpetra/Performance team (a) performs nightly testing across five DOE laboratories, (b) how we identify and remedy performance regressions, (c) how we make the validated build scripts available to Trilinos developers, and (d) how we interface with early users and vendors to ensure that vendor library updates work as advertised.
Speakers
CS

Chris Siefert

R&D Staff, Sandia National Laboratories
Wednesday May 7, 2025 4:20pm - 4:40pm CDT
Mississippi River 1 & 2

4:30pm CDT

Deploying Software on Frontier with NCCS Software Provisioning (NSP) - Fernando Posada, Oak Ridge National Laboratory
Wednesday May 7, 2025 4:30pm - 4:40pm CDT
Managing software stacks on high-performance computing (HPC) systems is a complex challenge, particularly when working within vendor-provided programming environments. To address this, the National Center for Computational Sciences (NCCS) at Oak Ridge National Laboratory has developed NCCS Software Provisioning (NSP), a unified framework for deploying and managing software stacks while monitoring their usage.

NSP leverages Spack and Ansible to automate the deployment of software environments. Using templated configurations, it streamlines Spack-based installations while also managing non-Spack software through custom Ansible roles. Additionally, NSP enhances LMOD-based module environments by incorporating hooks and Spack module projections, enabling a dynamic and responsive software layout that adapts seamlessly to changes in the programming environment.

This presentation will discuss the motivation behind NSP, its strategies for managing software complexity in vendor-provided environments, and its implementation on Frontier, the world’s fastest supercomputer for open science.
Speakers
avatar for Fernando Posada

Fernando Posada

Group Lead, System Acceptance and User Environment, Oak Ridge National Laboratory
Fernando Posada is the group leader of the Systems Acceptance and User Environment group at Oak Ridge National Laboratory. He earned his Ph.D. in Chemistry from the Universidad Nacional de Colombia. During his Ph.D., Fernando developed an interest in High-Performance Computing and... Read More →
Wednesday May 7, 2025 4:30pm - 4:40pm CDT
Salon E-G

4:40pm CDT

Organizational Approach to Spack Engagement: A Case Study - Phil Sakievich, Sandia National Laboratories
Wednesday May 7, 2025 4:40pm - 4:50pm CDT
The rapid growth in features, capabilities, and adoption of Spack has transformed it into a vital tool for managing software dependencies in scientific computing. However, one of the significant challenges faced by large organizations is the effective communication and information sharing across diverse teams. This talk will present a case study from Sandia National Laboratories, detailing the evolution and impact of strategies implemented to address these communication hurdles. We will explore the formation of a self-organized working group dedicated to fostering collaboration and knowledge exchange around Spack, as well as a recently funded initiative aimed at enhancing Spack collaboration and development efforts. By examining the successes and lessons learned from these organizational approaches, we aim to provide insights into best practices for engaging teams in the adoption of Spack, ultimately promoting a more cohesive and efficient development environment. Attendees will leave with a deeper understanding of how structured engagement strategies can facilitate the integration of Spack within large organizations and enhance collaborative software development efforts.
Speakers
avatar for Phil Sakievich

Phil Sakievich

Senior Computer Scientist R&D, Sandia National Laboratories
Phil comes from a high-performance computing and fluid mechanics background. He became involved with Spack during the ExaScale computing project and author of the Spack-Manager project. Phil is an active member of the Spack technical steering committee and currently leads several... Read More →
Wednesday May 7, 2025 4:40pm - 4:50pm CDT
Salon E-G

4:40pm CDT

Code Coupling of PeleLM and IRL for Liquid Fuel Combustion Simulation - Hyoungwoo Kim, Korea Advanced Institute of Science and Technology
Wednesday May 7, 2025 4:40pm - 5:00pm CDT
This research develops a framework to simulate multiphase flows with surface tension. Two open-sources are coupled to achieve this goal: interface reconstruction library (IRL), a library of the volume of fluid (VOF) scheme, and PeleLM, a reacting Navier-Stokes equation solver. In addition to the coupling, surface tension is implemented using the continuum surface model (CSF) along with the improved height function technique. The coupling produces spurious errors in the volume fraction field. The spurious error is corrected by geometric consideration, which improves the numerical stability and accuracy. With the developed framework, multiple validation simulations are conducted. The validation simulations are (i) translations and rotations of Zalesak’s disk, (ii) three-dimensional deformations of a sphere droplet, (iii) a stationary circular droplet with surface tension, and (iv) an oscillating elliptical droplet. Quantitative comparisons show that the shape errors of Zalesak’s disk and 3D deformation case show comparable errors with other solvers. The pressure inside the stationary droplet is well maintained within 3.6 % compared to the theoretical value. For the oscillating elliptical droplet case, the oscillation period is within 6.7% compared to the linear theory.
Speakers
HK

Hyoungwoo Kim

Ph.D. Student, Korea Advanced Institute of Science and Technology
Wednesday May 7, 2025 4:40pm - 5:00pm CDT
Rock River 1 & 2

4:40pm CDT

PyTrilinos2: Using Trilinos from Python - Christian Glusa, Sandia National Laboratories
Wednesday May 7, 2025 4:40pm - 5:00pm CDT
PyTrilinos2 is an auto-generated Python interface for several Trilinos packages. I will explain how additional C++ functionality can be exposed to Python and demonstrate existing solver capabilities.
Speakers
CG

Christian Glusa

Principal Member of the Technical Staff, Sandia National Laboratories
Wednesday May 7, 2025 4:40pm - 5:00pm CDT
Mississippi River 1 & 2

4:50pm CDT

Dynamic Resource Allocation for Continuous Integration Build Pipelines - Caetano Melone, Lawrence Livermore National Laboratory
Wednesday May 7, 2025 4:50pm - 5:00pm CDT
Spack team members manage a complex continuous integration (CI) infrastructure to compile and test software distributed via a centralized build cache. The workflows are executed in virtualized cloud environments, an approach that offers considerable flexibility in resource provisioning, resulting in quick feedback on pull requests submitted by the Spack community. However, fixed resource allocations for build jobs lead to job failures, slowdowns, and suboptimal resource usage.

spack-gantry aims to address these challenges by providing resource allocation predictions to build pipelines. We use historical build data to forecast future utilization and set resource limits. We will present strategies used to find optimal predictions, as well as experiments to create an online prediction service modeled after a genetic algorithm. The results of this work include improvements in usage efficiency and cost-per-job metrics.

Because spack-gantry involves modifications to the current CI system, we will discuss ramifications for users and developers and solicit feedback about how to make the infrastructure more user-friendly, reliable, and functional for the Spack community.
Speakers
avatar for Caetano Melone

Caetano Melone

Software Developer, Lawrence Livermore National Laboratory
Caetano Melone is a software developer at Lawrence Livermore National Laboratory working on open-source tools for HPC developer productivity.
Wednesday May 7, 2025 4:50pm - 5:00pm CDT
Salon E-G
 
Thursday, May 8
 

8:30am CDT

Registration & Badge Pick-Up
Thursday May 8, 2025 8:30am - 4:00pm CDT
Thursday May 8, 2025 8:30am - 4:00pm CDT
Ballroom Meeting Foyer

9:00am CDT

Breaking Charliecloud News - Reid Priedhorsky, Los Alamos National Laboratory
Thursday May 8, 2025 9:00am - 9:20am CDT
This session will cover late-breaking developments in Charliecloud, such as recent/upcoming new features, notable bugs, and requests for feedback. The specific agenda is TBD.

LA-UR-25-22140
Speakers
RP

Reid Priedhorsky

Scientist, Los Alamos National Laboratory
I am a staff scientist at Los Alamos National Laboratory. Prior to Los Alamos, I was a research staff member at IBM Research. I hold a Ph.D. in computer science from the University of Minnesota and a B.A., also in computer science, from Macalester College.My work focuses on large-scale... Read More →
Thursday May 8, 2025 9:00am - 9:20am CDT
Illinois River

9:00am CDT

Creating Reproducible Performance Optimization Pipelines with Spack and Ramble - Doug Jacobsen, Google LLC
Thursday May 8, 2025 9:00am - 9:20am CDT
In this talk, I will present about how we use Spack at Google both with customers and internally. This portion will include an overview of our public cache, along with our use of Spack for benchmarking.

Additionally, this talk will discuss some open-sourced tooling we have developed, Ramble, which was built on top of Spack's infrastructure. This talk will discuss how we use Ramble and Spack together to construct complex end-to-end performance benchmarking studies for HPC and AI applications, and how this process can use these tools to create reproducible experiments that are shareable with external users.
Speakers
DJ

Doug Jacobsen

HPC Software Engineer, Google Cloud
Thursday May 8, 2025 9:00am - 9:20am CDT
Salon E-G

9:00am CDT

Broader Kokkos Ecosystem
Thursday May 8, 2025 9:00am - 10:20am CDT
1. kokkos-fft Updates – Yuuichi Asahi, CEA (10 minutes)
kokkos-fft implements local interfaces between Kokkos and de facto standard FFT libraries, including fftw, cufft, hipfft (rocfft), and oneMKL. We are inclined to implement the numpy.fft-like interfaces adapted for Kokkos. A key concept is that "As easy as numpy, as fast as vendor libraries". In the talk, we will introduce the basic APIs and typical use cases. We will also present future development plans.

2. Fortran Porting Wish List for Kokkos – Matthew Norman, Oak Ridge National Laboratory (10 minutes)
This presentation covers the beginnings of the Yet Another Kernel Launcher (YAKL) C++ portability library, its evolution alongside Kokkos, the use of Kokkos in its current form, and remaining issues before it can be retired in lieu of using Kokkos instead. The primary outstanding issues are the inclusion of arbitrary lower bounds for Fortran-like View behavior and the ability to use an underlying pool allocator for Views for cheap frequent device allocation and deallocation so that Views can be locally created and destroyed only where needed rather than existing for the global lifetime of simulations. This may improve readability and reduce the memory high water mark in simulations. A few performance related issues will be covered as well, mainly limited to MDRangePolicy and parallel_for register usage.

3. Custom Layout and Tiling for Multi-Dimensional Data – Cedric Chevalier & Gabriel Dos Santos, CEA (10 minutes)
Performance optimizations for exascale HPC applications primarily rely on fine-tuning implementations, requiring comprehensive knowledge of heterogeneous hardware architectures that domain experts often lack. One of Kokkos' biggest successes is tying the memory layout of multi-dimensional arrays to the execution backend. It allows the exploitation of coalescence or cache, depending on the hardware. Here, we propose to go further and design custom tiled layouts that are generic for C++23's std::mdspan. Instead of running tile algorithms on flat data, like Kokkos' mdrange, we want to explore how running flat algorithms on tiled data performs. On CPU, the first experimental results with std::mdspan on a naive dense matrix multiplication demonstrate that, by replacing standard layouts with our proposed solution, we achieve an average speedup of over 2.2x, with peak performance improvements of up to 7.8x. Then, we will discuss how external indexing can improve efficiency. We will present how to exploit it with Kokkos' mdrange algorithm, and how it can behave on GPU.

4. Runtime Auto-Tuning for Kokkos Applications with APEX – Kevin Huck, University of Oregon (10 minutes)
Traditional GPU programming with libraries like CUDA or HIP requires tuning parameters exposed to the user, for example block sizes or number of teams. Kokkos also exposes portable parameters to the Kokkos user. How can Kokkos application programmers easily tune these Kokkos parameters for their application’s deployment when using any given Kokkos backend, without incurring large overheads? In particular, how do we ensure the tuning itself is portable across platforms? We propose using online, i.e., runtime, autotuning, utilizing the APEX Kokkos Tools connector to tune exposed parameters. Specifically, we discuss the Kokkos Tools Tuning Interface, tuning contexts, variable definition, the APEX runtime auto-tuning library utilizing Kokkos Tools, and distributed Kokkos auto-tuning. Applying our auto-tuning approaches to Kokkos sample kernels on Perlmutter and Frontier, we have obtained promising performance results. These results suggest Kokkos online auto-tuning is beneficial for production applications, and we invite Kokkos users to try these features and for Kokkos developers to contribute.

5. Unifying the HPC Ecosystem with std::execution – Mikael Simberg, Swiss National Supercomputing Centre (20 minutes)
Asynchronous programming models are becoming increasingly essential for fully leveraging modern hardware. In the C++ ecosystem, projects typically provide ad-hoc and varying interfaces, making interoperability difficult. Recently approved for C++26, the std::execution library promises to unify the ecosystem by providing a standard, composable interface for asynchronous operations. This talk briefly introduces the motivation and design principles of std::execution, and shares our experiences using it prior to standardization at CSCS in various projects, including Kokkos, HPX, and more. We'll discuss challenges, successes, and opportunities encountered while adopting std::execution.

6. PyKokkos: Performance Portability for Python Developers – Milos Gligoric, The University of Texas at Austin (20 minutes)
Kokkos is a programming model for writing performance portable applications for all major high performance computing platforms. It provides abstractions for data management and common parallel operations, allowing developers to write portable high performance code with minimal knowledge of architecture-specific details. Kokkos is implemented as a heavily-templated C++ library. However, C++ is not ideal for rapid prototyping and quick algorithmic exploration. An increasing number of developers use Python for scientific computing, machine learning, and data analytics. In this talk, I will present a new Python framework, PyKokkos, for writing performance portable applications entirely in Python. PyKokkos provides Kokkos-like abstractions that are easier to use and more concise than the C++ interface. We implemented PyKokkos by building a translator from a subset of Python to C++ Kokkos and bridging necessary function calls via automatically generated Python bindings. I will also cover our recent work on automatic kernel fusion with the goal to optimize PyKokkos applications. The talk will also cover our experience on developing PyKokkos, its current limitations, and future plans.
Speakers
avatar for Kevin Huck

Kevin Huck

Senior Research Associate, University of Oregon
Kevin Huck is a Senior Research Associate in the Oregon Advanced Computing Institute for Science and Society (OACISS) at the University of Oregon. He is interested in the unique problems of performance analysis of large HPC applications as well as automated methods for diagnosing... Read More →
avatar for Cedric Chevalier

Cedric Chevalier

Research Scientist, CEA
Cédric Chevalier is a research scientist at CEA in France. He is interested in developing libraries for HPC simulation codes, particularly in Linear Algebra and Mesh/Graph partitioning. His work at CEA is led by providing practical ways to exploit newer hardware, use new programming... Read More →
avatar for Gabriel Dos Santos

Gabriel Dos Santos

PhD Student, CEA
PhD student on the management of data structures representations in heterogeneous architecture for exascale-class HPC workloads, with a strong background in performance optimization, CPU microarchitectures and vectorization.
avatar for Matthew Norman

Matthew Norman

Climate Scientist, Oak Ridge National Laboratory
Matt Norman leads the Advanced Computing for Life Sciences and Engineering group in the Oak Ridge Leadership Computing Facility (OLCF). He works with weather and climate simulation, urban and wind turbine simulation, PDE discretizations for the Navier-Stokes Equations, GPU acceleration... Read More →
avatar for Mikael Simberg

Mikael Simberg

HPC Application Engineer, Swiss National Supercomputing Centre
Mikael Simberg holds a master's degree in operations research and computer science from Aalto University in Finland. He joined the Swiss National Supercomputing Centre in 2017 where he works as a software developer helping scientific projects make the best use of modern hardware through... Read More →
avatar for Milos Gligoric

Milos Gligoric

Associate Professor, The University of Texas at Austin
Milos Gligoric is an Associate Professor in Electrical and Computer Engineering at The University of Texas at Austin where he holds the Archie W. Straiton Endowed Faculty Fellowship in Engineering. His research interests are in software engineering, especially in designing techniques... Read More →
avatar for Yuuichi Asahi

Yuuichi Asahi

Research Scientist, CEA
His recent interests are HPC and AI with NVIDIA, AMD and Intel GPUs. He has a rich experience in GPU programming models including CUDA, HIP, SYCL, Kokkos, OpenMP, OpenACC, thrust, stdpar, and senders/receivers. For exascale computing, he is highly interested in improving performance... Read More →
Thursday May 8, 2025 9:00am - 10:20am CDT
Salon A-C

9:20am CDT

Democratizing Access to Optimized HPC Software Through Build Caches - Stephen Sachs & Heidi Poxon, AWS
Thursday May 8, 2025 9:20am - 9:40am CDT
This talk presents our implementation of a build cache of pre-optimized HPC applications using Spack. By implementing architecture-specific enhancements for both x86 and ARM platforms during the build process, we created a set of stacks of optimized software accessible through build caches. Using application builds from the cache, users can reduce compute resource requirements without requiring specialized tuning expertise.
We'll demonstrate how teams can quickly deploy HPC clusters using these stacks and discuss the substantial advantages compared to building from source. We'll present comparisons to traditional builds, showing significant time-to-solution improvements. This work represents a step toward enabling the HPC community to focus on scientific discovery rather than software compilation and tuning.
Speakers
HP

Heidi Poxon

Principal Member of Technical Staff, AWS
avatar for Stephen Sachs

Stephen Sachs

Principal HPC Application Engineer, AWS
Dr. Stephen Sachs is a Principal HPC Application Engineer on the HPC Performance Engineering team at AWS. With over 15 years of domain specific experience, he specializes in application optimization and cloud-based HPC solutions. Previously, he worked as an Application Analyst at... Read More →
Thursday May 8, 2025 9:20am - 9:40am CDT
Salon E-G

9:20am CDT

Deploying AI Chatbot Assistants with Charliecloud - Jemma Stachelek, Los Alamos National Laboratory
Thursday May 8, 2025 9:20am - 10:00am CDT
Additional Authors: Tolulope Olatunbosun, Phil Romero & Mike Mason, Los Alamos National Laboratory

Retrieval Augmented Generation (RAG) systems improve the response relevance of LLMs (Large Language Models) by limiting the context to a document corpus. RAG systems have seen broad deployment as document summarization engines and AI chatbots. However, deploying these systems often assumes a privileged and “cloudy” environment with multi-container orchestration (i.e. docker compose) and unfettered internet access to pull resources (e.g. software, data, and models) on-the-fly. As an alternative, we leveraged Charliecloud’s NVIDIA GPU support capabilities to deploy a RAG chatbot in an unprivileged HPC environment where resources are pre-staged. We demonstrate the deployment of AI Chatbots using Charliecloud on a variety of hardware and software versioning.

LA-UR-25-21968
Speakers
JS

Jemma Stachelek

Scientist, Los Alamos National Laboratory
Thursday May 8, 2025 9:20am - 10:00am CDT
Illinois River

9:40am CDT

Spack, Containers, CMake: The Good, The Bad & The Ugly in the CI & Distribution of the PDI Library - Julien Bigot, CEA
Thursday May 8, 2025 9:40am - 10:00am CDT
The PDI data interface is a library that supports loose coupling of simulation codes with data handling libraries: the simulation code is annotated in a library-agnostic way, and data management through external libraries is described in a YAML "data handling specification tree". Access to each data handling tool or library (HDF5, NetCDF, Python, compiled functions, Dask/Deisa, libjson, MPI, etc.) is provided through a dedicated plugin. Testing, packaging and distributing PDI is a complex problem as each plugin comes with its own dependencies, some of wich are typically not provided by supercomputer administrators. In the last five years, we have managed to devise solutions to test & validate, package & distribute the library and its plugins, largely based on spack.

In this talk, we will describe PDI, the specific problems we encounter and how we tackled them with a mix of cmake, spack, and containers. We specifically focus on the creation of a large family of spack-based container images used as test environment of the library, and on the efforts deployed to ensure easy installation on the wide range of supercomputers our downstream application rely on.
Speakers
avatar for Julien Bigot

Julien Bigot

Permanent Research Scientist, CEA
Julien is a permanent computer scientist at Maison de la Simulation at CEA. He leads the Science of Computing team. His research focuses on programming models for high-performance computing. He is especially interested in the question of separation of concerns between the simulated... Read More →
Thursday May 8, 2025 9:40am - 10:00am CDT
Salon E-G

10:00am CDT

Using Charliecloud to Wrap HTCondor Worker Nodes - Oliver Freyermuth, University of Bonn (Germany)
Thursday May 8, 2025 10:00am - 10:20am CDT
This talk will present a setup using Charliecloud to spawn virtual HTCondor compute nodes inside of jobs submitted to a SLURM cluster. The actual containers are distributed via CernVM-FS and mounted unprivilegedly at the HPC site using cvmfsexec. The spawned HTCondor nodes integrate into a larger overlay batch system to run High-Throughput compute jobs from the Worldwide LHC Computing Grid community.

Charliecloud allows to make this setup very portable with its lightweight design, minimal system dependencies and simplicity of use. CernVM-FS which is optimized for distribution of large numbers of small files of which only few might be accessed proves an ideal fit for distribution of directory-format container images. In combination with HTCondor which focuses on optimizing the total throughput and easily handles large numbers of jobs, compute resources can be used opportunistically by integrating them fully unprivilegedly into an overlay batch system. The workloads themselves can again use unprivileged containers to enable the use of user-defined software stacks.
Speakers
OF

Oliver Freyermuth

Research Scientist for IT Operations and High Throughput Computing, University of Bonn (Germany)
Thursday May 8, 2025 10:00am - 10:20am CDT
Illinois River

10:00am CDT

Spack Deployment Story at LBNL/UC Berkeley - Abhiram Chintangal, Lawrence Berkeley National Lab
Thursday May 8, 2025 10:00am - 10:20am CDT
The High-Performance Computing Services group at Lawrence Berkeley National Laboratory delivers extensive computing resources to Berkeley Lab and the University of California at Berkeley, supporting approximately 4,000 users and nearly 600 research projects across diverse scientific disciplines.

Over the past year and a half, we have modernized our primarily manual software build process using Spack, enabling us to meet the growing application and workflow demands of the HPC software stack.

This presentation will highlight how we leverage Spack’s features—such as environments, views, and module sets—to meet our specific needs and requirements. Additionally, we will discuss how, over the past year, our Spack pipeline, integrated with Reframe (a testing framework), has enabled our larger infrastructure team to efficiently plan and execute large-scale OS migrations across multiple scientific clusters in a short timeframe.
Speakers
avatar for Abhiram Chintangal

Abhiram Chintangal

Site Reliability Engineer, Lawrence Berkeley National Lab
Abhiram is a Systems Engineer with over nine years of experience specializing in meeting the computational and IT demands of scientific labs. He has a deep understanding of the complexities of software in the data-driven landscape of modern science and recognizes its critical role... Read More →
Thursday May 8, 2025 10:00am - 10:20am CDT
Salon E-G

10:20am CDT

Coffee Break
Thursday May 8, 2025 10:20am - 10:45am CDT
Thursday May 8, 2025 10:20am - 10:45am CDT

10:45am CDT

Lessons Learned from Developing and Shipping Advanced Scientific Compressors with Spack - Robert Underwood, Argonne National Laboratory
Thursday May 8, 2025 10:45am - 11:05am CDT
Modern scientific applications increasingly produce extremely large volumes of data while the scalability of I/O systems has not increased at the same rate. Lossy data compression has helped many applications address these limitations, but to meet the needs of the most demanding applications, specialized compression pipelines are needed. The FZ project helps users and compression scientists collaborate to meet the I/O needs of exascale applications by making it easier to implement custom compression tools and integrate them with applications. However, to fulfill the complex needs of this diverse ecosystem of software and systems, the FZ project uses Spack to manage the complexity of developing, distributing, and deploying specialized compression pipelines to meet the needs of its developers and users.

Spoken from the perspective of someone who has tried nearly every new spack feature in the last 5 years, and who maintains over 50 packages. This talk tells the story of how the FZ project tackled that complexity with spack, and where spack can grow to meet its future challenges coupled with tips and tricks we've learned along the way.
Speakers
avatar for Robert Underwood

Robert Underwood

Assistant Computer Scientist, Argonne National Laboratory
Assistant Computer Scientist in the Mathematics and Computer Science Division at Argonne National Laboratory focusing on data and I/O for large-scale scientific apps including AI for Science using lossy compression techniques and data management. Robert developed LibPressio, which... Read More →
Thursday May 8, 2025 10:45am - 11:05am CDT
Salon E-G

10:45am CDT

Charliecloud + Gitlab-CI: Building and Using System-Representative Base Containers - Nick Sly, Lawerence Livermore National Laboratory
Thursday May 8, 2025 10:45am - 11:25am CDT
Charliecloud is used in conjunction with Gitlab-CI to build out a matrix of system-representative containers that can be used for building target system-compatible binaries for automated building and testing production codes on NNSA lab machines. This presentation covers the method of generating the base containers as well as a couple of use cases where they have proven helpful.
Speakers
NS

Nick Sly

Scientist, Lawerence Livermore National Laboratory
Thursday May 8, 2025 10:45am - 11:25am CDT
Illinois River

10:45am CDT

Tuning and Performance
Thursday May 8, 2025 10:45am - 12:05pm CDT
1. Leveraging the C Configuration Space and Tuning Library (CCS) in Kokkos Tools - Brice Videau, Argonne National Laboratory (20 minutes)
Online autotuning of runtime and applications presents untapped opportunities to increase HPC application performance and efficiency. During ECP, in order to exploit this potential, the autotuning working group at Argonne National Laboratory and the Kokkos team co-designed the Kokkos Tools tuning API and the C Configuration Space and Tuning Library (CCS). The Kokkos Tools tuning API would create a framework to plug tuners inside Kokkos and expose tuning regions to them, while the CCS library would offer an API to both capture Kokkos configuration spaces and implement tuners to optimize them. This effort led to the creation of the CCS Kokkos connector, a Kokkos tool that leverages both APIs to offer a baseline tuner for Kokkos regions. In this presentation, we will present the result of this collaboration from the perspective of CCS, the abstractions it offers and how they map to Kokkos tuning model. We will describe the capabilities of the CCS library and how it fulfills the goal of offering a standard interface to bridge the gap between tuners and applications/runtimes. We will also discuss the perspectives and future works around the CCS Kokkos connector.

2. Bottlenecks in High-Dimensional Simulations - Nils Schild, Max Planck Institute for Plasma Physics (20 minutes)
The Vlasov-Maxwell system, which describes the motion for charged particles of matter in a plasma state using a particle distribution function, is based on a 6-D phase space defined through configuration and velocity coordinates.
Considering an Eulerian grid for this system with only 32^6 degrees of freedom, the distribution function requires already 8.5 GB of memory. This implies that high-resolution simulations can only be executed on large compute clusters.
In this talk, we focus on two aspects of the open-source code BSL6D to solve a reduced version of the Vlasov-Maxwell system. The shared memory parallelization based on Kokkos applies a stencil algorithm to data, which is non-contiguous in memory, to reduce memory requirements. The inter-node communication bottleneck poses a challenge due to the large halo domain to compute domain ratio. Finally, we discuss the advantages of RAII-managed MPI communicators for distributed domains, simplifying the implementation of parallel algorithms with distributed memory concepts.

3. Accelerating SPECFEM++ with Explicit SIMD and Cache-Optimized Layouts - Rohit Kakodkar, Princeton University (20 minutes)
SPECFEM++ is a suite of computational tools based on the spectral element method used to simulate wave propagation through heterogeneous media. The project aims to unify the legacy SPECFEM codes - three separate Fortran packages (SPECFEM2D, SPECFEM3D, and SPECFEM3D_globe) - into a single C++ package. This new package aims to deliver optimal performance across different architectures by leveraging the Kokkos library. In this presentation, I will outline our efforts to enhance CPU performance using explicit SIMD types (Kokkos::Experimental::simd). High vectorization throughput can be challenging, particularly because the data involved in spectral element assembly is not always organized cache-friendly. To address this, we have implemented a strategy that prefetches the data into cache-optimized scratch views of SIMD types before executing the SIMD operations. Additionally, we have optimized data layouts using custom-defined tiled layouts that improve cache locality. As a result of these optimizations, we have achieved approximately a 2.5x speed-up compared to auto-vectorized implementations.

4. Managing Kokkos Callbacks for Benchmarking, Profiling, and Unit Testing - Maarten Arnst & Romin Tomasetti, University of Liège (20 minutes)
Many Kokkos functions have instrumentation hooks defined within the framework of Kokkos::Tools. These instrumentation hooks allow Kokkos::Tools as well as third-party tracing, profiling and testing tools to register callbacks to monitor and interact with the runtime behavior of the program. In this presentation, we will describe several utilities that we have designed to help manage such callbacks. We have implemented a manager class that can register function objects that can listen to such callbacks. And we have implemented several such function objects, such as an event recorder, an event counter, and a kernel timer that uses event stream synchronization markers on device backends. We will illustrate these utilities through their use in benchmarking, profiling, and unit testing of a Kokkos-based finite-element code.
Speakers
avatar for Brice Videau

Brice Videau

Computer Scientist, Argonne National Laboratory
Brice is a computer scientist, co-leading the performance engineering team at Argonne Leadership Computing Facility. Brice's research topics include heterogeneous programming models, system software, auto-tuning, code generation, and code transformation.
avatar for Maarten Arnst

Maarten Arnst

Associate professor, University of Liege
Associate Professor at University of Liege.
avatar for Nils Schild

Nils Schild

PhD student, Max Planck Institute for Plasma Physics
After studying physics and working on solvers for sparse eigenvalue problems in quantum mechanics at the University of Bayreuth, he moved to the Max Planck Institute for Plasma Physics in Garching (Germany). During his Ph.D., he started implementing the software BSL6D, a solver for... Read More →
avatar for Rohit Kakodkar

Rohit Kakodkar

Research Software Engineer II, Princeton University
Rohit is a Research Software Engineer in Princeton University's Research Computing department. He is focused on rewriting SPECFEM, a spectral element solver designed to simulate wave propagation through heterogeneous media. SPECFEM is extensively used within the computational seismology... Read More →
avatar for Romin Tomasetti

Romin Tomasetti

PhD student, University of Liège
PhD student at University of Liège.
Thursday May 8, 2025 10:45am - 12:05pm CDT
Salon A-C

11:05am CDT

Challenges Mixing Spack-Optimized Hardware Accelerator Libraries on Personal Scientific Computers - Pariksheet Nanda, University of Pittsburgh
Thursday May 8, 2025 11:05am - 11:25am CDT
Personal computing devices sold today increasingly include AI hardware accelerators such as neural processing units and graphics cards with compute capability. However, scientific libraries packaged for laptop and desktop computers focus first on broad instruction set compatibility. Yet, hardware optimized libraries and behaviors can be applied at runtime as widely used by Intel MPI environmental variables. This session discusses the unique use case of the R package system for vendor neutral hardware acceleration using vendor agnostic SYCL / Kokkos. The goal is to allow scientific package developers to quickly and easily write vendor independent accelerator code with deep control and tuning capabilities that use hardware acceleration capabilities as well on laptop / desktop machines as HPC clusters. Although R is specifically discussed, ideas from this session translate to Python and other high-level language packages used in scientific computing. Additionally, this session raises technical challenges directly using Kokkos as well as Apptainer for continuous integration that would greatly benefit from early-stage feedback of those audience members at this conference.
Speakers
avatar for Pariksheet Nanda

Pariksheet Nanda

Postdoctoral Fellow, University of Pittsburgh
Pariksheet first learned about Spack from his university research HPC supervisor who returned from Supercomuting and told him about the "cool new project we need to start using" and has been hooked ever since. When not working on research manuscripts, he enjoys reading and writing... Read More →
Thursday May 8, 2025 11:05am - 11:25am CDT
Salon E-G

11:25am CDT

An Opinionated-Default Approach to Enhance Spack Developer Experience - Kin fai Tse, The Hong Kong University of Science and Technology
Thursday May 8, 2025 11:25am - 11:45am CDT
Despite Spack's strengths as a feature-rich HPC package manager generating fast executables for HPC apps, its adoption remains limited partly due to a steep learning curve and its perception as primarily a sysadmin tool.

We propose a set of opinionated defaults that help new users quickly adopt best practices with guaranteed reproducibility and architecture compatibilities. The approach draws from conventions used in popular userspace Python package managers like pip and conda which was proven to be effective

Unlike Python, Spack is a source-distribution system, compilation errors are a common challenge. We experimented with smoke-testing compatibility compatibility across compilers, libraries, and x86_64 architectures. Results are encoded into conflict rules into the defaults, such practice can be helpful to avoid many common build failures.

We successfully deployed this approach on x86_64 platforms with substantially different purpose (DL vs HPC), demonstrating its transferability and proving current Spack features sufficient for implementation. Additional DX enhancements will be discussed. The defaults are available as an open-source repository.
Speakers
avatar for Kin fai Tse

Kin fai Tse

IT Manager (Research Computing), The Hong Kong University of Science and Technology
Dr. Kin Fai TSE now serves to overseeing DGX cluster operations and HPC migration at HKUST. After his Physics Ph.D., he led MLOps at a voicebot startup (2021). Co-founding Flying Milktea (2022), he built a marketplace with ~2-week onboarding for new interns. He was lead coach for... Read More →
Thursday May 8, 2025 11:25am - 11:45am CDT
Salon E-G

11:25am CDT

Maintaining the Debian Charliecloud Package - Peter Wienemann, Independent
Thursday May 8, 2025 11:25am - 12:05pm CDT
Charliecloud has been in the Debian archive since the early development days of Charliecloud. It was initially packaged by Lucas Nussbaum at the end of 2017/beginning of 2018. The speaker joined the packaging effort in January 2018 and continuously contributed to it since then. This talk will give a brief introduction into Debian and describe how its tool set was useful to improve the Debian Charliecloud package and feed improvements back into the upstream project. But also the information flow from upstream authors to package maintainers has been exemplary. This presentation will provide a few examples showing this fruitful interplay.
Speakers
PW

Peter Wienemann

Independent
Thursday May 8, 2025 11:25am - 12:05pm CDT
Illinois River

11:45am CDT

Developing and Managing Data Acquisition Software Using Spack - Eric Flumerfelt, Fermi National Accelerator Laboratory
Thursday May 8, 2025 11:45am - 12:05pm CDT
The Data Acquisition systems of particle physics experiments regularly push the boundaries of high-throughput computing, demanding low-latency collection of data from thousands of devices, collating data into time-sliced events, processing these events and making trigger decisions, and writing the selected data streams to disk. To accomplish these tasks, the DAQ Engineering and Operations department at Fermilab leverages multiple software libraries and builds reusable DAQ frameworks on top. These libraries must be delivered in well-defined bundles and are thoroughly tested for compatibility and functionality before being deployed to live detectors. We have several techniques used to ensure that a consistent set of dependencies can be delivered and re-created at need. We must also support active development of DAQ software components, ideally in an environment as close as possible to that of the detectors. This development often occurs across multiple packages which have to be built in concert and features tested in a consistent and reproducible manner.
I will present our scheme for accomplishing these goals using Spack environments, bundle packages, and Github Actions-based CI.
Speakers
avatar for Eric Flumerfelt

Eric Flumerfelt

Computational Physics Developer, Fermi National Accelerator Laboratory
I have been developing data acquisition systems at Fermilab since 2014. I have worked with a number of particle physics experiments, from small test-beam experiments which run for two weeks to large international collaborations.
Thursday May 8, 2025 11:45am - 12:05pm CDT
Salon E-G

12:05pm CDT

Lunch (Provided for Attendees)
Thursday May 8, 2025 12:05pm - 1:35pm CDT
Thursday May 8, 2025 12:05pm - 1:35pm CDT
Atrium

1:35pm CDT

Key Charliecloud Innovation - Kubernetes - Angelica Loshak, Los Alamos National Laboratory
Thursday May 8, 2025 1:35pm - 1:55pm CDT
Kubernetes automates container deployment and management across environments. HPC users can benefit from Kubernetes to support the increasing demand for novel workflows, especially in AI. Kubernetes' declarative approach allows users to schedule, scale, and maintain metrics on containers while supporting multiple container runtimes. Charliecloud can enhance HPC workloads when integrated with Kubernetes. However, Kubernetes only supports container runtimes that implement the Container Runtime Interface (CRI), which Charliecloud does not. To address this, we developed a prototype CRI-compatible server for Charliecloud, allowing Kubernetes to manage pods and to create, start, and track Charliecloud containers. Despite Kubernetes expecting certain features that Charliecloud does not use, such as network namespaces, we show that the two systems can still communicate effectively. Our implementation requires 700 lines of new code, fewer than 50 lines of modification to Charliecloud, and no changes to Kubernetes. This demonstrates that Kubernetes and Charliecloud are compatible tools, advancing scientific workflows that require large compute power.

LA-UR-24-28252
Speakers
AL

Angelica Loshak

Student, Los Alamos National Laboratory
Thursday May 8, 2025 1:35pm - 1:55pm CDT
Illinois River

1:35pm CDT

From Complexity to Efficiency: SPACK’s Impact on NSM Supercomputers - Samir Shaikh, Harshitha Ugave, Centre for Developement of Advanced Computing (C-DAC)
Thursday May 8, 2025 1:35pm - 1:55pm CDT
The National Supercomputing Mission (NSM) advances India’s research by providing HPC infrastructure across institutions. However, managing software on diverse HPC systems is challenging due to hardware variations, dependencies, and version control.

Spack, a flexible package manager, addresses these issues by enabling seamless software deployment and dependency management across clusters. This study examines Spack’s implementation on 17 NSM HPC systems, improving software availability and consistency.

Spack simplifies this through customized installations, automated dependency handling, and reproducible builds, ensuring compatibility.

Implementation involved a centralized repository, automated builds, user training, software optimization, and continuous refinement. This improved research productivity, reduced support overhead, and standardized environments.

Key benefits include reproducibility, faster issue resolution, and better collaboration. Future plans involve expanding Spack repositories, integrating containers, automating updates, and training. This presentation covers our implementation, challenges, and best practices.
Speakers
avatar for Samir Shaikh

Samir Shaikh

Scientist, Centre for Developement of Advanced Computing (C-DAC)
Samir Shaikh is an HPC specialist at C-DAC, Pune, optimizing large-scale workloads, parallel computing, and system architecture. As a Scientist C, he enhances HPC performance for AI/ML, scientific computing, and NSM supercomputers. An IIT Guwahati M.Tech graduate, he has contributed... Read More →
Thursday May 8, 2025 1:35pm - 1:55pm CDT
Salon E-G

1:35pm CDT

Algorithms
Thursday May 8, 2025 1:35pm - 3:15pm CDT
1. Gyselalib++: A Portable, Kokkos-Based Library for Exascale Gyrokinetic Simulations - Etienne Malaboeuf, CINES (10 minutes)
The development of fusion energy in magnetic confinement devices relies heavily on simulations of plasma behavior. Gyselalib++ is a new open-source C++ library under active development by a European distributed and multidisciplinary team of physicists, mathematicians, and computer scientists at EPFL, CEA/IRFM, Maison de la Simulation, IPP Garching, and CINES. Gyselalib++ is itself built on top of PDI, DDC and Kokkos and provides mathematical tools for gyrokinetic semi-Lagrangian codes for tokamak plasma simulations. This presentation will introduce the library, its design and the rationale behind its development, and will highlight its key features. It will showcase how the choice of Kokkos made it possible to achieve high performance on modern hardware with performance portability over a wide range of hardware, and will explain the need to introduce DDC to improve development safety. We will discuss feedback from this experience, analyze our successes and the limitations of the approach, especially when it comes to performance, performance portability, and programmability of the code by a highly diverse team in terms of background.

2. Expression Templates with Kokkos for Lattice QCD - Travis Whyte, Jülich Supercomputing Centre (10 minutes)
Lattice quantum chromodynamics (QCD) is a first principles approach to studying the interaction of quarks and gluons. The calculation of observables in lattice QCD requires many different operations between multidimensional arrays of various ranks. In this talk, I will describe an implementation of expression templates using Kokkos that allows for lattice QCD practitioners to simply implement linear algebra operations while avoiding temporaries for views of arbitrary rank. This abstraction has the potential to promote high productivity in the development process. The performance of various benchmarks on different architectures will also be discussed.

3. Bridging Parallel Communication and On-Node Computation with Kokkos - Evan Suggs, Tennessee Technological University (20 minutes)
Although MPI and Kokkos have long been used together, there were no well-defined methods for integrating them effectively. The only approach is to point the underlying Kokkos View buffers to MPI functions.
This causes several major pain points: handling non-contiguous Views, asynchronous operations in both models, and how MPI interacts with Kokkos Profiling. Kokkos Comm is an experimental MPI interface for the Kokkos C++ Performance Portability Programming ecosystem that aims to address these concerns and improve the productivity of Kokkos users.
Currently, Kokkos Comm integrates point-to-point collectives, handling of non-contiguous Views, and Kokkos Tools Profiling. Kokkos Comm also aims to be a springboard for new and improved features that go beyond MPI and Kokkos, allowing Kokkos to work with MPI, stream-triggered MPIs, and other non-MPI communication libraries (e.g., NCCL and RCCL). This presentation will cover the Kokkos Comm API, conversion of existing code, best practices, how Kokkos Comm can help address common issues in Kokkos/MPI, and upcoming additions to Kokkos Comm, such as persistent communication and device-initiated communication.

4. Integration of PETSc, Kokkos Core, and Kernels for Performance Portability in the Age of Accelerators - Junchao Zhang, Argonne National Laboratory (20 minutes)
PETSc, the Portable, Extensible Toolkit for Scientific Computation, provides an extensive suite of scalable parallel solvers for linear and nonlinear equations, ordinary differential equation (ODE) integrators, and optimization algorithms. Widely adopted in both industry and academia, PETSc historically achieved performance portability through the C programming language and the Message Passing Interface (MPI) programming model. It used single-threaded MPI processes for both shared and distributed memory systems. This strategy had served us very well in the microprocessor age. However, the recent proliferation of accelerator-based architectures, particularly graphics processing units (GPUs), has posed new challenges to this performance portability. To address these challenges, we have integrated PETSc with the Kokkos ecosystem, specifically Kokkos-Core and Kokkos-Kernels. In this presentation, we describe our integration approach, highlight our experiences—both effective strategies and encountered challenges—and outline future developments aimed at further enhancing performance portability across evolving computational architectures.

5. Parallel Sweep Algorithms for Cartesian and Honeycomb Grids - Ansar Calloo, CEA (20 minutes)
The linear Boltzmann transport equation (BTE) is the governing equation for expressing the behaviour of neutral particles in a system such as a nuclear reactor. BTE can be solved for the flux of particles using deterministic methods whereby the equation is discretised in the phase space of is fundamental variables. This discrete equation is then usually solved using the source iteration. In this talk, we will present how the sweep algorithm which is based upon a wavefront pattern has been optimised in the context of SMP for CPU and also some preliminary results on GPU. The goal is to show how to adapt the sweep algorithm to be efficient on new supercomputer architectures.We will briefly introduce DONUT (Discrete Ordinates NeUtron Transport), a modern C++ miniapp for solving BTE based on the discrete ordinates and discontinuous Galerkin discretisations for Cartesian and honeycomb grids.
Speakers
avatar for Ansar Calloo

Ansar Calloo

Research engineer, CEA
Ansar obtained his PhD in deterministic neutron transport at CEA. For the past fifteen years, he has been working on improving simulations for reactor physics applications first at EDF R&D, then CEA. His research interests involve nuclear reactor model, numerical methods to solve... Read More →
avatar for Etienne Malaboeuf

Etienne Malaboeuf

HPC Engineer, CINES/CEA
I focus on improving the performance of projects related to real-time and high-performance computing, while providing various forms of support to researchers using French supercomputers. I have worked on numerical simulation software in an HPC context, on supercomputers and on game... Read More →
avatar for Evan Suggs

Evan Suggs

Staff Researcher, Tennessee Technological University
Evan Drake Suggs is a Research Scientist at Tennessee Technological University in Cookeville, Tennessee. In 2023, Suggs graduated with a Master's degree in Data Science from the University of Tennessee at Chattanooga and presented his thesis work on MPI+Kokkos using the ExaMPI implementation... Read More →
avatar for Junchao Zhang

Junchao Zhang

Principal Specialist, Research Software Engineering, Argonne National Laboratory
Junchao Zhang is a software developer at Argonne. He currently works on the Portable, Extensible Toolkit for Scientific Computation (PETSc) project. Before joining PETSc, he was an MPICH developer at Argonne and developed the MPI Fortran 2008 binding and MPI tool interface of MPI-3.0... Read More →
avatar for Travis Whyte

Travis Whyte

Postdoc, Jülich Supercomputing Centre
I graduated from Baylor University with a Ph.D. in Physics, focusing on algorithmic improvements for lattice QCD simulations. Since then, I have continued to work in the field, focusing on improving iterative solvers, scattering simulations and HPC software development.
Thursday May 8, 2025 1:35pm - 3:15pm CDT
Salon A-C

1:55pm CDT

BEE: Orchestrating Workflows with Containerized Applications Leveraging Charliecloud - Krishna Chilleri, Los Alamos National Laboratory
Thursday May 8, 2025 1:55pm - 2:15pm CDT
Build and Execution Environment (BEE) is a workflow orchestration system designed to build containerized HPC applications and orchestrate workflows across HPC and cloud systems. BEE integrates with existing tools and infrastructure in the scientific computing ecosystem, making it an ideal choice for large-scale scientific simulations. The use of these tools and standards allows for efficient management of the provisioning, scheduling, and monitoring of individual tasks, as well as providing flexibility, portability, and reproducibility of workflows. This presentation will highlight how BEE leverages Charliecloud—as one of its container runtimes—to facilitate unprivileged builds, pull images from registries, and run containerized applications.

LA-UR-25-22166
Speakers
KC

Krishna Chilleri

Student, Los Alamos National Laboratory
Thursday May 8, 2025 1:55pm - 2:15pm CDT
Illinois River

1:55pm CDT

Deploying a Large HPC Software Stack - Challenges and Experiences - Jose Gracia, HLRS, University Stuttgart
Thursday May 8, 2025 1:55pm - 2:15pm CDT
We aim to use Spack to deploy a large software stack at a German national HPC center. In this talk, we will give some background related to the size of the software stack, its deployment frequency, and constraints arising from the operational environment. Next, we will briefly outline some of the challenges and obstacles that we encountered such as configuration issues, interaction with Cray Programming Environment, and unexpected outcomes of the concretizer. We end the talk with a current status and next steps.
Speakers
avatar for Jose Gracia

Jose Gracia

Senior Researcher, HLRS, University Stuttgart
Together with his group, José Gracia does research into topics related to scalable programming models such as new approaches to MPI or task-based programming models and their interoperability at scale. He also works on performance analysis tools, characterization of application performance... Read More →
Thursday May 8, 2025 1:55pm - 2:15pm CDT
Salon E-G

2:15pm CDT

Building and Maintaining OSS on Fugaku: RIKEN’s Experience with Spack - Yuchi Otsuka, RIKEN R-CCS
Thursday May 8, 2025 2:15pm - 2:35pm CDT
Fugaku, Japan’s flagship supercomputer, serves a diverse range of scientific disciplines, requiring extensive open-source software (OSS) support. However, managing OSS on Fugaku presents unique challenges due to its A64FX-based Arm architecture and Fujitsu’s proprietary compilers and libraries. Our team has been leveraging Spack to efficiently manage and maintain OSS. In this talk, we will share our experience using Spack on Fugaku, highlighting how it has enabled a robust and up-to-date OSS environment. We will discuss the practical benefits of Spack, including streamlined software deployment and simplified package management, and reflect on lessons learned from maintaining software in a large-scale HPC system. By sharing our insights, we aim to contribute to the broader Spack community and reinforce its role as a key tool for HPC software management.
Speakers
avatar for Yuchi Otsuka

Yuchi Otsuka

Technical Scientist, RIKEN R-CCS
I have a long-standing background in computational condensed-matter physics research and have been involved in managing and maintaining OSS on Fugaku since 2022. My role is to ensure a robust and up-to-date OSS environment on Fugaku to support a wide range of scientific applicati... Read More →
Thursday May 8, 2025 2:15pm - 2:35pm CDT
Salon E-G

2:15pm CDT

Charliecloud Office Hours - Reid Priedhorsky, Los Alamos National Laboratory
Thursday May 8, 2025 2:15pm - 3:15pm CDT
Members of the Charliecloud team will be available for office hours to listen to feedback/suggestions, answer questions, and/or help debug issues. 

LA-UR-25-22140
Speakers
RP

Reid Priedhorsky

Scientist, Los Alamos National Laboratory
I am a staff scientist at Los Alamos National Laboratory. Prior to Los Alamos, I was a research staff member at IBM Research. I hold a Ph.D. in computer science from the University of Minnesota and a B.A., also in computer science, from Macalester College.My work focuses on large-scale... Read More →
Thursday May 8, 2025 2:15pm - 3:15pm CDT
Illinois River

2:35pm CDT

Using Spack to Build and Maintain a Facility-Specific Programming Environment - Nicholas Sly, Lawrence Livermore National Laboratory
Thursday May 8, 2025 2:35pm - 2:55pm CDT
The trials and tribulations of using Spack to construct, build, and maintain a facility-specific programming environment at LLNL. Working with the Spack developers to ensure that Spack is able to do what it claims it can do in a large production environment.
Speakers
NS

Nick Sly

Scientist, Lawerence Livermore National Laboratory
Thursday May 8, 2025 2:35pm - 2:55pm CDT
Salon E-G

2:55pm CDT

Aurora PE: Rethinking Software Integration in the Exascale Era - Sean Koyama, Argonne National Laboratory
Thursday May 8, 2025 2:55pm - 3:15pm CDT
The exascale Aurora supercomputer at the Argonne Leadership Computing Facility posed numerous challenges during its development due to its novel scale. One such challenge was in creating a scalable and maintainable scientific software environment. Typical software deployment methods failed to scale and were difficult to maintain over time, necessitating a new way of thinking about software integration. In this talk we present our work on the Aurora Programming Environment, a bespoke scientific programming environment which optimizes for scale and leverages Spack for its strengths in reproducibility, automation, and multiplicative build combinations. We discuss details of the containerized build process and read-only image deployment strategy as well as existing pain points and workarounds. We also examine the future possibilities that our approach opens up, including tightly integrated CI/CD flows and portable containerized access to the PE. We believe this approach is generalizeable and may benefit facilities where traditional software integration methods fall short of needs.
Speakers
avatar for Sean Koyama

Sean Koyama

Systems Integration Admin, Argonne National Laboratory
Sean Koyama is a Systems Integration Administrator at the Argonne National Laboratory's Leadership Computing Facility. Sean integrates scientific software stacks into the user environments on ALCF machines, including Aurora, the ALCF's exascale supercomputer. Their work includes developing... Read More →
Thursday May 8, 2025 2:55pm - 3:15pm CDT
Salon E-G

3:15pm CDT

Coffee Break
Thursday May 8, 2025 3:15pm - 3:40pm CDT
Thursday May 8, 2025 3:15pm - 3:40pm CDT

3:40pm CDT

Driving Continuous Integration and Developer Workflows with Spack - Richard Berger, Los Alamos National Laboratory
Thursday May 8, 2025 3:40pm - 4:00pm CDT
Spack makes it easy to install dependencies for our software on multiple HPC platforms. However, there is little guidance on how to structure Spack environments for larger projects, share common Spack installations with code teams and utilize them in an effective way for continuous integration and development.

This presentation will share some of the lessons learned from deploying chained Spack installations for multiple code teams at LANL on various HPC platforms both on site and on other Tri-Lab systems, how to structure such deployments for reusability and upgradability, and make them deployable even on air-gapped systems. It will also show how we utilize Spack's build facilities to drive CMake-based projects on GitLab for continuous integration, without having to replicate build configuration logic in GitLab files, while giving developers an easy-to-follow workflow for recreating CI runs in various configurations.
Speakers
avatar for Richard Berger

Richard Berger

Scientist, Los Alamos National Laboratory
Richard is a research software engineer in the Applied Computer Science Group (CCS-7) at Los Alamos National Laboratory (LANL) with a background in Mechatronics, high-performance computing, and software engineering. He is currently contributing to the core development of LAMMPS, FleCSI... Read More →
Thursday May 8, 2025 3:40pm - 4:00pm CDT
Salon E-G

3:40pm CDT

Panel Discussion to be Announced
Thursday May 8, 2025 3:40pm - 5:00pm CDT
Thursday May 8, 2025 3:40pm - 5:00pm CDT
Salon A-C

4:00pm CDT

Implementing a Security Conscious Build Configuration Relay with a Shared Build Cache - Chris White, Lawrence Livermore National Laboratory
Thursday May 8, 2025 4:00pm - 4:20pm CDT
In large-scale software development efforts, effective communication between projects is essential to ensure consistency, reproducibility, and efficiency. This presentation explores strategies to improve coordination among software teams by leveraging Continuous Integration (CI) for relaying crucial build configurations while maintaining security for proprietary project sources. We will demonstrate best practices for sharing build configurations with upstream projects without exposing proprietary code.

A key focus will be optimizing the use of Spack, particularly in reducing the number of Spack Package Repositories utilized across multiple teams. This will simplify maintenance, harden builds, and avoiding duplication. Additionally, we will highlight the benefits of heavily integrating Spack CI to generate build caches, which will reduce rebuild times, and enhance software portability. By adopting these approaches, teams can achieve better collaboration, streamlined workflows, and improved software sustainability.
Speakers
avatar for Chris White

Chris White

WSC DevOps Coordinator, Lawrence Livermore National Laboratory
Chris White is the WSC DevOps Coordinator at Lawrence Livermore National Laboratory. He advises multi-disciplinary teams on software best practices with a focus on unifying complex DevOps workflows across multiple teams. Chris specializes in improving collaboration while ensuring... Read More →
Thursday May 8, 2025 4:00pm - 4:20pm CDT
Salon E-G

4:20pm CDT

Spack-Based WEAVE Environment at LLNL - Lina Muryanto, Lawrence Livermore National Security, LLC
Thursday May 8, 2025 4:20pm - 4:50pm CDT
The WEAVE team at LLNL has created a Spack-based virtual environment (accessible to Livermore Computing community) that has a rich set of open-source tools to create a workflow for any HPC applications, commonly used Python packages and several commonly used ML and AI packages.
The goal is to provide a stable, well tested environment that users can activate and use directly across Livermore Computing's vast array of machines/OSes and networks.
We also provide the capability for users to create a local environment based on the WEAVE environment.

Using Spack allows us to install the same set of software across different platforms in LC. It also allows us to use the same Spack environment file to recreate the exact same ecosystem across networks within the Lab.
We leverage GitLab CI for our DevOps platform to automate Spack install, creating test environment, running tests and deploying the environment.
We also leverage LLNL Nexus Repository to sync our build files across networks within the lab.

WEAVE team also has implemented the "WEAVE Badging Program" where the community can submit a request to have a tool to be integrated into the WEAVE environment.
Speakers
avatar for Lina Muryanto

Lina Muryanto

Software Engineer, Lawrence Livermore National Security, LLC
Lina joined LLNL in 2018 as a DevOps engineer for the ESGF Project. In 2021, Lina joined the SD program and joined the WEAVE team in 2022. She has implemented CI/CD from scratch. Lina is passionate about achieving high software quality and reliability through software test development... Read More →
Thursday May 8, 2025 4:20pm - 4:50pm CDT
Salon E-G

4:40pm CDT

DevOps for Monolithic Repositories Using Spack - Phil Sakievich, Sandia National Laboratories
Thursday May 8, 2025 4:40pm - 5:00pm CDT
In the realm of large software projects, the choice between a monolithic repository and several distributed repositories presents significant trade-offs that can impact development efficiency, collaboration, and maintainability. Monolithic repositories, while offering centralized management and streamlined dependency handling, can become unwieldy as project size increases. Conversely, distributed repositories provide modularity and flexibility but may lead to challenges in integration and version control. This presentation will delve into the ongoing research conducted by Sandia National Laboratories, where researchers are exploring innovative solutions to harness the strengths of both repository models through the use of Spack, a package manager designed for scientific computing. We will outline the methodology employed in this exploration, highlighting the performance trade-offs i dentified thus far, including aspects such as build times, dependency resolution, and ease of collaboration. Attendees will gain insights into the implications of repository structure on software development practices and the potential for hybrid approaches to optimize project outcomes.
Speakers
avatar for Phil Sakievich

Phil Sakievich

Senior Computer Scientist R&D, Sandia National Laboratories
Phil comes from a high-performance computing and fluid mechanics background. He became involved with Spack during the ExaScale computing project and author of the Spack-Manager project. Phil is an active member of the Spack technical steering committee and currently leads several... Read More →
Thursday May 8, 2025 4:40pm - 5:00pm CDT
Salon E-G
 
  • Filter By Date
  • Filter By Venue
  • Filter By Type
  • Timezone

Share Modal

Share this link via

Or copy link

Filter sessions
Apply filters to sessions.