Flexible Package Manager Automates the Deployment of Software on Supercomputers


A podcast with Todd Gamblin, a computer scientist in the Center for Applied Scientific Computing at Lawrence Livermore National Laboratory (LLNL) and the creator of Spack, the deployment tool for the Exascale Computing Project’s (ECP) software stack and winner of a 2019 R&D 100 Award.


Reducing the Memory Footprint and Data Movement on Exascale Systems


A podcast with Peter Lindstrom, a computer scientist in the Center for Applied Scientific Computing at Lawrence Livermore National Laboratory (LLNL). Lindstrom manages all activities related to the ZFP project, which has been folded into ECP’s ALPINE in-situ visualization project. ZFP is a compressed representation of multidimensional floating-point arrays that are ubiquitous in high-performance computing.


Delivering Exascale Machine Learning Algorithm and Tools for Scientific Research


A podcast, including an interview with LLNL’s Brian Van Essen, who has implemented surrogate models using deep neural networks and the Sierra and Summit resources.


Optimizing Math Libraries to Prepare Applications for Exascale Computing


A podcast featuring LLNL’s Ulrike Meier Yang, principal investigator of the xSDK4ECP project and lead of hypre’s ECP effort. Her expertise spans more than three decades in numerical algorithms, particularly linear system solvers, parallel computing, high-performance computing, and scientific software design.


R&D100 Award-Winning Software Enables I/O Performance Portability and Code Progress


A podcast featuring two LLNL researchers from the Exascale Computing Project (ECP) involved in the development of Scalable Checkpoint/Restart (SCR) Framework 2.0, Elsa Gonsiorowski and Kathryn Mohror, discuss what SCR is and does, the challenges involved in creating it, and the impact it is expected to have in high-performance computing. 


Software Enables Use of Distributed In-System Storage and Parallel File System


LLNL’s Kathryn Mohror contributes to a podcast discussing UnifyFS and how it can provide I/O performance portability for applications, enabling them to use distributed in-system storage and the parallel file system.


First Annual Exascale Day Celebrates Discoveries on the Fastest Supercomputers in the World


LLNL is participating in the new National Exascale Day, a registered holiday to be celebrated annual on October 10.


Experts Weigh in on the Game-Changing Nature of Exascale Computing


LLNL’s Lori Diachin is one of three experts featured in a Q&A by writer Mae Rice from the social network and blogging platform called Built In.


CANDLE Illuminates New Pathways in Fight Against Cancer


An article on the U.S. Department of Energy (DOE) website written by Andrea Peterson takes a look at the CANcer Distributed Learning Environment (CANDLE), a cross-cutting initiative of the Joint Design of Advanced Computing Solutions for Cancer (JDACS4C) collaboration that is supported by DOE’s Exascale Computing Project.


Cray Wins NNSA-Livermore ‘El Capitan’ Exascale Contract


DOE, NNSA, and LLNL have announced the signing of contracts with Cray Inc. to build the NNSA’s first exascale supercomputer, "El Capitan." El Capitan will have a peak performance of more than 1.5 exaflops (1.5 quintillion calculations per second) and an anticipated delivery in late 2022. 


Robustly Delivering Highly Accurate Computer Simulations of Complex Materials


LLNL is part of a team developing a QMC methods software named QMCPACK to find, predict, and control materials and properties at the quantum level. The ultimate aim is to achieve an unprecedented and systematically improvable accuracy by leveraging the memory and power capabilities of the forthcoming exascale computing systems.


CEED’s Impact on Exascale Computing Project (ECP) Efforts is Wide-Ranging


The Center for Efficient Exascale Discretizations (CEED) is a research partnership involving more than 30 computational scientists from two DOE labs (Lawrence Livermore and Argonne) and five universities (the University of Illinois at Urbana-Champaign, Virginia Tech, the University of Colorado Boulder, the University of Tennessee, and Rensselaer Polytechnic Institute). Led by LLNL’s Tzanio Kolev, CEED is helping applications leverage future architectures by developing state-of-the-art discretization algorithms that better exploit the hardware and deliver a significant performance gain over conventional methods.


Offering More Detailed Simulations By Solving Problems Faster and at Larger Scales

TUESDAY, MAY 21, 2019

An LLNL team is preparing hypre to take advantage of the capabilities of the upcoming exascale computing systems. Specifically, the hypre project is delivering scalable performance on massively parallel computer architectures to positively impact a variety of applications that need to solve linear systems.


LLNL's Lori Diachin Receives High Performance Computing Industry Publication Honor


Lori Diachin, Computation's deputy associate director for science and technology, has been named one of the "People to Watch" in High Performance Computing (HPC) for 2019 by HPCwire, the online news service covering supercomputing. Diachin, who has served in numerous leadership positions in LLNL's Computation Directorate, is the deputy director for the Department of Energy's Exascale Computing Project (ECP).


Demonstrating In-Situ Computer Simulation Visualization and Analysis on Sierra


ALPINE Ascent is the first in-situ visualization and analysis tool meant for next-generation supercomputers, said Cyrus Harrison of Lawrence Livermore National Laboratory (LLNL) and the Department of Energy’s Exascale Computing Project (ECP). ALPINE Ascent is referred to as a flyweight in-situ visualization and analysis library for ECP applications. 


ECP Receives HPCwire Editors’ Choice Award for Best HPC Collaboration of Government, Academia, and Industry


The U.S. Department of Energy’s (DOE) Exascale Computing Project (ECP) announced it has been recognized by HPCwire with an Editor’s Choice Award for the project’s extensive collaborative engagement with government, academia and industry in support of the ECP’s effort to accelerate delivery of a capable exascale computing ecosystem, as the nation prepares for the next era of supercomputers capable of a quintillion operations per second.


Spack logo

Spack: The Deployment Tool for ECP’s Software Stack


Audio chat with Todd Gamblin, a computer scientist in the Center for Applied Scientific Computing at Lawrence Livermore National Laboratory (LLNL). His research focuses on scalable tools for measuring, analyzing, and visualizing parallel performance data. He leads LLNL’s DevRAMP (Reproducibility, Analysis, Monitoring, and Performance) team, and he is the creator of Spack, the deployment tool for the Exascale Computing Project’s (ECP) software stack.


Lori Diachin

Exascale Computing Project Names Lori Diachin as New Deputy Director

TUESDAY, JULY 31, 2018

The Department of Energy’s Exascale Computing Project (ECP) has named Lori Diachin as its new Deputy Director effective August 7, 2018. Lori replaces Stephen Lee who has retired from Los Alamos National Laboratory. Lori has been serving as the Deputy Associate Director for Science and Technology in the Computation Directorate at Lawrence Livermore National Laboratory (LLNL) since 2017.  


ECP logo

ECP Announces New Co-Design Center to Focus on Exascale Machine Learning Technologies

FRIDAY, JULY 20, 2018

The Exascale Computing Project has initiated its sixth Co-Design Center, ExaLearn, to be led by Principal Investigator Francis J. Alexander, Deputy Director of the Computational Science Initiative at the U.S. Department of Energy’s (DOE) Brookhaven National Laboratory. ExaLearn is a co-design center for Exascale Machine Learning (ML) Technologies and is a collaboration initially consisting of experts from eight multipurpose DOE labs. Brian Van Essen is the LLNL Lead.


CEED logo

Scaling the Unknown: The CEED Co-Design Center


CEED is one of five co-design centers in the Exascale Computing Project, a collaboration between DOE’s Office of Science and National Nuclear Security Administration. The centers facilitate cooperation between the ECP’s supercomputer vendors, application scientists and hardware and software specialists. Led by Livermore, the CEED collaboration alone encompasses more than 30 researchers at two DOE national laboratories – LLNL and Argonne – and five universities.


earthquake simulation showing a map of the Hayward Fault

New Simulations Break Down Potential Impact of a Major Quake by Building Location and Size


With unprecedented resolution, scientists and engineers are simulating precisely how a large-magnitude earthquake along the Hayward Fault would affect different locations and buildings across the San Francisco Bay Area. A team from Berkeley Lab and Lawrence Livermore National Laboratory, both U.S. Department of Energy (DOE) national labs, is leveraging powerful supercomputers to portray the impact of high-frequency ground motion on thousands of representative different-sized buildings spread out across the California region.


Terri Quinn

Audio Update: Hardware and Integration

TUESDAY, MAY 15, 2018

In an audio discussion, ECP’s Hardware and Integration (HI) Director Terri Quinn (Lawrence Livermore National Laboratory) describes how the HI focus area performs its mission and what its top goals are.


Tapasya Patki

Efficiently Using Power and Optimizing Performance of Scientific Applications at Exascale


Tapasya Patki of Lawrence Livermore National Laboratory leads the Power Steering project within the ECP. Her project provides a job-level power management system that can optimize performance under power and/or energy constraints. She has expertise in the areas of power-aware supercomputing, large-scale job scheduling, and performance modeling.