A-Z Index | Directory | Careers

Deep Learning at 15 PFlops Enables Training for Extreme Weather Identification at Scale

March 27, 2018

Contact: Rob Farber, [email protected]

Reprinted with permission from HPCWire.

Petaflop per second deep learning training performance on the Cori supercomputer at Berkeley Lab’s National Energy Research Scientific Computing Center (NERSC) has given climate scientists the ability to use machine learning to identify extreme weather events in huge climate simulation datasets. Predictive accuracies ranging from 89.4% to as high as 99.1% show that trained deep learning neural networks (DNNs) can identify weather fronts, tropical cyclones, and long narrow air flows that transport water vapor from the tropics called atmospheric rivers (Figure 1).

Screen Shot 2018 03 29 at 2.39.00 PM

Figure 1: Relation between ground truth (green boxes) and classification plus regression results (red boxes) of the DNN trained to recognize atmospheric phenomena.

 

The strong relationship between ground truth and the neural network prediction can be seen in the classification plus regression results reported by Berkeley Lab climate scientist Michael Wehner at the Intel Developer Conference held during SC17 last November in Denver, Colorado.

Supercomputers like NERSC’s Cori system provide scientists with an extraordinary tool to model climate change significantly faster and far more accurately than was possible on previous generation supercomputers. For example, simulated storms can run 300x to 10,000x faster than real time, according to Wehner in his presentation at the Intel conference. This meets the needs of climate scientists who need to run many-century long simulations to evaluate the impact of climate change far into the future.

Screen Shot 2018 03 29 at 2.45.27 PM

Figure 2: Comparative results showing the additional detail that is modeled by a 25 Km spatial resolution model as opposed to a 200 Km model.

While humans can (and do) perform well in identifying and tracking extreme weather events in real time, they simply cannot keep up when climate models run two to five orders of magnitude faster, he added. Thus machine learning has to be used to identify and track extreme weather events. Further, these machine learning based results can be used to validate the climate models so we have confidence in the future predictions of these models.

Figure 2 represents an assessment of what will happen to the number and intensity of hurricanes as the climate warms.

According to Prabhat, who leads NERSC’s Data and Analytics Services group and is director the Big Data Center at NERSC, identifying phenomena in climate data is analogous to commercial vision applications (Figure 3). In his Intel Developer Conference keynote presentation, he noted that initial supervised training results show that this analogy is correct because machine learning was able to train and recognize each of three desired atmospheric phenomena with high accuracy.

Researchers from MILA, NERSC and Microsoft teamed up to create a novel semi-supervised convolutional DNN architecture that was able to do the work of all three individual supervised DNNs at the same time. Essentially, this novel neural network finds the bounding box size and location when it classifies the atmospheric phenomena. Further, the neural network also associates a probability with the classification.

Screen Shot 2018 03 29 at 2.50.07 PM

Figure 3. Intuition showing that conventional DNNs could potentially recognize atmospheric phenomena

 

Powerful leadership class supercomputers like the Cori supercomputer have made fast, accurate global climate simulations possible. Innovations such as the petascale capable hybrid machine learning technique pioneered by Intel, NERSC and Stanford means those same machines can also train DNNs to evaluate the tens to hundreds of terabytes of data created by these faster and more accurate climate simulations.

Click here for the full version of this article, which was originally published on HPCWire on March 19, 2018.

About the Author: Rob Farber is a global technology consultant and author with an extensive background in HPC and in developing machine learning technology that he applies at national labs and commercial organizations.


About Computing Sciences at Berkeley Lab

High performance computing plays a critical role in scientific discovery. Researchers increasingly rely on advances in computer science, mathematics, computational science, data science, and large-scale computing and networking to increase our understanding of ourselves, our planet, and our universe. Berkeley Lab’s Computing Sciences Area researches, develops, and deploys new foundations, tools, and technologies to meet these needs and to advance research across a broad range of scientific disciplines.

Founded in 1931 on the belief that the biggest scientific challenges are best addressed by teams, Lawrence Berkeley National Laboratory and its scientists have been recognized with 13 Nobel Prizes. Today, Berkeley Lab researchers develop sustainable energy and environmental solutions, create useful new materials, advance the frontiers of computing, and probe the mysteries of life, matter, and the universe. Scientists from around the world rely on the Lab’s facilities for their own discovery science. Berkeley Lab is a multiprogram national laboratory, managed by the University of California for the U.S. Department of Energy’s Office of Science.

DOE’s Office of Science is the single largest supporter of basic research in the physical sciences in the United States, and is working to address some of the most pressing challenges of our time. For more information, please visit energy.gov/science.