A-Z Index | Phone Book | Careers

Novel Data Processing Technique Aids XFEL Protein Structure Analysis

GPU computing resources at NERSC are helping a Berkeley Lab-led research team gain more accurate insights

January 26, 2021

This science highlight was originally published by Lawrence Berkeley National Laboratory’s Biosciences Area on Dec. 21, 2020. It has been modified to include more information about the National Energy Research Scientific Computing Center’s role in supporting this research.

X-ray free-electron lasers (XFELs) came into use in 2010 for protein crystallography, allowing scientists to study fully hydrated specimens at room temperature without radiation damage. Researchers have developed many new experimental and computational techniques to optimize the technology and draw the most accurate picture of proteins from crystals.

SauterXFELfigure2

Collage of partial Bragg spots recorded during experiments at the XFEL. For each pixel in the "shoe-box" of a given Bragg spot, diffBragg applies precise modeling of the various parameters affecting their shape and intensity.

Now scientists in the Molecular Biophysics and Integrated Bioimaging (MBIB) Division at Lawrence Berkeley National Laboratory (Berkeley Lab) have developed a new program, diffBragg, that can process every pixel collected from an XFEL for a protein structure independently. In a recent IUCrJ paper, the team – led by MBIB Senior Scientist Nicholas Sauter – proposed a new processing framework for more accurate determination of protein structures.

Scientists use X-ray crystallography to determine the three-dimensional structures of proteins by measuring the pattern of Bragg spots produced by diffracting an X-ray beam through a stream of protein crystals. However, traditional image algorithms do not precisely fit the Bragg spots from X-ray free-electron lasers, and so the full potential of the data is not realized.

“Cutting-edge discovery in structural biology has always been about extending the resolution, in both space and time, of what can be visualized,” said Sauter, lead author of the paper. “By extracting more accurate information from XFELs we get a better spatial resolution that reveals more about enzyme mechanisms, and potentially allows us to see tiny changes – less than the length of an atomic bond – to follow chemical reactions over time.”

One longstanding project that has captivated the attention of Sauter and his team has been understanding photosystem II, the key macromolecule in photosynthesis, and how it collects energy from 4 photons from sunlight and channels it into the creation of a single molecule of oxygen (O2) from water. Sauter explained, “XFEL light sources allow us to follow this reaction on a microsecond time scale, and the new data processing offers us a chance to improve the picture from this exceedingly difficult experiment. However, our current work is at the theoretical proof-of-principle stage and will require a dedicated effort to apply the methods to our photosystem II experiments.”

This data processing approach is computational, and its success relies on decades of instrumentation development to record the X-ray diffraction pattern at high sensitivity and frequency. “We were able to get an early preview of the latest imaging detectors just now coming online, and they are sensitive to single X-ray photons per pixel, with framing rates of 2 kHz. That amounts to a 10-fold improvement in measurement accuracy and a 10-fold improvement in framing rate, compared to where we were 10 years ago,” said Derek Mendez, project scientist and first author of the IUCrJ article. “This is impressive technology that is complemented by our new approach that can computationally model the data at the level of a single X-ray photon.”

While this work is at a very early stage, eventually it will allow researchers to model more detail in protein crystals. Although the same protein is encased in each crystal, the individual proteins can have many different backbone and side chain conformations, which leads to internal disorder, and is currently difficult to model. diffBragg will also allow scientists to correctly account for correlated uncertainties, for example, one particular pixel might be relatively unreliable and the effect of that pixel in the final model may be minimized.

Multi-year, Multi-facility Effort

The MBIB researchers worked with engineers at the SLAC Linac Coherent Light Source (LCLS) in their efforts to improve the data processing pipeline for X-ray crystallography. The team also collaborated with the National Energy Research Scientific Computing Center (NERSC) to implement their approach.

“This latest paper is one step in a multi-year effort to improve data processing for X-ray crystallography,” Sauter said. “Our application will implement a large inverse-modeling problem, but the steps to get there are to implement the ‘forward simulation,’ (already published), plus a first-derivative based parameter estimation (the current paper), and then apply these techniques to experimental data.” NERSC’s Johannes Blaschke has been critical to moving the entire project forward, he added.

Another critical part of this research is the longstanding collaboration between NERSC and LCLS. “We have begun to routinely transfer LCLS data to NERSC during live experiments for real-time feedback,” Sauter said. “The new inverse modeling (diffBragg) will eventually be part of this pipeline.”

Given that the diffBragg program makes calculations that involve every pixel collected, which is 100 billion pixels per experiment, the computational power that is needed for these types of analyses is rather large. “One of the peer reviewers for our paper took note of the large computational footprint and wanted to know if there were simpler algorithms that would be more tractable,” said Sauter. “Instead, we revised our paper with performance numbers for GPU acceleration, showing that this is truly a problem that can be solved on the Perlmutter supercomputer when it comes on line in 2021.”

The work was funded by the Exascale Computing Project and NIH/NIGMS. In addition to Sauter and Mendez, MBIB researchers involved in this work were project scientists Robert Bolotovsky and Asmit Bhowmick, Aaron Brewster (research scientist), Jan Kern (staff scientist), Junko Yano (senior staff scientist), and James M. Holton (faculty scientist).

NERSC is a U.S. Department of Energy Office of Science user facility.


About Computing Sciences at Berkeley Lab

High performance computing plays a critical role in scientific discovery, and researchers increasingly rely on advances in computer science, mathematics, computational science, data science, and large-scale computing and networking to increase our understanding of ourselves, our planet, and our universe. Berkeley Lab’s Computing Sciences Area researches, develops, and deploys new foundations, tools, and technologies to meet these needs and to advance research across a broad range of scientific disciplines.

Founded in 1931 on the belief that the biggest scientific challenges are best addressed by teams, Lawrence Berkeley National Laboratory and its scientists have been recognized with 13 Nobel Prizes. Today, Berkeley Lab researchers develop sustainable energy and environmental solutions, create useful new materials, advance the frontiers of computing, and probe the mysteries of life, matter, and the universe. Scientists from around the world rely on the Lab’s facilities for their own discovery science. Berkeley Lab is a multiprogram national laboratory, managed by the University of California for the U.S. Department of Energy’s Office of Science.

DOE’s Office of Science is the single largest supporter of basic research in the physical sciences in the United States, and is working to address some of the most pressing challenges of our time. For more information, please visit energy.gov/science.