Adaptive Mesh Refinement Algorithms Create Computational Microscope
January 26, 1998
By Jon Bashor
NOTE: This archived news story is made available as-is. It may contain references to programs, people, and research activities that are no longer active at Berkeley Lab. It may include links to web pages that no longer exist or refer to documents no longer available.
Computer modeling algorithms that break large problems into small pieces, and then focus computing power on those areas of the most scientific interest are helping scientists better understand such everyday situations as running an internal combustion engine, flying in an airplane or predicting the weather.
The axiom that big problems are much easier to tackle when divided into much smaller tasks is especially true when solving problems in fluid dynamics.
Called adaptive mesh refinement capability, this applied math research by the Lab’s Center for Computational Science and Engineering provides tools for computer modeling to allow researchers to automatically apply their computing resources to the most intriguing problems. The end result is that researchers get better answers at a lower cost.
The system works by taking a problem and dividing it into smaller parts, much like putting a mesh over the subject in question and then looking at each segment. Areas of interest are treated with a finer mesh to give scientists more detailed information. The smaller the piece, the higher the accuracy (Imagine wanting to accurately measure the temperature in an auditorium. Taking eight measurements over two hours would be much less reliable than measuring a thousand points every minute).
However, analyzing more points requires more computing capacity. The adaptive mesh refinement capability allows scientists to focus their existing computer power on a narrower part of the overall problem, rather than requiring bigger, more powerful computers.
For example, a typical computer modeling program may provide a “big picture” image of how a diesel engine operates. This turbulent process, a strong interplay of chemistry and fluid dynamics, is not well understood, even by the experts who design the engines. Better computer models will allow them to “see” inside the engine and gain a better understanding. But researchers studying ways to make diesel combustion more efficient and less polluting want to look only at the point where fuel is burning. And because the process occurs very quickly, accurate modeling requires that the burn time be divided into thousands of time steps to provide an accurate model.
The computer simulations are then checked by comparing the results with actual experiments. The model can then be tweaked by changing any of various factors, such as temperature, time, chemistry, and then compared again with data from an engine.
Creating workable algorithms to achieve this requires the solving of both mathematical and computer science problems. For example, the computer needs to be programmed to keep track of cells of both fine and coarse data, the interfaces between them and the movement of the cells over time. This requires a whole set of data structures to keep track of the various elements and allocate appropriate portions of the computer architecture to do the job.
Mathematically, everything has to add up. The law of conservation applies, even to computer modeling. In the case of weather modeling, a cloud in a coarse-grain cell has to correspond with the same cloud as it passes through a fine-grained cell. To make the algorithms work properly, mathematical formulas must be written to maintain the balance between the appropriate cells even as the number of data points changes.
The overall task is made easier by the library of existing algorithms. In many cases, portions of these can be plugged in to help solve new problems. As with the problems they address, this allows the computer scientists to focus their efforts on the most challenging aspects of the work at hand.
Adaptive Mesh Refinement Capability
The Center for Computational Science and Engineering has developed Adaptive Mesh Refinement Capability for improved computer modeling of low Mach-number flows in which all the motion is at speeds slower than the speed of sound. This capability allows researchers to model flows for less computational cost (less computer time and memory) and achieve the same accuracy.
This capability, which is being developed with funding from DOE MICS and the DSWA, is being used by Lab scientists to study a broad range of phenomena, from turbulent flow of water through a channel to combustion to atmospheric dispersal.
Many everyday situations we take for granted, such as running an internal combustion engine, flying in an airplane or predicting the weather, actually involve very complex problems which scientists are only now beginning to understand, thanks to computer modeling. At Berkeley Lab, computer scientists and mathematicians are developing specialized algorithms for computer modeling of fluid dynamics problems.
The way that CCSE tackles a problem is by first looking at what’s known. In the case of a diesel engine, this would include fuel sprays from injectors, the heating and vaporization of the fuel mixture, ignition and burning, and generation of soot. Then, basic algorithms for studying fluid dynamics are hooked up to models of those processes and the shortcomings identified.
The algorithms work by applying a computer grid, or mesh, to the problem and breaking the overall process into smaller components. This mesh then automatically adapts to the problem being studied and determines the areas of most interest to the scientist. By focusing primarily on a specific area, more computing power can be brought to bear on the problem and more accurate solutions devised.
The end result is that problems are solved more quickly and efficiently.
Although there are different ways of approaching the problem, the Berkeley group uses a technique called structured hierarchical adaptivity. Because supercomputers work on highly structured grids, this approach is a natural for adapting to massive parallel processor computers.
About Computing Sciences at Berkeley Lab
The Computing Sciences Area at Lawrence Berkeley National Laboratory provides the computing and networking resources and expertise critical to advancing Department of Energy Office of Science research missions: developing new energy sources, improving energy efficiency, developing new materials, and increasing our understanding of ourselves, our world, and our universe.
Founded in 1931 on the belief that the biggest scientific challenges are best addressed by teams, Lawrence Berkeley National Laboratory and its scientists have been recognized with 13 Nobel Prizes. Today, Berkeley Lab researchers develop sustainable energy and environmental solutions, create useful new materials, advance the frontiers of computing, and probe the mysteries of life, matter, and the universe. Scientists from around the world rely on the Lab’s facilities for their own discovery science. Berkeley Lab is a multiprogram national laboratory, managed by the University of California for the U.S. Department of Energy’s Office of Science.
DOE’s Office of Science is the single largest supporter of basic research in the physical sciences in the United States, and is working to address some of the most pressing challenges of our time. For more information, please visit energy.gov/science.