A Computer for the Clouds
A proposed supercomputer would do just one job--model global climate--but consume far less electricity than a general-purpose machine
August 1, 2008
However pleasing to the eye, clouds are a computational headache because their numerous features can either accelerate or retard warming. These altocumulus clouds form in midatmosphere, at 2000 to 6000 meters.
In May an IBM-built supercomputer called Roadrunner crunched through a quadrillion floating-point operations per second, officially becoming the first supercomputer to break the petaflop barrier. But unofficially, that barrier had fallen two years before, when MDGRAPEâ''3, a machine at Japan's Riken Institute, in Wako, powered up. Accepted benchmarking methods ruled out that performance because MDGRAPE-3 is a purpose-built computer, able to model molecular interactions and little else. Yet the machine cost Riken just oneâ''tenth of Roadrunner's price--more than US$100million--and consumes just one-tenth the power.
That power-saving potential is convincing many people who have belittled special-purpose machines to give them a second look. Electricity already accounts for more than half the lifetime cost of owning and operating a supercomputer--or any large server farm, for that matter--and power's share is expected to increase.
”We think scientific computing is ripe for a change,” says Michael Wehner, a climatologist at Lawrence Berkeley National Laboratory. ”Instead of getting a big computer and saying, ’What can we do?' we want to do what particle physicists do and say, ’We want to do this kind of science--what kind of machine do we need to do it?' ”
Wehner and two engineers, Lenny Oliker and John Shalf, also of Lawrence Berkeley, have proposed perhaps the most powerful special-purpose computer yet. It is intended to model changes in climatic patterns over periods as long as a century. Specifically, it should be able to remedy today's inability to model clouds well enough to tell whether their net effect is to warm the world or cool it. To solve the problem, climatologists figure they need to deal in chunks of the atmosphere measuring 1kilometer on a side--a job for an exaflop machine, one with 1000 times more computing power than even Roadrunner can provide.
Wehner, Oliker, and Shalf estimate that a general-purpose machine using today's technology would cost $1 billion to build and 200 megawatts to power--enough for a small city. By comparison, they estimate, a specialized machine would cost just $75 million and consume just 4 MW.
The researchers are now trying to validate their claims with a hardware mock-up, which they are building in collaboration with Tensilica, a custom-chip supplier in Santa Clara, Calif. The plan is to bench-test a single processor by November and a parallel array of processors by the middle of 2009. If the claims are vindicated, the researchers hope to get government funding for a full-size machine.
Critics of special-purpose machines say they've heard it all before. ”The problem is that when we devise a new way to solve a problem, the machine designed for the old way will no longer be asgood,” says Jack Dongarra, a professor of electrical engineering and computer science at the University of Tennessee.
But according to Horst Simon, who heads the Lawrence Berkeley lab's research computing center, the proposed machine would not be so specialized that a new algorithm would render it instantly obsolete.
”We are building hardware that runs not just one algorithm but a large class of related algorithms,” he says. ”We are trying to eliminate unessential features of the architecture, much of it developed for desktop applications, and to optimize it for a class of applications that is scientifically focused.”
Not that there wouldn't still be room for superspecialized machines. As IEEE Spectrum went to press, D.E.Shaw Research of New York City said that by the end of the year it will have a specialized machine, called Anton, that can simulate molecular interactions hundreds of times as fast as anything now available.
About Computing Sciences at Berkeley Lab
The Lawrence Berkeley National Laboratory (Berkeley Lab) Computing Sciences organization provides the computing and networking resources and expertise critical to advancing the Department of Energy's research missions: developing new energy sources, improving energy efficiency, developing new materials and increasing our understanding of ourselves, our world and our universe.
ESnet, the Energy Sciences Network, provides the high-bandwidth, reliable connections that link scientists at 40 DOE research sites to each other and to experimental facilities and supercomputing centers around the country. The National Energy Research Scientific Computing Center (NERSC) powers the discoveries of 7,000-plus scientists at national laboratories and universities, including those at Berkeley Lab's Computational Research Division (CRD). CRD conducts research and development in mathematical modeling and simulation, algorithm design, data storage, management and analysis, computer system architecture and high-performance software implementation. NERSC and ESnet are Department of Energy Office of Science User Facilities.
Lawrence Berkeley National Laboratory addresses the world's most urgent scientific challenges by advancing sustainable energy, protecting human health, creating new materials, and revealing the origin and fate of the universe. Founded in 1931, Berkeley Lab's scientific expertise has been recognized with 13 Nobel prizes. The University of California manages Berkeley Lab for the DOE’s Office of Science.
DOE’s Office of Science is the single largest supporter of basic research in the physical sciences in the United States, and is working to address some of the most pressing challenges of our time. For more information, please visit science.energy.gov.