Modeling Climate at Warp Speed
NASA scientists have managed to squeeze about 10 times more power out of cutting-edge "parallel" supercomputers through innovations in software and memory design. This leap in effective computing "muscle" -- together with the data from NASA's Earth-observing satellites -- enables greater realism and statistical confidence in simulations of global climate.
Above: New SGI® Origin® 3000 Series supercomputers, shown here, together with software tools developed at NASA's Ames Research Center are producing a 10-fold improvement in computing power. The computer on the right is the size of a tall filing cabinet. Image courtesy SGI.
Whether the question concerns the path of an approaching hurricane or the rise in global temperatures over the next century, predictions always carry some amount of uncertainty. But the computer "models" that produce the simulations can be improved so that this uncertainty is reduced.
Making these improvements will require all the computing power scientists can get their hands on.
To provide the immense "number crunching" power needed for demanding scientific applications such as climate simulation, some computer makers are turning to "parallel" designs with hundreds or thousands of processors in one supercomputer.
The numbers on these machines make even the fastest desktop computers look like pocket calculators. For example, the newest supercomputer at NASA's Goddard Space Flight Center boasts 512 processors each running at 400 MHz, 128 GB of RAM, 2,800 GB of disk space, and a peak performance of 409 gigaflops! (A "gigaflop" is a billion calculations per second.) A newer machine at NASA's Ames Research Center will top even this with 1,024 processors.
Above: The bottom line shows how the increase in computing power (gigaflops) normally tapers off as the number of processors increases. The top line shows the performance on the same processors using the software tools developed by NASA. [more]
But simply adding more processors doesn't guarantee a proportionate increase in effective power. In fact, the full potential of these parallel supercomputers still has not been tapped.
"So what's the problem? Each node (i.e., processor) has certain performance," Asrar explains. "Individually they perform well, but as you add them all together, as the number of nodes goes up, the overall efficiency degrades." For example, a system with 100 processors would not have 100 times the power of a single processor -- the overall performance would be somewhat lower.
This loss of computing efficiency is a bit like what happens when people must work together to get a task done. Some effort must go into managing and coordinating the people involved -- effort that's diverted away from producing anything -- and even the productive workers must spend some amount of time communicating with each other. In a similar way, a supercomputer with more processors must use more of its power to coordinate those processors, and the increased communication between all the processors bogs the system down.
Left: Simulations of Earth's climate are among the most computationally intensive problems in science. Click on the image to view an MPG movie.
For the past four years, scientists at NASA's Ames Research Center have been working in partnership with computer maker Silicon Graphics, Inc., to tackle this problem. The fruits of their labor are two new technologies that increase the effective power of these machines by roughly an order of magnitude (that is, a factor of 10). Both technologies are freely available to the supercomputing community, are computer vendor independent, and are not specific to climate modeling.
The first of these technologies is a memory architecture called "single-image shared memory." In this design, all of the supercomputer's memory is used as one continuous memory space by all of the processors. (Other architectures distribute the memory among the processors.) This lets the processors exchange the messages needed to coordinate their efforts by accessing this "common ground" of memory. This scheme is more efficient than passing the messages directly between the processors, as most parallel supercomputers do.
Below: Using faster computers, forecasters will be able to narrow the estimated path of hurricanes and perhaps save millions of dollars in unneeded evacuations. Image courtesy NOAA.
The extra computing power milked from the processors by these technologies will help NASA's Earth Science Enterprise make better models of Earth's climate.
These models work by dividing the atmosphere and oceans up into a 3-dimensional grid of boxes. These boxes are assigned values for temperature, moisture content, chemical content, and so on, and then the interactions between the boxes are calculated using equations from physics and chemistry. The result is an approximation of the real system.
With more computing power available, more of the physics of the real climate system can be incorporated into the models, and the atmosphere can be divided into more, smaller boxes. This makes the models more realistic, and the predictions they will produce will be of more interest on a regional scale.
Also, the ability to run these models faster will mean that more simulations can be performed, which will produce a larger pool of results. In statistical terms, this larger "population" will allow for a better analysis of the strength of the conclusions.
Above: The software tools developed by NASA and SGI can be used for other simulations, too. Show here is a supercomputer model of a human protein. [more]
NASA's suite of Earth-observing satellites, together with a global network of meteorological stations, provide the dose of real-world data that is needed to keep the models on track. And the archives of this data provide the ultimate proving grounds for the models: Can the computers accurately "predict" the real weather observed in the past?
Asrar says that the computer models are already quite good at this, but there's still room for improvement. As supercomputers continue to advance -- along with the software that taps that power -- climate models will become more and more accurate, offering better answers to the vexing questions of climate change.
New 512-processor supercomputer -- short article about the 512-processor SGI machine and some of the science applications for which it will be used
NASA Advanced Supercomputing Division -- home page, contains lots of information about the supercomputing resources at NASA and the many science and engineering challenges that make use of them.
SGI 1,024-processor supercomputer -- short article from SGI about the newest member of the Origin® 3000 Series of supercomputers
NASA's Earth Science Enterprise -- home page of the program at NASA that uses satellite and space technology to better understand the Earth's natural processes
Supercomputing resources at NASA -- information on NASA's major supercomputers
What is data assimilation? -- explanation and links related to the important process of merging satellite data into computer simulations of climate
Join our growing list of subscribers - sign up for our express news delivery and you will receive a mail message every time we post a new story!!!