麻豆淫院

July 30, 2009

Seeking efficiency, scientists run visualizations directly on supercomputers

Argonne scientists are working to improve techniques for capturing complex visualizations like this astrophysics simulation, which seeks to discover the mechanism behind the violent death of short-lived, massive stars. The image shows entropy values in the core of the supernova; different colors and transparencies are assigned to different values of entropy. By selectively adjusting the color and transparency, the scientist can peel away outer layers and see values in the interior of the 3-D volume. Image courtesy Hongfeng Yu.
× close
Argonne scientists are working to improve techniques for capturing complex visualizations like this astrophysics simulation, which seeks to discover the mechanism behind the violent death of short-lived, massive stars. The image shows entropy values in the core of the supernova; different colors and transparencies are assigned to different values of entropy. By selectively adjusting the color and transparency, the scientist can peel away outer layers and see values in the interior of the 3-D volume. Image courtesy Hongfeng Yu.

(麻豆淫院Org.com) -- If you wanted to perform a single run of a current model of the explosion of a star on your home computer, it would take more than three years just to download the data. In order to do cutting-edge astrophysics research, scientists need a way to more quickly compile, execute and especially visualize these incredibly complex simulations.

These days, many scientists generate quadrillions of data points for visualizations of everything from supernovas to protein structures鈥攁nd they鈥檙e quickly overwhelming current computing capabilities. Scientists at the U.S. Department of Energy's Argonne National Laboratory are exploring other ways to speed up the process, using a technique called .

Volume rendering is a technique that can be used to make sense of the billions of tiny points of data collected from an X-ray, MRI, or a researcher鈥檚 simulation. For example, bone is denser than muscle, so an MRI measuring the densities of every square millimeter of your arm will register the higher readings for the radius bone in your forearm.

Argonne scientists are trying to find better, quicker ways to form a recognizable image from all of these points of data. Equations can be written to search for sudden density changes in the dataset that might set bone apart from muscle, and researchers can create a picture of the entire arm, with bone and muscle tissue in different colors.

鈥淏ut on the scale that we鈥檙e working, creating a movie would take a very long time on your laptop鈥攋ust rotating the image one degree could take days,鈥 said Mark Hereld, who leads the and analysis efforts at the Argonne Leadership .

Get free science updates with Science X Daily and Weekly Newsletters 鈥 to customize your preferences!

First, researchers divide the data among many processing cores so that they can all work at once, a technique that鈥檚 called parallel computing. On Argonne鈥檚 Blue Gene/P , 160,000 computing cores all work together in parallel. Today鈥檚 typical laptop, by comparison, has two cores.

Usually, the supercomputer鈥檚 work stops once the data ha颅s been gathered, and the data is sent to a set of graphics processors (GPUs), which create the final visualizations. But the driving commercial force behind developing GPUs has been the video game industry, so GPUs aren鈥檛 always well suited for scientific tasks. In addition, the sheer amount of data that has to be transferred between computers eats up valuable time and disk space.

鈥淚t鈥檚 so much data that we can鈥檛 easily ask all of the questions that we want to ask: each new answer creates new questions and it just takes too much time to move the data from one calculation to the next,鈥 said Hereld. 鈥淭hat drives us to look for better and more efficient ways to organize our computational work.鈥

Argonne researchers wanted to know if they could improve performance by skipping the transfer to the GPUs and instead performing the visualizations right there on the supercomputer. They tested the technique on a set of astrophysics data and found that they could indeed increase the efficiency of the operation.

鈥淲e were able to scale up to large problem sizes of over 80 billion voxels per time step and generated images up to 16 megapixels,鈥 said Tom Peterka, a postdoctoral appointee in Argonne鈥檚 Mathematics and Computer Science Division.

Because the Blue Gene/P's main processor can visualize data as they are analyzed, Argonne's scientists can investigate physical, chemical, and biological phenomena with much more spatial and temporal detail.

According to Hereld, this new visualization method could enhance research in a wide variety of disciplines. 鈥淚n astrophysics, studying how stars burn and explode pulls together all kinds of physics: hydrodynamics, gravitational physics, nuclear chemistry and energy transport,鈥 he said. 鈥淥ther models study the migration of dangerous pollutants through complex structures in the soil to see where they鈥檙e likely to end up; or combustion in cars and manufacturing plants鈥攚here fuel is consumed and whether it鈥檚 efficient.鈥

鈥淭hose kinds of problems often lead to questions that are very complicated to pose mathematically,鈥 Hereld said. 鈥淏ut when you can simply watch a star explode through visualization of the simulation, you can gain insight that鈥檚 not available any other way.鈥

Argonne鈥檚 work in advanced computing is supported by the Department of Energy鈥檚 Office of Advanced Scientific Computing Research.

Provided by Argonne National Laboratory ( : )

Load comments (3)

This article has been reviewed according to Science X's and . have highlighted the following attributes while ensuring the content's credibility:

Get Instant Summarized Text (GIST)

This summary was automatically generated using LLM.