Âé¶¹ÒùÔº

May 19, 2025

Novel data streaming software chases light speed from accelerator to supercomputer

Under construction at Berkeley Lab, the Gamma-Ray Energy Tracking Array, or GRETA, is a cutting-edge instrument for nuclear physics. Software developed at ORNL will enable GRETA to send data to the nation's fastest supercomputers for immediate analysis, radically reducing data storage needs and enabling real-time experiment adjustments for optimal results. Credit: Jason Smith / ORNL, Berkeley Lab, U.S. Dept of Energy
× close
Under construction at Berkeley Lab, the Gamma-Ray Energy Tracking Array, or GRETA, is a cutting-edge instrument for nuclear physics. Software developed at ORNL will enable GRETA to send data to the nation's fastest supercomputers for immediate analysis, radically reducing data storage needs and enabling real-time experiment adjustments for optimal results. Credit: Jason Smith / ORNL, Berkeley Lab, U.S. Dept of Energy

Analyzing massive datasets from nuclear physics experiments can take hours or days to process, but researchers are working to radically reduce that time to mere seconds using special software being developed at the Department of Energy's Lawrence Berkeley and Oak Ridge national laboratories.

DELERIA—short for Distributed Event-Level Experiment Readout and Integrated Analysis—is a novel software platform designed specifically to support the GRETA spectrometer, a cutting-edge instrument for nuclear physics experiments. The Gamma Ray Energy Tracking Array (GRETA), is currently under construction at Berkeley Lab and is scheduled to be installed in 2026 at the Facility for Rare Isotope Beams (FRIB), at Michigan State University.

The software will enable GRETA to stream data directly to the nation's leading computing centers with the goal of analyzing large datasets in seconds. The data will be sent via the Energy Sciences Network, or ESnet. This will allow researchers to make critical adjustments to the experiment as it is taking place, leading to increased scientific productivity with significantly faster, more accurate results.

"Our primary goal is to establish a data pipeline for future experimental needs without having to build a local computing infrastructure. The pipeline will also allow us to increase the amount of analysis that is performed online," said Gustav Jansen, an ORNL computational nuclear physicist who has been developing DELERIA for the last two years. "But we also intend to show that DELERIA can be scaled to work with other research facilities and a broader range of scientific applications."

GRETA is a spherical array of 120 hyper-pure germanium crystals. Experimental samples are placed inside the instrument and bombarded with charged particles from the radioactive ion accelerator at the FRIB facility.

The collisions emit hundreds of thousands of photons every second. Each time a photon hits a crystal, the interaction is recorded and transmitted over the pipeline for analysis. This process is ultimately used to reveal how protons and neutrons in the nuclei of a sample are arranged and how they behave.

Significant computing resources are required to calculate the energies and coordinates of the photons and process the information in a way that is meaningful to researchers.

"Establishing a data pipeline allows us to offload the heavy computing to at the national labs. So, instead of building a whole new computing system for a potential upgrade to GRETA, we can simply connect to one," Jansen said. "And from a nuclear physics standpoint, this presents us with options for upgrading GRETA, as well as instruments at other facilities that will also be able to leverage the DELERIA software platform."

The GRETA data pipeline has five steps (left to right): Filter boards detect events, the forward buffer queues them, and Defiant at OLCF processes and reduces the data before sending it back to Berkeley Lab, where it’s aggregated for storage or further analysis. Credit: Jason Smith / ORNL, U.S. Dept of Energy
× close
The GRETA data pipeline has five steps (left to right): Filter boards detect events, the forward buffer queues them, and Defiant at OLCF processes and reduces the data before sending it back to Berkeley Lab, where it’s aggregated for storage or further analysis. Credit: Jason Smith / ORNL, U.S. Dept of Energy

Get free science updates with Science X Daily and Weekly Newsletters — to customize your preferences!

Testing, testing

Together, Jansen and his colleagues at the Oak Ridge Leadership Computing Facility, a DOE Office of Science user facility located at ORNL, and Berkeley Lab built a testbed of the GRETA data pipeline. At Berkeley Lab, a virtual version of the GRETA detector array simulates photon collision events and transmits the data via ESnet to a computing cluster at OLCF more than 2,000 miles away.

"The crystals inside the detector have lots of electrical contacts connected to them. When a photon hits one of the crystals, an electrical signal goes out on each contact point," Jansen said. "We need to figure out where inside the crystal the photons hit—the X, Y, Z coordinates. And we need to do that within 10 seconds of the collision. That's what we designed the pipeline to do."

The pipeline's processing power is provided by OLCF's Defiant, a 36-node AMD GPU-accelerated computing cluster. Defiant's job is to do the number crunching to determine the X, Y, Z coordinates of the photon collisions and send the processed data back to Berkeley Lab for storage or additional analysis. DELERIA also significantly reduces the volume of stored data by 97.5%, which is a factor of 40.

The testbed can generate and analyze approximately 480,000 photon collision events every second. The data must travel more than 4,000 miles there and back again and be processed all in 10 seconds or less.

The GRETA testbed is part of OLCF's Advanced Computing Ecosystem (ACE), a testbed that provides computing and data resources for a wide range of system architectures.

Tricking time

In a series of demonstrations, the team has been able to process up to 35 gigabits (Gbps) of data per second, and transmit data at a rate of 35 gigabits per second, which actually surpasses GRETA's required maximum rate and is easily accommodated by ESnet's ESnet6 network, which has multiple 400 Gbps paths between Berkeley Lab and ORNL. The team intends to achieve even higher rates. According to Jansen, the big problem with moving data across the country is latency.

Currently, their system can process a single photon collision event in 5 milliseconds, or 0.005 seconds. But, traveling at , it takes 120 milliseconds (0.12 seconds) to transmit data from Berkeley Lab to OLCF and back—more than an order of magnitude longer than it takes to process a single event.

"Of course, only in can you bypass the speed of light," Jansen said. "What we're working on now is to get a 10-times speedup by tricking latency, so to speak. Instead of running one event at a time, the solution is to run events in parallel so that we can process one event while others are being transferred. By finding the right balance, we can make sure that the computing cluster is always busy. Otherwise, our analysis will take 10 times longer than it should."

When it comes to moving faster than the speed of light, "Only Gustav can do that," said Tom Beck, jokingly. Beck is the section head for science engagement in ORNL's National Center for Computation Sciences. In addition to supporting Jansen, Beck is also working on other IRI projects that connect major DOE research facilities with leadership-class computing resources at the OLCF.

"This is really pioneering work in the IRI theme that's setting the standards for how we network the nation's leading centers of research. And the more use cases we have, the better we're going to get," Beck said. "One case at a time. And so far, this is really the first case that's up and working at this level of maturity that I'm aware of, which means we're off to a really great start."

In addition to Jansen, the project team members include Mario Cromaz (Berkeley Lab), Eli Dart, Eric Pouyoul, Ezra Kissel, Kiran Vasu, Seyoung Yu (ESnet), and Ross Miller (OLCF).

Load comments (0)

This article has been reviewed according to Science X's and . have highlighted the following attributes while ensuring the content's credibility:

fact-checked
trusted source
proofread

Get Instant Summarized Text (GIST)

A new software platform, DELERIA, enables real-time streaming and analysis of large nuclear physics datasets from the GRETA spectrometer to remote supercomputers via ESnet. The system processes photon collision events in milliseconds and reduces stored data volume by 97.5%, allowing rapid, accurate experimental adjustments and scalable application to other research facilities.

This summary was automatically generated using LLM.