It's almost a rite of passage in physics and astronomy. Scientists spend years scrounging up money to build a fantastic new instrument. Then, when the long-awaited device finally approaches completion, the panic begins: How will they handle the torrent of data?
That's the situation now, at least, with the Square Kilometer Array (SKA), a radio telescope planned for Africa and Australia that will have an unprecedented ability to deliver data -- lots of data points, with lots of details -- on the location and properties of stars, galaxies and giant clouds of hydrogen gas.
In a study published in The Astronomical Journal, a team of scientists at the University of Wisconsin-Madison has developed a new, faster approach to analyzing all that data.
Hydrogen clouds may seem less flashy than other radio telescope targets, like exploding galaxies. But hydrogen is fundamental to understanding the cosmos, as it is the most common substance in existence and also the "stuff" of stars and galaxies.
As astronomers get ready for SKA, which is expected to be fully operational in the mid-2020s, "there are all these discussions about what we are going to do with the data," says Robert Lindner, who performed the research as a postdoctoral fellow in astronomy and now works as a data scientist in the private sector. "We don't have enough servers to store the data. We don't even have enough electricity to power the servers. And nobody has a clear idea how to process this tidal wave of data so we can make sense out of it."
Lindner worked in the lab of Associate Professor Snezana Stanimirovic, who studies how hydrogen clouds form and morph into stars, in turn shaping the evolution of galaxies like our own Milky Way.
In many respects, the hydrogen data from SKA will resemble the vastly slower stream coming from existing radio telescopes. The smallest unit, or pixel, will store every bit of information about all hydrogen directly behind a tiny square in the sky. At first, it is not clear if that pixel registers one cloud of hydrogen or many -- but answering that question is the basis for knowing the actual location of all that hydrogen.
People are visually oriented and talented in making this interpretation, but interpreting each pixel requires 20 to 30 minutes of concentration using the best existing models and software. So, Lindner asks, how will astronomers interpret hydrogen data from the millions of pixels that SKA will spew? "SKA is so much more sensitive than today's radio telescopes, and so we are making it impossible to do what we have done in the past."
In the new study, Lindner and colleagues present a computational approach that solves the hydrogen location problem with just a second of computer time.
For the study, UW-Madison postdoctoral fellow Carlos Vera-Ciro helped write software that could be trained to interpret the "how many clouds behind the pixel?" problem. The software ran on a high-capacity computer network at UW-Madison called HTCondor. And "graduate student Claire Murray was our 'human,'" Lindner says. "She provided the hand-analysis for comparison."
Those comparisons showed that as the new system swallows SKA's data deluge, it will be accurate enough to replace manual processing.
Ultimately, the goal is to explore the formation of stars and galaxies, Lindner says. "We're trying to understand the initial conditions of star formation -- how, where, when do they start? How do you know a star is going to form here and not there?"
To calculate the overall evolution of the universe, cosmologists rely on crude estimates of initial conditions, Lindner says. By correlating data on hydrogen clouds in the Milky Way with ongoing star formation, data from the new radio telescopes will support real numbers that can be entered into the cosmological models.
"We are looking at the Milky Way, because that's what we can study in the greatest detail," Lindner says, "but when astronomers study extremely distant parts of the universe, they need to assume certain things about gas and star formation, and the Milky Way is the only place we can get good numbers on that."
With automated data processing, "suddenly we are not time-limited," Lindner says. "Let's take the whole survey from SKA. Even if each pixel is not quite as precise, maybe, as a human calculation, we can do a thousand or a million times more pixels, and so that averages out in our favor."
Robert Lindner | EurekAlert!
NASA detects solar flare pulses at Sun and Earth
17.11.2017 | NASA/Goddard Space Flight Center
Pluto's hydrocarbon haze keeps dwarf planet colder than expected
16.11.2017 | University of California - Santa Cruz
The formation of stars in distant galaxies is still largely unexplored. For the first time, astron-omers at the University of Geneva have now been able to closely observe a star system six billion light-years away. In doing so, they are confirming earlier simulations made by the University of Zurich. One special effect is made possible by the multiple reflections of images that run through the cosmos like a snake.
Today, astronomers have a pretty accurate idea of how stars were formed in the recent cosmic past. But do these laws also apply to older galaxies? For around a...
Just because someone is smart and well-motivated doesn't mean he or she can learn the visual skills needed to excel at tasks like matching fingerprints, interpreting medical X-rays, keeping track of aircraft on radar displays or forensic face matching.
That is the implication of a new study which shows for the first time that there is a broad range of differences in people's visual ability and that these...
Computer Tomography (CT) is a standard procedure in hospitals, but so far, the technology has not been suitable for imaging extremely small objects. In PNAS, a team from the Technical University of Munich (TUM) describes a Nano-CT device that creates three-dimensional x-ray images at resolutions up to 100 nanometers. The first test application: Together with colleagues from the University of Kassel and Helmholtz-Zentrum Geesthacht the researchers analyzed the locomotory system of a velvet worm.
During a CT analysis, the object under investigation is x-rayed and a detector measures the respective amount of radiation absorbed from various angles....
The quantum world is fragile; error correction codes are needed to protect the information stored in a quantum object from the deteriorating effects of noise. Quantum physicists in Innsbruck have developed a protocol to pass quantum information between differently encoded building blocks of a future quantum computer, such as processors and memories. Scientists may use this protocol in the future to build a data bus for quantum computers. The researchers have published their work in the journal Nature Communications.
Future quantum computers will be able to solve problems where conventional computers fail today. We are still far away from any large-scale implementation,...
Pillared graphene would transfer heat better if the theoretical material had a few asymmetric junctions that caused wrinkles, according to Rice University...
15.11.2017 | Event News
15.11.2017 | Event News
30.10.2017 | Event News
17.11.2017 | Physics and Astronomy
17.11.2017 | Health and Medicine
17.11.2017 | Studies and Analyses