Reynolds collaborated with astrophysicists at the University of California at San Diego as part of a National Science Foundation project to simulate cosmic reionization, the time from 380,000 years to 400 million years after the universe was born.
Together the scientists built a computer model of events during the "Dark Ages" when the first stars emitted radiation that altered the surrounding matter, enabling light to pass through. The team tested its model on two of the largest existing NSF supercomputers, "Ranger" at the University of Texas at Austin and "Kraken" at the University of Tennessee.
The new mathematical model tightly couples a myriad of physical processes present during cosmic reionization, such as gas motion, radiation transport, chemical kinetics and gravitational acceleration due to star clustering and dark matter dynamics, Reynolds says.
The key characteristic of the model that differentiates it from competing work is that the researchers focused on enforcing a very tight coupling in the model between the different physical processes.
"By forcing the computational methods to tightly bind these processes together, our new model allows us to generate simulations that are highly accurate, numerically stable and computationally scalable to the largest supercomputers available," Reynolds says.
They presented their research at a Texas Cosmology Network Meeting at UT in late October. Reynolds' mathematical research also was published as "Self-Consistent Solution of Cosmological Radiation-Hydrodynamics and Chemical Ionization" in the October issue of the "Journal of Computational Physics."
Simulation models typically consist of a complex bundle of mathematical equations representing physical processes. The equations are integrated to reflect interaction of the physical processes. Only supercomputers can simultaneously solve the equations. Scientific intuition and creativity come into play by developing the base model with equations with the best parameters, Reynolds says. Variables can be altered to describe different scenarios that might have occurred. The objective is to develop a simulation model with results that most closely resemble telescope observations and that predict a universe that looks like what we have. If that happens, scientists have discovered the set of physical processes that existed at the birth of the universe as it was evolving from one instant to the next.Physical processes include the heating of various gases, gravity, the conservation of mass, the conservation of momentum, the conservation of energy, expansion of the universe, the transport of radiation, and the chemical ionization of different species such as Hydrogen and Helium, the primary elements present at the beginning of the universe. An additional equation running in the background describes and models the dynamics of dark matter — the majority of the matter in the universe — which gives rise to gravity and is attributed with helping the universe form stars, planets and galaxies.
"Supercomputers are so big, they hold so much data, you can build models that work with many processes at one time," Reynolds says. "A lot of these processes behave nonlinearly. When they are put together, they inhibit each other, feed off each other, so you end up with many different processes when they are put together."
A direct consequence of the tight coupling that the researchers enforce in their model is that the resulting system of equations is much more complex than those that must be solved by other models, Reynolds says.
"This paper describes both how we form the coupled model, as well as the mathematical methods that enable us to solve the systems of equations that result. These include methods that accurately track the different time scales of each process, which often occur at rates that vary by orders of magnitude," he says. "However, perhaps the most important contribution of this paper is our description of how we pose the complex interaction of different models as a nonlinear problem with potentially billions of equations and unknowns, and solve that problem using new algorithms designed for next-generation supercomputers. We conclude by demonstrating that the new model lives up to the ideal, providing an approach that allows high accuracy, stability and scalability on a suite of difficult test problems."
Only recently have mathematics algorithms been invented to solve basic problems — like diffusion of heat — using resources as large as those available on modern supercomputers, Reynolds says. There have been simple analytical solutions to many problems from mathematical physics for hundreds of years. However, those analytical solutions only work when scientists simplify the problem in some way or another. For example, he says, they may approximate the shape of a planet as a sphere, instead of an ellipsoid, or may assume that ocean water is incompressible, which only works for very shallow water, or assume the Earth is homogeneous, instead of formed using widely differing layers of rock.
"Scientists have been able to approximate a great many physical processes in such idealized situations. But the true frontier nowadays is to let go of these simplifying approximations and treat the problems as they really are, by modeling all of the geometric structure and the in-homogeneity," Reynolds says. "To do that, you need to solve harder equations with lots of data, which is ideally suited to using supercomputers. The numerical methods that can allow us to use larger and larger computers have only just come out. The problems are getting more challenging and harder to solve, but the numerical methods are reaching greater capability, so you can really start moving them forward. These new computers make everything a new frontier."
Besides Reynolds, other researchers were John C. Hayes, Lawrence Livermore National Laboratory, Livermore, Calif.; Pascal Paschos, Center for Astrophysics and Space Sciences, University of California at San Diego, La Jolla, Calif.; and Michael L. Norman, Center for Astrophysics and Space Sciences, and physics department, the University of California at San Diego, La Jolla. — Margaret Allen
Kim Cobb | EurekAlert!
Scientific achievements during the operation of Lomonosov satellite
18.12.2017 | Lomonosov Moscow State University
Quantum memory with record-breaking capacity based on laser-cooled atoms
18.12.2017 | Faculty of Physics University of Warsaw
A study carried out by an international team of researchers and published in the journal Physical Review X shows that ion-trap technologies available today are suitable for building large-scale quantum computers. The scientists introduce trapped-ion quantum error correction protocols that detect and correct processing errors.
In order to reach their full potential, today’s quantum computer prototypes have to meet specific criteria: First, they have to be made bigger, which means...
Since 2016, German and Spanish researchers, among them scientists from the University of Göttingen, have been hunting for exoplanets with the “Carmenes”...
DNA molecules that follow specific instructions could offer more precise molecular control of synthetic chemical systems, a discovery that opens the door for engineers to create molecular machines with new and complex behaviors.
Researchers have created chemical amplifiers and a chemical oscillator using a systematic method that has the potential to embed sophisticated circuit...
MPQ scientists achieve long storage times for photonic quantum bits which break the lower bound for direct teleportation in a global quantum network.
Concerning the development of quantum memories for the realization of global quantum networks, scientists of the Quantum Dynamics Division led by Professor...
Researchers have developed a water cloaking concept based on electromagnetic forces that could eliminate an object's wake, greatly reducing its drag while...
11.12.2017 | Event News
08.12.2017 | Event News
07.12.2017 | Event News
18.12.2017 | Information Technology
18.12.2017 | Physics and Astronomy
18.12.2017 | Agricultural and Forestry Science