While most personal computers today can process a few hundred thousand calculations per second, computer scientists are laying the groundwork for exascale machines that will process more than a million trillion – or 10^18 – calculations per second. Just a few months ago, scientists reached the long-sought-after high-performance computing milestone of one petaflop by processing more than a thousand trillion – or 10^15 – calculations per second.
“The need for exascale-sized machines is well-established,” said Karsten Schwan, a professor in the School of Computer Science in the College of Computing at the Georgia Institute of Technology. “With exascale machines, weather simulations will be able to operate at finer resolution, biologists will be able to model more complex systems, and businesses will be able to run and manage many applications at the same time on a single large machine.”
Schwan recently received a 2008 HP Labs Innovation Research Award to work with HP Labs, HP’s central research arm, to help solve some of the key problems in developing exascale machines. The high-impact research award, one of only two granted for exascale research and 41 granted overall to professors around the world, encourages open collaboration with HP Labs. The award amount is renewable for a total of three years based on research progress and HP business requirements.
With the petaflop barrier broken, researchers like Schwan are focusing on the next goal – improving that processing power a thousandfold to reach the exascale. Schwan’s expertise in high performance and enterprise computing will help him solve some of the challenges surrounding exascale systems.
“We believe that machines will reach exascale size only by combining common chips – such as quad core processors – with special purpose chips – such as graphics accelerators,” said Schwan, who is also director of the Georgia Tech Center for Experimental Research in Computer Systems (CERCS).
A challenge that arises from this scenario is how to efficiently run programs on these heterogeneous many-core chips. To investigate possible methods for doing this, Schwan will team with Georgia Tech School of Electrical and Computer Engineering professor Sudhakar Yalamanchili, an expert in heterogeneous many-core platforms.
Exascale machines must also be able to run multiple systems and applications on a single platform at the same time, while guaranteeing that they will not interfere with each other. An approach called virtualization may help solve this challenge by hiding some of the underlying computer architecture issues from applications.
“With virtualization, decisions have to be made about where, when and for how long certain programs should run, but there are many ways of determining what might be appropriate because there might be multiple goals,” explained Schwan. “For instance, one might want to minimize the exascale machine’s power consumption while at the same time meet some performance goal for the application. In other words, virtualized systems must be actively ‘managed’ to attain end user, institutional or corporate goals.”
Ada Gavrilovska, a specialist in virtualization and multi-core operation and research scientist in the School of Computer Science in the College of Computing, will collaborate with Schwan to determine how to manage multiple programs on exascale machines that consist of hundreds of thousands of processors.
Though exascale machines are high-performance computing systems, the vision for these future systems goes beyond the typical vision painted for high performance computing. Instead of scaling a single program to run on hundreds of thousands of cores, exascale systems will also be used to run multiple programs on a single large machine.
“This future virtualized and managed exascale system will guarantee some level of service even when parts of the machine get too loaded or too hot or fail, since applications can be moved while they are running,” said Schwan.
Though it will be several years before exascale systems are developed, scientists at Georgia Tech will use the HP Labs Innovation Research Award to lay the foundation for solving emerging science and engineering challenges in national defense, energy assurance, advanced materials and climate.
“Around the world, HP partners with the best and the brightest in industry and academia to drive open innovation and set the agenda for breakthrough technologies that are designed to change the world,” said Prith Banerjee, senior vice president of research at HP and director of HP Labs. “HP Labs’ selection of Karsten Schwan for a 2008 Innovation Award demonstrates outstanding achievement and will help accelerate HP Labs’ global research agenda in pursuit of scientific breakthroughs.”
Abby Vogel | Newswise Science News
A novel hybrid UAV that may change the way people operate drones
28.03.2017 | Science China Press
Timing a space laser with a NASA-style stopwatch
28.03.2017 | NASA/Goddard Space Flight Center
The Institute of Semiconductor Technology and the Institute of Physical and Theoretical Chemistry, both members of the Laboratory for Emerging Nanometrology (LENA), at Technische Universität Braunschweig are partners in a new European research project entitled ChipScope, which aims to develop a completely new and extremely small optical microscope capable of observing the interior of living cells in real time. A consortium of 7 partners from 5 countries will tackle this issue with very ambitious objectives during a four-year research program.
To demonstrate the usefulness of this new scientific tool, at the end of the project the developed chip-sized microscope will be used to observe in real-time...
Astronomers from Bonn and Tautenburg in Thuringia (Germany) used the 100-m radio telescope at Effelsberg to observe several galaxy clusters. At the edges of these large accumulations of dark matter, stellar systems (galaxies), hot gas, and charged particles, they found magnetic fields that are exceptionally ordered over distances of many million light years. This makes them the most extended magnetic fields in the universe known so far.
The results will be published on March 22 in the journal „Astronomy & Astrophysics“.
Galaxy clusters are the largest gravitationally bound structures in the universe. With a typical extent of about 10 million light years, i.e. 100 times the...
Researchers at the Goethe University Frankfurt, together with partners from the University of Tübingen in Germany and Queen Mary University as well as Francis Crick Institute from London (UK) have developed a novel technology to decipher the secret ubiquitin code.
Ubiquitin is a small protein that can be linked to other cellular proteins, thereby controlling and modulating their functions. The attachment occurs in many...
In the eternal search for next generation high-efficiency solar cells and LEDs, scientists at Los Alamos National Laboratory and their partners are creating...
Silicon nanosheets are thin, two-dimensional layers with exceptional optoelectronic properties very similar to those of graphene. Albeit, the nanosheets are less stable. Now researchers at the Technical University of Munich (TUM) have, for the first time ever, produced a composite material combining silicon nanosheets and a polymer that is both UV-resistant and easy to process. This brings the scientists a significant step closer to industrial applications like flexible displays and photosensors.
Silicon nanosheets are thin, two-dimensional layers with exceptional optoelectronic properties very similar to those of graphene. Albeit, the nanosheets are...
20.03.2017 | Event News
14.03.2017 | Event News
07.03.2017 | Event News
28.03.2017 | Life Sciences
28.03.2017 | Information Technology
28.03.2017 | Physics and Astronomy