HLRS Supercomputer Successfully Executed Extreme-Scale Simulation Projects
Supercomputer Hornet of the High Performance Computing Center Stuttgart (HLRS) is ready for extreme-scale computing challenges. The newly installed HPC system (High Performance Computing) successfully finished extensive simulation projects that by far exceeded the calibre of previously performed simulation runs at HLRS:
Der HLRS-Supercomputer Hornet, ein Cray XC40 System, liefert eine Rechenhöchstleistung von 3,8 Petaflops (3,8 Billiarden Rechenoperationen pro Sekunde).
(c) Boris Lehner für HLRS/Universität Stuttgart
Six so called XXL-Projects from computationally demanding scientific fields such as planetary research, climatology, environmental chemistry, aerospace, and scientific engineering were recently applied on the HLRS supercomputer. With each application scaling up to all of Hornet’s available 94,646 compute cores, the machine was put through a demanding endurance test. The achieved results more than satisfied the HLRS HPC experts as well as the scientific users: Hornet lived up to the challenge and passed the simulation “burn-in runs” with flying colors.
The new HLRS supercomputer Hornet, a Cray XC40 system which in its current configuration delivers a peak performance of 3.8 PetaFlops (1 PetaFlops = 1 quadrillion floating point operations per second), was declared “up and running” in late 2014.
In its early installation phase, prior to making the machine available for general use, HLRS had invited national scientists and researchers from various fields to apply large-scale simulation projects on Hornet. The goal was to deliver evidence that all related HPC hardware and software components required to smoothly run highly complex and extreme-scale compute jobs are up and ready for top-notch challenges. Six perfectly suited XXL-Projects were identified and implemented on the HLRS supercomputer:
(1) “Convection Permitting Channel Simulation”, Institute of Physics and Meteorology, Universität Hohenheim
(84,000 compute cores, 84 compute hours, 330 TB of data + 120 TB for pre-processing)
Objective: Run a latitude belt simulation around the Earth at a resolution of a few km for a time period long enough to cover various extreme events on the Northern hemisphere and to study the model performance.
(2) “Direct Numerical Simulation of a Spatially-Developing Turbulent Boundary Along a Flat Plate”, Institute of Aerodynamics and Gas Dynamics (IAG), Universität Stuttgart
(93,840 compute cores, 70 machine hours, 30 TB of data)
Objective: To conduct a direct numerical simulation of the complete transition of a boundary layer flow to fully-developed turbulence along a flat plate up to high Reynolds numbers.
(3) “Prediction of the Turbulent Flow Field Around a Ducted Axial Fan”, Institute of Aerodynamics, RWTH Aachen University
(92,000 compute cores, 110 machine hours, 80 TB of data)
Objective: To better understand the development of vortical flow structures and the turbulence intensity in the tip-gap of a ducted axial fan.
(4) “Large-Eddy Simulation of a Helicopter Engine Jet”, Institute of Aerodynamics, RWTH Aachen University
(94,646 compute cores, 300 machine hours, 120 TB of data)
Objective: Analysis of the impact of internal perturbations due to geometric variations on the flow field and the acoustic field of a helicopter engine jet.
(5) “Ion Transport by Convection and Diffusion“, Institute of Simulation Techniques and Scientific Computing, Universität Siegen
(94.080 compute cores, 5 machine hours, 1.1 TB of data)
Objective: To better understand and optimize the electrodialysis desalination process.
(6) “Large Scale Numerical Simulation of Planetary Interiors”, German Aerospace Center/Technische Universität Berlin
(54,000 compute cores, 3 machine hours, 2 TB of data)
Objective: To study the effect of heat driven convection within planets on the evolution of a planet (how is the surface influenced, how are conditions for life maintained, how do plate-tectonics work, and how quickly can a planet cool).
Demand for High Performance Computing on the Rise
Demand for High Performance Computing is unbroken. Scientists continue to crave for ever increasing computing power. They are eagerly awaiting the availability of even faster systems and better scalable software enabling them to attack and puzzle out the most challenging scientific and engineering problems. “Supply generates demand”, states Prof. Dr.-Ing. Michael M. Resch, Director of HLRS. “With the abilities of ultra-fast machines like Hornet both industry and researchers are quickly realizing that fully leveraging the vast capabilities of such a supercomputer opens unprecedented opportunities and helps them deliver results previously impossible to obtain. We are positive that our HPC infrastructure will be leveraged to its full extent. Hornet will be an invaluable tool in supporting researchers in their pursuit for answers to the most pressing subjects of today’s time, leading to scientific findings and knowledge of great and enduring value,” adds Professor Resch.
Following its ambitious technology roadmap, HLRS is currently striving to implement a planned system expansion which is scheduled to be completed by the end of 2015. The HLRS supercomputing infrastructure will then deliver a peak performance of more than seven PetaFlops (quadrillion mathematical calculations per second) and feature 2.3 petabytes of additional file system storage.
More information about the HLRS XXL-Projects can be found at http://www.gauss-centre.eu/gauss-centre/EN/Projects/XXL_Projects_Hornet/XXL_Proj...
About HLRS: The High Performance Computing Center Stuttgart (HLRS) of the University of Stuttgart is one of the three German supercomputer institutions forming the national Gauss Centre for Supercomputing. HLRS supports German and pan-European researchers as well as industrial users with leading edge supercomputing technology, HPC trainings, and support.
About GCS: The Gauss Centre for Supercomputing (GCS) combines the three national supercom-puting centres HLRS (High Performance Computing Center Stuttgart), JSC (Jülich Supercomputing Centre), and LRZ (Leibniz Supercomputing Centre, Garching near Munich) into Germany’s Tier-0 supercomputing institution. Concertedly, the three centres provide the largest and most powerful supercomputing infrastructure in all of Europe to serve a wide range of industrial and research activities in various disciplines. They also provide top-class training and education for the national as well as the European High Performance Computing (HPC) community. GCS is the German member of PRACE (Partnership for Advance Computing in Europe), an international non-profit association consisting of 25 member countries, whose representative organizations create a pan-European supercomputing infrastructure, providing access to computing and data management resources and services for large-scale scientific and engineering applications at the highest performance level.
GCS has its headquarters in Berlin/Germany.
Andrea Mayer-Grenu | idw - Informationsdienst Wissenschaft
Cutting edge research for the industries of tomorrow – DFKI and NICT expand cooperation
21.03.2017 | Deutsches Forschungszentrum für Künstliche Intelligenz GmbH, DFKI
Molecular motor-powered biocomputers
20.03.2017 | Technische Universität Dresden
Astronomers from Bonn and Tautenburg in Thuringia (Germany) used the 100-m radio telescope at Effelsberg to observe several galaxy clusters. At the edges of these large accumulations of dark matter, stellar systems (galaxies), hot gas, and charged particles, they found magnetic fields that are exceptionally ordered over distances of many million light years. This makes them the most extended magnetic fields in the universe known so far.
The results will be published on March 22 in the journal „Astronomy & Astrophysics“.
Galaxy clusters are the largest gravitationally bound structures in the universe. With a typical extent of about 10 million light years, i.e. 100 times the...
Researchers at the Goethe University Frankfurt, together with partners from the University of Tübingen in Germany and Queen Mary University as well as Francis Crick Institute from London (UK) have developed a novel technology to decipher the secret ubiquitin code.
Ubiquitin is a small protein that can be linked to other cellular proteins, thereby controlling and modulating their functions. The attachment occurs in many...
In the eternal search for next generation high-efficiency solar cells and LEDs, scientists at Los Alamos National Laboratory and their partners are creating...
Silicon nanosheets are thin, two-dimensional layers with exceptional optoelectronic properties very similar to those of graphene. Albeit, the nanosheets are less stable. Now researchers at the Technical University of Munich (TUM) have, for the first time ever, produced a composite material combining silicon nanosheets and a polymer that is both UV-resistant and easy to process. This brings the scientists a significant step closer to industrial applications like flexible displays and photosensors.
Silicon nanosheets are thin, two-dimensional layers with exceptional optoelectronic properties very similar to those of graphene. Albeit, the nanosheets are...
Enzymes behave differently in a test tube compared with the molecular scrum of a living cell. Chemists from the University of Basel have now been able to simulate these confined natural conditions in artificial vesicles for the first time. As reported in the academic journal Small, the results are offering better insight into the development of nanoreactors and artificial organelles.
Enzymes behave differently in a test tube compared with the molecular scrum of a living cell. Chemists from the University of Basel have now been able to...
20.03.2017 | Event News
14.03.2017 | Event News
07.03.2017 | Event News
24.03.2017 | Materials Sciences
24.03.2017 | Physics and Astronomy
24.03.2017 | Physics and Astronomy