The grid is changing as the big, centralized providers of the past are replaced by smaller, distributed suppliers. Keeping such complex networks running stable requires high-resolution sensor technology – AI provides a way to make accurate predictions and automatically detect any disturbances or anomalies in real time. Here is how Fraunhofer researchers developed the compression techniques, algorithms and neural networks to make a power supply fit for the future.
The way power is generated is in transition: Whereas, before, all our power came from big power plants, these days it comes from a range of distributed sources as well, including wind turbines, photovoltaic systems and other similar facilities.
This shift has a big impact on our grid – with particular challenges for operators of transmission grids. How to monitor the proper functioning of grid parameters such as phase angle and frequencies? Might there be discrepancies or anomalies in the proper functioning of the grid? Or are there lines or power plants down?
Today’s standard measurement technology is no longer able to reliably furnish answers to these sorts of questions. More and more operators are, therefore, turning to additional phasor measurement units (PMUs) and other digital solutions. These systems measure the amplitude and phase of current and voltage up to 50 times a second. This process generates huge volumes of data, easily several gigabytes a day.
Data compression saves 80 percent of data
In response, researchers at the Advanced System Technology (AST) branch of the Fraunhofer Institute for Optronics, System Technologies and Image Exploitation IOSB in Ilmenau are looking for ways to optimize the data processing using artificial intelligence, with a view to improving grid reliability and establishing a power supply system fit for the future.
“We can use AI to automatically log, compress and process up to 4.3 million data sets per day,” says Prof. Peter Bretschneider, head of the Energy department at the AST branch of the Fraunhofer IOSB.
In the first phase of their work, the researchers have come up with a compression technique that saves 80 percent of the data. Not only is it easier to store the data, but faster and more efficient to process it too.
Automated data processing in real time
In the second phase, the researchers went on to utilize the phasor measurement data they had collected to apply neural networks – one of the key components for today’s artificial intelligence. More specifically, they “fed” the neural networks with examples of typical system outages. This way, the algorithms gradually learn to distinguish – and precisely categorize – normal operating data from defined system malfunctions.
Following the training phase, the researchers applied the neural networks to current data generated from phasor measurements – data that previously had to be taken and manually processed. This is where the algorithm made its first leap into real-time application, making split-second decisions on where there is an anomaly or fault, as well as the type and location of that disturbance. To take an example, if one power plant should fail, an abrupt spike can be expected in the load placed on the other power plants.
The increased load slows down the generators, and the frequency decreases. This calls for rapid countermeasures because if the frequency sinks below a threshold value, the operator may be forced to cut off sections of the grid for the sake of system stability. And by rapid, we are talking about less than 500 milliseconds. Since the algorithm is capable of reaching a decision within 20–50 milliseconds, that leaves sufficient time to implement the appropriate fully automated countermeasures.
The algorithm is ready to be implemented, as the researchers continue to work on the control and regulation of the relevant countermeasures. The development is of interest not only to the big operators of power transmission grids, but also to regional distribution grids. “To make an analogy with the road network, what’s the point of having clear motorways when the smaller regional roads are permanently blocked?” says Bretschneider.
Power to predict problems of the future
All the same, the researchers are not restricting themselves to the problems of today, but also want to factor in anomalies that have not even occurred so far. “If we continue to pursue renewables, it may lead to situations we don’t even know about yet,” says Bretschneider. Here, too, the researchers have turned to artificial intelligence, where they work on categorizing these sorts of unknown phenomena and developing the appropriate algorithms using digital network maps.
Martin Käßler | Fraunhofer Research News
Accelerating quantum technologies with materials processing at the atomic scale
15.05.2019 | University of Oxford
A step towards probabilistic computing
15.05.2019 | University of Konstanz
A new assessment of NASA's record of global temperatures revealed that the agency's estimate of Earth's long-term temperature rise in recent decades is accurate to within less than a tenth of a degree Fahrenheit, providing confidence that past and future research is correctly capturing rising surface temperatures.
The most complete assessment ever of statistical uncertainty within the GISS Surface Temperature Analysis (GISTEMP) data product shows that the annual values...
Physicists at the University of Basel are able to show for the first time how a single electron looks in an artificial atom. A newly developed method enables them to show the probability of an electron being present in a space. This allows improved control of electron spins, which could serve as the smallest information unit in a future quantum computer. The experiments were published in Physical Review Letters and the related theory in Physical Review B.
The spin of an electron is a promising candidate for use as the smallest information unit (qubit) of a quantum computer. Controlling and switching this spin or...
Engineers at the University of Tokyo continually pioneer new ways to improve battery technology. Professor Atsuo Yamada and his team recently developed a...
With a quantum coprocessor in the cloud, physicists from Innsbruck, Austria, open the door to the simulation of previously unsolvable problems in chemistry, materials research or high-energy physics. The research groups led by Rainer Blatt and Peter Zoller report in the journal Nature how they simulated particle physics phenomena on 20 quantum bits and how the quantum simulator self-verified the result for the first time.
Many scientists are currently working on investigating how quantum advantage can be exploited on hardware already available today. Three years ago, physicists...
'Quantum technologies' utilise the unique phenomena of quantum superposition and entanglement to encode and process information, with potentially profound benefits to a wide range of information technologies from communications to sensing and computing.
However a major challenge in developing these technologies is that the quantum phenomena are very fragile, and only a handful of physical systems have been...
29.04.2019 | Event News
17.04.2019 | Event News
15.04.2019 | Event News
24.05.2019 | Physics and Astronomy
24.05.2019 | Medical Engineering
24.05.2019 | Life Sciences