To ease the use of these techniques, the Remote Sensing Group has built add-ons for ImageJ. ImageJ is public domain Java image processing software. These add-ons are necessary to process multi-spectral images, and there is now an image read/write module with specific formats for multi-band images (usable for any number of spectral bands).
Additionally, the Remote Sensing Group has developed a module of generic utilities. Another module implements the image transforms most commonly used in image fusion. And yet another includes different multi-spectral and panchromatic image fusion algorithms, including algorithms developed by the Remote Sensing Group, already published in specialized journals.
Some of our planet’s most pressing problems, like population growth, the need for sustainable intensive agriculture, safe food and energy production and distribution, soil management, climate change, public health and social conflicts, cannot be solved without immediate information on a local and a global scale.
Remote sensing is the only available technology at present for acquiring most of this information almost immediately and on any scale, as it supplies a tremendous amount of high quality (accurate, consistent and reliable) geographic, spatial and spectral data, etc.
One of the most commonly used products in remote sensing are images from remote optical sensors: multi-spectral and panchromatic images. A multi-spectral image is an image recorded by a sensor that picks up information contained in different bands of the electromagnetic scale. A multi-spectral image is composed of as many images of the same area of the earth’s surface as spectral bands recorded by the sensor (spectral resolution). Each of these images actually provides information about the amount of radiation reflected by the different surfaces in the area being observed within each of the spectral bands. Another feature of these images is their spatial resolution. Spatial resolution determines the size of the smallest detail that can be observed in the image.
On the other hand, panchromatic images provide the same type of information for a single, normally the broadest, spectral band.
Information extraction methods
The volume of data to be managed and interpreted can vary significantly depending on the sensor type used and the features of the images they record. Even so, data throughput is always very high and calls for objective and precise information extraction methods.
Traditionally these methods have been based on automatic classification and interpretation techniques. On the whole, these techniques output qualitative information that is not always easily quantifiable. Spectral indices or model inversion are other approaches that have recently proved to be more effective for extracting quantitative information from satellite images.
Whatever approach is taken, the precision of the results provided by the remote sensing techniques depends on a series of factors. Current sensor technology (ETM+, ASTER, MERIS, MOS, SPOT, IKONOS, QUICKBIRD...) and, consequently, the features (spatial, spectral and temporal resolution) of the images they record are two key factors. Due to space platform information transmission technology limitations, there is at present a trade-off between these three types of resolution. Sensors that provide a high spatial resolution tend to have a lower spectral resolution and vice versa. This is why panchromatic images have a higher spatial resolution than multi-spectral images. Note, for example, that multi-spectral images from space programmes like LANDSAT, SPOT or IRS, have a high spectral resolution compared with the respective panchromatic images, but their spatial resolution is lower.
Image fusion techniques
Still, there is a wide range of remote sensing applications that require satellite images combining both features (high spatial and spectral resolution).
Even though some of today’s satellites, like QUICKBIRD and IKONOS, do actually deliver such high resolution images, they are costly and not always to be had by ordinary users. One way of getting high resolution spatial and spectral images at fairly reasonable costs is to use image fusion techniques. What’s more, image fusion techniques can output very high resolution images if applied to data recorded by latest generation sensors. The goal of image fusion techniques is to consistently integrate information from different images, assuring that the final fused image retains the key information from the source images.
There are now many optical image fusion methodologies and algorithms. The most common are based on a number of different transforms. Some, like the methodologies based on the Brovey transform, principal component analysis or the IHS transformation method (intensity, hue, saturation), are conceptually very simple.
As demonstrated in many papers, however, the colour of the fused images output by these methodologies is quite distorted compared with the original multi-spectral images. This means that they cannot be used in a variety of routine tasks in the field of remote sensing, like image classification or change detection.
On the other hand, there are many methods based on multi-resolution analysis techniques. Most of these techniques use the discrete wavelet transform (DWT). Generally, these methods provide only a slight distortion of the colour of the fused images with respect to the multi-spectral image. They outperform the techniques described above, but there is no way of controlling the trade-off between the spatial and spectral quality of the fused images.
Remote Sensing Group TechniquesThe School of Computing’s Remote Sensing Group has proposed a weighted version of DWT-mediated fusion calculated using the à trous algorithm. This algorithm can efficiently separate the background information from the detail of an image, avoiding the decimation process characteristic of other algorithms used to calculate the DWT.
The Remote Sensing Group has also proposed other fusion techniques based on multidirection-multiresolution transforms (MDMR). These methods output better quality fused images than DWT-based techniques and, at the same time, have a built-in facility for controlling the trade-off between the images’ spatial quality and colour distortion.
With the aim of easing the use of the proposed algorithms already published in reputed international journals, the Remote Sensing Group set itself the goal of integrating the methods into a tool.
To do this, it set out to build the tools required to process multi-spectral images into public domain image processing software developed in Java: ImageJ.
As mentioned earlier, the modules that are now up and running are:• Image read/write module with specific formats for multi-band images (any number of spectral bands)
All these ImageJ add-ons have been packaged as IJFusion.
The TU Ilmenau develops tomorrow’s chip technology today
27.04.2017 | Technische Universität Ilmenau
Five developments for improved data exploitation
19.04.2017 | Deutsches Forschungszentrum für Künstliche Intelligenz GmbH, DFKI
More and more automobile companies are focusing on body parts made of carbon fiber reinforced plastics (CFRP). However, manufacturing and repair costs must be further reduced in order to make CFRP more economical in use. Together with the Volkswagen AG and five other partners in the project HolQueSt 3D, the Laser Zentrum Hannover e.V. (LZH) has developed laser processes for the automatic trimming, drilling and repair of three-dimensional components.
Automated manufacturing processes are the basis for ultimately establishing the series production of CFRP components. In the project HolQueSt 3D, the LZH has...
Reflecting the structure of composites found in nature and the ancient world, researchers at the University of Illinois at Urbana-Champaign have synthesized thin carbon nanotube (CNT) textiles that exhibit both high electrical conductivity and a level of toughness that is about fifty times higher than copper films, currently used in electronics.
"The structural robustness of thin metal films has significant importance for the reliable operation of smart skin and flexible electronics including...
The nearby, giant radio galaxy M87 hosts a supermassive black hole (BH) and is well-known for its bright jet dominating the spectrum over ten orders of magnitude in frequency. Due to its proximity, jet prominence, and the large black hole mass, M87 is the best laboratory for investigating the formation, acceleration, and collimation of relativistic jets. A research team led by Silke Britzen from the Max Planck Institute for Radio Astronomy in Bonn, Germany, has found strong indication for turbulent processes connecting the accretion disk and the jet of that galaxy providing insights into the longstanding problem of the origin of astrophysical jets.
Supermassive black holes form some of the most enigmatic phenomena in astrophysics. Their enormous energy output is supposed to be generated by the...
The probability to find a certain number of photons inside a laser pulse usually corresponds to a classical distribution of independent events, the so-called...
Microprocessors based on atomically thin materials hold the promise of the evolution of traditional processors as well as new applications in the field of flexible electronics. Now, a TU Wien research team led by Thomas Müller has made a breakthrough in this field as part of an ongoing research project.
Two-dimensional materials, or 2D materials for short, are extremely versatile, although – or often more precisely because – they are made up of just one or a...
28.04.2017 | Event News
20.04.2017 | Event News
18.04.2017 | Event News
28.04.2017 | Medical Engineering
28.04.2017 | Earth Sciences
28.04.2017 | Life Sciences