Digital images can open our eyes to the most extraordinary detail and beauty. But there is one major drawback. The information in an image is purely visual. It tells us nothing about when or where the image was taken.
It tells us nothing about the people in the image. We do all that interpretation ourselves. That makes cataloguing and searching for particular images difficult – whether you work for an art gallery or you are updating your family album.
Scientists on a major European research project called aceMedia have taken important steps towards a solution to this problem. They are building an information layer into digital image files. Their vision is that image files will come with content information, metadata (background information for use on the internet) and an intelligence layer that automatically generates word-searchable data about the image.
An extra ‘information layer’, that adds both automatically generated and manually generated information to images, would revolutionise image searching on the internet as well as on your home computer or mobile phone. The technologies developed in the EU-funded aceMedia project have sparked interest from a range of commercial companies, looking to exploit the ideas in a host of directions.
Building a picture puzzle
The project re-used, developed and combined a series of technologies that provide greatly enriched content information about an image.
One of the technologies exploited by an aceMedia team uses software that can identify low-level visual descriptors, such as consistent areas of colour that may be sky, sea, sand or possibly snow, and information about texture, edge and shape.
Combining the low-level descriptors with sets of contextual rules held in domain ontologies (such as the fact that consistent areas of blue at the top of an image are likely to be sky, or that beach and snow are unlikely to appear in the same picture) turns data into a rich information source.
“Turning low-level descriptors into useful information is a very difficult step,” according to Yiannis Kompatsiaris, Head of the Multimedia Knowledge Laboratory at the Informatics and Telematics Institute in Thessaloniki, Greece and one of the lead researchers on aceMedia. His team was involved in structuring knowledge and adding it to the domain ontologies that classified and identified the information provided by the low-level descriptors.
Data from low-level descriptors was also combined with the results from specific detectors, such as the kinds of face detectors that are commercially available on some cameras today. All add further clues or searchable data for image users.
Another layer of information can be added by the individual user. They can add rules defining their personal preferences, profiles and policies to create a personalised filing system. ‘Inferencing’ techniques, filtering and similarity algorithms were used to make that personal filing simpler.
To enable easier searching, some of the aceMedia researchers incorporated natural language processing techniques into the mix, which mean you can use everyday language when searching for an image.
The ace in your hand
AceMedia researchers drew together their full range of technologies in an Autonomous Content Entity (ACE) Framework. The ACE Framework defines APIs to support networking, database management, scalable coding, content pre-processing, content visualisation, knowledge-assisted content analysis, as well as context analysis and modelling modules.
Using the framework, ACEs can be created that contain all of the rules, metadata and content information. They become a part of the image file.
For video, aceMedia researchers developed a scalable video codec, the aceSVC. Pical scalable video coding chain consists of three main modules – an encoder, extractor and decoder. The aceSVC enables video playing, reviewing and video analysis in the compressed domain.
As part of the project, aceMedia researchers demonstrated the benefits of automated content sharing and easier content management that ACEs could provide on a series of home network devices, including PCs, mobile phones and set-top boxes.
While the vision of the aceMedia project was to combine technologies, each delivering a piece to the overall information puzzle, they are not interdependent according to Kompatsiaris.
“The tools we developed in aceMedia are scalable to many concepts and many environments,” he confirms.
Adding time and location
“In five years time, a good number of these technologies will be in common use – combined with a number of technologies that have grown in popularity since the aceMedia project started, such as geo-tagging using GPS receivers. I think cameras in the future will know their position and be able to combine that information with content analysis to give much better results than we are capable of at the moment. For example, if the camera knows it is in a mountainous environment then it can analyse the content of the image much more efficiently.
Christian Nielsen | alfa
Seeing the forest through the trees with a new LiDAR system
28.06.2017 | The Optical Society
Drones that drive
27.06.2017 | Massachusetts Institute of Technology, CSAIL
An international team of scientists has proposed a new multi-disciplinary approach in which an array of new technologies will allow us to map biodiversity and the risks that wildlife is facing at the scale of whole landscapes. The findings are published in Nature Ecology and Evolution. This international research is led by the Kunming Institute of Zoology from China, University of East Anglia, University of Leicester and the Leibniz Institute for Zoo and Wildlife Research.
Using a combination of satellite and ground data, the team proposes that it is now possible to map biodiversity with an accuracy that has not been previously...
Heatwaves in the Arctic, longer periods of vegetation in Europe, severe floods in West Africa – starting in 2021, scientists want to explore the emissions of the greenhouse gas methane with the German-French satellite MERLIN. This is made possible by a new robust laser system of the Fraunhofer Institute for Laser Technology ILT in Aachen, which achieves unprecedented measurement accuracy.
Methane is primarily the result of the decomposition of organic matter. The gas has a 25 times greater warming potential than carbon dioxide, but is not as...
Hydrogen is regarded as the energy source of the future: It is produced with solar power and can be used to generate heat and electricity in fuel cells. Empa researchers have now succeeded in decoding the movement of hydrogen ions in crystals – a key step towards more efficient energy conversion in the hydrogen industry of tomorrow.
As charge carriers, electrons and ions play the leading role in electrochemical energy storage devices and converters such as batteries and fuel cells. Proton...
Scientists from the Excellence Cluster Universe at the Ludwig-Maximilians-Universität Munich have establised "Cosmowebportal", a unique data centre for cosmological simulations located at the Leibniz Supercomputing Centre (LRZ) of the Bavarian Academy of Sciences. The complete results of a series of large hydrodynamical cosmological simulations are available, with data volumes typically exceeding several hundred terabytes. Scientists worldwide can interactively explore these complex simulations via a web interface and directly access the results.
With current telescopes, scientists can observe our Universe’s galaxies and galaxy clusters and their distribution along an invisible cosmic web. From the...
Temperature measurements possible even on the smallest scale / Molecular ruby for use in material sciences, biology, and medicine
Chemists at Johannes Gutenberg University Mainz (JGU) in cooperation with researchers of the German Federal Institute for Materials Research and Testing (BAM)...
19.06.2017 | Event News
13.06.2017 | Event News
13.06.2017 | Event News
28.06.2017 | Physics and Astronomy
28.06.2017 | Physics and Astronomy
28.06.2017 | Health and Medicine