When you think of the solar system you probably picture a textbook diagram: nine planets, different sizes and colours, all circling the bright yellow sun. But how can a visually impaired child take in this information? How can they grasp how the solar system works?
European researchers have developed a computer program that lets sighted and visually impaired pupils explore space together. Using a special device similar to a robotic arm, a visually impaired child can move around the solar system and hear about the planets.
Sighted children can guide their visually impaired partners. As they move the mouse, the pressure and resistance in the robot arm gently nudges the visually impaired user in the right direction.
Students colloborate in hands on learning the Micole way.
The space explorer is just one of many applications developed by the EU-funded MICOLE project which is bringing multi-modal interfaces into the classroom.
The software is one of the first releases of multi-modal educational computer programs. It allows sighted children and the visually impaired to collaborate, helping to remove some of the barriers that the latter group faces every day in mainstream education.
Multi-modal computing is on the rise. Soon a third sense could be added to the sight and sound that computers currently use as their mode of output. Haptic devices, such as the robot arm, enable users to control computers as if they were using a mouse. Such devices also provide output – of movement, pressures or even raised bumps to deliver information to the user via the sense of touch.
This output is exploited in another MICOLE application. Current software drawing software allows pupils to create pictures, but the lines are in relief.
Using the haptic device, they can actually feel the picture. As they move a virtual stylus over the picture on the screen the pupils can sense the raised lines through the device communicated to their fingertips.
“These applications have been very welcomed in our test classrooms,” says project coordinator Roope Raisamo. “It has been wonderful to see visually impaired and sighted children really working together and sharing the same experiences and opportunities.
“Our aim was to allow blind people to create, manipulate and share information using hearing and touch to replace use of vision, and to work with other blind and sighted people. We want to see visually impaired people take a full role in society and work at an equal level with sighted people.”
Before MICOLE, very little was known about how multi-modality could be used when blind people work alongside those with full sight. But through its user-centred approach – lots of observation, focus groups, consultation and field-testing – the MICOLE partners have greatly added to our understanding about such collaborative work.
“Adding the sense of touch to information and communication technology is just getting to the point where it can be commercialised,” Raisamo continues. “The first people to benefit are people with disabilities, especially people who are blind or have visual impairment. The more senses you can use, the more multi-modal your computer interface, the more inclusive the technology can be.”
Much of the knowledge that the team has built up during the project's three years have been incorporated into guidelines for software developers on how to develop successful applications for collaborative learning.
Raisamo believes the project has increased the potential for successful commercialisation of multi-modal software in all fields.
A feeling for the future
MICOLE’s researchers have done far more than publish theoretical documents and produce a few prototype applications. Their main output comes in the form of the ‘MICOLE architecture’ and an accompanying software development toolkit. The architecture is like a high level programming language. It provides the virtual environment – all the processes and mechanisms for software applications to function in a multi-modal fashion.
It is designed so that software developers can quickly build new multi-modal applications for the visually impaired and can be used for applications in widespread disciplines including medicine, education or ‘edutainment’.
“We've made it an open source architecture to encourage more people to access it, spread it, and adapt it for their own requirements without worrying about licenses or intellectual property,” says Raisamo.
So far 14 organisations from outside the project have downloaded the architecture and toolkit files in the first six months since these were made available for download on the project's website.
“This is encouraging,” Raisamo says. “We know that several groups are working on new applications and we hope that the market will take off soon.”
Reachin Technology, a key MICOLE partner and a distributor of the haptic robot device, could be one of the first to sell multi-modal software based on MICOLE’s research.
Multi-modal interfaces can support people with many disabilities because if one of the modalities is missing, the system remains usable.
“If multi-modal applications are used widely and match the needs of different users, they could narrow the gap between sighted and impaired children,” says Raisamo.
MICOLE, which received funding from the EU's Sixth Framework Programme for research, ended in August 2007. However, the partners involved continue to push research and development in the field.
Raisamo's team is currently working with local schools to develop new applications based on the architecture – one that explores physical phenomena, another the natural sciences.
Ahmed ElAmin | alfa
29.06.2017 | Institute of Science and Technology Austria
Seeing the forest through the trees with a new LiDAR system
28.06.2017 | The Optical Society
Computer scientists use wave packet theory to develop realistic, detailed water wave simulations in real time. Their results will be presented at this year’s SIGGRAPH conference.
Think about the last time you were at a lake, river, or the ocean. Remember the ripples of the water, the waves crashing against the rocks, the wake following...
An international team of scientists has proposed a new multi-disciplinary approach in which an array of new technologies will allow us to map biodiversity and the risks that wildlife is facing at the scale of whole landscapes. The findings are published in Nature Ecology and Evolution. This international research is led by the Kunming Institute of Zoology from China, University of East Anglia, University of Leicester and the Leibniz Institute for Zoo and Wildlife Research.
Using a combination of satellite and ground data, the team proposes that it is now possible to map biodiversity with an accuracy that has not been previously...
Heatwaves in the Arctic, longer periods of vegetation in Europe, severe floods in West Africa – starting in 2021, scientists want to explore the emissions of the greenhouse gas methane with the German-French satellite MERLIN. This is made possible by a new robust laser system of the Fraunhofer Institute for Laser Technology ILT in Aachen, which achieves unprecedented measurement accuracy.
Methane is primarily the result of the decomposition of organic matter. The gas has a 25 times greater warming potential than carbon dioxide, but is not as...
Hydrogen is regarded as the energy source of the future: It is produced with solar power and can be used to generate heat and electricity in fuel cells. Empa researchers have now succeeded in decoding the movement of hydrogen ions in crystals – a key step towards more efficient energy conversion in the hydrogen industry of tomorrow.
As charge carriers, electrons and ions play the leading role in electrochemical energy storage devices and converters such as batteries and fuel cells. Proton...
Scientists from the Excellence Cluster Universe at the Ludwig-Maximilians-Universität Munich have establised "Cosmowebportal", a unique data centre for cosmological simulations located at the Leibniz Supercomputing Centre (LRZ) of the Bavarian Academy of Sciences. The complete results of a series of large hydrodynamical cosmological simulations are available, with data volumes typically exceeding several hundred terabytes. Scientists worldwide can interactively explore these complex simulations via a web interface and directly access the results.
With current telescopes, scientists can observe our Universe’s galaxies and galaxy clusters and their distribution along an invisible cosmic web. From the...
19.06.2017 | Event News
13.06.2017 | Event News
13.06.2017 | Event News
29.06.2017 | Physics and Astronomy
29.06.2017 | Life Sciences
29.06.2017 | Health and Medicine