Change is coming to the IT world, says Dave Robertson, coordinator of the EU-funded OpenKnowledge research project.
Just as individuals are now storing, editing and sharing photos and videos on the web, other users from small businesses to CERN’s Large Hadron Collider are moving their data, computation, and collaboration into “the cloud” – the internet’s worldwide network of servers and computers, but also the millions of handheld devices, monitors, sensors and other components linked to it.
“More and more companies are pushing much of what they do out into the cloud,” says Robertson. “If that’s the way things are going, and if it’s going to be very large, then society needs some way to be able to take control of how that gets coordinated.”
Creating a toolkit to access, coordinate and exploit the cloud’s dynamic resources is what OpenKnowledge set out to accomplish.
If the IT world embraces OpenKnowledge, says Robertson, users will no longer have to rely on a small number of big companies to access interactive internet services.
“You would see lots of people who weren’t so specialised writing and using and sharing lots of specific keys that would unlock what’s available on the internet for themselves and others,” says Robertson.
Roles, rules, results
Suppose, says Robertson, a number of potential partners want to create a new service or product. One might manage a database, another can analyse the data, a third can package and present the results, while the fourth has marketing and management skills.
These potential partners could be anywhere in the world and could be using a wide variety of software, natural and computer languages, internet interfaces and devices.
The OpenKnowledge researchers – who, as members of an EU-funded project are themselves scattered from Scotland to Spain – set out to create a user-friendly system that would let these virtual partners find each other, define their respective roles, figure out the rules and sequences that will let them interact smoothly, and get their new enterprise up and running.
To accomplish that goal, the OpenKnowledge team created a new language for specifying the kind of processes that let different systems interact with each other. The language is called LCC, for Lightweight Coordination Calculus.
“We’ve gone for the simplest way to understand a process that we could possibly devise,” says Robertson.
The researchers also found a way to deal with the fact that the same step in a process is likely to be labelled in different ways by different components of the system.
For example, a handheld device might use an asterisk to signal that it is about to send a number while the database where that number is needed might expect to receive an input labelled “price.”
System engineers often approach this semantic problem by building what are called global ontologies – essentially dictionaries that specify the labels and properties of all the objects or events within a system.
In situations where such rules of interaction already exist, OpenKnowledge will find and use them.
Most of the time, however, that approach will not work because there is no way of knowing in advance what devices or systems will be interacting in a particular exchange.
In that case, OpenKnowledge uses statistical regularities to build a much smaller dictionary that defines only the steps that are needed for the purpose at hand.
“You know that you’re at some specific point,” says Robertson, “and you look to see what other people were doing at the same point. As the system gets used, you have a lot of interactions, possibly thousands or millions. That’s where your mapping comes from.”
But can I trust you?
Like anyone using the internet, OpenKnowledge clients are vulnerable. For example, a partner might provide poor quality services, or not be who he claims to be.
The researchers believe they have solved that problem to some degree by building measures of reputation into their software package. One approach is to measure how often interactions with a potential partner have gone well. Another is to see how often they have interacted with other trustworthy partners.
“We do exactly the same things that are used to rate web pages, but with these more complicated forms of information,” says Robertson.
All of the key OpenKnowledge functions – discovering and interpreting interactions, ontology matching, and reputation checking – reside in the OpenKnowledge kernel, an open-source software package that can be downloaded from the project’s website.
Robertson and his colleagues have tested OpenKnowledge in three real-world areas: healthcare coordination, proteomics research, and emergency response. These applications will be featured in a subsequent ICT Results feature on 29 December.
In the meantime, they are eager for others to use OpenKnowledge to unlock the cloud’s capabilities and choreograph their own ideas.
“It will only become revolutionary,” Robertson writes, “if clever people invent interactions that are really useful for lots of other people.”
The OpenKnowledge project received funding from ICT strand of the Sixth Framework Programme for research.
This is the first of a two-part feature on OpenKnowledge.
Christian Nielsen | alfa
Terahertz spectroscopy goes nano
20.10.2017 | Brown University
New software speeds origami structure designs
12.10.2017 | Georgia Institute of Technology
University of Maryland researchers contribute to historic detection of gravitational waves and light created by event
On August 17, 2017, at 12:41:04 UTC, scientists made the first direct observation of a merger between two neutron stars--the dense, collapsed cores that remain...
Seven new papers describe the first-ever detection of light from a gravitational wave source. The event, caused by two neutron stars colliding and merging together, was dubbed GW170817 because it sent ripples through space-time that reached Earth on 2017 August 17. Around the world, hundreds of excited astronomers mobilized quickly and were able to observe the event using numerous telescopes, providing a wealth of new data.
Previous detections of gravitational waves have all involved the merger of two black holes, a feat that won the 2017 Nobel Prize in Physics earlier this month....
Material defects in end products can quickly result in failures in many areas of industry, and have a massive impact on the safe use of their products. This is why, in the field of quality assurance, intelligent, nondestructive sensor systems play a key role. They allow testing components and parts in a rapid and cost-efficient manner without destroying the actual product or changing its surface. Experts from the Fraunhofer IZFP in Saarbrücken will be presenting two exhibits at the Blechexpo in Stuttgart from 7–10 November 2017 that allow fast, reliable, and automated characterization of materials and detection of defects (Hall 5, Booth 5306).
When quality testing uses time-consuming destructive test methods, it can result in enormous costs due to damaging or destroying the products. And given that...
Using a new cooling technique MPQ scientists succeed at observing collisions in a dense beam of cold and slow dipolar molecules.
How do chemical reactions proceed at extremely low temperatures? The answer requires the investigation of molecular samples that are cold, dense, and slow at...
Scientists from the Max Planck Institute of Quantum Optics, using high precision laser spectroscopy of atomic hydrogen, confirm the surprisingly small value of the proton radius determined from muonic hydrogen.
It was one of the breakthroughs of the year 2010: Laser spectroscopy of muonic hydrogen resulted in a value for the proton charge radius that was significantly...
17.10.2017 | Event News
10.10.2017 | Event News
10.10.2017 | Event News
20.10.2017 | Information Technology
20.10.2017 | Materials Sciences
20.10.2017 | Interdisciplinary Research