The system works by taking advantage of the multiple sensors on a mobile phone, as well as those of other mobile phones in the vicinity.
Dubbed TagSense, the new app was developed by students from Duke University and the University of South Carolina (USC) and unveiled at the ninth Association for Computing Machinery's International Conference on Mobile Systems, Applications and Services (MobiSys), being held in Washington, D.C.
"In our system, when you take a picture with a phone, at the same time it senses the people and the context by gathering information from all the other phones in the area," said Xuan Bao, a Ph.D. student in computer science at Duke who received his master's degree at Duke in electrical and computer engineering.
Bao and Chuan Qin, a visiting graduate student from USC, developed the app working with Romit Roy Choudhury, assistant professor of electrical and computer engineering at Duke's Pratt School of Engineering. Qin and Bao are currently involved in summer internships at Microsoft Research.
"Phones have many different kinds of sensors that you can take advantage of," Qin said. "They collect diverse information like sound, movement, location and light. By putting all that information together, you can sense the setting of a photograph and describe its attributes."
By using information about the environment of a photograph, the students believe they can achieve a more accurate tagging of a particular photograph than could be achieved by facial recognition alone. Such information about a photograph's entirety provides additional details that can then be searched at a later time.
For example, the phone's built-in accelerometer can tell if a person is standing still for a posed photograph, bowling or even dancing. Light sensors in the phone's camera can tell if the shot is being taken indoors or outdoors on a sunny or cloudy day. The sensors can also approximate environmental conditions – such as snow or rain -- by looking up the weather conditions at that time and location. The microphone can detect whether or not a person in the photograph is laughing, or quiet. All of these attributes are then assigned to each photograph, the students said.
Bao pointed out that with multiple tags describing more than just a particular person's name, it would be easier to not only organize an album of photographs for future reference, but find particular photographs years later. With the exploding number of digital pictures in the cloud and in our personal computers, the ability to easily search and retrieve desired pictures will be valuable in the future, he said.
"So, for example, if you've taken a bunch of photographs at a party, it would be easy at a later date to search for just photographs of happy people dancing," Qin said. "Or more specifically, what if you just wanted to find photographs only of Mary dancing at the party and didn't want to look through all the photographs of Mary?"
These added details of automatic tagging could help complement existing tagging applications, according to senior researcher Roy Choudhury.
"While facial recognition programs continue to improve, we believe that the ability to identify photographs based on the setting of the photograph can lead to a richer, more detailed way to tag photographs," Roy Choudhury said. "TagSense was compared to Apple's iPhoto and Google's Picasa, and showed that it can provide greater sophistication in tagging photographs."
The students envision that TagSense would most likely be adopted by groups of people, such as friends, who would "opt in," allowing their mobile phone capabilities to be harnessed when members of the group were together. Importantly, Roy Choudhury added, TagSense would not request sensed data from nearby phones that do not belong to this group, thereby protecting users' privacy.
The experiments were conducted using eight Google Nexus One mobile phones on more than 200 photos taken at various locations across the Duke campus, including classroom buildings, gyms and the art museum.
The current application is a prototype, and the researchers believe that a commercial product could be available in a few years.
Srihari Nelakuditi, associate professor of computer science and engineering at USC, was also a member of the research team. The research is supported by the National Science Foundation. Roy Choudhury's Systems Networking Research Group is also supported by Microsoft, Nokia, Verizon, and Cisco.
Richard Merritt | EurekAlert!
Controlling robots with brainwaves and hand gestures
20.06.2018 | Massachusetts Institute of Technology, CSAIL
Innovative autonomous system for identifying schools of fish
20.06.2018 | IMDEA Networks Institute
In a recent publication in the renowned journal Optica, scientists of Leibniz-Institute of Photonic Technology (Leibniz IPHT) in Jena showed that they can accurately control the optical properties of liquid-core fiber lasers and therefore their spectral band width by temperature and pressure tuning.
Already last year, the researchers provided experimental proof of a new dynamic of hybrid solitons– temporally and spectrally stationary light waves resulting...
Scientists from the University of Freiburg and the University of Basel identified a master regulator for bone regeneration. Prasad Shastri, Professor of...
Moving into its fourth decade, AchemAsia is setting out for new horizons: The International Expo and Innovation Forum for Sustainable Chemical Production will take place from 21-23 May 2019 in Shanghai, China. With an updated event profile, the eleventh edition focusses on topics that are especially relevant for the Chinese process industry, putting a strong emphasis on sustainability and innovation.
Founded in 1989 as a spin-off of ACHEMA to cater to the needs of China’s then developing industry, AchemAsia has since grown into a platform where the latest...
The BMBF-funded OWICELLS project was successfully completed with a final presentation at the BMW plant in Munich. The presentation demonstrated a Li-Fi communication with a mobile robot, while the robot carried out usual production processes (welding, moving and testing parts) in a 5x5m² production cell. The robust, optical wireless transmission is based on spatial diversity; in other words, data is sent and received simultaneously by several LEDs and several photodiodes. The system can transmit data at more than 100 Mbit/s and five milliseconds latency.
Modern production technologies in the automobile industry must become more flexible in order to fulfil individual customer requirements.
An international team of scientists has discovered a new way to transfer image information through multimodal fibers with almost no distortion - even if the fiber is bent. The results of the study, to which scientist from the Leibniz-Institute of Photonic Technology Jena (Leibniz IPHT) contributed, were published on 6thJune in the highly-cited journal Physical Review Letters.
Endoscopes allow doctors to see into a patient’s body like through a keyhole. Typically, the images are transmitted via a bundle of several hundreds of optical...
13.06.2018 | Event News
08.06.2018 | Event News
05.06.2018 | Event News
22.06.2018 | Life Sciences
22.06.2018 | Physics and Astronomy
22.06.2018 | Earth Sciences