Walking around the streets searching for a place to eat will be no hassle when a head-mounted display (HMD) becomes affordable and ubiquitous. Researchers at the Korea Advanced Institute of Science and Technology (KAIST) developed K-Glass, a wearable, hands-free HMD that enables users to find restaurants while checking out their menus.
K-Glass Developed by KAIST
Copyright : KAIST
If the user of K-Glass walks up to a restaurant and looks at the name of the restaurant, today’s menu and a 3D image of food pop up. The Glass can even show the number of tables available inside the restaurant. K-Glass makes this possible because of its built-in augmented reality (AR) processor.
Unlike virtual reality which replaces the real world with a computer-simulated environment, AR incorporates digital data generated by the computer into the reality of a user. With the computer-made sensory inputs such as sound, video, graphics or GPS data, the user’s real and physical world becomes live and interactive. Augmentation takes place in real-time and in semantic context with surrounding environments, such as a menu list overlain on the signboard of a restaurant when the user passes by it, not an airplane flight schedule, which is irrelevant information, displayed.
Most commonly, location-based or computer-vision services are used in order to generate AR effects. Location-based services activate motion sensors to identify the user’s surroundings, whereas computer-vision uses algorithms such as facial, pattern, and optical character recognition, or object and motion tracking to distinguish images and objects. Many of the current HMDs deliver augmented reality experiences employing location-based services by scanning the markers or bar-codes printed on the back of objects. The AR system tracks the codes or markers to identify objects and then align them with virtual reality. However, this AR algorithm is difficult to use for the objects or spaces which do not have bar-codes, QR codes, or markers, particularly those in outdoor environments and thus cannot be recognized.
To solve this problem, Hoi-Jun Yoo, Professor of Electrical Engineering at KAIST and his team developed, for the first time in the world, an AR chip that works just like human vision. This processor is based on the Visual Attention Model (VAM) that duplicates the ability of human brain to process visual data. VAM, almost unconsciously or automatically, disentangles the most salient and relevant information about the environment in which human vision operates, thereby eliminating unnecessary data unless they must be processed. In return, the processor can dramatically speed up the computation of complex AR algorithms.
The AR processor has a data processing network similar to that of a human brain’s central nervous system. When the human brain perceives visual data, different sets of neurons, all connected, work concurrently on each fragment of a decision-making process; one group’s work is relayed to other group of neurons for the next round of the process, which continues until a set of decider neurons determines the character of the data. Likewise, the artificial neural network allows parallel data processing, alleviating data congestion and reducing power consumption significantly.
KAIST’s AR processor, which is produced using the 65 nm (nanometers) manufacturing process with the area of 32 mm2, delivers 1.22 TOPS (tera-operations per second) peak performance when running at 250 MHz and consumes 778 miliWatts on a 1.2V power supply. The ultra-low power processor shows 1.57 TOPS/W high efficiency rate of energy consumption under the real-time operation of 30fps/720p video camera, a 76% improvement in power conservation over other devices. The HMDs, available on the market including the Project Glass whose battery lasts only for two hours, have revealed so far poor performance. Professor Yoo said, “Our processor can work for long hours without sacrificing K-Glass’s high performance, an ideal mobile gadget or wearable computer, which users can wear for almost the whole day.”
He further commented:
“HMDs will become the next mobile device, eventually taking over smartphones. Their markets have been growing fast, and it’s really a matter of time before mobile users will eventually embrace an optical see-through HMD as part of their daily use. Through augmented reality, we will have richer, deeper, and more powerful reality in all aspects of our life from education, business, and entertainment to art and culture.”
The KAIST team presented a research paper at the International Solid-State Circuits Conference (ISSCC) held on February 9-13, 2014 in San Francisco, CA, which is entitled “1.22TOPS and 1.52mW/MHz Augmented Reality Multi-Core Processor with Neural Network NoC for HMD Applications.”For further inquires:
Miniscule Mirrored Cavities Connect Quantum Memories
24.06.2015 | Department of Energy, Office of Science
SASER-Siegfried – Record-breaking Transmission Field Trial
24.06.2015 | EURESCOM European institute for research and strategic studies in telecommunications
New technique combines electron microscopy and synchrotron X-rays to track chemical reactions under real operating conditions
A new technique pioneered at the U.S. Department of Energy's Brookhaven National Laboratory reveals atomic-scale changes during catalytic reactions in real...
Think of an object made of iron: An I-beam, a car frame, a nail. Now imagine that half of the iron in that object owes its existence to bacteria living two and a half billion years ago.
Think of an object made of iron: An I-beam, a car frame, a nail. Now imagine that half of the iron in that object owes its existence to bacteria living two and...
A team of scientists including PhD student Friedrich Schuler from the Laboratory of MEMS Applications at the Department of Microsystems Engineering (IMTEK) of...
The three-year clinical trial results of the retinal implant popularly known as the "bionic eye," have proven the long-term efficacy, safety and reliability of...
On June 23, the second Sentinel mission was launched from the space mission launch center in Kourou. A critical component of Aachen is on board. Researchers at the Fraunhofer Institute for Laser Technology ILT and Tesat-Spacecom have jointly developed the know-how for space-qualified laser components. For the Sentinel mission the diode laser pump module of the Laser Communication Terminal LCT was planned and constructed in Aachen in cooperation with the manufacturer of the LCT, Tesat-Spacecom, and the Ferdinand Braun Institute.
After eight years of preparation, in the early morning of June 23 the time had come: in Kourou in French Guiana, the European Space Agency launched the...
25.06.2015 | Event News
16.06.2015 | Event News
11.06.2015 | Event News
01.07.2015 | Press release
01.07.2015 | Awards Funding
01.07.2015 | Physics and Astronomy