This opens up a wider world of personal robotics, in which machines are doing anything their owners can program them to do—without actually being programmers.
Laying some helpful groundwork for this world is, a new study by researchers in Georgia Tech’s Center for Robotics & Intelligent Machines (RIM), who have identified the types of questions a robot can ask during a learning interaction that are most likely to characterize a smooth and productive human-robot relationship.
These questions are about certain features of tasks, more so than labels of task components or real-time demonstrations of the task itself, and the researchers identified them not by studying robots, but by studying the everyday (read: non-programmer) people who one day will be their masters. The findings were detailed in the paper, “Designing Robot Learners that Ask Good Questions,” presented this week in Boston at the 7th ACM/IEEE Conference on Human-Robot Interaction (HRI).
“People are not so good at teaching robots because they don’t understand the robots’ learning mechanism,” said lead author Maya Cakmak, Ph.D. student in the School of Interactive Computing. “It’s like when you try to train a dog, and it’s difficult because dogs do not learn like humans do. We wanted to find out the best kinds of questions a robot could ask to make the human-robot relationship as ‘human’ as it can be.”
Cakmak’s study attempted to discover the role “active learning” concepts play in human-robot interaction. In a nutshell, active learning refers to giving machine learners more control over the information they receive. Simon, a humanoid robot created in the lab of Andrea Thomaz (assistant professor in the Georgia Tech’s School of Interactive Computing, and co-author), is well acquainted with active learning; Thomaz and Cakmak are programming him to learn new tasks by asking questions.
Cakmak designed two separate experiments (see video): first, she asked human volunteers to assume the role of an inquisitive robot attempting to learn a simple task by asking questions of a human instructor. Having identified the three main question types (feature, label and demonstration), Cakmak tagged each of the participants’ questions as one of the three. The overwhelming majority (about 82 percent) of questions were feature queries, showing a clear cognitive preference in human learning for this query type.
Type of question Example
Label query “Can I pour salt like this?"
Demonstration query “Can you show me how to pour salt from here?”
Feature query “Can I pour salt from any height?”
Next, Cakmak recruited humans to teach Simon new tasks by answering the robot’s questions and then rating those questions on how “smart” they thought they were. Feature queries once again were the preferred interrogatory, with 72 percent of participants calling them the smartest questions.
“These findings are important because they help give us the ability to teach robots the kinds of questions that humans would ask,” Cakmak said. “This in turn will help manufacturers produce the kinds of robots that are most likely to integrate quickly into a household or other environment and better serve the needs we’ll have for them.”
Georgia Tech is fielding five of the 38 papers accepted for HRI’s technical program, making it the largest academic contributor to the conference. Those five include:
“Designing Robot Learners that Ask Good Questions,” by Maya Cakmak and Andrea L. Thomaz
“Real World Haptic Exploration for Telepresence of the Visually Impaired,” by Chung Hyuk Park and Ayanna M. Howard
“The Domesticated Robot: Design Guidelines for Assisting Older Adults to Age in Place,” by Jenay Beer, Cory-Ann Smarr, Tiffany Chen, Akanksha Prakash, Tracy Mitzner, Charles Kemp and Wendy Rogers
“Enhancing Interaction Through Exaggerated Motion Synthesis,” by Michael Gielniak and Andrea Thomaz
“Trajectories and Keyframes for Kinesthetic Teaching: A Human-Robot Interaction Perspective,” by Baris Akgun, Maya Cakmak, Jae Wook Yoo and Andrea L. Thomaz
All five papers describe research geared toward the realization of in-home robots assisting humans with everyday activities. Ph.D. student Baris Akgun’s paper, for example, assumes the same real-life application scenario as Cakmak’s—a robot learning new tasks from a non-programmer—and examines whether robots learn more quickly from continuous, real-time demonstrations of a physical task, or from isolated key frames in the motion sequence. The research is nominated for Best Paper at HRI 2012.
“Georgia Tech is certainly a leader in the field of human-robot interaction; we have more than 10 faculty across campus for whom HRI is a primary research area,” Thomaz said. “Additionally, the realization of ‘personal robots’ is a shared vision of the whole robotics faculty—and a mission of the RIM research center.”Contacts
Michael Terrazas | EurekAlert!
Researchers produce synthetic Hall Effect to achieve one-way radio transmission
13.09.2019 | University of Illinois College of Engineering
Penn engineers' new topological insulator reroutes photonic 'traffic' on the fly
13.09.2019 | University of Pennsylvania
Researchers from the Department of Atomically Resolved Dynamics of the Max Planck Institute for the Structure and Dynamics of Matter (MPSD) at the Center for Free-Electron Laser Science in Hamburg, the University of Hamburg and the European Molecular Biology Laboratory (EMBL) outstation in the city have developed a new method to watch biomolecules at work. This method dramatically simplifies starting enzymatic reactions by mixing a cocktail of small amounts of liquids with protein crystals. Determination of the protein structures at different times after mixing can be assembled into a time-lapse sequence that shows the molecular foundations of biology.
The functions of biomolecules are determined by their motions and structural changes. Yet it is a formidable challenge to understand these dynamic motions.
At the International Symposium on Automotive Lighting 2019 (ISAL) in Darmstadt from September 23 to 25, 2019, the Fraunhofer Institute for Organic Electronics, Electron Beam and Plasma Technology FEP, a provider of research and development services in the field of organic electronics, will present OLED light strips of any length with additional functionalities for the first time at booth no. 37.
Almost everyone is familiar with light strips for interior design. LED strips are available by the metre in DIY stores around the corner and are just as often...
Later during this century, around 2060, a paradigm shift in global energy consumption is expected: we will spend more energy for cooling than for heating....
Researchers from the Department of Atomically Resolved Dynamics of the Max Planck Institute for the Structure and Dynamics of Matter (MPSD) at the Center for Free-Electron Laser Science in Hamburg, the University of Potsdam (both in Germany) and the University of Toronto (Canada) have pieced together a detailed time-lapse movie revealing all the major steps during the catalytic cycle of an enzyme. Surprisingly, the communication between the protein units is accomplished via a water-network akin to a string telephone. This communication is aligned with a ‘breathing’ motion, that is the expansion and contraction of the protein.
This time-lapse sequence of structures reveals dynamic motions as a fundamental element in the molecular foundations of biology.
Two research teams have succeeded simultaneously in measuring the long-sought Thorium nuclear transition, which enables extremely precise nuclear clocks. TU Wien (Vienna) is part of both teams.
If you want to build the most accurate clock in the world, you need something that "ticks" very fast and extremely precise. In an atomic clock, electrons are...
10.09.2019 | Event News
04.09.2019 | Event News
29.08.2019 | Event News
17.09.2019 | Materials Sciences
17.09.2019 | Health and Medicine
17.09.2019 | Ecology, The Environment and Conservation