Or at least that’s not too far off from what University of Vermont roboticist Josh Bongard has discovered, as he reports in the January 10 online edition of the Proceedings of the National Academy of Sciences.
In a first-of-its-kind experiment, Bongard created both simulated and actual robots that, like tadpoles becoming frogs, change their body forms while learning how to walk. And, over generations, his simulated robots also evolved, spending less time in “infant” tadpole-like forms and more time in “adult” four-legged forms.
These evolving populations of robots were able to learn to walk more rapidly than ones with fixed body forms. And, in their final form, the changing robots had developed a more robust gait -- better able to deal with, say, being knocked with a stick -- than the ones that had learned to walk using upright legs from the beginning.
“This paper shows that body change, morphological change, actually helps us design better robots,” Bongard says. “That’s never been attempted before.”
Bongard’s research, supported by the National Science Foundation, is part of a wider venture called evolutionary robotics. “We have an engineering goal,” he says “to produce robots as quickly and consistently as possible.” In this experimental case: upright four-legged robots that can move themselves to a light source without falling over.
“But we don’t know how to program robots very well,” Bongard says, because robots are complex systems. In some ways, they are too much like people for people to easily understand them.
“They have lots of moving parts. And their brains, like our brains, have lots of distributed materials: there’s neurons and there’s sensors and motors and they’re all turning on and off in parallel,” Bongard says, “and the emergent behavior from the complex system which is a robot, is some useful task like clearing up a construction site or laying pavement for a new road.” Or at least that’s the goal.
But, so far, engineers have been largely unsuccessful at creating robots that can continually perform simple, yet adaptable, behaviors in unstructured or outdoor environments.
Which is why Bongard, an assistant professor in UVM’s College of Engineering and Mathematical Sciences, and other robotics experts have turned to computer programs to design robots and develop their behaviors -- rather than trying to program the robots’ behavior directly.
His new work may help.
Using a sophisticated computer simulation, Bongard unleashed a series of synthetic beasts that move about in a 3-dimensional space. “It looks like a modern video game,” he says. Each creature -- or, rather, generations of the creatures -- then run a software routine, called a genetic algorithm, that experiments with various motions until it develops a slither, shuffle, or walking gait -- based on its body plan -- that can get it to the light source without tipping over.
“The robots have 12 moving parts,” Bongard says. “They look like the simplified skeleton of a mammal: it’s got a jointed spine and then you have four sticks -- the legs -- sticking out.”
Some of the creatures begin flat to the ground, like tadpoles or, perhaps, snakes with legs; others have splayed legs, a bit like a lizard; and others ran the full set of simulations with upright legs, like mammals.
And why do the generations of robots that progress from slithering to wide legs and, finally, to upright legs, ultimately perform better, getting to the desired behavior faster?
“The snake and reptilian robots are, in essence, training wheels,” says Bongard, “they allow evolution to find motion patterns quicker, because those kinds of robots can’t fall over. So evolution only has to solve the movement problem, but not the balance problem, initially. Then gradually over time it’s able to tackle the balance problem after already solving the movement problem.”
Sound anything like how a human infant first learns to roll, then crawl, then cruise along the coffee table and, finally, walk?
“Yes,” says Bongard, “We’re copying nature, we’re copying evolution, we’re copying neural science when we’re building artificial brains into these robots.” But the key point is that his robots don’t only evolve their artificial brain -- the neural network controller -- but rather do so in continuous interaction with a changing body plan. A tadpole can’t kick its legs, because it doesn’t have any yet; it’s learning some things legless and others with legs.
And this may help to explain the most surprising -- and useful -- finding in Bongard’s study: the changing robots were not only faster in getting to the final goal, but afterward were more able to deal with new kinds of challenges that they hadn’t before faced, like efforts to tip them over.
Bongard is not exactly sure why this is, but he thinks it’s because controllers that evolved in the robots whose bodies changed over generations learned to maintain the desired behavior over a wider range of sensor-motor arrangements than controllers evolved in robots with fixed body plans. It seem that learning to walk while flat, squat, and then upright, gave the evolving robots resilience to stay upright when faced with new disruptions. Perhaps what a tadpole learns before it has legs makes it better able to use its legs once they grow.
“Realizing adaptive behavior in machines has to date focused on dynamic controllers, but static morphologies,” Bongard writes in his PNAS paper “This is an inheritance from traditional artificial intelligence in which computer programs were developed that had no body with which to affect, and be affected by, the world.”
“One thing that has been left out all this time is the obvious fact that in nature it’s not that the animal’s body stays fixed and its brain gets better over time,” he says, “in natural evolution animals bodies and brains are evolving together all the time.” A human infant, even if she knew how, couldn’t walk: her bones and joints aren’t up to the task until she starts to experience stress on the foot and ankle.
That hasn’t been done in robotics for an obvious reason: “it’s very hard to change a robot’s body,” Bongard says, “it’s much easier to change the programming inside its head.”
Still, Bongard gave it a try. After running 5000 simulations, each taking 30 hours on the parallel processors in UVM’s Vermont Advanced Computing Center -- “it would have taken 50 or 100 years on a single machine,” Bongard says—he took the task into the real world.
“We built a relatively simple robot, out of a couple of Lego Mindstorm kits, to demonstrate that you actually could do it,” he says. This physical robot is four-legged, like in the simulation, but the Lego creature wears a brace on its front and back legs. “The brace gradually tilts the robot,” as the controller searches for successful movement patterns, Bongard says, “so that the legs go from horizontal to vertical, from reptile to quadruped.
“While the brace is bending the legs, the controller is causing the robot to move around, so it’s able to move its legs, and bend its spine,” he says, “it’s squirming around like a reptile flat on the ground and then it gradually stands up until, at the end of this movement pattern, it’s walking like a coyote.”
“It’s a very simple prototype,” he says, “but it works; it’s a proof of concept.”
Joshua Brown | Newswise Science News
New approach: Researchers succeed in directly labelling and detecting an important RNA modification
30.04.2018 | Westfälische Wilhelms-Universität Münster
Start of work for the world's largest electric truck
20.04.2018 | Empa - Eidgenössische Materialprüfungs- und Forschungsanstalt
A research team led by physicists at the Technical University of Munich (TUM) has developed molecular nanoswitches that can be toggled between two structurally different states using an applied voltage. They can serve as the basis for a pioneering class of devices that could replace silicon-based components with organic molecules.
The development of new electronic technologies drives the incessant reduction of functional component sizes. In the context of an international collaborative...
At the LASYS 2018, from June 5th to 7th, the Laser Zentrum Hannover e.V. (LZH) will be showcasing processes for the laser material processing of tomorrow in hall 4 at stand 4E75. With blown bomb shells the LZH will present first results of a research project on civil security.
At this year's LASYS, the LZH will exhibit light-based processes such as cutting, welding, ablation and structuring as well as additive manufacturing for...
There are videos on the internet that can make one marvel at technology. For example, a smartphone is casually bent around the arm or a thin-film display is rolled in all directions and with almost every diameter. From the user's point of view, this looks fantastic. From a professional point of view, however, the question arises: Is that already possible?
At Display Week 2018, scientists from the Fraunhofer Institute for Applied Polymer Research IAP will be demonstrating today’s technological possibilities and...
So-called quantum many-body scars allow quantum systems to stay out of equilibrium much longer, explaining experiment | Study published in Nature Physics
Recently, researchers from Harvard and MIT succeeded in trapping a record 53 atoms and individually controlling their quantum state, realizing what is called a...
The historic first detection of gravitational waves from colliding black holes far outside our galaxy opened a new window to understanding the universe. A...
02.05.2018 | Event News
13.04.2018 | Event News
12.04.2018 | Event News
24.05.2018 | Ecology, The Environment and Conservation
24.05.2018 | Medical Engineering
24.05.2018 | Physics and Astronomy