The fourth decade of this century could see Europe participating in a manned mission to Mars in what would be one of humanity's grandest space expeditions ever.
Aurora is ESA's programme aimed at the long-term robotic and human exploration of the Solar System, with Mars and the Moon as the main targets.A human mission to the Red Planet would be a major, multi-year undertaking requiring fantastic, entirely new capabilities such as automated cargo vessels, prepositioned supplies and tools, and communication and navigation satellites in Mars orbit similar to Earth's current GPS systems.
ExoMars will explore the biological environment on Mars in preparation for further robotic and, later, human activity. Data from the mission will also provide invaluable input for broader studies of exobiology – the search for life on other planets.
The main element of the mission is a wheeled, robotic rover vehicle, similar in concept to NASA's current Mars Rover mission, but having different scientific objectives and improved capabilities.
ExoMars: a wheeled rover delivered in a dramatic direct approach
The mission will likely consist of a carrier spacecraft, a descent module, some sort of landing system, and the surface rover, and the mission profile is likely to include a dramatic direct approach to Mars, with the carrier spacecraft discarded after the rover detaches itself for descent to the surface.
The rover will use solar arrays to generate electricity, and will travel over the rocky orange-red surface of Mars, transporting an up to 12-kilogram experimental payload including a first-ever lightweight drilling system, as well as a sampling and handling device, and a set of scientific instruments to search for signs of past or present life.
Due to distance time-lag and complexity, ExoMars will self-navigate using 'smart' electro-optics to visually sense and interpret the surrounding terrain and will be capable of operating autonomously using intelligent onboard software.
Automated control a major advance
This automated mode of operation is a major advance for ESA, long used to controlling spacecraft directly using human controllers. And not only will the rover's onboard control systems be new.
"ExoMars will require entirely new techniques and technology for several aspects of the Earth-based rover control system, not just an upgrade of what we have today," says Mike McKay, a senior spacecraft controller and Mars expert based at ESOC, ESA's Spacecraft Operations Centre, in Darmstadt, Germany.
ESA spacecraft have long had some ability to make independent decisions based on external influences. For example, onboard instruments will automatically shut down if solar radiation suddenly rises, or the spacecraft will automatically switch into a diagnostic 'safe mode' if anything goes wrong. But for the most part, lengthy instructions still must be pre-programmed by mission controllers on Earth and sent up for later, step-by-step, execution.
And ESA controllers have never before operated a mission that moved about on the surface of another body; Huygens – which touched down successfully on Titan in 2005 – was an atmospheric probe and not a lander, although it functioned briefly after reaching Titan's surface.
Robotic task: traverse kilometres of terrain in search of life
In one typical example of the rover's autonomous operation, ground controllers might radio up a high-level command telling it to drive to a scientifically interesting spot anywhere from 500 to 2000 metres away and conduct science operations, such as drilling beneath the surface to sample soil for life signs. But the vehicle would handle the details of the move on its own.
It would survey the ground with a 3D camera, create a digital terrain model, verify its present location, run internal simulations and then make an autonomous decision on the best path to follow, based on obstacles, the rover's current status and risk/resource considerations.
"Then it will drive itself to the target. We expect its target accuracy to be within one-half metre over a traverse of 20 metres," says Bob Chesson, head of the Human Spaceflight and Exploration Operations Department in ESA’s Operations directorate.
ExoMars profits from current robotic explorers
As the next generation of robot, ExoMars will profit from lessons learned from the current generation, including NASA's Mars Explorer Rover (MER) mission, including the need for improved locomotion ability, improved local terrain sensing – to avoid ground slippage – and the need for higher autonomy to transverse cluttered terrain.
Earlier missions, such as NASA's Sojourner rover in 1997, used an even less sophisticated approach, with Sojourner sensing its surrounding terrain, but then with all processing and path planning being done on Earth. "We're not shy in trying to learn from the experiences of our sister agencies," says Chesson.
Innovative ground control to enable autonomous functioningFor ExoMars, the controllers on Earth would most likely be located in a 'rover dedicated control room', similar in concept to the dedicated control rooms (DCR) that ESA now sets up for individual missions that orbit planets.
ESOC will serve as the overall mission operations control centre (MOCC), controlling the launch and early orbit phase (LEOP), the cruise to Mars, the separation and landing of the Descent Module and the Rover egress, with management of rover surface operations likely to be conducted from the Rover Operation Centre located at ALTEC, the Advanced Logistic Technology Engineering Center, in Turin, Italy.
"The design of the rover ground control system, or ground segment, depends on the scientific and operational goals of the rover, which are not yet final, so the ground system is still evolving," says Chesson. "In principle, the basic telemetry and telecommand functions would be essentially the same as now, but it will have significantly new capabilities to allow for the rover's autonomous functioning."
The ground control system will at least require computing facilities to enable high-level mission planning tools and to allow monitoring of the rover's digital terrain and 3D modelling, ground path and trajectory planning, on-ground simulation and tight integration with the payload control and scientific operations.
"Classic direct control methods just won't work when we operate on the surface of Mars in an unstructured environment and with a significant signal time delay, says Reinhold Bertrand," a planning engineer and robotics expert at ESOC. "ExoMars will require a change in culture; we have to 'let the child walk on its own' while we develop a truly interdisciplinary operations concept."
Jocelyne Landeau-Constantin | alfa
First results of NSTX-U research operations
26.10.2016 | DOE/Princeton Plasma Physics Laboratory
Scientists discover particles similar to Majorana fermions
25.10.2016 | Chinese Academy of Sciences Headquarters
Ultrafast lasers have introduced new possibilities in engraving ultrafine structures, and scientists are now also investigating how to use them to etch microstructures into thin glass. There are possible applications in analytics (lab on a chip) and especially in electronics and the consumer sector, where great interest has been shown.
This new method was born of a surprising phenomenon: irradiating glass in a particular way with an ultrafast laser has the effect of making the glass up to a...
Terahertz excitation of selected crystal vibrations leads to an effective magnetic field that drives coherent spin motion
Controlling functional properties by light is one of the grand goals in modern condensed matter physics and materials science. A new study now demonstrates how...
Researchers from the Institute for Quantum Computing (IQC) at the University of Waterloo led the development of a new extensible wiring technique capable of controlling superconducting quantum bits, representing a significant step towards to the realization of a scalable quantum computer.
"The quantum socket is a wiring method that uses three-dimensional wires based on spring-loaded pins to address individual qubits," said Jeremy Béjanin, a PhD...
In a paper in Scientific Reports, a research team at Worcester Polytechnic Institute describes a novel light-activated phenomenon that could become the basis for applications as diverse as microscopic robotic grippers and more efficient solar cells.
A research team at Worcester Polytechnic Institute (WPI) has developed a revolutionary, light-activated semiconductor nanocomposite material that can be used...
By forcefully embedding two silicon atoms in a diamond matrix, Sandia researchers have demonstrated for the first time on a single chip all the components needed to create a quantum bridge to link quantum computers together.
"People have already built small quantum computers," says Sandia researcher Ryan Camacho. "Maybe the first useful one won't be a single giant quantum computer...
14.10.2016 | Event News
14.10.2016 | Event News
12.10.2016 | Event News
26.10.2016 | Power and Electrical Engineering
26.10.2016 | Awards Funding
26.10.2016 | Power and Electrical Engineering