Researchers develop the first-ever quantum device that detects and corrects its own errors
When scientists develop a full quantum computer, the world of computing will undergo a revolution of sophistication, speed and energy efficiency that will make even our beefiest conventional machines seem like Stone Age clunkers by comparison.
But, before that happens, quantum physicists like the ones in UC Santa Barbara's physics professor John Martinis' lab will have to create circuitry that takes advantage of the marvelous computing prowess promised by the quantum bit ("qubit"), while compensating for its high vulnerability to environmentally-induced error.
In what they are calling a major milestone, the researchers in the Martinis Lab have developed quantum circuitry that self-checks for errors and suppresses them, preserving the qubits' state(s) and imbuing the system with the highly sought-after reliability that will prove foundational for the building of large-scale superconducting quantum computers.
It turns out keeping qubits error-free, or stable enough to reproduce the same result time and time again, is one of the major hurdles scientists on the forefront of quantum computing face.
"One of the biggest challenges in quantum computing is that qubits are inherently faulty," said Julian Kelly, graduate student researcher and co-lead author of a research paper that was published in the journal Nature. "So if you store some information in them, they'll forget it."
Unlike classical computing, in which the computer bits exist on one of two binary ("yes/no", or "true/false") positions, qubits can exist at any and all positions simultaneously, in various dimensions. It is this property, called "superpositioning," that gives quantum computers their phenomenal computational power, but it is also this characteristic which makes qubits prone to "flipping," especially when in unstable environments, and thus difficult to work with.
"It's hard to process information if it disappears," said Kelly.
However, that obstacle may just have been cleared by Kelly, postdoctoral researcher Rami Barends, staff scientist Austin Fowler and others in the Martinis Group.
The error process involves creating a scheme in which several qubits work together to preserve the information, said Kelly. To do this, information is stored across several qubits.
"And the idea is that we build this system of nine qubits, which can then look for errors," he said. Qubits in the grid are responsible for safeguarding the information contained in their neighbors, he explained, in a repetitive error detection and correction system that can protect the appropriate information and store it longer than any individual qubit can.
"This is the first time a quantum device has been built that is capable of correcting its own errors," said Fowler. For the kind of complex calculations the researchers envision for an actual quantum computer, something up to a hundred million qubits would be needed, but before that a robust self-check and error prevention system is necessary.
Key to this quantum error detection and correction system is a scheme developed by Fowler, called the surface code. It uses parity information -- the measurement of change from the original data (if any) -- as opposed to the duplication of the original information that is part of the process of error detection in classical computing. That way, the actual original information that is being preserved in the qubits remains unobserved.
Why? Because quantum physics.
"You can't measure a quantum state, and expect it to still be quantum," explained Barends. The very act of measurement locks the qubit into a single state and it then loses its superpositioning power, he said. Therefore, in something akin to a Sudoku puzzle, the parity values of data qubits in a qubit array are taken by adjacent measurement qubits, which essentially assess the information in the data qubits by measuring around them.
"So you pull out just enough information to detect errors, but not enough to peek under the hood and destroy the quantum-ness," said Kelly.
This development represents a meeting of the best in the science behind the physical and the theoretical in quantum computing -- the latest in qubit stabilization and advances in the algorithms behind the logic of quantum computing.
"It's a major milestone," said Barends. "Because it means that the ideas people have had for decades are actually doable in a real system."
The Martinis Group continues to refine its research to develop this important new tool. This particular quantum error correction has been proved to protect against the "bit-flip" error, however the researchers have their eye on correcting the complimentary error called a "phase-flip," as well as running the error correction cycles for longer periods to see what behaviors might emerge.
Martinis and the senior members of his research group have, since this research was performed, entered into a partnership with Google.
Sonia Fernandez | EurekAlert!
Terahertz spectroscopy goes nano
20.10.2017 | Brown University
New software speeds origami structure designs
12.10.2017 | Georgia Institute of Technology
Salmonellae are dangerous pathogens that enter the body via contaminated food and can cause severe infections. But these bacteria are also known to target...
University of Maryland researchers contribute to historic detection of gravitational waves and light created by event
On August 17, 2017, at 12:41:04 UTC, scientists made the first direct observation of a merger between two neutron stars--the dense, collapsed cores that remain...
Seven new papers describe the first-ever detection of light from a gravitational wave source. The event, caused by two neutron stars colliding and merging together, was dubbed GW170817 because it sent ripples through space-time that reached Earth on 2017 August 17. Around the world, hundreds of excited astronomers mobilized quickly and were able to observe the event using numerous telescopes, providing a wealth of new data.
Previous detections of gravitational waves have all involved the merger of two black holes, a feat that won the 2017 Nobel Prize in Physics earlier this month....
Material defects in end products can quickly result in failures in many areas of industry, and have a massive impact on the safe use of their products. This is why, in the field of quality assurance, intelligent, nondestructive sensor systems play a key role. They allow testing components and parts in a rapid and cost-efficient manner without destroying the actual product or changing its surface. Experts from the Fraunhofer IZFP in Saarbrücken will be presenting two exhibits at the Blechexpo in Stuttgart from 7–10 November 2017 that allow fast, reliable, and automated characterization of materials and detection of defects (Hall 5, Booth 5306).
When quality testing uses time-consuming destructive test methods, it can result in enormous costs due to damaging or destroying the products. And given that...
Using a new cooling technique MPQ scientists succeed at observing collisions in a dense beam of cold and slow dipolar molecules.
How do chemical reactions proceed at extremely low temperatures? The answer requires the investigation of molecular samples that are cold, dense, and slow at...
23.10.2017 | Event News
17.10.2017 | Event News
10.10.2017 | Event News
24.10.2017 | Physics and Astronomy
24.10.2017 | Physics and Astronomy
24.10.2017 | Life Sciences