Forum for Science, Industry and Business

Sponsored by:     3M 
Search our Site:

 

Quantum computers may be easier to build than predicted

03.03.2005


The new NIST architecture for quantum computing relies on several levels of error checking to ensure the accuracy of quantum bits (qubits). The image above illustrates how qubits are grouped in blocks to form the levels. To implement the architecture with three levels, a series of operations is performed on 36 qubits (bottom row)each one representing either a 1, a 0, or both at once. The operations on the nine sets of qubits produce two reliably accurate qubits (top row). The purple spheres represent qubits that are either used in error detection or in actual computations. The yellow spheres are qubits that are measured to detect or correct errors but are not used in final computations.


A full-scale quantum computer could produce reliable results even if its components performed no better than today’s best first-generation prototypes, according to a paper in the March 3 issue in the journal Nature* by a scientist at the Commerce Department’s National Institute of Standards and Technology (NIST).

In theory, such a quantum computer could be used to break commonly used encryption codes, to improve optimization of complex systems such as airline schedules, and to simulate other complex quantum systems.

A key issue for the reliability of future quantum computers--which would rely on the unusual properties of nature’s smallest particles to store and process data--is the fragility of quantum states. Today’s computers use millions of transistors that are switched on or off to reliably represent values of 1 or 0. Quantum computers would use atoms, for example, as quantum bits (qubits), whose magnetic and other properties would be manipulated to represent 1 or 0 or even both at the same time. These states are so delicate that qubit values would be unusually susceptible to errors caused by the slightest electronic "noise."



To get around this problem, NIST scientist Emanuel Knill suggests using a pyramid-style hierarchy of qubits made of smaller and simpler building blocks than envisioned previously, and teleportation of data at key intervals to continuously double-check the accuracy of qubit values. Teleportation was demonstrated last year by NIST physicists, who transferred key properties of one atom to another atom without using a physical link.

"There has been a tremendous gap between theory and experiment in quantum computing," Knill says. "It is as if we were designing today’s supercomputers in the era of vacuum tube computing, before the invention of transistors. This work reduces the gap, showing that building quantum computers may be easier than we thought. However, it will still take a lot of work to build a useful quantum computer."

Use of Knill’s architecture could lead to reliable computing even if individual logic operations made errors as often as 3 percent of the time--performance levels already achieved in NIST laboratories with qubits based on ions (charged atoms). The proposed architecture could tolerate several hundred times more errors than scientists had generally thought acceptable.

Knill’s findings are based on several months of calculations and simulations on large, conventional computer workstations. The new architecture, which has yet to be validated by mathematical proofs or tested in the laboratory, relies on a series of simple procedures for repeatedly checking the accuracy of blocks of qubits. This process creates a hierarchy of qubits at various levels of validation.

For instance, to achieve relatively low error probabilities in moderately long computations, 36 qubits would be processed in three levels to arrive at one corrected pair. Only the top-tier, or most accurate, qubits are actually used for computations. The more levels there are, the more reliable the computation will be.

Knill’s methods for detecting and correcting errors rely heavily on teleportation. Teleportation enables scientists to measure how errors have affected a qubit’s value while transferring the stored information to other qubits not yet perturbed by errors. The original qubit’s quantum properties would be teleported to another qubit as the original qubit is measured.

The new architecture allows trade-offs between error rates and computing resource demands. To tolerate 3 percent error rates in components, massive amounts of computing hardware and processing time would be needed, partly because of the "overhead" involved in correcting errors. Fewer resources would be needed if component error rates can be reduced further, Knill’s calculations show.

Laura Ost | EurekAlert!
Further information:
http://www.nist.gov

More articles from Physics and Astronomy:

nachricht Unconventional superconductor may be used to create quantum computers of the future
19.02.2018 | Chalmers University of Technology

nachricht Hubble sees Neptune's mysterious shrinking storm
16.02.2018 | NASA/Goddard Space Flight Center

All articles from Physics and Astronomy >>>

The most recent press releases about innovation >>>

Die letzten 5 Focus-News des innovations-reports im Überblick:

Im Focus: In best circles: First integrated circuit from self-assembled polymer

For the first time, a team of researchers at the Max-Planck Institute (MPI) for Polymer Research in Mainz, Germany, has succeeded in making an integrated circuit (IC) from just a monolayer of a semiconducting polymer via a bottom-up, self-assembly approach.

In the self-assembly process, the semiconducting polymer arranges itself into an ordered monolayer in a transistor. The transistors are binary switches used...

Im Focus: Demonstration of a single molecule piezoelectric effect

Breakthrough provides a new concept of the design of molecular motors, sensors and electricity generators at nanoscale

Researchers from the Institute of Organic Chemistry and Biochemistry of the CAS (IOCB Prague), Institute of Physics of the CAS (IP CAS) and Palacký University...

Im Focus: Hybrid optics bring color imaging using ultrathin metalenses into focus

For photographers and scientists, lenses are lifesavers. They reflect and refract light, making possible the imaging systems that drive discovery through the microscope and preserve history through cameras.

But today's glass-based lenses are bulky and resist miniaturization. Next-generation technologies, such as ultrathin cameras or tiny microscopes, require...

Im Focus: Stem cell divisions in the adult brain seen for the first time

Scientists from the University of Zurich have succeeded for the first time in tracking individual stem cells and their neuronal progeny over months within the intact adult brain. This study sheds light on how new neurons are produced throughout life.

The generation of new nerve cells was once thought to taper off at the end of embryonic development. However, recent research has shown that the adult brain...

Im Focus: Interference as a new method for cooling quantum devices

Theoretical physicists propose to use negative interference to control heat flow in quantum devices. Study published in Physical Review Letters

Quantum computer parts are sensitive and need to be cooled to very low temperatures. Their tiny size makes them particularly susceptible to a temperature...

All Focus news of the innovation-report >>>

Anzeige

Anzeige

VideoLinks
Industry & Economy
Event News

2nd International Conference on High Temperature Shape Memory Alloys (HTSMAs)

15.02.2018 | Event News

Aachen DC Grid Summit 2018

13.02.2018 | Event News

How Global Climate Policy Can Learn from the Energy Transition

12.02.2018 | Event News

 
Latest News

Contacting the molecular world through graphene nanoribbons

19.02.2018 | Materials Sciences

When Proteins Shake Hands

19.02.2018 | Materials Sciences

Cells communicate in a dynamic code

19.02.2018 | Life Sciences

VideoLinks
Science & Research
Overview of more VideoLinks >>>