Scientists find quantum computing via error correction
In what could be called as a big leap in quantum mechanics, a research team at Yale have achieved the entanglement of three solid-state qubits, or quantum bits, for the first time.
London: In what could be called as a big leap in quantum mechanics, a research team at Yale have achieved the entanglement of three solid-state qubits, or quantum bits, for the first time.
Their accomplishment is a first step towards quantum error correction, a crucial aspect of future quantum computing.
"Entanglement between three objects has been demonstrated before with photons and charged particles. But this is the first three-qubit, solid-state device that looks and feels like a conventional microprocessor," Nature quoted Steven Girvin, the Eugene Higgins Professor of Physics & Applied Physics at Yale and an author of the paper, as saying.
The new result builds on the team``s development last year of the world``s first rudimentary solid-state quantum processor, which they demonstrated was capable of executing simple algorithms using two qubits.
The team, led by Robert Schoelkopf, the William A. Norton Professor of Applied Physics and Physics at Yale, used artificial "atoms"—actually made up of a billion aluminum atoms that behave as a single entity—as their qubits.
These "atoms" can occupy two different energy states, akin to the "1" and "0" or "on" and "off" states of regular bits used in conventional computers.
The strange laws of quantum mechanics, however, allow for qubits to be placed in a "superposition" of these two states at the same time, resulting in far greater information storage and processing power.
In this new study, the team was able to achieve an entangled state by placing the three qubits in a superposition of two possibilities—all three were either in the 0 state or the 1 state.
They were able to attain this entangled state 88 percent of the time.
With the particular entangled state the team achieved, they also demonstrated for the first time the encoding of quantum information from a single qubit into three qubits using a so-called repetition code.
"This is the first step towards quantum error correction, which, as in a classical computer, uses the extra qubits to allow the computer to operate correctly even in the presence of occasional errors," said Girvin.
Such errors might include a cosmic ray hitting one of the qubits and switching it from a 0 to a 1 state, or vice versa.
By replicating the qubits, the computer can confirm whether all three are in the same state (as expected) by checking each one against the others.
"Error correction is one of the holy grails in quantum computing today. It takes at least three qubits to be able to start doing it, so this is an exciting step," said Schoelkopf.
The study has been published in the latest issue of the journal Nature.