Download but quantum computing is in its infancy.

Survey
yes no Was this document useful for you?
   Thank you for your participation!

* Your assessment is very important for improving the workof artificial intelligence, which forms the content of this project

Document related concepts

Scalar field theory wikipedia , lookup

Basil Hiley wikipedia , lookup

Relativistic quantum mechanics wikipedia , lookup

Probability amplitude wikipedia , lookup

Bohr–Einstein debates wikipedia , lookup

Delayed choice quantum eraser wikipedia , lookup

Quantum decoherence wikipedia , lookup

Wave–particle duality wikipedia , lookup

Density matrix wikipedia , lookup

Measurement in quantum mechanics wikipedia , lookup

Theoretical and experimental justification for the Schrödinger equation wikipedia , lookup

Double-slit experiment wikipedia , lookup

Quantum electrodynamics wikipedia , lookup

Particle in a box wikipedia , lookup

Quantum field theory wikipedia , lookup

Coherent states wikipedia , lookup

Max Born wikipedia , lookup

Copenhagen interpretation wikipedia , lookup

Bell's theorem wikipedia , lookup

Quantum dot wikipedia , lookup

Path integral formulation wikipedia , lookup

Hydrogen atom wikipedia , lookup

Quantum entanglement wikipedia , lookup

Quantum fiction wikipedia , lookup

Many-worlds interpretation wikipedia , lookup

Orchestrated objective reduction wikipedia , lookup

Symmetry in quantum mechanics wikipedia , lookup

EPR paradox wikipedia , lookup

History of quantum field theory wikipedia , lookup

Interpretations of quantum mechanics wikipedia , lookup

Quantum group wikipedia , lookup

Quantum key distribution wikipedia , lookup

Quantum state wikipedia , lookup

Quantum machine learning wikipedia , lookup

Quantum teleportation wikipedia , lookup

Quantum cognition wikipedia , lookup

Quantum computing wikipedia , lookup

Hidden variable theory wikipedia , lookup

Canonical quantization wikipedia , lookup

T-symmetry wikipedia , lookup

Transcript
Quantum computing the ‘Holy Grail’ for University of Waterloo scientists
“Before we make one chip that works, we make 1,000 that don’t,” says physicist
Raymond Laflamme at the University of Waterloo
WATERLOO, ONT.—Say hello to Raymond Laflamme — and forget everything
you thought you knew about computers.
Forget about bits and bytes and silicon chips. Forget about mega-this and giga-that.
Discard the impatience you sometimes feel while waiting for your laptop to
perform some supposedly “complicated” function. Rid your mind of all those annoying little icons — from
primitive-looking wristwatches to twirling coloured spheres — that indicate time is passing, and passing, while
your doddering digital device stumbles from one slowpoke application to another.
Be done with all of that and, instead, start thinking about quantum mechanics. Where computers are concerned,
quantum mechanics are the future.
“It’s not an instantaneous transition,” says Laflamme, 50, formerly of Quebec City and now executive director
of the Institute for Quantum Computing at the University of Waterloo. “It’s a process. But a quantum computer
is the Holy Grail.”
Comparing a fully realized quantum computing device to even the most powerful classical computers now in
existence is a bit like comparing a vast galaxy of swirling planets and radiant stars to a basement bachelor
apartment in a transit-starved neighbourhood of Toronto. In other words, there is really no comparison at all.
But there are two formidable obstacles facing anyone in the fledgling quantum computing business. First,
quantum computers are extremely difficult to build, so difficult that we haven’t actually built one yet, or not
really.
Second, no matter how difficult they might be to fabricate, quantum computers are even more difficult to
explain.
Despite these obstacles, quantum computing, or the prospect of quantum computing, is very much of the
moment. Last month, two scientists — Frenchman Serge Haroche and American David J. Wineland —
gathered in Stockholm to receive this year’s Nobel Prize for physics, honouring work they have done separately
to advance the development of quantum computers. They are far from alone.
In offices and laboratories scattered around the globe — in Russia, China, Singapore, Europe, the United States
and Canada — some of the finest scientific minds are busily pursuing the same goal, a computing device whose
fundamental components are so small as to be invisible to humans but whose calculating power would far
outstrip any conventional computer you could imagine, even one as big as the universe itself.
Here in Canada, the main centre for this area of research — there are others — was founded 10 years ago at the
University of Waterloo. Now the Institute for Quantum Computing is in the process of moving into palatial
new quarters, the Mike & Ophelia Lazaridis Quantum-Nano Centre, designed by Toronto architectural firm
KPMB and built at a cost of $160 million.
Headed by Laflamme, the institute brings 25 faculty members from roughly half a dozen countries together with
about 40 post-doctoral fellows and roughly 90 students in the quest for something that is at once almost
unimaginably small — on the scale of atoms and electrons — and yet colossally powerful.
Perhaps the best way to begin an explanation of what a quantum computer is might be to explain what it is not.
It is not a classical computer.
“A small quantum computer can do things no classical computer can do,” says Laflamme, a wiry, animated man
with mobile facial features and a changeable yet nearly constant grin. “It’s a totally different class.”
Let’s face it, a classical computer is essentially a pretty stupid device. Its guts are a network of electrical circuits
that can do just two things. They can be on or they can be off. They can be set at either “one” or “zero.”
Essentially, that’s all that classical computing is, a highly developed narrative of zeros or ones.
Here’s how it works.
A single computer circuit, called a bit, has just two possible states (0 or 1). Double that to two circuits and you
get four states (00, 11, 01 or 10). With three circuits, you double the possible number of combinations yet again,
this time to eight (001, 011, 111, 110, 100, 000, 101, 010). When you get to eight circuits, you have 256
possible combinations, and this arrangement is called a “byte.”
The processing power and memory capacity of laptop computers are nowadays calculated in gigabytes,
meaning billions of bytes, and these numbers just keep getting bigger — but the trend is not going to last
forever, not for classical computers.
At least since the 1970s, computer scientists have been familiar with a phenomenon known as Moore’s law, the
principle that technological innovation causes computing power to double on average every two years. Or, to
put the same idea another way, a computer circuit decreases in size, on average, by 50 per cent every two years,
meaning you can fit twice as many into the same space.
At this rate, the computer industry is on track to run head-on into an immovable wall in approximately 10 years.
Sometime between 2020 and 2030, processing circuits will have become so small — as small as atoms — that
they will be regulated by the same physical imperative that governs atoms: the inability to get any smaller.
A laptop in 2025 will be vastly more powerful than any laptop today, but that will be the end of it. Classical
computing will have reached a natural limit and will cease to evolve.
With any luck, however, this is where quantum computing will come in.
Put in its sparest terms, the difference between classical computing and quantum computing is a simple matter
of a single conjunction — the substitution of “and” for “or.”
Recall that a conventional computer circuit, or a bit, can represent a 0 or a 1. It can be switched off or it can
switched on. One or the other. But a quantum bit is a device of an entirely different kind.
Owing to the bizarre properties that govern the microscopic realm of quantum mechanics, a quantum bit, known
as a “qubit”, can represent a 0 and a 1, both of them at the same time. It can be simultaneously off and on.
Known as the uncertainty principle, this seemingly impossible attribute has no equivalent in the world that we
humans can see and touch and feel, the domain of automobiles, maple trees, wide-screen TVs, bachelor
apartments, cufflinks, pumpkin pie and classical computers. But it seems to be critical to the nature of atomic
and subatomic particles (including electrons, photons, protons, neutrinos and other minuscule dots of nature)
that they exist in a haze of possibility rather than in any fixed state.
A microscopic particle can be both here and there at the same instant of time, which means that it can also be in
every intermediate location between here and there, also at the same time.
“It turns out that this is incredibly powerful,” says Laflamme. “Just changing an ‘or’ to an ‘and.’”
If you don’t understand how a thing can be in two places at once, don’t worry. Nobody does. Nonetheless, it’s
true.
Even though an atom-sized classical bit would be extremely small, it would still be restricted to just two
possible states — off and on. By contrast, an atom-sized qubit would possess a near infinity of states,
simultaneously.
What this means, by a common calculation, is that a quantum computer consisting of just 300 atoms would
possess more sheer calculating power than a conventional machine harnessing every last atom in the universe,
yielding an almost unimaginable increase in our ability to solve problems of immense complexity.
Such a machine could perform in an instant certain difficult calculations — for example, the factoring of very
large numbers — that the largest conventional computer would need billions of years to figure out, if it could
figure them out at all.
It turns out that classical computers are not very good at factoring large numbers, a weakness that has long been
exploited by cryptographers to safeguard data on the Internet. It is easy to multiply two prime numbers in order
to produce a much larger number, but it turns out to be horrendously difficult to engineer the same process in
reverse, to find the two prime divisors of a large number, a process called factoring.
The only way classical computers can address the challenge is by systematic trial and error — trying out two
numbers to see if they work, discarding them, trying out two different numbers, and so on. There’s no shortcut.
This defect in conventional computers is used to secure your banking information on the Internet, along with
much else. Even armed with powerful computers, would-be hackers still cannot find a way to expose the key —
the two original prime numbers used to secure the code that protects your data. The only method available is
trial and error, examining every possible combination of divisors, one pair at a time, which could take forever.
By contrast, a quantum computer could crack such privacy barriers in an instant, by the dazzling expedient of
testing every possible combination of divisors, not one by one, but all at once, something no conventional
computer could do. The right answer would reveal itself almost immediately.
No wonder governments around the world are pouring huge amounts of money into quantum computing.
Whoever is first to build a sophisticated quantum machine would suddenly be able to crack just about every
secret code in cyberspace. But that is far from all.
According to Laflamme, quantum computers will eventually ignite a revolution in a variety of fields other than
cryptography. These might well include the development of new drugs, or the creation of new superconducting
materials, or a range of other innovations currently impossible to predict. “We are just scratching at the surface
of this,” he says. “Today, we are just dipping our big toe in the shallow end.”
At least in a very basic form, quantum computers do now exist, but they are extremely limited both in
complexity and capability, restricted to no more than a dozen qubits and able to perform only the simplest of
mathematical feats. The highest number so far factored by a quantum device, using an algorithm developed in
the 1990s by U.S. scientist Peter Shor, is 21.
This calculation might not seem very impressive — the answer is three and seven, as anybody with a Grade 8
education already knows — but quantum computing is in its infancy.
Laflamme takes a visitor on a tour of the laboratory facilities at IQC’s existing premises — windowless rooms
cluttered with an endless variety of complicated machines and devices, pipes and wires, including equipment
that fabricates qubit chips from aluminum and niobium. The resulting slivers of metal are about the size of a
baby’s fingernail and are extremely difficult to produce.
“Before we make one chip that works, we make 1,000 that don’t,” says Laflamme.
Even the chips that do work aren’t up to much in computational terms.
“If you want to add two plus two, we can do it,” he says. “If you want to add very large numbers, we can’t.”
But the journey is made in stages, and Laflamme believes that a quantum machine able to outpace conventional
computers is no longer such a distant prospect.
“The boundary is about 50 qubits,” he says. “When we get there, we’re reaching a place where we’re departing
the classical world.”
One big obstacle involves the inherent instability of the quantum universe, a territory where superposition —
the ability of particles to occupy multiple locations at once — mysteriously collapses the very instant a human
observer interferes with the process in any way, even by looking at a particle through an electron microscope.
Somewhat like a performer struck by stage fright, an observed particle will instantly stop behaving in a quantum
fashion and instead adopt just one position or state. Somehow, researchers will have to overcome this
phenomenon.
Laflamme, for one, is certain they will. “In five to 10 years,” he says, “our goal is to get a quantum device to
the market.”