Quantum computing is the area of study focused on developing computer technology based on the principles of quantum theory, which explains the nature and behavior of energy and matter on the quantum (atomic and subatomic) level. Development of a quantum computer, if practical, would mark a leap forward in computing capability far greater than that from the abacus to a modern day supercomputer, with performance gains in the billion-fold realm and beyond. The quantum computer, following the laws of quantum physics, would gain enormous processing power through the ability to be in multiple states, and to perform tasks using all possible permutations simultaneously. Current centers of research in quantum computing include MIT, IBM, Oxford University, and the Los Alamos National Laboratory.
The essential elements of quantum computing originated with Paul Benioff, working at Argonne National Labs, in 1981. He theorized a classical computer operating with some quantum mechanical principles. But it is generally accepted that David Deutsch of Oxford University provided the critical impetus for quantum computing research. In 1984, he was at a computation theory conference and began to wonder about the possibility of designing a computer that was based exclusively on quantum rules, then published his breakthrough paper a few months later. With this, the race began to exploit his ideas. However, before we delve into what he started, it is beneficial to have a look at the background of the quantum world.
Quantum theory's development began in 1900 with a presentation by Max Planck to the German Physical Society, in which he introduced the idea that energy exists in individual units (which he called "quanta"), as does matter. Further developments by a number of scientists over the following thirty years led to the modern understanding of quantum theory.
Niels Bohr proposed the Copenhagen interpretation of quantum theory, which asserts that a particle is whatever it is measured to be (for example, a wave or a particle) but that it cannot be assumed to have specific properties, or even to exist, until it is measured. In short, Bohr was saying that objective reality does not exist. This translates to a principle called superposition that claims that while we do not know what the state of any object is, it is actually in all possible states simultaneously, as long as we don't look to check.
To illustrate this theory, we can use the famous and somewhat cruel analogy of Schrodinger's Cat. First, we have a living cat and place it in a thick lead box. At this stage, there is no question that the cat is alive. We then throw in a vial of cyanide and seal the box. We do not know if the cat is alive or if it has broken the cyanide capsule and died. Since we do not know, the cat is both dead and alive, according to quantum law - in a superposition of states. It is only when we break open the box and see what condition the cat is in that the superposition is lost, and the cat must be either alive or dead.
The second interpretation of quantum theory is the multiverse or many-worlds theory. It holds that as soon as a potential exists for any object to be in any state, the universe of that object transmutes into a series of parallel universes equal to the number of possible states in which that the object can exist, with each universe containing a unique single possible state of that object. Furthermore, there is a mechanism for interaction between these universes that somehow permits all states to be accessible in some way and for all possible states to be affected in some manner. Stephen Hawking and the late Richard Feynman are among the scientists who have expressed a preference for the many-worlds theory.
Which ever argument one chooses, the principle that, in some way, one particle can exist in numerous states opens up profound implications for computing.
Classical computing relies, at its ultimate level, on principles expressed by Boolean algebra, operating with a (usually) 7-mode logic gate principle, though it is possible to exist with only three modes (which are AND, NOT, and COPY). Data must be processed in an exclusive binary state at any point in time - that is, either 0 (off / false) or 1 (on / true). These values are binary digits, or bits. The millions of transistors and capacitors at the heart of computers can only be in one state at any point. While the time that the each transistor or capacitor need be either in 0 or 1 before switching states is now measurable in billionths of a second, there is still a limit as to how quickly these devices can be made to switch state. As we progress to smaller and faster circuits, we begin to reach the physical limits of materials and the threshold for classical laws of physics to apply. Beyond this, the quantum world takes over, which opens a potential as great as the challenges that are presented.
The Quantum computer, by contrast, can work with a two-mode logic gate: XOR and a mode we'll call QO1 (the ability to change 0 into a superposition of 0 and 1, a logic gate which cannot exist in classical computing). In a quantum computer, a number of elemental particles such as electrons or photons can be used (in practice, success has also been achieved with ions), with either their charge or polarization acting as a representation of 0 and/or 1. Each of these particles is known as a quantum bit, or qubit, the nature and behavior of these particles form the basis of quantum computing. The two most relevant aspects of quantum physics are the principles of superposition and entanglement .
Think of a qubit as an electron in a magnetic field. The electron's spin may be either in alignment with the field, which is known as a spin-up state, or opposite to the field, which is known as a spin-down state. Changing the electron's spin from one state to another is achieved by using a pulse of energy, such as from a laser - let's say that we use 1 unit of laser energy. But what if we only use half a unit of laser energy and completely isolate the particle from all external influences? According to quantum law, the particle then enters a superposition of states, in which it behaves as if it were in both states simultaneously. Each qubit utilized could take a superposition of both 0 and 1. Thus, the number of computations that a quantum computer could undertake is 2^n, where n is the number of qubits used. A quantum computer comprised of 500 qubits would have a potential to do 2^500 calculations in a single step. This is an awesome number - 2^500 is infinitely more atoms than there are in the known universe (this is true parallel processing - classical computers today, even so called parallel processors, still only truly do one thing at a time: there are just two or more of them doing it). But how will these particles interact with each other? They would do so via quantum entanglement.
Entanglement Particles (such as photons, electrons, or qubits) that have interacted at some point retain a type of connection and can be entangled with each other in pairs, in a process known as correlation . Knowing the spin state of one entangled particle - up or down - allows one to know that the spin of its mate is in the opposite direction. Even more amazing is the knowledge that, due to the phenomenon of superpostition, the measured particle has no single spin direction before being measured, but is simultaneously in both a spin-up and spin-down state. The spin state of the particle being measured is decided at the time of measurement and communicated to the correlated particle, which simultaneously assumes the opposite spin direction to that of the measured particle. This is a real phenomenon (Einstein called it "spooky action at a distance"), the mechanism of which cannot, as yet, be explained by any theory - it simply must be taken as given. Quantum entanglement allows qubits that are separated by incredible distances to interact with each other instantaneously (not limited to the speed of light). No matter how great the distance between the correlated particles, they will remain entangled as long as they are isolated.
Taken together, quantum superposition and entanglement create an enormously enhanced computing power. Where a 2-bit register in an ordinary computer can store only one of four binary configurations (00, 01, 10, or 11) at any given time, a 2-qubit register in a quantum computer can store all four numbers simultaneously, because each qubit represents two values. If more qubits are added, the increased capacity is expanded exponentially.
Perhaps even more intriguing than the sheer power of quantum computing is the ability that it offers to write programs in a completely new way. For example, a quantum computer could incorporate a programming sequence that would be along the lines of "take all the superpositions of all the prior computations" - something which is meaningless with a classical computer - which would permit extremely fast ways of solving certain mathematical problems, such as factorization of large numbers, one example of which we discuss below.
There have been two notable successes thus far with quantum programming. The first occurred in 1994 by Peter Shor, (now at AT&T Labs) who developed a quantum algorithm that could efficiently factorize large numbers. It centers on a system that uses number theory to estimate the periodicity of a large number sequence. The other major breakthrough happened with Lov Grover of Bell Labs in 1996, with a very fast algorithm that is proven to be the fastest possible for searching through unstructured databases. The algorithm is so efficient that it requires only, on average, roughly N square root (where N is the total number of elements) searches to find the desired result, as opposed to a search in classical computing, which on average needs N/2 searches.
The above sounds promising, but there are tremendous obstacles still to be overcome. Some of the problems with quantum computing are as follows:
Even though there are many problems to overcome, the breakthroughs in the last 15 years, and especially in the last 3, have made some form of practical quantum computing not unfeasible, but there is much debate as to whether this is less than a decade away or a hundred years into the future. However, the potential that this technology offers is attracting tremendous interest from both the government and the private sector. Military applications include the ability to break encryptions keys via brute force searches, while civilian applications range from DNA modeling to complex material science analysis. It is this potential that is rapidly breaking down the barriers to this technology, but whether all barriers can be broken, and when, is very much an open question.
- This Week in Tech: What on Earth Is a Quantum Computer? - The New York Times - December 7th, 2019
- Quantum Computers Are About to Forever Change Car Navigation - autoevolution - December 7th, 2019
- Global Quantum Computing Market Expected to Deliver Dynamic Progression until 2028| D-Wave Systems, Google, IBM, Intel, Microsoft, 1QB Information... - December 7th, 2019
- Amazon is now offering quantum computing as a service with Braket for AWS - The Verge - December 2nd, 2019
- ColdQuanta's Latest Ultracold Technology Heads to the International Space Station - Business Wire - December 2nd, 2019
- Researchers Discover New Way to Split and Sum Photons with Silicon - UT News | The University of Texas at Austin - December 2nd, 2019
- Archer Materials invited to chair quantum computing session at London conference - Proactive Investors Australia - December 2nd, 2019
- Archer Materials to chair Quantum Computing session at London Quantum.Tech Conference in 2020 - Proactive Investors Australia - November 30th, 2019
- The Future of Computing could be Magnetic - Robert Lea - Medium - November 30th, 2019
- Innovate, and grow - Economic Times - November 30th, 2019
- Tech news: The quantum internet is on the way - IOL - November 22nd, 2019
- Device Puts Photons in the Fast Lane - Optics & Photonics News - November 22nd, 2019
- Race is on to build quantum-proof encryption - Financial Times - November 21st, 2019
- Atos partners with Zapata to deliver complete quantum computing solution to the enterprise - Quantaneo, the Quantum Computing Source - November 20th, 2019
- NTT offers researchers $1 million salaries in bid to lure top talent in cryptography, quantum computing - The Japan Times - November 20th, 2019
- Information overload: The promise and risk of quantum computing - Bulletin of the Atomic Scientists - November 17th, 2019
- D-Wave sticks with its approach to quantum computing - TechCrunch - November 17th, 2019
- The Quantum Computing Threat to American Security - The Wall Street Journal - November 17th, 2019
- Dell Technologies on democratising 5G and the future of quantum computing - ZDNet - November 17th, 2019
- How Serious Is the Threat of Quantum Computing to Crypto? - Finance Magnates - November 17th, 2019
- Superconducting quantum computing - Wikipedia - October 6th, 2019
- Quantum computing | MIT News - October 6th, 2019
- How Do Quantum Computers Work? - sciencealert.com - October 2nd, 2019
- What is Quantum Computing? - Definition from Techopedia - October 2nd, 2019
- How Quantum Computers Work | HowStuffWorks - September 5th, 2019
- Quantum computing could change everything, and IBM is ... - May 15th, 2019
- Quantum Computing - Intel - April 29th, 2019
- IBM expands universities in its quantum computing research ... - April 25th, 2019
- Quantum computing is a marathon not a sprint | VentureBeat - April 22nd, 2019
- The CIO's Guide to Quantum Computing - Smarter With Gartner - April 19th, 2019
- This Startup Just Raised $21 Million To Bring Quantum ... - April 18th, 2019
- What is Quantum Computing ? Top 18 Quantum Computing ... - April 6th, 2019
- The promise of quantum computing - businessinsider.com - March 27th, 2019
- Quantum computing is coming: Heres why we need to get our ... - March 23rd, 2019
- Quantum computing will break your encryption in a few ... - March 21st, 2019
- Microsoft has formed a coalition to promote quantum computing ... - March 19th, 2019
- Quantum computing for everyone | Michael Nielsen - March 12th, 2019
- Ask a Techspert: What is quantum computing? - blog.google - March 6th, 2019
- IBM hits quantum computing milestone, may see 'Quantum ... - March 6th, 2019
- Its Time You Learned About Quantum Computing | WIRED - March 6th, 2019
- Microsofts quantum computing network takes a giant leap ... - March 2nd, 2019
- When Will Quantum Computing Have Real Commercial Value ... - February 25th, 2019
- The Case Against Quantum Computing - IEEE Spectrum - February 22nd, 2019
- How Does Quantum Computing Work? - ExtremeTech - January 31st, 2019
- Quantum technology - Wikipedia - January 23rd, 2019
- CES 2019: IBM's Q System One Is the Rock Star Quantum ... - January 13th, 2019
- Quantum Computing | The MIT Press - January 11th, 2019
- IBM thinks outside of the lab, puts quantum computer in a box - January 11th, 2019
- IBM unveils its first commercial quantum computer - January 9th, 2019
- A new type of quantum computer has smashed every record ... - December 21st, 2018
- China bet big on quantum computing. Now the US races to ... - October 26th, 2018
- US takes first step toward a quantum computing workforce ... - September 17th, 2018
- China bet big on quantum computing. Now the ... - money.cnn.com - September 17th, 2018
- The reality of quantum computing could be just three years ... - September 12th, 2018
- The quantum computing race the US cant afford to lose - September 3rd, 2018
- Quantum Computing | USRA - August 30th, 2018
- What Is Quantum Computing? The Complete WIRED Guide | WIRED - August 22nd, 2018
- Quantum Computing Market Research Report- Forecast 2022 | MRFR - August 1st, 2018
- Two Quantum Computing Bills Are Coming To Congress - July 5th, 2018
- Senate bills would make quantum computing a priority - June 10th, 2018
- The Era of Quantum Computing Is Here. Outlook: Cloudy ... - January 26th, 2018
- IBM puts its quantum computer to work in relaxing, nerdy ASMR ... - January 8th, 2018
- Quantum computing is going to change the world. Here's what ... - January 8th, 2018
- Is Quantum Computing an Existential Threat to Blockchain ... - December 25th, 2017
- What is Quantum Computing? | SAP News Center - December 23rd, 2017
- Quantum Computing Explained | What is Quantum Computing? - December 21st, 2017
- New silicon structure opens the gate to quantum computers - December 14th, 2017
- Microsoft offers developers a preview of its quantum ... - December 12th, 2017
- Quantum Computing Is the Next Big Security Risk | WIRED - December 8th, 2017
- Yale Professors Race Google and IBM to the First Quantum ... - November 16th, 2017
- IBM's processor pushes quantum computing ... - engadget.com - November 16th, 2017
- Quantum computing - news.microsoft.com - November 1st, 2017
- Intel Takes First Steps To Universal Quantum Computing - October 13th, 2017
- Qudits: The Real Future of Quantum Computing? - IEEE Spectrum - October 13th, 2017
- quantum computing - engadget.com - October 13th, 2017
- Quantum Computing | Intel Newsroom - October 13th, 2017
- What will you actually use quantum computing for? | ZDNet - October 11th, 2017
- Here's what quantum computing is and why it matters - October 6th, 2017
- Microsoft just upped its multi-million bet on quantum computing - ZDNet - September 7th, 2017
- Microsoft's Aussie quantum computing lab set to scale up next-gen ... - ARNnet - September 7th, 2017