Imperfections Lower the Simulation Cost of Quantum Computers – Physics

November 23, 2020• Physics 13, 183

Classical computers can efficiently simulate the behavior of quantum computers if the quantum computer is imperfect enough.

With a few quantum bits, an ideal quantum computer can process vast amounts of information in a coordinated way, making it significantly more powerful than a classical counterpart. This predicted power increase will be great for users but is bad for physicists trying to simulate on a classical computer how an ideal quantum computer will behave. Now, a trio of researchers has shown that they can substantially reduce the resources needed to do these simulations if the quantum computer is imperfect [1]. The arXiv version of the trios paper is one of the most Scited papers of 2020 and the result generated quite a stir when it first appeared back in FebruaryI overheard it being enthusiastically discussed at the Quantum Optics Conference in Obergurgl, Austria, at the end of that month, back when we could still attend conferences in person.

In 2019, Google claimed to have achieved the quantum computing milestone known as quantum advantage, publishing results showing that their quantum computer Sycamore had performed a calculation that was essentially impossible for a classical one [2]. More specifically, Google claimed that they had completed a three-minute quantum computationwhich involved generating random numbers with Sycamores 53 qubitsthat would take thousands of years on a state-of-the-art classical supercomputer, such as IBMs Summit. IBM quickly countered the claim, arguing that more efficient memory storage would reduce the task time on a classical computer to a couple of days [3]. The claims and counterclaims sparked an industry clash and an intense debate among supporters in the two camps.

Resolving the disparity between these estimates is one of the goals of the new work by Yiqing Zhou, of the University of Illinois at UrbanaChampaign, and her two colleagues [1]. In their study, they focused on algorithms for classically replicating imperfect quantum computers, which are also known as NISQ (noisy intermediate-scale quantum) devices [4]. Todays state-of-the-art quantum computersincluding Sycamoreare NISQ devices. The algorithms the team used are based on so-called tensor network methods, specifically matrix product states (MPS), which are good for simulating noise and so are naturally suited for studying NISQ devices. MPS methods approximate low-entangled quantum states with simpler structures, so they provide a data-compression-like protocol that can make it less computationally expensive to classically simulate imperfect quantum computers (see Viewpoint: Pushing Tensor Networks to the Limit).

Zhou and colleagues first consider a random 1D quantum circuit made of neighboring, interleaved two-qubit gates and single-qubit random unitary operations. The two-qubit gates are either Controlled-NOT gates or Controlled-Z (CZ) gates, which create entanglement. They ran their algorithm for NISQ circuits containing different numbers of qubits, N, and different depths, Da parameter that relates to the number of gates the circuit executes (Fig. 1). They also varied a parameter in the MPS algorithm. is the so-called bond dimension of the MPS and essentially controls how well the MPS capture entanglement between qubits.

The trio demonstrate that they can exactly simulate any imperfect quantum circuit if D and N are small enough and is set to a value within reach of a classical computer. They can do that because shallow quantum circuits can only create a small amount of entanglement, which is fully captured by a moderate . However, as D increases, the team finds that cannot capture all the entanglement. That means that they cannot exactly simulate the system, and errors start to accumulate. The team describes this mismatch between the quantum circuit and their classical simulations using a parameter that they call the two-qubit gate fidelity fn. They find that the fidelity of their simulations slowly drops, bottoming out at an asymptotic value f as D increases. This qualitative behavior persists for different values of N and . Also, while their algorithm does not explicitly account for all the error and decoherence mechanisms in real quantum computers, they show that it does produce quantum states of the same quality (perfection) as the experimental ones.

In light of Googles quantum advantage claims, Zhou and colleagues also apply their algorithm to 2D quantum systemsSycamore is built on a 2D chip. MPS are specifically designed for use in 1D systems, but the team uses well-known techniques to extend their algorithm to small 2D ones. They use their algorithm to simulate an N=54, D=20 circuit, roughly matching the parameters of Sycamore (Sycamore has 54 qubits but one is unusable because of a defect). They replace Googles more entangling iSWAP gates with less entangling CZ gates, which allow them to classically simulate the system up to the same fidelity as reported in Ref. [2] with a single laptop. The simulation cost should increase quadratically for iSWAP-gate circuits, and although the team proposes a method for performing such simulations, they have not yet carried them out because of the large computational cost it entails.

How do these results relate to the quantum advantage claims by Google? As they stand, they do not weaken or refute claimswith just a few more qubits, and an increase in D or f, the next generation of NISQ devices will certainly be much harder to simulate. The results also indicate that the teams algorithm only works if the quantum computer is sufficiently imperfectif it is almost perfect, their algorithm provides no speed up advantage. Finally, the results provide numerical insight into the values of N, D, f, and for which random quantum circuits are confined to a tiny corner of the exponentially large Hilbert space. These values give insight into how to quantify the capabilities of a quantum computer to generate entanglement as a function of f, for example.

So, whats next? One natural question is, Can the approach here be transferred to efficiently simulate other aspects of quantum computing, such as quantum error correction? The circuits the trio considered are essentially random, whereas quantum error correction circuits are more ordered by design [5]. That means that updates to the new algorithm are needed to study such systems. Despite this limitation, the future looks promising for the efficient simulation of imperfect quantum devices [6, 7].

Jordi Tura is an assistant professor at the Lorentz Institute of the University of Leiden, Netherlands. He also leads the institutes Applied Quantum Algorithms group. Tura obtained his B.Sc. degrees in mathematics and telecommunications and his M.Sc. in applied mathematics from the Polytechnic University of Catalonia, Spain. His Ph.D. was awarded by the Institute of Photonic Sciences, Spain. During his postdoctoral stay at the Max Planck Institute of Quantum Optics in Germany, Tura started working in the field of quantum information processing for near-term quantum devices.

A nanopatterned magnetic structure features an unprecedently strong coupling between lattice vibrations and quantized spin waves, which could lead to novel ways of manipulating quantum information. Read More

Go here to see the original:
Imperfections Lower the Simulation Cost of Quantum Computers - Physics

Related Post

Comments are closed.