**Title:** Information-theoretic bounds on quantum advantage in machine learning

**Authors:** Hsin-Yuan Huang, Richard Kueng, John Preskill

**First Author’s Institution:** California Institute of Technology

**Status:** Pre-print on arXiv

In the past few years, machine learning has proven to be an extremely useful computational tool in a variety of disciplines. Examples include natural language processing, image recognition, beating players at Go, quantum error correction, and shifting through massive quantities of data at CERN. As is often the case, we may choose to ponder whether replacing classical algorithms by their quantum counterparts will yield computational advantages.

The authors consider the complexity of training classical machine learning (CML) and quantum machine learning (QML) algorithms on the class of problems that map a classical input (bit string) to a real number output through any physical process (including quantum processes). The goal of the ML algorithms is to learn functions of the following form: . Unpacking the notation, we see that given some input bit string that designates an initial quantum pure state, we get an output by processing the input through a quantum channel and then determining the expectation value of an operator with respect to the output of the channel (i.e. quantum evolution of the initial state). The ML algorithms will attempt to predict the expectation value of after training on many input bit strings and channel uses. You may wonder whether a single bit string is a sufficiently complicated input, but recall that all of the movies you watch, the articles on Wikipedia you read, and the music you play on your computer are represented by bit strings. Given a sufficiently long input bit string, we can feed into our algorithm arbitrarily long and precise classical data. What is very interesting about this class of problems is that we can describe quantum experiments with many different input parameters in this language, meaning the ML algorithms are being trained to predict the outcome of (potentially very complicated) quantum physical experiments. This could include predicting the outcome of quantum chemistry experiments, determining the ground state energy of a system, or predicting a variety of other observables from a myriad of physical processes.

As is often the case, there isn’t a single clear answer to the question posed, but the authors discuss two scenarios of interest. First, a secnario in which QML models pose no significant advantage over CML models and another in which QML models provide exponential speedup. According to the paper, if one consider’s minimizing *average* prediction error (over a distribution of all the inputs), then QML **does** **not** **provide a significant advantage** over classical machine learning. However, if one is interested in minimizing *worst-case* prediction error (on the single input with the worst error), then quantum machine learning requires **exponentially fewer** runs of the quantum channel.

## What are the ML Models of Interest?

The goal of supervised machine learning is to use a large quantity of input data and a corresponding set of outputs to train a machine to generate accurate predictions from new data. In this paper, the objective is to see how many times it is necessary to run a physical process such that the machine can accurately predict the results of the experiments to some tolerance. The scaling of the number of times the quantum process is used in the training phase of the algorithm defines the complexity of the algorithm. Each of the ML algorithms generates a predictive function , where the prediction error on each input is given by . If given a probability distribution over all possible inputs, we would like to figure out the number of times that must be run to either produce an *average*-case prediction error of or a *worst*-case prediction error given by .

The quantum advantage arises only when one considers the *worst-case* prediction error instead of the *average* prediction error.

It is very reasonable to think quantum machine learning would be better than classical machine learning when trying to predict the outputs of quantum experiments, but the main result throws that assumption into question. We begin by explaining how QML and CML are defined.

Classical ML models are composed of a training phase that consists of taking a randomized set of inputs and performing quantum experiments defined by a physical map (a CPTP map) from a Hilbert space of qubits to one of qubits that yields an output quantum state for each input. However, CML needs to be trained on classical output and so a POVM measurement (which is a generalization of a projective measurement) is performed that yields an output . Any classical machine learning algorithm (neural networks, kernel methods, etc.) can then be applied to generate the predictive function that approximates (the true function) from the set of pairs of inputs and classic outputs . Here denotes the number of times that applying the channel is required for CML since it is only possible to obtain one input/output pair per channel usage. The authors present a further *restricted *classical ML algorithm, which is identical to the general case, but where instead of doing arbitrary measurements to generate the outputs, only the target observable is directly measured.

A quantum ML model has the same goal as the classical models, but has the added advantage that the algorithm can be trained directly on quantum data. The general flow is that given an initial state on many qubits, the quantum channel is applied times with quantum processing maps inserted in between applications. The quantum data processing maps take the place of the classical machine learning algorithms. The resultant state that is stored in quantum memory is the learned prediction model. To predict new results, the final state after all of the channel applications and processing () is measured differently based on the desired input bit strings . The outcomes of these measurements are the outputs of the predictive function that approximate the desired function .

## The Meat of the Paper

*Average*-case Prediction Error Result

The main results of the paper are presented in the form of theorems. The first theorem shows that if there exists a QML algorithm with an *average* prediction error bounded by that requires uses of the channel , then there exists a *restricted* CML algorithm with order *average *prediction error that requires channel uses. The training complexity of the classical algorithm is directly proportional to that of the quantum model up to a factor of where is the number of qubits at the output of the map .

### Sketch of First Theorem Proof

Consider an entire family of maps that contains the map of interest. Cover the space of the family of maps with a net by choosing the largest subset of maps such that the functions generated by different maps are sufficiently distinguishable over the distribution of inputs . (In the paper they take the *average*-case error between any two different elements of to be at least ).

The proof proceeds as a communication protocol that starts by having Alice choose a random element of the packing net (). Alice then prepares the final quantum state by using the chosen channel times and interleaving those uses with the quantum data processing maps . This results in the state the QML algorithm is supposed to generate to make predictions through measurements. Alice then sends to Bob and hopes he can determine what element of the packing net she initially chose by using the predictions of the QML model. Bob can then generate the function that approximates with low *average*-case prediction error by construction. Moreover, because the different elements of the packing net are guaranteed to be far enough apart, Bob can with high-probability determine which element Alice chose. Assuming Bob could perfectly decode which element Alice sent, Alice could transmit bits to Bob. With this scheme we expect the mutual information (a measure of the amount of information that one variable can tell you about another) between Alice’s selection and Bob’s decoding to be on the order of bits. The authors now make use of Holevo’s theorem, which provides an upper bound on the amount of information Bob is able to obtain about Alice’s selection by doing measurements on the state . This is on the order of the previously given mutual information, however, it is also possible to relate the Holevo quantity directly to the number of channel uses required to prepare the state Bob receives. Through an induction argument in the Appendix, the authors show the Holevo information is upper bounded by . From this it follows that the number of channel uses for the quantum ML algorithm is lower-bounded by .

All that’s left is to relate the number of channel uses required for the QML model to the necessary number needed for the classical machine learning algorithm. A *restricted *CML is constructed such that random elements are sampled from and the desired observable is measured after preparing each time. As mentioned earlier, experiments are performed. The pairs of data are used to perform a least-squares fit to the different functions of the packing net, which are sufficiently distinguishable as defined earlier. Therefore, given a sufficiently large number of data points, it is possible to determine which element of the packing net was used as the physical map from the larger family of maps. The authors show that if the number of points obtained is on the order of , it is possible to find a function that achieves an *average-*case* *prediction error on the order of . Relating and through directly yields that . **Hence, if there is a QML model that can approximately learn ,** **then there is also a restricted CML model that can approximately learn with a similar number of channel uses**. This means that there is no quantum advantage in terms of the training complexity when considering minimum *average*-case prediction error as any QML model could be replaced by a *restricted* CML model that achieves comparable results.

*Worst*-case Prediction Error Result

The second main result states that if we instead consider *worst-case* prediction error, then an exponential separation appears between the number of channel uses necessary in the QML and CML cases. This is shown in the paper through the example of trying to predict the expectation values of different Pauli operators for an unknown -qubit state. As a refresher, the Pauli operators () are 2×2 unitary matrices that form a basis for Hermitian operators, which correspond to observable operators in quantum mechanics. Since any Hermitian operator can be decomposed in terms of sums of Paulis, it is natural to wonder about the different expectation values of each Pauli operator given an unknown state. Given a -input bit string we can specify an -qubit Pauli operator and the channel , which both generates the unknown state of interest and maps to a fixed observable . Therefore, the function we are getting the ML model to learn is . The authors show that by cleverly breaking the QML model into two stages, the first of which is estimating the magnitude of the expectation value () and the second estimating the sign of the expectation value, only channel uses are necessary to predict expectation values of any Pauli observables. Therefore it is possible to predict all Pauli expectation values up to a constant error with only a linear scaling in the number of qubits (). On the other hand, the paper shows the lower bound for estimating all the expectation values of the Pauli observables using classical machine learning is . **Therefore, there is an exponential separation between QML and CML models in the number of channel uses necessary for predicting all Pauli observables** **in the worst-case prediction error scenario. **The authors numerically demonstrate the difference in complexity between the different types of algorithms and show the separation clearly exists for a class of mixed states, but vanishes for a class of product states.

## Takeaways

Classical machine learning is much more powerful than may be naively assumed. For *average*-case prediction error, CML models are capable of achieving a comparable training complexity to QML models. This means we may not need not to wait for QML to predict the outcomes of physical experiments. Performing quantum experiments and using the output measurements to train CML models is a process that can be implemented in the near-term. In the case of approximately learning functions it appears that using a fully quantum model does not provide a significant advantage. However, the authors did demonstrate an interesting class of problems in which QML models are capable of exponentially outperforming even the best possible classical algorithm. The question is whether there are interesting tasks that require having a minimum *worst-*case prediction error, where learning the function over a majority of inputs does not suffice. I leave it to the reader as an exercise to search out other interesting learning tasks where quantum advantage is achieved and where the classical world of computing does not suffice.

*Ariel Shlosberg is a PhD student in Physics at the University of Colorado/JILA. Ariel’s research focuses on quantum error correction and finding quantum communication bounds* *and protocols.*