Sei sulla pagina 1di 13

Quantum Machine Learning

Jacob Biamonte,1, 2, Peter Wittek, ,3, 4, Nicola Pancotti,5,


Patrick Rebentrost,6, Nathan Wiebe,7, and Seth Lloyd8,
1
Quantum Complexity Science Initiative
Department of Physics, University of Malta, MSD 2080 Malta
2
Institute for Quantum Computing
University of Waterloo, Waterloo, N2L 3G1 Ontario, Canada
3
ICFO-The Institute of Photonic Sciences
Castelldefels (Barcelona), 08860 Spain
4
University of Boras
Boras, 501 90 Sweden
arXiv:1611.09347v1 [quant-ph] 28 Nov 2016

5
Max Planck Insitute of Quantum Optics
Hans-Kopfermannstr. 1, D-85748 Garching, Germany
6
Massachusetts Institute of Technology, Research Laboratory of Electronics, Cambridge, MA 02139
7
Station Q Quantum Architectures and Computation Group, Microsoft Research, Redmond WA 98052
8
Massachusetts Institute of Technology, Department of Mechanical Engineering, Cambridge MA 02139 USA
Recent progress implies that a crossover between machine learning and quantum
information processing benefits both fields. Traditional machine learning has dramat-
ically improved the benchmarking and control of experimental quantum computing
systems, including adaptive quantum phase estimation and designing quantum com-
puting gates. On the other hand, quantum mechanics offers tantalizing prospects to en-
hance machine learning, ranging from reduced computational complexity to improved
generalization performance. The most notable examples include quantum enhanced
algorithms for principal component analysis, quantum support vector machines, and
quantum Boltzmann machines. Progress has been rapid, fostered by demonstrations
of midsized quantum optimizers which are predicted to soon outperform their classical
counterparts. Further, we are witnessing the emergence of a physical theory pinpoint-
ing the fundamental and natural limitations of learning. Here we survey the cutting
edge of this merger and list several open problems.

Machine learning has fundamentally changed the way tical or solid state systems have recently reached sizes
humans interact with and relate to data. Applications where optimization methods face unprecedented data-
range from self-driving cars to intelligent agents capable intensive landscapes. In addition, machine learning was
of exceeding the best humans at Jeopardy and Go. These employed in a variety of related fields, e.g. the discov-
applications exhibit large data sets and push current ery of the Higgs Boson [10], molecular energy predic-
algorithms and computational resources to their limit. tion trained using databases of known energy spectra [11]
Information is fundamentally governed by the laws of and gravitational wave detection [12]. In the computing
physics. The laws are quantum mechanical at the scales realm, this progress allows experimental breakthroughs
of present day information processing technology, in con- which probe the threshold of producing the first practical
trast to the more familiar classical physics at the human quantum computer [13], which in turn enables quantum
scale. The interface of quantum physics and machine enhanced versions of these very same learning algorithms.
learning naturally goes both ways: machine learning al- Quantum information has shown promising algorith-
gorithms find application in understanding and control- mic developments leading to quantum speedups of com-
ling quantum systems and, on the other hand, quantum putational problems such as prime number factoring and
computational devices promise enhancement of the per- searching an unstructured database. The underlying al-
formance of machine learning algorithms for problems gorithmic toolbox allows extensions to problems relevant
beyond the reach of classical computing. for machine learning and artificial intelligence. Recently,
Machine learning is rapidly being employed for the it was shown that quantum mechanics offers physical re-
benchmarking, control, and harnessing of quantum ef- sources to enhance machine learning with quantum algo-
fects [19]. State-of-the art quantum experiments in op- rithms [1421]. Quantum-enhanced versions of classical
machine learning algorithms include least-squares fitting,
support vector machines, principal component analysis,
and deep learning. Challenges that have to be addressed
jacob.biamonte@qubit.org; www.QuamPlexity.org

in this emerging field is the input of classical data into the
peterwittek.com
nicola.pancotti@mpq.mpg.de
quantum device, the efficient processing of the data, and
pr@patrickre.com subsequent readout of classically relevant information.
nawiebe@microsoft.com Beyond quantum algorithms for machine learning,
slloyd@mit.edu there has been progress in developing a physics based the-
2

ory pinpointing the fundamental and natural limitations further provides empirical evidence that their learning
of learning, quantum enhanced learning algorithms and algorithm will find an approximation that is maximally
the employment of learning algorithms to better control close to the true model when facing cases where the hypo-
and harness these same quantum effects [1421]. Quan- thetical model lacks terms present in the actual one. This
tum information theory sets the stage to understand how suggests that even imperfect quantum resources may be
fundamental laws of nature impact the ability of physical valuable when applying learning methods to characterize
agents to learn. The cutting edge of the intersection of quantum systems.
machine learning and quantum physics is reviewed here.
Sasaki et al. [1, 26] pioneered the approach framing the
We explain how the above areas interact and we list sev-
classification of unknown quantum states as a form of su-
eral open problems that are of contemporary research
pervised learning. The authors considered semiclassical
interest.
and fully coherent quantum strategies, proving that the
latter is optimal [1, 26]. Bisio et al. [2] considered learning
CONTENTS a unitary transformation from a finite number of exam-
ples. The best strategy for learning a unitary involves a
double optimization that requires both an optimal input
I. Classical learning in quantum systems 2 stateakin to active learning in the classical theory of
statistical learningand an optimal measurement, thus
II. Quantum enhanced learning 4
this protocol is incoherent and enables induction in the
III. Quantum learning experiments 8 classical sense [2]. In a separate study, Bisio et al. [3]
derived a learning algorithm for arbitrary von Neumann
IV. Frontiers in quantum machine learning 9 measurements such that, differently from the learning of
unitary gates, the optimal algorithm for learning of quan-
Acknowledgments 10 tum measurements was not able to be parallelized, and
required quantum memory for the storage of quantum
References 10 information [3].
The authors in [4] also devised a quantum learning ma-
I. CLASSICAL LEARNING IN QUANTUM chine for binary classification of qubit states that does not
SYSTEMS require/in no need of a quantum memory. The required
classical memory was found to grow only logarithmically
with the number of training qubits [4]. The binary dis-
Recent decades have seen a concerted effort to design,
crimination problem was considered in [5] specifically for
develop, benchmark, and control systems operating in a
the case of coherent states of light. They found that
quantum regime. Such systems range from condensed
a global measurement, performed jointly over the signal
phase systems such as Bose-Einstein condensates, quan-
and the training set, enhances identification rates com-
tum clocks, and quantum computers in optical, solid-
pared to learning strategies based on first estimating the
state, and other environments. For quantum computers,
unknown amplitude by means of Gaussian measurements
the goal is to achieve quantum supremacy when a quan-
on the training set, followed by an adaptive discrimina-
tum computer outperforms a conventional computer for
tion procedure on the signal [5].
a particular problem. Classical learning algorithms were
recently employed for several building blocks needed in Concept drift is an essential problem in machine learn-
such a quantum computational device. This is particu- ing: it refers to shifts in the distribution that is being
larly timely as the data size of these problems now makes sampled and learned [27]. A similar problem in quan-
exhaustive and greedy approaches either impossible or tum mechanics is detecting the change point, that is,
at best, highly non-optimal. Quantum computing gates identifying when a source changes its output quantum
can be optimized using machine learning and evolution- state(s). The work of [9] constructs strategies for mea-
ary algorithms. In addition, analyzing the data output suring the particles individually and provides an answer
from measurement of even small quantum devices bene- as soon as a new particle is emitted by the source, repli-
fits from modern data-processing algorithms. cating the overall scheme of online learning. The authors
a. Learning about quantum systems. Exper- also show that these strategies underperform the opti-
imental quantum systems must be characterized and mal strategy, which is a global measurement. Sasaki et
benchmarked under laboratory conditions in order for al. [1, 26] pioneered this approach by framing the classi-
them to be controlled. A tantamount task is then to fication of unknown quantum states as a form of super-
find a model (a.k.a. effective) Hamiltonian of the system vised learning. The authors considered semiclassical and
and to determine properties of the present noise sources. fully coherent quantum strategies, proving that the latter
By computing likelihood functions in an adaptation of is optimal [1, 26]. Learning the community structure
Bayesian inference, Wiebe et al. [2225] found that quan- of quantum states and walks was considered in [28] by
tum Hamiltonian learning can be performed using realis- means of maximizing modularity with hierarchical clus-
tic resources such as depolarizing noise. Wiebe et al. [24] tering.
3

quantum machine quantum information


machine learning learning processing

annealing
simulated annealing quantum annealing quantum gibbs sampling

quantum topological
markov chain monte-carlo quantum BM
algorithms

feed forward neural net quantum quantum PCA quantum SVM


perceptron quantum NN classification
neural nets quantum clustering Quantum ODE solvers
quantum data fitting

quantum rejection sampling / HHL

control and metrology


quantum control
reinforcement learning phase estimation tomography
hamiltonian
learning

FIG. 1. Conceptual depiction of mutual crossovers between quantum and traditional machine learning.

b. Controlling quantum systems Learning in the development of quantum computation and infor-
methods have also seen ample success in developing mation science) [3437]. In the presence of noise and
control sequences to optimize interferometric quantum by adapting a differential evolution scheme, Zahedine-
phase estimation which is a key quantum algorithmic jad, Ghosh and Sanders [34] considered nearest-neighbor-
building block [29, 30] that appears in quantum sim- coupled superconducting artificial atoms and employed
ulation algorithms and elsewhere [31], used as a key supervised learning, resulting in gate fidelity above 99.9%
component in [32] in a proposal for a quantum percep- and hence reaching an accepted threshold for fault-
tron. Having employed heuristic global optimization tolerant quantum computing. In a separate study [35],
algorithms, Hentschel and Sanders [29] optimized many- Zahedinejad, Ghosh and Sanders developed a quantum-
particle adaptive quantum metrology in a reinforcement control procedure to construct a single-shot Toffoli gate
learning scenario. Later Lovett et al. [30] extended their (a crucial building block of a universal quantum com-
procedure to several challenges including phase esti- puter), again reaching gate fidelity above 99.9%. Using
mation and coined quantum walks. Palittapongarnpim an alternative approach, Banchi, Pancotti and Bose [36]
et al. [33] optimized this latter approach by orders of also realized a Toffoli gate without time-dependent con-
magnitude while also improving on noise tolerance and trol using the natural dynamics of a quantum network.
robustness. Las et al. [38] used genetic algorithms to reduce digi-
tal and experimental errors in quantum gates. The au-
A similar heuristic methodology has been developed thors [38] added ancillary qubits to design a modular gate
to create quantum gates (a challenge for several decades
4

made out of imperfect gates, so that their fidelity is inter- systems, such as critical points of phase transitions [47]
estingly greater than the fidelity of any of the constituent or expectation values of observables [48], and can be em-
gates. To realize quantum gates, memories and protocols, ployed in other related simulation tasks [38, 49] leading to
contemporary methods to develop dynamical decoupling applications in several fields facing many-body problems.
sequences (a leading method to protect quantum states Making use of Googles deep-learning TensorFlow li-
from decoherence) can also be surpassed using recurrent brary [50], Carrasquilla and Melko [47] developed a learn-
neural networkssee for instance August and Ni [39]. ing procedure capable of determining the current phase
Common to these approaches in quantum gate de- of matter of a quantum system. The work is based on a
sign is that they work in a supervised learning setting, standard feed-forward neural network (for proposals that
in contrast to the quantum adaptive phase estimation realize neural networks in quantum dots, see [51, 52]),
which is closer to control theory and uses reinforcement and showed that it can be trained to detect multiple types
learning. One can also exploit reinforcement learning of order parameters directly from raw state configura-
in gate-based quantum systems. For instance, Tiersch, tions sampled with Monte Carlo methods. Interestingly
Ganahl and Briegel [40] laid out a path for adaptive this particular network in the work [47] is not aware of
controllers based on intelligent agents for quantum in- the model Hamiltonian which generated the data, or the
formation tasks, illustrating how to adapt to measure- length of the interactions. This analysis outputs non-
ment directions while corresponding to an external stray trivial results for a large variety of models, ranging from
field of unknown magnitude in a fixed direction can be the classical Ising model to Coulomb phases and topo-
overcomewhich they then applied to a measurement- logical phases [47].
based algorithm for Grovers search [40]. Mavadia et al. A simple recurrent neural network, a so-called Boltz-
also used a reinforcement learning scheme to predict and mann machine, is able to faithfully reproduce expectation
compensate for qubit decoherence [41]. values by creating a large set of configurations via Monte
Other quantum algorithms directly involve ideas from Carlo sampling from the partition function of an Ising
machine learning in their basic operation. Most notably, Hamiltonian at different temperatures [48]. Those con-
the iterative phase estimation algorithm uses concepts figurations are then used to train, test and validate the
from machine learning to infer eigenvalues of a given Boltzmann machine. Once the learning has converged,
unitary operator. These techniques allow the algorithm characteristic physical propertiessuch as energy, mag-
to be run using fewer qubits and also using far less ex- netization and specific heatare computed. Near the
perimental time than previous methods. This approach, transition point, one appears to experience more difficult
originally proposed by Kitaev, was further refined by Hig- learning when the associated number of neurons in the
gins, Berry et al [42, 43] who explored the use of adap- network are required to achieve the same level of preci-
tive methods to optimally learn the unknown eigenphase. sion [48].
Such use of adaptive policies to learn and infer eigen- Choosing a Boltzmann machine with hidden variables
phases was pioneered by Hentschell and Sanders [29]. as an ansatz for the wave function, Carleo and Troyer [49]
Wiebe and Granade provided efficient alternative meth- address the many-body problemcentral to physics, ma-
ods to policy based phase estimation methods by using terials science and chemistrythrough a search method
a form of adaptive Bayesian inference, itself based on as- for a lowest-energy state. Such a function is then trained
sumed density filtering [44]. These works illustrate that via a pseudo-gradient descent algorithm originally de-
the process of data extraction from quantum algorithms signed for Monte Carlo simulations in chemistry. Fur-
can be meaningfully influenced by ideas from machine thermore Carleo and Troyer [49] challenged their result,
learning. comparing it against tensor network algorithms (see Sec-
Future applications of supervised machine learning to tion II 0 g) in both one and two dimensions, and con-
tackle noise, tailor gates and develop core quantum infor- cluding that their own method systematically improves
mation processing building blocks is a direction of tan- the best known variational states for 2D finite lattice
tamount importance. Reinforcement learning in quan- systems. Deng et al. [53] extend this idea to topologi-
tum control should also be further exploredsee Rosi cal states with long-range entanglement, showing analyt-
et al. [45] for a prime example. Furthermore, quan- ically that a topological ground state can be represented
tum walks representing an established model that cap- exactly by a short-range Restricted Boltzmann Machine.
tures essential physics behind many natural and syn-
thetic phenomena, and proven to provide a universal
model of quantum computationwere briefly touched II. QUANTUM ENHANCED LEARNING
upon here [28, 30]. To date however, comparatively little
work [28, 30, 46] has been done towards a merger with Quantum mechanics can enhance machine learning in
machine learning, providing an interesting avenue of open two different ways. First, a quantum computational de-
problems for future research. vice could perform machine learning algorithms for prob-
c. Learning properties of quantum and statis- lems beyond the reach of classical computers. We dis-
tical physics. Classical machine learning has recently cuss recent developments in quantum techniques for big
unveiled properties of quantum and related statistical data, adiabatic optimization, and Gibbs sampling. Sec-
5

ond, techniques developed in quantum theory can im- trix A and a vector b, one is faced with finding a vector
prove machine learning algorithms. In this context, we x such that Ax = b). Matrix inversion represents a com-
discuss tensor networks, renormalization, and Bayesian monly employed subroutine in data science and learning
networks. algorithms. In their variant of the problem [14], one does
d. Quantum techniques for big data. Ex- not need to know the solution x itself, but rather an
tremely large data sets have become widespread and reg- approximation of the expectation value of some opera-
ularly analysed to reveal patterns, trends, and associa- tor associated with x. They recovered an exponential
tions, ranging from many areas of physical sciences to improvement over the best known classical algorithms
human behavior and economics. As quantum physics of- when A is sparse and well conditioned [14]. By develop-
fers certain enhancements in the storage and processing ing a state preparation routine that can initialize generic
of information, a clear research track is to develop and states, Clader, Jacobs and Sprouse [15] show how ele-
tailor these quantum methods to apply to problems when mentary ancilla measurements can be used to calculate
facing big data sets [14, 15, 1820, 48, 49, 5459]. quantities of interest, and hence integrate a quantum-
A quantum speedup is characterized in several different compatible preconditioner which expands the number of
ways. One characterization is by the query complexity, problems that can achieve exponential speedup over clas-
that is the number of queries to the information storage sical linear system solvers for constant precision solu-
medium for the classical or quantum algorithm, respec- tions. They further demonstrated that their algorithm
tively. The storage medium can be more abstractly con- can be used to compute the electromagnetic scattering
sidered to be an oracle and the algorithmic speedup is cross section of an arbitrary target exponentially faster
relative to that oracle [60]. Another way of character- than the best known classical algorithm [15]. Building
izing performance is the gate complexity, counting the on these linear systems results, a quantum algorithm
number of elementary gates, say single and two qubit discovered by Wiebe, Braun and Lloyd efficiently deter-
gates, required to obtain the desired results. Many recent mines the quality of a least-squares fit over an exponen-
quantum algorithms for machine learning rely on two tially large data set [16]. They further suggest that in
main types of speedups. First, amplitude amplification many cases their algorithms can also efficiently find a
is commonly used to quadratically reduce the number of concise function that approximates the data to be fit-
samples needed in sampling algorithms. Specifically, if N ted and bound the approximation error [16], particularly
samples would be required on average in a sampling algo- when the data is sparse. Wang [65] uses singular value
rithm thenamplitude amplification can be used to reduce decomposition for the same purpose, replacing sparsity
this to O( N ) samples on average. Grover search prob- by a low-rank condition employing the quantum princi-
lem is a well known example of amplitude amplification, pal component analysis of Lloyd et al. [20]. Keeping the
and so such quadratic speedups are often called Grover- same assumption, Schuld et al. [66] developed a protocol
like. Second, other types of are speedups are related for predicting labels for new points in regression.
to prime number factoring and finding eigenvalues and A quantum algorithm for the support vector machine
eigenvectors of large matrices. This speedup is enabled based on matrix inversion was provided by Rebentrost,
by quantum phase estimation, quantum Fourier trans- Mohseni and Lloyd [18]. Relying on a least-squares for-
form, and quantum simulation methods. In many cases, mulation of the support vector machine, this algorithm
the number of quantum gates is proportional to O(log N ) was shown to have run time logarithmic in the number
for preparing a quantum state encoding eigenvalues of an of features and training examples for both training of
N N matrix and the associated eigenstates, while clas- the classifier, and the classification of new data. In cases
sically O(N ) operations are required to find eigenvalues when classical sampling algorithms terminate in polyno-
and eigenvectors. mial time, an exponential quantum speed-up in queries to
Early work by Ventura and Martinez [61] applied quan- the training data can be achieved. Central to their quan-
tum computing to training associative memories that tum algorithm [18] is a non-sparse matrix exponentiation
built on discrete Grovers search. Their modification al- technique for efficient matrix inversion of the training
lows storing only a few patterns in a superposition, and data inner-product (kernel) matrix.
the retrieval protocol receives the most similar ones to a Returning to the problem of supervised vs. unsuper-
given new instance. Grovers search can be used for dis- vised learning, Lloyd, Mohseni and Rebentrost [17] dis-
crete optimization, and Anguita et al. [62] applied this covered quantum machine learning algorithms for cluster
variant to train support vector machines. Their idea was assignment and cluster finding providing a polynomial
later generalized to create building blocks of learning speedup over sampling based classical methods for k
algorithms using Grovers search [54, 63]. Common to means clustering [17, 19].
these approaches is discretization of the search space to Finding nearest-neighbors is an association problem
achieve a quadratic speedup over classical counterparts. faced in data-analysissome of these classical methods
By a similar technique, [64] proves rigorous bounds on have been applied to determine the so called community
the learning capacity of a quantum perceptron. structure of quantum transport problems [28]. Finding
Harrow, Hassidim and Lloyd [14] provided a quantum nearest-neighbors on a quantum computer was addressed
algorithm to solve linear systems (in which given a ma- with a quantum algorithm discovered by Wiebe, Kapoor
6

and Svore in [19]. Central to the algorithm are several tum enhancements in learning. They conclude that if
subroutines for computing distance metrics such as the the agent has quantum resources while the environment
inner product and Euclidean distance. Careful analysis is classical, the only improvements can be in terms of
revealed that even in the worst case, the quantum algo- computational complexity, and they show scenarios for a
rithms offer polynomial reductions in query complexity quadratic speedup by Grover-like protocols [8].
over classical sampling based methods. A challenge facing the application of many of these
In [20], Lloyd, Mohseni and Rebentrost devised a quan- methods for big data is the fact that the training set
tum algorithm for principal component analysis of an of classical data must be loaded into the quantum com-
unknown low-rank density matrix. The main idea is to puter, a step that can dominate the cost of the algorithm
take multiple copies of a possibly unknown density ma- in some cases [70]. This issue does not, however, occur
trix and apply it as a Hamiltonian to another quantum if the data are provided via an efficient quantum subrou-
state. As in quantum tomography, such a density matrix tine or a pretrained generative model. The alternative
can be prepared from an arbitrary quantum process not solution is to load the data into a QRAM, which is a low
necessarily involving QRAM. This allows large eigenval- depth circuit for accessing data in quantum superposi-
ues and corresponding eigenvectors of the density matrix tion. Work is ongoing to engineer inexpensive QRAMs
to be computed. If constant precision is required, this in both existing [71] and fault-tolerant hardware [72], as
method can accomplish the task by using exponentially well as benchmarking the performance of QRAM enabled
fewer accesses to the training data than any existing clas- algorithms against massively parallel classical machine
sical algorithm. In an oracular (or QRAM) setting, this learning algorithms.
effort was later extended to the singular value decom- e. Adiabatic quantum optimization. Adiabatic
position of non-sparse low-rank and non-positive matri- quantum computing relies on the idea of embedding a
ces, and applied to the Procrustes problem of finding problem instance into a physical system, such that the
the best orthogonal matrix mapping one matrix into an- systems lowest energy configuration stores the prob-
other [67]. Moreover, if class labels are also available, lin- lem instance solution [73]. Recent experimental progress
ear discriminant analysis is more advantageous than prin- has resulted in annealers with hundreds of spins [74]
cipal component analysis. Cong and Duan [68] adapted detailed further in Section III.
the quantum algorithm for solving linear equations [14] These annealers make use of a logical Ising model, pro-
to achieve an exponential reduction in the number of viding an immediate connection to Hopfield neural net-
queries made to the data for this task as well. These sce- works [75], as well as many other models phrased in terms
narios are special cases of manifold learning algorithms, of energy minimization of the Ising model. Indeed, at the
where it is assumed that the data points lie on some heart of many learning algorithms is a constrained opti-
high-dimensional manifold. Principal component analy- mization problem, which can be restated as an energy
sis and singular value decomposition ensure a global op- minimization problem of an Ising model.
timum, but often one is more interested in the topology Adiabatic quantum optimization relies on a physical
of the data instances, such as connected components and process to estimate the ground state energy of the Ising
voids. Lloyd, Garnerone and Zanardi [55] designed quan- modelresembling the widely used global optimization
tum algorithms for the approximation of Betti numbers heuristic that exploits both thermal fluctuations and
from the combinatorial Laplacian for a type of topological quantum tunneling to find the global energy minimum of
manifold learning known as persistent homology. Their a systemsee figure 2 A. In other words, given a discrete
algorithm provided an exponential speedup for comput- nonconvex optimization problem, we are able to find the
ing constant precision approximations to Betti numbers global optimum as long as we meet the criteria of the
relative to the current best known classical algorithms. adiabatic theorem that drives the physical process [76].
Dridi and Alghassi [69] also use quantum annealing for Adding non-commuting (so called, xx) interactions to
homology computation. While the empirical results of the Ising model is known to render it universal [77] for
their algorithm look encouraging, more work is needed adiabatic quantum computationyet programming this
to assess whether their approach truly can give an expo- universal model and understanding its connection to ma-
nential speedup. chine learning is an open problem.
Quantum mechanics was also shown in [6] to provide Denchev et al. developed robust, regularized boosting
a speed-up for reinforcement learning. A large class of algorithms using quantum annealing [78, 79]. Dulny III
learning agents was introduced, for which a quadratic and Kim [80] used a similar methodology in a range of
boost in learning efficiency over their classical analogues tasks, including natural language processing and testing
was recovered [6]. Development of learning agents in for linear separability, whereas Pudenz and Lidar [81]
quantum environments was further considered in [7, 8]. applied it to anomaly detection. Learning the structure
In [7] classical agents were upgraded to their quan- of a probabilistic graphical model, for instance that of a
tum counterparts by a nested process of adding coherent Bayesian network, is a notoriously hard task: OGorman
control, where the focus was on implementation in ion et al. [82] address this difficulty by quantum annealing.
traps. Further, in [8] the authors analyze the types of They map the posterior-probability scoring function on
classically specified environments which allow for quan- graphs to the Ising model: n random variables map to
7

thermal annealing

quantum
annealing

thermal state
A B

FIG. 2. A quantum state tunnels when approaching a resonance point before decoherence induces thermaliza-
tion. A. A quantum state must traverse a local minimum in thermal annealing whereas a coherent quantum state can tunnel
when brought close to resonance. B. Coherent eects decay through interaction with an environment, resulting in a probability
distribution in occupancy of a systems energy levels following a Gibbs distribution.

O(n2 ) qubits. called contrastive divergence can be employed. While


Limited connectivity in current quantum annealers is such contrastive divergence often suffices for machine
a recurrent problem in developing quantum optimization learning, it can fail to converge to the solution provided
algorithms. Zaribafiyan et. al [83] devised a generic, ef- by exact training and also it cannot be used directly
ficient graph-minor embedding methods to address this to efficiently train nonrestricted Boltzmann machines.
issue. Following a similar line of thought, Diridi and Al- Quantum Gibbs sampling replaces this heuristic.
ghassi [69] designed a quantum annealing algorithm for Wiebe et al. [88] developed a Gibbs state preparation
manifold learning, more specifically, for the computation and sampling protocol, also with the objective of train-
of homology of a large set of data points (see also Sec- ing deep belief networks. They achieve polynomial im-
tion II 0 d). Benedetti et al. [84] in turn developed an provements in computational complexity relative to its
embedding of arbitrary pairwise connectivity to train a classical analogue, and in some cases superpolynomial
maximum entropy model with at most quadratic number speedups relative to contrastive divergence training. Fur-
of qubits required to represent the nodes of the original thermore, their state preparation procedure is not spe-
graph. cific to any topology for Boltzmann machines, which al-
f. Gibbs sampling. Current quantum annealing lows deep networks that are not stacked restricted Boltz-
technology is seldom guaranteed to provide the global mann machines to be accurately trained. Further ad-
optimum. Instead, the energy levels after repeated steps vances in Gibbs state preparation methods [89] open
of annealing approximately follow a Gibbs distribution the door to improved methods for training other graph
see figure 2 B. Addressing the correct embedding on topologies, such as Markov logic networks [90].
the connectivity graph and estimating the temperature Taking these ideas further, Amin et al. [58] suggest an
can be used for training Boltzmann machines [85, 86]. approach based on quantum Boltzmann distribution of
Such machines appear in several variants in this review a transverse-field Ising Hamiltonian as the fundamental
[48, 49, 58, 8589], and are simple generative neural net- model for the data. While modest advantages are seen
works consisting of hidden and visible nodes. Typically, for small networks, more work is needed to understand
classical methods focus on training restricted Boltzmann the power that such quantum models possess.
machines that only have connectivity between the adja- g. Learning tensor networks and renormaliza-
cent layers of hidden and visible units but not within a tion. Tensor networks have taken a central role in mod-
layer. Deep belief networks, extensively used in speech ern quantum physics due to their ease of use as a graph-
and image recognition, can be formed by stacking many ical language to describe and pictorially reason about
restricted Boltzmann machines. Exact training of Boltz- quantum circuits and protocols, renormalization, and
mann machines requires Gibbs sampling, but given the numerical tensor contraction and simulation algorithms.
computational complexity thereof, a heuristic algorithm These tensor network algorithms are known to efficiently
8

represent the low-energy wave function for a vast fam- fect.


ily of Hamiltonians and they offer a means to efficiently In the absence of specific graph topologies, two chal-
simulate classes of quantum circuits. At the core of the lenges are addressed: (i) structure and weight learning
algorithms we find something roughly similar to prin- from correlations, and (ii) inference based on the struc-
ciple component analysis, yet in the case of quantum ture and partial observation (a.k.a. clamping) of nodes.
systems singular value decomposition is applied recur- The first problem is akin to the training phase of other
sively to factor stationary states, and often sequentially machine learning approaches. The second phase is about
repeated when modeling time-dependence. Accordingly, applying the learned model. The computational com-
various geometrical constructions have been shown to of- plexity is typically negligible of this second phase in other
fer advantages when compressing the data required to forms of learning, but for Bayesian and Markov networks,
represent different classes of quantum statessee figure this probabilistic inference is #P-hard.
1. These methods apply well to certain classical sys- The starting point is correlationsthis is already prob-
tems and problemsfor instance, to countingand can lematic if we study quantum correlations which may not
readily be merged with machine learning techniques to have a definite causal order [96], as it has been exper-
generally enhance their applicability [9194]. imentally probed [97]. Furthermore, intuition of cause
Anandkumar et al. [92] considers parameter estima- and effect fails with quantum correlations, making causal
tion for a wide class of latent variable modelsincluding discovery a challenge [98]. In classical Bayesian networks,
Gaussian mixture models, hidden Markov models, and the d-separation theorem clearly validates whether a
Latent Dirichlet Allocationwhich exploit a certain ten- given set of correlations can match a given directed graph
sor network structure found in their low-order observable structure. Progress has been made on the quantum gen-
moments. eralization of the d-deparation theorem [56, 99], but the
The aforementioned latent variable models are shal- fully general case is still an open issue.
low in the sense that there are not many layers of pa- To solve the first phase of learning such a model, an
rameters to be identifieddeep learning architectures are adiabatic quantum optimization method was introduced
the direct opposite of this way of thinking. Mehta and to learn the structure of a classical Bayesian network [82].
Schwab [93] merges ideas of mapping from the variational In this case, all correlations are classical.
renormalization group, first introduced by Kadanoff [95], For probabilistic inference, an early effort used
and deep learning models based on restricted Boltzmann complex-valued probability amplitudes [100]. The prob-
machines so that the results indicate that deep learning lems above were bypassed by requiring that the ampli-
algorithms might be viewed as employing a generalized tudes factorize according to classical conditions, restrict-
renormalization group-like scheme to learn relevant fea- ing the case to pure states. Leifer and Poulin devel-
tures from data. Also considering data analysis, [94] re- oped an inexact belief propagation protocol for mixed
lies on the matrix product state (MPS) decomposition states [101].
as a computational tool to likewise extract features of If we restrict the topology, learning weights and prob-
multidimensional data. abilistic inference can be polynomial or even linear in the
Probabilistic graphical models (see the next section) classical sense/case. This is the case with hidden Markov
can also form a hierarchical model akin to a deep learn- models. Monras and Winter [57] consider the situation
ing architecture. [91] compared ideas behind the renor- where the hidden variables are quantum, and ask whether
malization groupof which certain tensor network meth- there is a quantum instrument that could reproduce the
ods represent the modern incarnationand such mod- observable dynamics. Cholewa et al. [102] assume a se-
els. The multiscale entanglement renormalization ansatz ries of classical observations with underlying quantum
(MERA network) was converted into a learning algorithm dynamics, and generalize known classical algorithms for
based on a hierarchical Bayesian model. Under the as- training hidden Markov networks.
sumption that the distribution to be learned is fully char-
acterized by local correlations, this algorithm [91] does
not require sampling. III. QUANTUM LEARNING EXPERIMENTS
h. Causality and Bayesian Networks. Proba-
bilistic graphical modelsincluding Bayesian networks Quantum learning algorithms have been realized in a
and Markov networks, and their special topological vari- host of experimental systems and cover a range of appli-
ants such as hidden Markov models, Ising models, and cations. Brunner et al. used photonics to demonstrate
Kalman filtersoffer compactness of representation, cap- simultaneous spoken digit and speaker recognition and
turing the sparsity structure of the model and indepen- chaotic time-series prediction at data rates beyond a giga-
dence conditions among correlations. The graph struc- byte per second [103]. They were able to identify all dig-
ture of these encompasses the qualitative properties of its with very low classification errors and perform chaotic
the distribution. While causation does not directly ap- time-series prediction with 10% error. Using photonics,
pear in the representation, the edge structure of the a classifier was realized in [104] which worked on up to
graph indicates influence, opening the door to applica- eight-dimensional vectors. Facing a nonlinear photonic
tions which rely on determining causation of a given ef- delay system [105], classically employed methods of gra-
9

dient descent with backpropagation through time were Wolf [115], where it was found that quantum and clas-
employed on a model of the system to optimize input en- sical sample complexity are equal up to constant factors
coding. Physical experiments obtained show that the in- in both the probably approximately correct vs. agnostic
put encodings result in significantly better performance models. Arunachalam and de Wolfs results imply that,
than the common reservoir computing approach [105]. when restricted to sample complexity, the classical and
Also in nonlinear photonics [106], demonstrated an all- quantum cases are equal up to constant factors. Despite
optical linear classifier capable of learning the classifica- these differences, the ability of quantum computers to ex-
tion boundary iteratively from training data through a pediently search for the most informative samples can (in
feedback rule. Lau et al. [21] proposed building blocks some cases) polynomially improve the expected number
for learning algorithms in continuous-variable quantum of samples necessary to learn a concept [64]. Thus char-
systems with a matching photonic implementation, but acterizing the statistical efficiency of quantum machine
experiments to test this are lying still ahead. learning algorithms remains an open problem.
Neural networks have been realized using liquid state Another aspect of learning theory, model complexity
nuclear magnetic resonance (NMR), by Hopfield neural is central to establishing bounds on generalization per-
networks with simulated adiabatic quantum computation formance. Monras et al. [114] prove that a supervised in-
to recover basic pattern recognition tasks [107]. Hand- ductive learning protocol always splits into a training and
writing recognition was also explored by an NMR test testing phase, even with quantum resources, and thus es-
bench in [108], enabling the recognition of standard char- tablishes the theoretical foundations for defining model
acter fonts from a set with two candidates and hence re- complexity. In a similar vein, Wiebe and Granade [116]
alizing a quantum support vector machine. ask whether a logarithmically small quantum system can
Defaulting on a chain of trapped ions, [109] simulated learn at all in the sense of Bayesian updates, and their
a neural network with induced long range interactions. answer is affirmative if the system has access to classical
The storage capacity of such a network was possible to memory but can be negative otherwise. This is because
control by changing the external trapping potential. information stored in a quantum posterior distribution
Quantum annealing for machine learning by supercon- cannot be extracted without destroying the information
ducting systems is lead by the technology developed by that it spent so long learning. This indicates that prop-
D-Wave Systems. An early demonstration focused on erties of quantum mechanics, such as the no-cloning the-
regularized boosting with a nonconvex objective function orem, challenge us to think more broadly what it means
in a classification task [110]. The optimization problem for a quantum system to learn about its surroundings.
was discretized and mapped to a quadratic unconstrained A clear current use of machine learning in quantum
binary problem. Subsequent work developed this idea of physics builds on the dramatic success these techniques
discretization and mapping [78, 79, 82, 111]. Since the have had in learning to control experimental quantum
quantum annealer suffers from a number of implemen- systems [24, 29, 30, 3436, 3941, 117, 118]. On the other
tation issues that deviate from the underlying theory, in hand, a viable contemporary approach to quantum en-
general it cannot be guaranteed that by the end of the hanced machine learning is running a quantum annealer,
annealing process one will be able to read out the ground built for example from artificial flux-based qubits [74] as
state. The distribution of readouts of the final state ap- a subroutine to solve optimization problems. This of-
proximates a Gibbs distribution. This opened the way fers an advantage inasmuch as problems can be stated
to training Boltzmann machines [58, 8587]. in terms of objective functions which can then be mini-
mized remotely, with the further advantage that mid-size
versions of the technology are already available [74]. Al-
IV. FRONTIERS IN QUANTUM MACHINE though quantum optimizers have seen dramatic increases
LEARNING in scalability (at the time of this writing 2000 man-
ufactured spins are available on a single chip), quan-
Quantum computers can outperform classical ones in tum supremacyin which a quantum device outperforms
some machine learning tasks, but the full scope of their any existing classical device for a comparable algorithmic
power is unknown. Ameur et al. [112] asked what we taskhas not yet been achieved.
could expect by various combinations of quantum and Quantum computers, however, are thought to cross
classical data and objectives, and there are a few re- the quantum supremacy threshold and offer dramatic
sults in terms of complexity bounds for quantum machine runtime reductions for several classes of problems that
learning [113115]. Servedio and Gortler [113] establish can be used as subroutines in machine learning meth-
a polynomial relationship between the number of quan- ods [14, 15, 1820, 48, 49, 5458]. Yet at the same
tum versus classical queries required for certain classes of time, certain powerful and central ideas in the theory
learning problems [113]. Their work implies that, while of machine learningsuch as neural computing [119],
the sampling complexity of broad classes of quantum and quantum generalizations of Bayesian networks [56] and
classical machine learning algorithms are polynomially quantum causal models [99]seem to necessitate an in-
equivalent, their computational complexities need not be. creased effort to understand several fundamental ques-
This idea was also confirmed by Arunachalam and de tions attached to their quantum mechanical generaliza-
10

tions [56, 57, 96, 99]. knowledges financial support from the ERC (Consol-
From the papers and research directions reviewed here, idator Grant QITBOX), MINECO (Severo Ochoa grant
we see that machine learning and quantum computing SEV-2015-0522 and FOQUS), Generalitat de Catalunya
became intertwined on many different levels: quantum (SGR 875), and Fundacio Privada Cellex. NP acknowl-
control using classical reinforcement learning, learning edges ExQM (Exploring Quantum Matter) for finan-
unitaries with optimal strategies, and speedup in various cial support. SL and PR were supported by ARO and
learning algorithms are notable examples. This leads to AFOSR.
a frontier where both quantum computing and machine
intelligence will co-evolve, and they will become enabling
technologies for each other.

ACKNOWLEDGMENTS

JDB acknowledges IARPA and the Foundational Ques-


tions Institute (FQXi) for financial support. PW ac-

[1] Masahide Sasaki, Alberto Carlini, and Richard Jozsa. cation of machine learning algorithms to the study of
Quantum template matching. Phys. Rev. A, 64:022317, noise artifacts in gravitational-wave data. Phys. Rev.
July 2001. D, 88:062003, September 2013.
[2] Alessandro Bisio, Giulio Chiribella, Giacomo Mauro [13] Rami Barends, Julian Kelly, Anthony Megrant, An-
DAriano, Stefano Facchini, and Paolo Perinotti. Op- drzej Veitia, Daniel Sank, Evan Jerey, Ted C. White,
timal quantum learning of a unitary transformation. Josh Mutus, Austin G. Fowler, Brooks Campbell,
Phys. Rev. A, 81:032324, March 2010. Yu Chen, Zijun Chen, Ben Chiaro, Andrew Dunsworth,
[3] Alessandro Bisio, Giacomo Mauro DAriano, Paolo Charles Neill, Peter OMalley, Pedram Roushan, Amit
Perinotti, and Michal Sedlak. Quantum learning al- Vainsencher, Jim Wenner, Alexander N. Korotkov, An-
gorithms for quantum measurements. Phys. Lett. A, drew N. Cleland, and John M. Martinis. Superconduct-
375(39):34253434, 2011. ing quantum circuits at the surface code threshold for
[4] G. Sents, J. Calsamiglia, R. Munoz-Tapia, and fault tolerance. Nature, 508(7497):500503, April 2014.
E. Bagan. Quantum learning without quantum memory. Letter.
Sci. Rep., 2, October 2012. Article. [14] Aram W. Harrow, Avinatan Hassidim, and Seth Lloyd.
[5] Gael Sents, Madalin Guta, and Gerardo Adesso. Quan- Quantum algorithm for linear systems of equations.
tum learning of coherent states. EPJ Quantum Tech- Phys. Rev. Lett., 103:150502, October 2009.
nology, 2(1):17, 2014. [15] B. David Clader, Bart C. Jacobs, and Chad R. Sprouse.
[6] Giuseppe Davide Paparo, Vedran Dunjko, Adi Mak- Preconditioned quantum linear system algorithm. Phys.
mal, Miguel Angel Martin-Delgado, and Hans J. Briegel. Rev. Lett., 110:250504, June 2013.
Quantum speedup for active learning agents. Phys. Rev. [16] Nathan Wiebe, Daniel Braun, and Seth Lloyd. Quan-
X, 4:031002, July 2014. tum algorithm for data tting. Phys. Rev. Lett.,
[7] Vedran Dunjko, Nicolai Friis, and Hans J. Briegel. 109:050505, August 2012.
Quantum-enhanced deliberation of learning agents us- [17] Seth Lloyd, Masoud Mohseni, and Patrick Rebentrost.
ing trapped ions. New J. Phys., 17(2):023006, 2015. Quantum algorithms for supervised and unsupervised
[8] Vedran Dunjko, Jacob M. Taylor, and Hans J. Briegel. machine learning. arXiv:1307.0411, July 2013.
Quantum-enhanced machine learning. Physical Review [18] Patrick Rebentrost, Masoud Mohseni, and Seth Lloyd.
Letters, 117(13), September 2016. Quantum support vector machine for big data classi-
[9] Gael Sents, Emilio Bagan, John Calsamiglia, Giulio cation. Phys. Rev. Lett., 113:130503, September 2014.
Chiribella, and Ramon Munoz Tapia. Quantum change [19] Nathan Wiebe, Ashish Kapoor, and Krysta M. Svore.
point. Phys. Rev. Lett., 117(15), October 2016. Quantum algorithms for nearest-neighbor methods for
[10] Pierre Chiappetta, Pietro Colangelo, P. De Felice, supervised and unsupervised learning. Quantum Info.
Giuseppe Nardulli, and Guido Pasquariello. Higgs Comput., 15(3-4):316356, March 2015.
search by neural networks at LHC. Phys. Lett. B, [20] Seth Lloyd, Masoud Mohseni, and Patrick Rebentrost.
322(3):219223, 1994. Quantum principal component analysis. Nat. Phys.,
[11] Matthias Rupp. Machine learning for quantum mechan- 10(9):631633, September 2014. Letter.
ics in a nutshell. Int. J. Quantum Chem., 115(16):1058 [21] Hoi-Kwan Lau, Raphael Pooser, George Siopsis, and
1073, 2015. Christian Weedbrook. Quantum machine learning over
[12] Rahul Biswas, Lindy Blackburn, Junwei Cao, Reed Es- innite dimensions. arXiv:1603.06222, March 2016.
sick, Kari Alison Hodge, Erotokritos Katsavounidis, [22] Christopher E Granade, Christopher Ferrie, Nathan
Kyungmin Kim, Young-Min Kim, Eric-Olivier Le Bigot, Wiebe, and David G. Cory. Robust online Hamiltonian
Chang-Hwan Lee, John J. Oh, Sang Hoon Oh, Edwin J. learning. New J. Phys., 14(10):103013, 2012.
Son, Ye Tao, Ruslan Vaulin, and Xiaoge Wang. Appli- [23] Nathan Wiebe, Christopher Granade, Christopher Fer-
11

rie, and David G. Cory. Hamiltonian learning and cer- networks to optimize dynamical decoupling for quantum
tication using quantum resources. Phys. Rev. Lett., memory. arXiv:1604.00279, April 2016.
112:190501, May 2014. [40] Markus Tiersch, E. J. Ganahl, and Hans J. Briegel.
[24] Nathan Wiebe, Christopher Granade, Christopher Fer- Adaptive quantum computation in changing environ-
rie, and David Cory. Quantum Hamiltonian learn- ments using projective simulation. Sci. Rep., 5:12874,
ing using imperfect quantum resources. Phys. Rev. A, August 2015. Article.
89:042314, April 2014. [41] Sandeep Mavadia, Virginia Frey, Jarrah Sastrawan,
[25] Nathan Wiebe, Christopher Granade, and D G Stephen Dona, and Michael J. Biercuk. Prediction
Cory. Quantum bootstrapping via compressed quan- and real-time compensation of qubit decoherence via
tum Hamiltonian learning. New J. Phys., 17(2):022005, machine-learning. arXiv:1604.03991, April 2016.
2015. [42] Brendon L. Higgins, Dominic W. Berry, Stephen D.
[26] Masahide Sasaki and Alberto Carlini. Quantum learn- Bartlett, Howard M. Wiseman, and Geo J. Pryde.
ing and universal quantum matching machine. Phys. Entanglement-free Heisenberg-limited phase estimation.
Rev. A, 66:022303, August 2002. Nature, 450(7168):393396, November 2007.
[27] Georey I. Webb, Roy Hyde, Hong Cao, Hai Long [43] Dominic W. Berry, Brendon L. Higgins, Stephen D.
Nguyen, and Francois Petitjean. Characterizing con- Bartlett, Morgan W. Mitchell, Geo J. Pryde, and
cept drift. arXiv:1511.03816, December 2015. Howard M. Wiseman. How to perform the most ac-
[28] Mauro Faccin, Piotr Migdal, Tomi H. Johnson, Ville curate possible phase measurements. Phys. Rev. A,
Bergholm, and Jacob D. Biamonte. Community de- 80:052114, November 2009.
tection in quantum complex networks. Phys. Rev. X, [44] Nathan Wiebe and Chris Granade. Ecient Bayesian
4:041012, October 2014. phase estimation. Phys. Rev. Lett., 117:010503, June
[29] Alexander Hentschel and Barry C. Sanders. Machine 2016.
learning for precise quantum measurement. Phys. Rev. [45] S. Rosi, A. Bernard, Nicole Fabbri, Leonardo Fallani,
Lett., 104:063603, February 2010. Chiara Fort, Massimo Inguscio, Tommasco Calarco, and
[30] Neil B. Lovett, Cecile Crosnier, Mart Perarnau-Llobet, Simone Montangero. Fast closed-loop optimal control
and Barry C. Sanders. Dierential evolution for many- of ultracold atoms in an optical lattice. Phys. Rev. A,
particle adaptive quantum metrology. Phys. Rev. Lett., 88:021601, August 2013.
110:220501, May 2013. [46] Maria Schuld, Ilya Sinayskiy, and Francesco Petruc-
[31] Benjamin P. Lanyon, James D. Whiteld, Geo G. cione. Quantum walks on graphs representing the ring
Gillet, Michael E. Goggin, Marcelo P. Almeida, Ivan patterns of a quantum neural network. Phys. Rev. A,
Kassal, Jacob D. Biamonte, Masoud Mohseni, Ben J. 89:032333, March 2014.
Powell, Marco Barbieri, Alan Aspuru-Guzik, and An- [47] Juan Carrasquilla and Roger G. Melko. Machine learn-
drew G. White. Towards quantum chemistry on a quan- ing phases of matter. arXiv:1605.01735, May 2016.
tum computer. Nat. Chem., 2(2):106111, February [48] Giacomo Torlai and Roger G. Melko. Learning thermo-
2010. dynamics with Boltzmann machines. arXiv:1606.02718,
[32] Maria Schuld, Ilya Sinayskiy, and Francesco Petruc- June 2016.
cione. Simulating a perceptron on a quantum computer. [49] Giuseppe Carleo and Matthias Troyer. Solving the
Phys. Lett. A, 379(7):660663, March 2015. quantum many-body problem with articial neural net-
[33] Pantita Palittapongarnpim, Peter Wittek, and Barry C. works. arXiv:1606.02318, June 2016.
Sanders. Controlling adaptive quantum phase estima- [50] Martn Abadi, Paul Barham, Jianmin Chen, Zhifeng
tion with scalable reinforcement learning. In Proceedings Chen, Andy Davis, Jerey Dean, Matthieu Devin, San-
of ESANN-16, 24th European Symposium on Artificial jay Ghemawat, Georey Irving, Michael Isard, Man-
Neural Networks, Computational Intelligence and Ma- junath Kudlur, Josh Levenberg, Rajat Monga, Sherry
chine Learning, pages 327332, April 2016. Moore, Derek G. Murray, Benoit Steiner, Paul Tucker,
[34] Ehsan Zahedinejad, Joydip Ghosh, and Barry C. Vijay Vasudevan, Pete Warden, Martin Wicke, Yuan
Sanders. Designing high-delity single-shot three-qubit Yu, and Xiaoqiang Zhang. TensorFlow: A system for
gates: A machine learning approach. arXiv:1511.08862, large-scale machine learning. arXiv:1605.08695, May
December 2015. 2016.
[35] Ehsan Zahedinejad, Joydip Ghosh, and Barry C. [51] Elizabeth C. Behrman, John Niemel, James E. Steck,
Sanders. High-delity single-shot Tooli gate via quan- and Steve R. Skinner. A quantum dot neural network. In
tum control. Phys. Rev. Lett., 114:200502, 2015. Proceedings of PhysComp-96, 4th Workshop on Physics
[36] Leonardo Banchi, Nicola Pancotti, and Sougato Bose. of Computation, pages 2228, 1996.
Quantum gate learning in qubit networks: Tooli gate [52] Mikhail V. Altaisky, Nadezhda N. Zolnikova, Natalia E.
without time-dependent control. npj Quantum Inf., Kaputkina, Victor A. Krylov, Yurii E. Lozovik, and
2:16019, July 2016. Nikesh S. Dattani. Towards a feasible implementation
[37] Pantita Palittapongarnpim, Peter Wittek, Ehsan Za- of quantum neural networks using quantum dots. Appl.
hedinejad, and Barry C. Sanders. Learning in quantum Phys. Lett., 108(10), 2016.
control: High-dimensional global optimization for noisy [53] Dong-Ling Deng, Xiaopeng Li, and S. Das Sarma. Exact
quantum dynamics. arXiv:1607.03428, July 2016. machine learning topological states. arXiv:1609.09060,
[38] Urtzi Las Heras, Unai Alvarez-Rodriguez, Enrique September 2016.
Solano, and Mikel Sanz. Genetic algorithms for digi- [54] Esma Ameur, Gilles Brassard, and Sebastien Gambs.
tal quantum simulations. Phys. Rev. Lett., 116:230504, Quantum speed-up for unsupervised learning. Machine
June 2016. Learning, 90(2):261287, 2013.
[39] Moritz August and Xiaotong Ni. Using recurrent neural [55] Seth Lloyd, Silvano Garnerone, and Paolo Zanardi.
12

Quantum algorithms for topological and geometric anal- Bunyk, Erin M. Chapple, C. Enderud, Jeremy P. Hilton,
ysis of data. Nat. Commun., 7, January 2016. Article. Kamran Karimi, Eric Ladizinsky, N. Ladizinsky, T. Oh,
[56] Joe Henson, Raymond Lal, and Matthew F Pusey. Ilya Perminov, C. Rich, M. C. Thom, Elena Tolkacheva,
Theory-independent limits on correlations from gener- Colin J. S. Truncik, Sergey Uchaikin, J. Wang, B. Wil-
alized Bayesian networks. New J. Phys., 16(11):113043, son, and Geordie Rose. Quantum annealing with man-
2014. ufactured spins. Nature, 473(7346):194198, May 2011.
[57] Alex Monras and Andreas Winter. Quantum learning of [75] John J Hopeld. Neural networks and physical systems
classical stochastic processes: The completely positive with emergent collective computational abilities. Proc.
realization problem. J. Math. Phys., 57(1):015219, 2016. Natl. Acad. Sci., 79(8):25542558, 1982.
[58] Mohammad H. Amin, Evgeny Andriyash, Jason Rolfe, [76] Donny Cheung, Peter Hyer, and Nathan Wiebe. Im-
Bohdan Kulchytskyy, and Roger Melko. Quantum proved error bounds for the adiabatic approximation. J.
Boltzmann machine. arXiv:1601.02036, January 2016. Phys. A: Math. Theor., 44(41):415302, 2011.
[59] Keisuke Fujii and Kohei Nakajima. Harnessing [77] Jacob D. Biamonte and Peter J. Love. Realizable
disordered quantum dynamics for machine learning. Hamiltonians for universal adiabatic quantum comput-
arXiv:1602.08159, February 2016. ers. Phys. Rev. A, 78:012352, July 2008.
[60] Scott Aaronson. Bqp and the polynomial hierarchy. [78] Vasil S. Denchev, Nan Ding, S.V.N. Vishwanathan, and
In Proceedings of the forty-second ACM symposium on Hartmut Neven. Robust classication with adiabatic
Theory of computing, pages 141150. ACM, 2010. quantum optimization. In Proceedings of ICML-2012,
[61] Dan Ventura and Tony Martinez. Quantum associative 29th International Conference on Machine Learning,
memory. Inform. Sciences, 124(1):273296, 2000. June 2012.
[62] Davide Anguita, Sandro Ridella, Fabio Rivieccio, and [79] Vasil S. Denchev, Nan Ding, Shin Matsushima,
Rodolfo Zunino. Quantum optimization for training S. V. N. Vishwanathan, and Hartmut Neven. To-
support vector machines. Neural Netw., 16(56):763 tally corrective boosting with cardinality penalization.
770, 2003. Advances in Neural Networks Research: arXiv:1504.01446, April 2015.
{IJCNN} 03. [80] Joseph Dulny III and Michael Kim. Developing quan-
[63] Esma Ameur, Gilles Brassard, and Sebastien Gambs. tum annealer driven data discovery. arXiv:1603.07980,
Quantum clustering algorithms. In Proceedings of March 2016.
ICML-07, 24th International Conference on Machine [81] Kristen L. Pudenz and Daniel A. Lidar. Quantum
Learning, pages 18, Corvallis, OR, USA, June 2007. adiabatic machine learning. Quantum Inf. Process.,
[64] Nathan Wiebe, Ashish Kapoor, and Krysta M Svore. 12:20272070, May 2013.
Quantum perceptron models. arXiv:1602.04799, Febru- [82] Bryan A. OGorman, Alejandro Perdomo-Ortiz, Ryan
ary 2016. Babbush, Alan Aspuru-Guzik, and Vadim Smelyanskiy.
[65] Guoming Wang. Quantum algorithms for curve tting. Bayesian network structure learning using quantum an-
arXiv:1402.0660, 2014. nealing. The European Physical Journal Special Topics,
[66] Maria Schuld, Ilya Sinayskiy, and Francesco Petruc- 224(1):163188, 2015.
cione. Prediction by linear regression on a quantum [83] Arman Zaribayan, Dominic J. J. Marchand, and Seyed
computer. Phys. Rev. A, 94(2), August 2016. Saeed Changiz Rezaei. Systematic and determinis-
[67] Patrick Rebentrost, Adrian Steens, and Seth Lloyd. tic graph-minor embedding for cartesian products of
Quantum singular value decomposition of non-sparse graphs. arXiv:1602.04274, February 2016.
low-rank matrices. arXiv:1607.05404, July 2016. [84] Marcello Benedetti, John Realpe-Gomez, Rupak
[68] Iris Cong and Luming Duan. Quantum discriminant Biswas, and Alejandro Perdomo-Ortiz. Quantum-
analysis for dimensionality reduction and classication. assisted learning of graphical models with arbitrary
New J. Phys., 18(7):073011, July 2016. pairwise connectivity. arXiv:1609.02542, September
[69] Raouf Dridi and Hedayat Alghassi. Homology compu- 2016.
tation of large point clouds using quantum annealing. [85] Steven H. Adachi and Maxwell P. Henderson. Appli-
arXiv:1512.09328, December 2015. cation of quantum annealing to training of deep neural
[70] Scott Aaronson. Read the ne print. Nat. Phys., networks. arXiv:1510.06356, November 2015.
11(4):291293, April 2015. Commentary. [86] Marcello Benedetti, John Realpe-Gomez, Rupak
[71] Francesco De Martini, Vittorio Giovannetti, Seth Lloyd, Biswas, and Alejandro Perdomo-Ortiz. Estimation of ef-
Lorenzo Maccone, Eleonora Nagali, Linda Sansoni, and fective temperatures in quantum annealers for sampling
Fabio Sciarrino. Experimental quantum private queries applications: A case study with possible applications in
with linear optics. Phys. Rev. A, 80:010302, July 2009. deep learning. Phys. Rev. A, 94:022308, August 2016.
[72] Srinivasan Arunachalam, Vlad Gheorghiu, Tomas [87] Alejandro Perdomo-Ortiz, Bryan OGorman, Joseph
Jochym-OConnor, Michele Mosca, and Priyaa Varshi- Fluegemann, Rupak Biswas, and Vadim N. Smelyan-
nee Srinivasan. On the robustness of bucket brigade skiy. Determination and correction of persistent biases
quantum RAM. New J. Phys., 17(12):123010, 2015. in quantum annealers. Sci. Rep., 6:18628, January 2016.
[73] Edward Farhi, Jerey Goldstone, Sam Gutmann, Article.
Joshua Lapan, Andrew Lundgren, and Daniel Preda. A [88] Nathan Wiebe, Ashish Kapoor, and Krysta M. Svore.
quantum adiabatic evolution algorithm applied to ran- Quantum deep learning. arXiv:1412.3489, 2014.
dom instances of an NP-complete problem. Science, [89] Anirban Narayan Chowdhury and Rolando D. Somma.
292(5516):472475, 2001. Quantum algorithms for Gibbs sampling and hitting-
[74] Mark W. Johnson, Mohammad H. S. Amin, Suzanne time estimation. arXiv:1603.02940, March 2016.
Gildert, Trevor Lanting, Firas Hamze, Neil Dickson, [90] Peter Wittek and Christian Gogolin. Quan-
R. Harris, Andrew J. Berkley, Jan Johansson, Paul tum enhanced inference in markov logic networks.
13

arXiv:1611.08104, November 2016. tron for all-optical learning. EPJ Quantum Technol.,
[91] Cedric Beny. Deep learning and the renormalization 2(1), April 2015.
group. arXiv:1301.3124, January 2013. [107] Rodion Neigovzen, Jorge L. Neves, Rudolf Sollacher,
[92] Anima Anandkumar, Rong Ge, Daniel Hsu, Sham M. and Steen J. Glaser. Quantum pattern recognition
Kakade, and Matus Telgarsky. Tensor Decomposi- with liquid-state nuclear magnetic resonance. Phys.
tions for Learning Latent Variable Models, pages 19 Rev. A, 79:042321, April 2009.
38. Springer International Publishing, Cham, October [108] Zhaokai Li, Xiaomei Liu, Nanyang Xu, and Jiangfeng
2015. Du. Experimental realization of a quantum support vec-
[93] Pankaj Mehta and David J. Schwab. An exact mapping tor machine. Phys. Rev. Lett., 114:140504, April 2015.
between the variational renormalization group and deep [109] Marisa Pons, Veronica Ahunger, Christof Wunder-
learning. arXiv:1410.3831, November 2014. lich, Anna Sanpera, Sibylle Braungardt, Aditi Sen(De),
[94] Johann A. Bengua, Ho N. Phien, Hoang D. Tuan, and Ujjwal Sen, and Maciej Lewenstein. Trapped ion chain
Minh N. Do. Matrix product state for feature extraction as a neural network: Error resistant quantum computa-
of higher-order tensors. arXiv:1503.00516, March 2015. tion. Phys. Rev. Lett., 98:023003, January 2007.
[95] Leo P. Kadano, Anthony Houghton, and Mehmet C. [110] Hartmut Neven, Vasil S Denchev, Marshall Drew-
Yalabik. Variational approximations for renormaliza- Brook, Jiayong Zhang, William G Macready, and
tion group transformations. J. Stat. Phys., 14(2):171 Geordie Rose. Binary classication using hardware im-
203, 1976. plementation of quantum annealing. In Demonstrations
[96] Ognyan Oreshkov, Fabio Costa, and Caslav Brukner. at NIPS-09, 24th Annual Conference on Neural Infor-
Quantum correlations with no causal order. Nat. Com- mation Processing Systems, pages 117, December 2009.
mun., 3:1092, October 2012. [111] Kamran Karimi, Neil G Dickson, Firas Hamze, M HS
[97] Giulia Rubino, Lee A. Rozema, Adrien Feix, Mateus Amin, Marshall Drew-Brook, Fabian A Chudak, Paul I
Araujo, Jonas M. Zeuner, Lorenzo M. Procopio, Caslav Bunyk, William G Macready, and Geordie Rose. Inves-
Brukner, and Philip Walther. Experimental verication tigating the performance of an adiabatic quantum opti-
of an indenite causal order. arXiv:1608.01683, August mization processor. Quantum Inf. Process., 11(1):7788,
2016. 2012.
[98] Christopher J. Wood and Robert S. Spekkens. The les- [112] Esma Ameur, Gilles Brassard, and Sebastien Gambs.
son of causal discovery algorithms for quantum correla- Machine Learning in a Quantum World, pages 431442.
tions: Causal explanations of Bell-inequality violations Springer Berlin Heidelberg, Berlin, Heidelberg, 2006.
require ne-tuning. New J. Phys., page 033002, 2015. [113] Rocco A. Servedio and Steven J. Gortler. Quantum
[99] Jacques Pienaar and Caslav Brukner. A graph- versus classical learnability. In Proceedings of CCC-01,
separation theorem for quantum causal models. New 16th Annual IEEE Conference on Computational Com-
J. Phys., 17(7):073020, 2015. plexity, pages 138148, June 2001.
[100] Robert R. Tucci. Quantum Bayesian Nets. International [114] Alex Monras, Gael Sents, and Peter Wittek. Inductive
Journal of Modern Physics B, 09(03):295337, 1995. quantum learning: Why you are doing it almost right.
[101] Matt. S Leifer and David Poulin. Quantum graph- arXiv:1605.07541, May 2016.
ical models and belief propagation. Ann. Phys., [115] Srinivasan Arunachalam and Ronald de Wolf. Opti-
323(8):18991946, 2008. mal quantum sample complexity of learning algorithms.
[102] MichalCholewa, Piotr Gawron, Przemyslaw Glomb, and arXiv:1607.00932, July 2016.
Dariusz Kurzyk. Quantum hidden Markov models based [116] Nathan Wiebe and Christopher Granade. Can small
on transition operation matrices. arXiv:1503.08760, quantum systems learn? arXiv:1512.03145, December
March 2015. 2015.
[103] Daniel Brunner, Miguel C. Soriano, Claudio R. Mirasso, [117] Ehsan Zahedinejad, Sophie Schirmer, and Barry C.
and Ingo Fischer. Parallel photonic information process- Sanders. Evolutionary algorithms for hard quantum
ing at gigabyte per second data rates using transient control. Phys. Rev. A, 90:032310, September 2014.
states. Nat. Commun., 4:1364, January 2013. [118] Paul B. Wigley, Patrick J. Everitt, Anton van den Hen-
[104] Xin-Dong Cai, Dian Wu, Zu-En Su, Ming-Cheng Chen, gel, J. W. Bastian, Mahasen A. Sooriyabandara, Gor-
Xi-Ling Wang, Li Li, Nai-Le Liu, Chao-Yang Lu, and don D. McDonald, Kyle S. Hardman, C. D. Quinli-
Jian-Wei Pan. Entanglement-based machine learning van, P. Manju, Carlos C. N. Kuhn, Ian R. Petersen,
on a quantum computer. Phys. Rev. Lett., 114:110504, Andre N. Luiten, Joseph J. Hope, Nicholas P. Robins,
March 2015. and Michael R. Hush. Fast machine-learning online op-
[105] Michiel Hermans, Miguel C. Soriano, Joni Dambre, Pe- timization of ultra-cold-atom experiments. Sci. Rep.,
ter Bienstman, and Ingo Fischer. Photonic delay sys- 6:25890, May 2016. Article.
tems as machine learning implementations. J. Mach. [119] Maria Schuld, Ilya Sinayskiy, and Francesco Petruc-
Learn. Res., 16(1):20812097, January 2015. cione. The quest for a quantum neural network. Quan-
[106] Nikolas Tezak and Hideo Mabuchi. A coherent percep- tum Inf. Process., 13(11):25672586, November 2014.

Potrebbero piacerti anche