scispace - formally typeset
Open AccessJournal Article

The power of quantum neural networks

Reads0
Chats0
TLDR
This work is the first to demonstrate that well-designed quantum neural networks offer an advantage over classical neural networks through a higher effective dimension and faster training ability, which is verified on real quantum hardware.
Abstract
Fault-tolerant quantum computers offer the promise of dramatically improving machine learning through speed-ups in computation or improved model scalability. In the near-term, however, the benefits of quantum machine learning are not so clear. Understanding expressibility and trainability of quantum models-and quantum neural networks in particular-requires further investigation. In this work, we use tools from information geometry to define a notion of expressibility for quantum and classical models. The effective dimension, which depends on the Fisher information, is used to prove a novel generalisation bound and establish a robust measure of expressibility. We show that quantum neural networks are able to achieve a significantly better effective dimension than comparable classical neural networks. To then assess the trainability of quantum models, we connect the Fisher information spectrum to barren plateaus, the problem of vanishing gradients. Importantly, certain quantum neural networks can show resilience to this phenomenon and train faster than classical models due to their favourable optimisation landscapes, captured by a more evenly spread Fisher information spectrum. Our work is the first to demonstrate that well-designed quantum neural networks offer an advantage over classical neural networks through a higher effective dimension and faster training ability, which we verify on real quantum hardware.

read more

Citations
More filters
Posted Content

Noise-Induced Barren Plateaus in Variational Quantum Algorithms

TL;DR: This work rigorously proves a serious limitation for noisy VQAs, in that the noise causes the training landscape to have a barren plateau, and proves that the gradient vanishes exponentially in the number of qubits n if the depth of the ansatz grows linearly with n.
Journal ArticleDOI

Noise-induced barren plateaus in variational quantum algorithms.

TL;DR: In this article, the authors prove that the gradient vanishes exponentially in the number of qubits n if the depth of the ansatz grows linearly with n. And they prove the same result for a generic ansatz that includes as special cases the Quantum Alternating Operator Ansatz and the Unitary Coupled Cluster Ansatz.
Journal ArticleDOI

Effect of barren plateaus on gradient-free optimization

TL;DR: It is shown that gradient-free optimizers do not solve the barren plateau problem, and the main result proves that cost function differences, which are the basis for making decisions in a gradient- free optimization, are exponentially suppressed in a barren plateau.
References
More filters
Book ChapterDOI

I and J

Book

Elements of information theory

TL;DR: The author examines the role of entropy, inequality, and randomness in the design of codes and the construction of codes in the rapidly changing environment.
Book

Deep Learning

TL;DR: Deep learning as mentioned in this paper is a form of machine learning that enables computers to learn from experience and understand the world in terms of a hierarchy of concepts, and it is used in many applications such as natural language processing, speech recognition, computer vision, online recommendation systems, bioinformatics, and videogames.
Book ChapterDOI

On the Uniform Convergence of Relative Frequencies of Events to Their Probabilities

TL;DR: This chapter reproduces the English translation by B. Seckler of the paper by Vapnik and Chervonenkis in which they gave proofs for the innovative results they had obtained in a draft form in July 1966 and announced in 1968 in their note in Soviet Mathematics Doklady.
Related Papers (5)

Quantum supremacy using a programmable superconducting processor

Frank Arute, +85 more
- 24 Oct 2019 -