Characterizing the loss landscape of variational quantum circuits
Machine learning techniques enhanced by noisy intermediate-scale quantum (NISQ) devices and especially variational quantum circuits (VQC) have recently attracted much interest and have already been benchmarked for certain problems. Inspired by classical deep learning, VQCs are trained by gradient de...
Saved in:
Published in | Quantum science and technology Vol. 6; no. 2; pp. 25011 - 25023 |
---|---|
Main Authors | , |
Format | Journal Article |
Language | English |
Published |
IOP Publishing
01.04.2021
|
Subjects | |
Online Access | Get full text |
Cover
Loading…
Summary: | Machine learning techniques enhanced by noisy intermediate-scale quantum (NISQ) devices and especially variational quantum circuits (VQC) have recently attracted much interest and have already been benchmarked for certain problems. Inspired by classical deep learning, VQCs are trained by gradient descent methods which allow for efficient training over big parameter spaces. For NISQ sized circuits, such methods show good convergence. There are however still many open questions related to the convergence of the loss function and to the trainability of these circuits in situations of vanishing gradients. Furthermore, it is not clear how 'good' the minima are in terms of generalization and stability against perturbations of the data and there is, therefore, a need for tools to quantitatively study the convergence of the VQCs. In this work, we introduce a way to compute the Hessian of the loss function of VQCs and show how to characterize the loss landscape with it. The eigenvalues of the Hessian give information on the local curvature and we discuss how this information can be interpreted and compared to classical neural networks. We benchmark our results on several examples, starting with a simple analytic toy model to provide some intuition about the behaviour of the Hessian, then going to bigger circuits, and also train VQCs on data. Finally, we show how the Hessian can be used to adjust the learning rate for faster convergence during the training of variational circuits. |
---|---|
Bibliography: | QST-100976.R2 |
ISSN: | 2058-9565 2058-9565 |
DOI: | 10.1088/2058-9565/abdbc9 |