Quantum Brain
← Back to papers

Stochastic noise can be helpful for variational quantum algorithms

Junyu Liu, Frederik Wilde, A. A. Mele, Xin Jin, Liang Jiang, J. Eisert·October 13, 2022·DOI: 10.1103/PhysRevA.111.052441
PhysicsComputer Science

AI Breakdown

Get a structured breakdown of this paper — what it's about, the core idea, and key takeaways for the field.

Abstract

Saddle points constitute a crucial challenge for first-order gradient descent algorithms. In notions of classical machine learning, they are avoided, for example, by means of stochastic gradient descent methods. In this work, we provide evidence that the saddle-points problem can be naturally avoided in variational quantum algorithms by exploiting the presence of stochasticity. We prove convergence guarantees and present practical examples in numerical simulations and on quantum hardware. We argue that the natural stochasticity of variational algorithms can be beneficial for avoiding strict saddle points, i.e., those saddle points with at least one negative Hessian eigenvalue. This insight that some levels of shot noise could help is expected to add a new perspective to notions of near-term variational quantum algorithms. Published by the American Physical Society 2025

Related Research

Quantum Intelligence

Ask about quantum research, companies, or market developments.