Quantum Brain
← Back to papers

Qubit-Efficient Quantum Algorithm for Linear Differential Equations

Di Fang, D. L. George, Yu Tong·July 22, 2025·DOI: 10.48550/arXiv.2507.16995
Computer SciencePhysicsMathematics

AI Breakdown

Get a structured breakdown of this paper — what it's about, the core idea, and key takeaways for the field.

Abstract

As quantum hardware rapidly advances toward the early fault-tolerant era, a key challenge is to develop quantum algorithms that are not only theoretically sound but also hardware-friendly on near-term devices. In this work, we propose a quantum algorithm for solving linear ordinary differential equations (ODEs) with a provable runtime guarantee. Our algorithm uses only a single ancilla qubit, and is locality preserving, i.e., when the coefficient matrix of the ODE is $k$-local, the algorithm only needs to implement the time evolution of $(k+1)$-local Hamiltonians. We also discuss the connection between our proposed algorithm and Lindbladian simulation as well as its application to the interacting Hatano-Nelson model, a widely studied non-Hermitian model with rich phenomenology.

Related Research

Quantum Intelligence

Ask about quantum research, companies, or market developments.