Training the quantum approximate optimization algorithm without access to a quantum processing unit
AI Breakdown
Get a structured breakdown of this paper — what it's about, the core idea, and key takeaways for the field.
Abstract
In this paper, we eliminate the classical outer learning loop of the quantum approximate optimization algorithm (QAOA) and present a strategy to find good parameters for QAOA based on topological arguments of the problem graph and tensor network techniques. Starting from the observation of the concentration of control parameters of QAOA, we find a way to classically infer parameters which scales polynomially in the number of qubits and exponentially with the depth of the circuit. Using this strategy, the quantum processing unit (QPU) is only needed to sample from the final state of QAOA. This method paves the way for a variation-free version of QAOA and makes QAOA more practical for applications on NISQ devices. We investigate the performance of the proposed approach for the initial assumptions and its resilience with respect to situations where they are not fulfilled. Moreover, we investigate the applicability of our method beyond the scope of QAOA, in improving schedules for quantum annealing.