Learning high-dimensional quantum entanglement through physics-guided neural networks
AI Breakdown
Get a structured breakdown of this paper — what it's about, the core idea, and key takeaways for the field.
Abstract
High-gain spontaneous parametric down-conversion (SPDC) produces bright squeezed vacuum with rich high-dimensional entanglement, but its output is inherently multimodal and non-perturbative, making the full modal characterization a major computational bottleneck. We propose a physics-guided deep neural network that reconstructs the source's modal fingerprint: the high-dimensional correlation signature across radial and azimuthal indices. We designed a FiLM-modulated convolutional architecture that predicts the joint (m,l) distribution, and training is driven by a hybrid loss that couples data-driven metrics (JSD, KL, MSE, Wasserstein) with a soft orbital-angular-momentum (OAM) conservation term, providing an essential inductive bias toward physically consistent solutions. Across gain regimes, our method achieves high-fidelity reconstruction with average JSD of 1.96e-3, WEMD of 1.54e-3, and KL divergence of 7.85e-3, delivering an approximate 128-fold speedup over full numerical simulation and more than 30% accuracy gains over U-Net baselines. These results demonstrate that physics-guided learning, via a soft OAM-conservation regularizer and physically generated training targets, enables rapid and data-efficient modal characterization. Compared with traditional numerical simulation, our mesh-free method has demonstrated good generalization with limited or contaminated training data and has enabled fast "online" prediction of the quantum dynamics of a high-dimensional entanglement system for real-world experimental implementation.