Fast Variational Knowledge Graph Embedding
AI Breakdown
Get a structured breakdown of this paper — what it's about, the core idea, and key takeaways for the field.
Abstract
Embedding of a knowledge graph(KG) entities and relations in the form of vectors is an important aspect for the manipulation of the KG database for several downstream tasks, such as link prediction, knowledge graph completion, and recommendation. Because of the growing size of the knowledge graph databases, it has become a daunting task for the classical computer to train a model efficiently. Quantum computer can help speedup the embedding process of the KGs by encoding the entities into a variational quantum circuit of polynomial depth. Usually, the time complexity for such variational circuit-dependent quantum classical algorithms for each epoch is O(Npoly(logM)), where $N$ is number of elements in the knowledge graph and $M$ is the number of features of each entities of the knowledge graph. In this article we exploit additional quantum advantage by training multiple elements of KG in superpositions, thereby reducing the computing time further for the knowledge graph embedding model.