GRL: Knowledge graph completion with GAN-based reinforcement learning

作者:

Highlights:

摘要

Knowledge graph completion intends to infer the entities that need to be queried through the entities and relations known in the knowledge graphs. It is used in many applications, such as question and answer systems, and searching engines. As the completion process can be represented as a Markov process, existing works would solve this problem with reinforcement learning. However, there are three issues blocking them from achieving high accuracy, which are reward sparsity, missing specific domain rules, and ignoring the generation of knowledge graphs. In this paper, we design a generative adversarial net (GAN)-based reinforcement learning model, named GRL, for knowledge graph completion. First, GRL employs the graph convolutional network to embed the knowledge graphs into the low-dimensional space. Second, GRL employs both GAN and long short-term memory (LSTM) to record trajectory sequences obtained by the agent from traversing the knowledge graph and generate new trajectory sequences if needed. At the same time, GRL applies domain-specific rules accordingly. Finally, GRL employs the deep deterministic policy gradient method to optimize both rewards and adversarial loss. The experiments show that GRL is able to both generate better policies and outperform traditional methods for several tasks.

论文关键词:Knowledge graph,Knowledge graph completion,Reinforcement learning,Deep learning

论文评审过程:Received 20 February 2020, Revised 15 August 2020, Accepted 1 September 2020, Available online 23 September 2020, Version of Record 29 September 2020.

论文官网地址:https://doi.org/10.1016/j.knosys.2020.106421