Learning multimodal word representation with graph convolutional networks

作者:

Highlights:

摘要

Multimodal models have been proven to outperform text-based models on learning semantic word representations. According to psycholinguistic theory, there is a graphical relationship among the modalities of language, and in recent years, the graph convolution network (GCN) has been proven to have substantial advantages in the extraction of non-European spatial features. This inspires us to propose a new multimodal word representation model, namely, GCNW, which uses the graph convolutional network to incorporate the phonetic and syntactic information into the word representation. We use a greedy strategy to update the modality-relation matrix in the GCN, and we train the model through unsupervised learning. We evaluated the proposed model on multiple downstream NLP tasks, and various experimental results demonstrate that the GCNW outperforms strong unimodal baselines and state-of-the-art multimodal models. We make the source code of both models available to encourage reproducible research.

论文关键词:Natural language processing,Word representation,Multimodal word representation,Graph convolutional network

论文评审过程:Received 19 October 2020, Revised 11 July 2021, Accepted 27 July 2021, Available online 12 August 2021, Version of Record 12 August 2021.

论文官网地址:https://doi.org/10.1016/j.ipm.2021.102709