Word n-gram attention models for sentence similarity and inference

作者:

Highlights:

• Bag-of-Words models boost their performance when making use of context.

• Context in the form of arbitrary n-grams outperforms informed baselines.

• Trainable attention over arbitrary n-grams further improves results.

• Improvements of up to 41% error reduction in language inference.

• Improvements of up to 38 % error reduction in textual similarity.

摘要

•Bag-of-Words models boost their performance when making use of context.•Context in the form of arbitrary n-grams outperforms informed baselines.•Trainable attention over arbitrary n-grams further improves results.•Improvements of up to 41% error reduction in language inference.•Improvements of up to 38 % error reduction in textual similarity.

论文关键词:Attention models,Deep learning,Natural language understanding,Natural Language Inference,Semantic textual similarity

论文评审过程:Available online 22 April 2019, Version of Record 4 May 2019.

论文官网地址:https://doi.org/10.1016/j.eswa.2019.04.054