Integrating information theory and adversarial learning for cross-modal retrieval
作者:
Highlights:
• Combining information theory and adversarial learning into an end-to-end framework.
• Reducing the heterogeneity gap for cross-modal retrieval by reducing the measured modality uncertainty.
• Introducing a term based on KL-divergence to address the data imbalance issue in visual and textual modality.
摘要
•Combining information theory and adversarial learning into an end-to-end framework.•Reducing the heterogeneity gap for cross-modal retrieval by reducing the measured modality uncertainty.•Introducing a term based on KL-divergence to address the data imbalance issue in visual and textual modality.
论文关键词:Cross-modal retrieval,Shannon information theory,Adversarial learning,Modality uncertainty,Data imbalance
论文评审过程:Received 20 November 2019, Revised 1 November 2020, Accepted 31 March 2021, Available online 8 April 2021, Version of Record 17 April 2021.
论文官网地址:https://doi.org/10.1016/j.patcog.2021.107983