Multi-network contrastive learning of visual representations
作者:
Highlights:
•
摘要
Contrastive learning, as an important way of self-supervised learning, has achieved great success in visual representations, which significantly reduces its gap with supervised learning. The essential strategy is to maximize the similarities between two augmented views of the same image (positive pairs) and to make such image easily distinguishable from other images of different types (negative pairs). Previous methods rely heavily on a large number of negative samples, such as SimCLR and MoCo. However, some recently proposed methods, such as BYOL and SimSiam, discard negative samples by introducing asymmetric structures. This paper proposes a multi-network contrastive learning methodology for visual representations (MNCLR), which integrates the end-to-end and the momentum encoder mechanisms to introduce more negative samples under a multi-network framework. The classification results on three benchmark image datasets demonstrate that the proposed MNCLR algorithm outperforms some classic contrastive learning methods.
论文关键词:Self-supervised learning,Contrastive learning,Multi-network,End-to-end,Momentum encoder
论文评审过程:Received 26 March 2022, Revised 21 September 2022, Accepted 5 October 2022, Available online 13 October 2022, Version of Record 22 October 2022.
论文官网地址:https://doi.org/10.1016/j.knosys.2022.109991