Complementarity-aware cross-modal feature fusion network for RGB-T semantic segmentation

作者:

Highlights:

• We propose a Complementarity-aware Encoder, which can capture discriminative cross-modal features by modeling and fusing complementary information from RGB and thermal features.

• We design a Three-Path Fusion and Supervision to further fuse features from both modalities and supervise the training of our model.

• Our fully-equipped model (CCFFNet), achieves the new state-of-the-art performance in RGB-T semantic segmentation. And the CAE in our model is plug-and-play.

摘要

•We propose a Complementarity-aware Encoder, which can capture discriminative cross-modal features by modeling and fusing complementary information from RGB and thermal features.•We design a Three-Path Fusion and Supervision to further fuse features from both modalities and supervise the training of our model.•Our fully-equipped model (CCFFNet), achieves the new state-of-the-art performance in RGB-T semantic segmentation. And the CAE in our model is plug-and-play.

论文关键词:RGB-T,Cross-modal fusion,Multi-supervision,Semantic segmentation

论文评审过程:Received 22 March 2022, Revised 11 June 2022, Accepted 29 June 2022, Available online 7 July 2022, Version of Record 12 July 2022.

论文官网地址:https://doi.org/10.1016/j.patcog.2022.108881