Deep contextual recurrent residual networks for scene labeling
作者:
Highlights:
•
摘要
Designed as extremely deep architectures, deep residual networks which provide a rich visual representation and offer robust convergence behaviors have recently achieved exceptional performance in numerous computer vision problems. Being directly applied to a scene labeling problem, however, they were limited to capture long-range contextual dependence, which is a critical aspect. To address this issue, we propose a novel approach, Contextual Recurrent Residual Networks (CRRN) which is able to simultaneously handle rich visual representation learning and long-range context modeling within a fully end-to-end deep network. Furthermore, our proposed end-to-end CRRN is completely trained from scratch, without using any pre-trained models in contrast to most existing methods usually fine-tuned from the state-of-the-art pre-trained models, e.g. VGG-16, ResNet, etc. The experiments are conducted on four challenging scene labeling datasets, i.e. SiftFlow, CamVid, Stanford background and SUN datasets, and compared against various state-of-the-art scene labeling methods.
论文关键词:Recurrent network,Residual learning,Visual representation,Context modeling,Scene labeling
论文评审过程:Received 20 May 2017, Revised 25 October 2017, Accepted 7 January 2018, Available online 31 January 2018, Version of Record 8 March 2018.
论文官网地址:https://doi.org/10.1016/j.patcog.2018.01.005