RoI Tanh-polar transformer network for face parsing in the wild
作者:
Highlights:
• We transform the whole image to Tanh-polar space to preserve context and to make the convolutions rotational- equivariant.
• We use Hybrid Residual Representation Learning Blocks to extract features in both Tanh-polar and Tanh-Cartesian space.
• We present iBugMask dataset, a novel in-the-wild face parsing bench-mark that consists of more than 22 thousand images.
• We conduct extensive experiments and show that the overall framework, RTNet, improves the state-of-the-art on all benchmarks.
摘要
•We transform the whole image to Tanh-polar space to preserve context and to make the convolutions rotational- equivariant.•We use Hybrid Residual Representation Learning Blocks to extract features in both Tanh-polar and Tanh-Cartesian space.•We present iBugMask dataset, a novel in-the-wild face parsing bench-mark that consists of more than 22 thousand images.•We conduct extensive experiments and show that the overall framework, RTNet, improves the state-of-the-art on all benchmarks.
论文关键词:Face parsing,In-the-wild dataset,Head pose augmentation,Tanh-polar representation
论文评审过程:Received 26 April 2021, Accepted 27 April 2021, Available online 6 May 2021, Version of Record 25 May 2021.
论文官网地址:https://doi.org/10.1016/j.imavis.2021.104190